Model

Baize Homepage, Documentation and Downloads – Large Language Model Trained Using LoRA – News Fast Delivery

Baize is an open-source chat model trained with LoRA, which improves upon the open-source large-scale language model LLaMA by fine-tuning LLaMA with a newly generated chat corpus, which runs on a single GPU, making it available to a wider range of researchers . Bai Ze currently includes four models in English: Bai Ze-7B, 13B, and […]

Baize Homepage, Documentation and Downloads – Large Language Model Trained Using LoRA – News Fast Delivery Read More »

Dialog generation large model PanGu-Dialog

Pengcheng·Pangu dialog generation large model (PanGu-Dialog). PanGu-Dialog is a large-scale open-domain dialogue generation model characterized by big data and large models. It makes full use of the knowledge and language capabilities of large-scale pre-trained language models to build a controllable, reliable, intelligent and natural human-computer dialogue. Model. The main features are as follows: Dialogue intelligence

Dialog generation large model PanGu-Dialog Read More »

Linly Homepage, Documentation and Downloads – Large Scale Chinese Language Model – News Fast Delivery

This project provides the community withChinese dialogue model Linly-ChatFlow, Chinese basic model Linly-Chinese-LLaMA and their training data.model based on Tencent Pretrain The pre-training framework is realized, and the full-tuning on 32 * A100 GPU will gradually open up the Chinese model weights of 7B, 13B, 33B, and 65B scales. The Chinese basic model is based

Linly Homepage, Documentation and Downloads – Large Scale Chinese Language Model – News Fast Delivery Read More »

Huatuo Homepage, Documentation and Downloads- LLaMA fine-tuning model based on Chinese medical knowledge- News Fast Delivery

HuaTuo is an LLaMA fine-tuning model based on Chinese medical knowledge. This project has open sourced the LLaMA-7B model that has undergone Chinese medical instruction fine-tuning/instruct-tuning. A Chinese medical instruction dataset was constructed through the medical knowledge graph and GPT3.5 API, and on this basis, LLaMA was fine-tuned to improve the question-answering effect of LLaMA

Huatuo Homepage, Documentation and Downloads- LLaMA fine-tuning model based on Chinese medical knowledge- News Fast Delivery Read More »

Multimodal Large Language Model mPLUG-Owl

The multimodal GPT model proposed by Alibaba Dharma Institute: mPLUG-Owl, a multimodal large language model based on mPLUG modularization. It can understand not only the content of inference text, but also visual information, and has excellent cross-modal alignment ability. Paper: https://arxiv.org/abs/2304.14178 DEMO: https://huggingface.co/spaces/MAGAer13/mPLUG-Owl Example highlights a modular training paradigm for multimodal language models. It can

Multimodal Large Language Model mPLUG-Owl Read More »

VLE Homepage, Documentation and Downloads- Vision-Language Multimodal Pre-Training Model- News Fast Delivery

VLE (Vision-Llanguage E.ncoder) is an image-text multimodal understanding model based on pre-trained text and image encoders, which can be applied to multimodal discriminative tasks such as visual question answering and image-text retrieval. In particular, VLE achieves the best performance among public models in the Visual Commonsense Reasoning (VCR) task, which has stronger requirements on language

VLE Homepage, Documentation and Downloads- Vision-Language Multimodal Pre-Training Model- News Fast Delivery Read More »

CINO Homepage, Documentation and Downloads- Minority Language Pre-training Model- News Fast Delivery

In the field of natural language processing, the pre-trained language model (Pre-trained Language Model, PLM) has become an important basic technology, and the use of pre-trained models is becoming more and more common in multilingual research. In order to promote the research and development of Chinese minority language information processing,Harbin Institute of Technology Xunfei Joint

CINO Homepage, Documentation and Downloads- Minority Language Pre-training Model- News Fast Delivery Read More »

Chinese small pre-training model MiniRBT

In the field of natural language processing, pre-trained language models (Pre-trained Language Models) have become a very important basic technology. In order to further promote the research and development of Chinese information processing, Harbin Institute of Technology Xunfei Joint Laboratory (HFL) based on the self-developed knowledge distillation tool TextBrewer, combined with Whole Word Masking (Whole

Chinese small pre-training model MiniRBT Read More »

Bark Homepage, Documentation and Downloads – Text Prompt to Generate Audio Model – News Fast Delivery

Bark is a transformer-based text-to-audio model created by Suno. Bark can generate highly realistic multilingual speech as well as other audio – including music, background noise and simple sound effects. The model can also produce non-verbal communication such as laughing, sighing and crying. To support the research community, we provide access to checkpoints of pretrained

Bark Homepage, Documentation and Downloads – Text Prompt to Generate Audio Model – News Fast Delivery Read More »

MOSS Homepage, Documentation and Downloads – Dialogue Large Language Model – News Fast Delivery

MOSS is an open source dialogue language model that supports Chinese-English bilingualism and various plug-ins.moss-moonThe series models have 16 billion parameters, and can run on a single A100/A800 or two 3090 graphics cards at FP16 precision, and can run on a single 3090 graphics card at INT4/8 precision. The MOSS pedestal language model is pre-trained

MOSS Homepage, Documentation and Downloads – Dialogue Large Language Model – News Fast Delivery Read More »

StableLM Homepage, Documentation and Downloads – Language Model Developed by Stability AI – News Fast Delivery

The StableLM project repository contains Stability AI’s ongoing development of the StableLM series of language models, and currently Stability AI has released the initial StableLM-alpha model set with 3 billion and 7 billion parameters. Models with 15 billion and 30 billion parameters are in development. Stability AI is currently at Hugging Face Spaces A demo

StableLM Homepage, Documentation and Downloads – Language Model Developed by Stability AI – News Fast Delivery Read More »

LLaVA, a Large Multimodal Model for End-to-End Training

LLaVA is a large-scale language and vision assistant built for multimodal GPT-4 level features. represents a large multimodal model trained end-to-end, concatenating a visual encoder and LLM for general vision and language understanding. Demo Early experiments show that LLaVA demonstrates excellent multi-modal chat capabilities, sometimes exhibiting multi-model GPT-4 behavior on unseen images/instructions, compared to GPT-4

LLaVA, a Large Multimodal Model for End-to-End Training Read More »

RedPajama Homepage, Documentation and Downloads – Large Language Model – News Fast Delivery

The RedPajama project aims to create a leading set of fully open source large language models. Currently, the project has completed the first step, successfully replicating more than 1.2 trillion data tokens from the LLaMA training dataset. The project is jointly developed by Together, Ontocord.ai, ETH DS3Lab, Stanford University CRFM, Hazy Research and MILA Quebec

RedPajama Homepage, Documentation and Downloads – Large Language Model – News Fast Delivery Read More »