-
miqu-1-70b - miqudev 🤗
-
H2O-Danube-1.8B Technical Report,
arXiv, 2401.16818
, arxiv, pdf, cication: -1Philipp Singer, Pascal Pfeiffer, Yauhen Babakhin, Maximilian Jeblick, Nischay Dhankhar, Gabor Fodor, Sri Satish Ambati
-
Smaug-34B-v0.1 - abacusai 🤗
-
bagel-34b-v0.2 - jondurbin 🤗
-
TinyLlama: An Open-Source Small Language Model,
arXiv, 2401.02385
, arxiv, pdf, cication: -1Peiyuan Zhang, Guangtao Zeng, Tianduo Wang, Wei Lu · (TinyLlama - jzhang38)
-
TigerBot: An Open Multilingual Multitask LLM,
arXiv, 2312.08688
, arxiv, pdf, cication: -1Ye Chen, Wei Cai, Liangmin Wu, Xiaowei Li, Zhanxuan Xin, Cong Fu
-
DeciLM-7B - Deci 🤗
-
DeciLM-7B-instruct - Deci 🤗
· (huggingface)
-
LLM360: Towards Fully Transparent Open-Source LLMs,
arXiv, 2312.06550
, arxiv, pdf, cication: -1Zhengzhong Liu, Aurick Qiao, Willie Neiswanger, Hongyi Wang, Bowen Tan, Tianhua Tao, Junbo Li, Yuqi Wang, Suqi Sun, Omkar Pangarkar
-
GPT4All: An Ecosystem of Open Source Compressed Language Models,
arXiv, 2311.04931
, arxiv, pdf, cication: -1Yuvanesh Anand, Zach Nussbaum, Adam Treat, Aaron Miller, Richard Guo, Ben Schmidt, GPT4All Community, Brandon Duderstadt, Andriy Mulyar · (gpt4all - nomic-ai)
-
OpenChat: Advancing Open-source Language Models with Mixed-Quality Data,
arXiv, 2309.11235
, arxiv, pdf, cication: -1Guan Wang, Sijie Cheng, Xianyuan Zhan, Xiangang Li, Sen Song, Yang Liu · (openchat - imoneoi)
· (huggingface) · (openchat)
-
Zephyr: Direct Distillation of LM Alignment,
arXiv, 2310.16944
, arxiv, pdf, cication: 1Lewis Tunstall, Edward Beeching, Nathan Lambert, Nazneen Rajani, Kashif Rasul, Younes Belkada, Shengyi Huang, Leandro von Werra, Clémentine Fourrier, Nathan Habib · (alignment-handbook - huggingface)
-
H2O Open Ecosystem for State-of-the-art Large Language Models,
arXiv, 2310.13012
, arxiv, pdf, cication: -1Arno Candel, Jon McKinney, Philipp Singer, Pascal Pfeiffer, Maximilian Jeblick, Chun Ming Lee, Marcos V. Conde
-
BTLM-3B-8K: 7B Parameter Performance in a 3B Parameter Model,
arXiv, 2309.11568
, arxiv, pdf, cication: -1Nolan Dey, Daria Soboleva, Faisal Al-Khateeb, Bowen Yang, Ribhu Pathria, Hemant Khachane, Shaheer Muhammad, Zhiming, Chen, Robert Myers
-
OpenBA: An Open-sourced 15B Bilingual Asymmetric seq2seq Model Pre-trained from Scratch,
arXiv, 2309.10706
, arxiv, pdf, cication: -1Juntao Li, Zecheng Tang, Yuyang Ding, Pinzheng Wang, Pei Guo, Wangjie You, Dan Qiao, Wenliang Chen, Guohong Fu, Qiaoming Zhu · (openba - opennlg)
-
XGen-7B Technical Report,
arXiv, 2309.03450
, arxiv, pdf, cication: 3Erik Nijkamp, Tian Xie, Hiroaki Hayashi, Bo Pang, Congying Xia, Chen Xing, Jesse Vig, Semih Yavuz, Philippe Laban, Ben Krause
-
FLM-101B: An Open LLM and How to Train It with $100K Budget,
arXiv, 2309.03852
, arxiv, pdf, cication: 3Xiang Li, Yiqun Yao, Xin Jiang, Xuezhi Fang, Xuying Meng, Siqi Fan, Peng Han, Jing Li, Li Du, Bowen Qin · (huggingface)
-
adept-inference - persimmon-ai-labs
Inference code for Persimmon-8B
-
WizardLM - nlpxucan
Family of instruction-following LLMs powered by Evol-Instruct: WizardLM, WizardCoder
-
FreeWilly2 - stabilityai 🤗
-
xgen - salesforce
Salesforce open-source LLMs with 8k sequence length.
-
PolyLM: An Open Source Polyglot Large Language Model,
arXiv, 2307.06018
, arxiv, pdf, cication: 5Xiangpeng Wei, Haoran Wei, Huan Lin, Tianhao Li, Pei Zhang, Xingzhang Ren, Mei Li, Yu Wan, Zhiwei Cao, Binbin Xie
-
A Technical Report for Polyglot-Ko: Open-Source Large-Scale Korean Language Models,
arXiv, 2306.02254
, arxiv, pdf, cication: -1Hyunwoong Ko, Kichang Yang, Minho Ryu, Taekyoon Choi, Seungmu Yang, Jiwung Hyun, Sungho Park, Kyubyong Park
-
Examining User-Friendly and Open-Sourced Large GPT Models: A Survey on Language, Multimodal, and Scientific GPT Models,
arXiv, 2308.14149
, arxiv, pdf, cication: -1Kaiyuan Gao, Sunan He, Zhenyu He, Jiacheng Lin, QiZhi Pei, Jie Shao, Wei Zhang · (gpt_alternatives - GPT-Alternatives)
· (jiqizhixin)
-
OLMo: Accelerating the Science of Language Models,
arXiv, 2402.00838
, arxiv, pdf, cication: -1Dirk Groeneveld, Iz Beltagy, Pete Walsh, Akshita Bhagia, Rodney Kinney, Oyvind Tafjord, Ananya Harsh Jha, Hamish Ivison, Ian Magnusson, Yizhong Wang · (OLMo - allenai)
· (allenai) · (allenai)
-
OLMo-7B - allenai 🤗
-
phi-2 - microsoft 🤗
-
phi-1_5 - microsoft 🤗
-
phi-1 - microsoft 🤗
-
phi-2 - randomblock1 🤗
-
phixtral-4x2_8 - mlabonne 🤗
-
mistral-src - mistralai
Reference implementation of Mistral AI 7B v0.1 model. · (jiqizhixin)
-
mixtral - 🤗
-
llama-mistral - dzhulgakov
Inference code for Mistral and Mixtral hacked up into original Llama implementation
-
DiscoLM-mixtral-8x7b-v2 - DiscoResearch 🤗
-
Mixtral-8x7B-Instruct-v0.1 - mistralai 🤗
-
mixtral-7b-8expert - DiscoResearch 🤗
· (huggingface)
-
mixtral-8x7b-32kseqlen - someone13574 🤗
-
mixtral-46.7b-chat - openskyml 🤗
-
Mixtral-8x7B-v0.1-GPTQ - TheBloke 🤗
-
MixtralKit - open-compass
A toolkit for inference and evaluation of 'mixtral-8x7b-32kseqlen' from Mistral AI
-
mistral-playground - marcofrodl 🤗
-
Mixtral-8x7B-Instruct-v0.1-bnb-4bit - ybelkada 🤗
-
notux-8x7b-v1 - argilla 🤗
-
mixtral-offloading - dvmazur
Run Mixtral-8x7B models in Colab or consumer desktops
-
mixtral-test-46.7b-chat - johann22 🤗
-
Nous-Hermes-2-Mixtral-8x7B-SFT - NousResearch 🤗
· (jiqizhixin)
-
Nous-Hermes-2-Mixtral-8x7B-DPO - NousResearch 🤗
-
Nous-Hermes-2-Mixtral-8x7B-DPO-adapter - NousResearch 🤗
-
miqu-1-70b - miqudev 🤗
-
StripedHyena-Hessian-7B - togethercomputer 🤗
-
StripedHyena-Nous-7B - togethercomputer 🤗
· (together)
- BLOOMChat-176B-v1-GPTQ - TheBloke 🤗
GitHub - mosaicml/llm-foundry: LLM training code for MosaicML foundation models
- mpt-30b-chat - mosaicml 🤗
-
h2oGPT: Democratizing Large Language Models,
arXiv, 2306.08161
, arxiv, pdf, cication: -1Arno Candel, Jon McKinney, Philipp Singer, Pascal Pfeiffer, Maximilian Jeblick, Prithvi Prabhu, Jeff Gambera, Mark Landry, Shivam Bansal, Ryan Chesler
-
LiteLlama-460M-1T - ahxt 🤗
· (jiqizhixin)
-
Llama-2-7b-chat-mlx - mlx-llama 🤗
-
TinyLlama - jzhang38
The TinyLlama project is an open endeavor to pretrain a 1.1B Llama model on 3 trillion tokens.
-
llama-recipes - facebookresearch
Examples and recipes for Llama 2 model · (mp.weixin.qq) · (jiqizhixin) · (mp.weixin.qq) · (d7mv45xi4m.feishu)
-
llama2-13b-orca-8k-3319 - OpenAssistant 🤗
-
pyllama - juncongmoo
LLaMA: Open and Efficient Foundation Language Models
-
llama-gpt - getumbrel
A self-hosted, offline, ChatGPT-like chatbot. Powered by Llama 2. 100% private, with no data leaving your device.
-
LLongMA-2-13b-16k - conceptofmind 🤗
-
LLongMA-2-13b - conceptofmind 🤗
-
LLongMA-2-7b-16k - conceptofmind 🤗
-
llama2-webui - liltom-eth
Run Llama 2 locally with gradio UI on GPU or CPU from anywhere (Linux/Windows/Mac). Supporting Llama-2-7B/13B/70B with 8-bit, 4-bit. Supporting GPU inference (6 GB VRAM) and CPU inference.
-
Flan-Open-Llama-13b - conceptofmind 🤗
-
Llama-2 - amitsangani
All the projects related to Llama
-
Falcon-LLM - Sentdex
Helper scripts and examples for exploring the Falcon LLM models · (huggingface) · (huggingface)
[2304.01373] Pythia: A Suite for Analyzing Large Language Models Across Training and Scaling
· ([pythia](https://github.com/EleutherAI/pythia) - EleutherAI) 
-
The History of Open-Source LLMs: Imitation and Alignment (Part Three)
· (mp.weixin.qq)
-
os-llms - blog 🤗
-
Flacuna: Unleashing the Problem Solving Power of Vicuna using FLAN Fine-Tuning,
arXiv, 2307.02053
, arxiv, pdf, cication: 3Deepanway Ghosal, Yew Ken Chia, Navonil Majumder, Soujanya Poria
-
FastChat - lm-sys
An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5.
-
stanford_alpaca - tatsu-lab
Code and documentation to train Stanford's Alpaca models, and generate the data. · (crfm.stanford)
-
dolly - databrickslabs
Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform · (huggingface) · (databricks)
-
Beagle14-7B - mlabonne 🤗
-
Improving Open-Source LLMs - Datasets, Merging and Stacking - The Abacus.AI Blog
-
CrystalChat - LLM360 🤗
-
btlm-3b-8k-chat - cerebras 🤗
-
stablelm-zephyr-3b - stabilityai 🤗
· (huggingface)
-
smol-7b - rishiraj 🤗
-
LLM-Blender: Ensembling Large Language Models with Pairwise Ranking and Generative Fusion,
arXiv, 2306.02561
, arxiv, pdf, cication: -1Dongfu Jiang, Xiang Ren, Bill Yuchen Lin · (huggingface) · (LLM-Blender - yuchenlin)
-
Intel Neural-Chat 7b: Fine-Tuning on Gaudi2 for Top LLM Performance
-
Starling-7B: Increasing LLM Helpfulness & Harmlessness with RLAIF
-
sparse-llama-gsm8k - neuralmagic 🤗
-
DeciLM-6b - Deci 🤗
-
GOAT-7B-Community - GOAT-AI 🤗
-
openchat - imoneoi
OpenChat: Less is More for Open-source Models · (mp.weixin.qq)
-
GPT-4-LLM - Instruction-Tuning-with-GPT-4
Instruction Tuning with GPT-4
-
Instruction Tuning with GPT-4,
arXiv, 2304.03277
, arxiv, pdf, cication: 182Baolin Peng, Chunyuan Li, Pengcheng He, Michel Galley, Jianfeng Gao · (instruction-tuning-with-gpt-4.github)
-
deepseek-coder-7b-instruct - deepseek-ai 🤗
-
UltraChat - thunlp
Large-scale, Informative, and Diverse Multi-round Chat Data (and Models) · (mp.weixin.qq) · (qbitai)
-
How Far Can Camels Go? Exploring the State of Instruction Tuning on Open Resources,
arXiv, 2306.04751
, arxiv, pdf, cication: 40Yizhong Wang, Hamish Ivison, Pradeep Dasigi, Jack Hessel, Tushar Khot, Khyathi Raghavi Chandu, David Wadden, Kelsey MacMillan, Noah A. Smith, Iz Beltagy · (jiqizhixin) · (open-instruct - allenai)
-
MiniCPM - OpenBMB
MiniCPM-2.4B: An end-side LLM outperforms Llama2-13B.
· (huggingface)
-
Orion-14B: Open-source Multilingual Large Language Models,
arXiv, 2401.12246
, arxiv, pdf, cication: -1Du Chen, Yi Huang, Xiaopu Li, Yongqiang Li, Yongqiang Liu, Haihui Pan, Leichao Xu, Dacheng Zhang, Zhipeng Zhang, Kun Han
-
Orion - OrionStarAI
Orion-14B is a family of models includes a 14B foundation LLM, and a series of models: a chat model, a long context model, a quantized model, a RAG fine-tuned model, and an Agent fine-tuned model. Orion-14B 系列模型包括一个具有140亿参数的多语言基座大模型以及一系列相关的衍生模型,包括对话模型,长文本模型,量化模型,RAG微调模型,Agent微调模型等。 · (Orion - OrionStarAI)
-
TeleChat Technical Report,
arXiv, 2401.03804
, arxiv, pdf, cication: -1Zihan Wang, Xinzhang Liu, Shixuan Liu, Yitong Yao, Yuyao Huang, Zhongjiang He, Xuelong Li, Yongxiang Li, Zhonghao Che, Zhaoxi Zhang
-
YAYI 2: Multilingual Open-Source Large Language Models,
arXiv, 2312.14862
, arxiv, pdf, cication: -1Yin Luo, Qingchao Kong, Nan Xu, Jia Cao, Bao Hao, Baoyu Qu, Bo Chen, Chao Zhu, Chenyang Zhao, Donglei Zhang
-
SeaLLMs -- Large Language Models for Southeast Asia,
arXiv, 2312.00738
, arxiv, pdf, cication: -1Xuan-Phi Nguyen, Wenxuan Zhang, Xin Li, Mahani Aljunied, Qingyu Tan, Liying Cheng, Guanzheng Chen, Yue Deng, Sen Yang, Chaoqun Liu
· (SeaLLMs - DAMO-NLP-SG)
-
YUAN 2.0: A Large Language Model with Localized Filtering-based Attention,
arXiv, 2311.15786
, arxiv, pdf, cication: -1Shaohua Wu, Xudong Zhao, Shenling Wang, Jiangang Luo, Lingjun Li, Xi Chen, Bing Zhao, Wei Wang, Tong Yu, Rongguo Zhang · (Yuan-2.0 - IEIT-Yuan)
-
Ziya2: Data-centric Learning is All LLMs Need,
arXiv, 2311.03301
, arxiv, pdf, cication: -1Ruyi Gan, Ziwei Wu, Renliang Sun, Junyu Lu, Xiaojun Wu, Dixiang Zhang, Kunhao Pan, Ping Yang, Qi Yang, Jiaxing Zhang
· (huggingface)
-
Skywork: A More Open Bilingual Foundation Model,
arXiv, 2310.19341
, arxiv, pdf, cication: 1Tianwen Wei, Liang Zhao, Lichang Zhang, Bo Zhu, Lijie Wang, Haihua Yang, Biye Li, Cheng Cheng, Weiwei Lü, Rui Hu · (jiqizhixin) · (qbitai) · (skywork - skyworkai)
-
Aquila2 - FlagAI-Open
The official repo of Aquila2 series proposed by BAAI, including pretrained & chat large language models. · (mp.weixin.qq)
-
ColossalAI - hpcaitech
Making large AI models cheaper, faster and more accessible · (qbitai)
-
VisCPM - OpenBMB
基于CPM基础模型的中英双语多模态大模型系列 · (jiqizhixin)
-
Yi - 01-ai
A series of large language models trained from scratch by developers @01-ai
· (jiqizhixin)
-
InternLM - InternLM
InternLM has open-sourced a 7 billion parameter base model, a chat model tailored for practical scenarios and the training system. · (qbitai) · (qbitai)
· (mp.weixin.qq) · (huggingface)
-
internlm2-chat-7b - internlm 🤗
-
DeepSeek-Coder: When the Large Language Model Meets Programming -- The Rise of Code Intelligence,
arXiv, 2401.14196
, arxiv, pdf, cication: -1Daya Guo, Qihao Zhu, Dejian Yang, Zhenda Xie, Kai Dong, Wentao Zhang, Guanting Chen, Xiao Bi, Y. Wu, Y. K. Li
-
DeepSeek-MoE - deepseek-ai
· (huggingface)
-
DeepSeek-LLM - deepseek-ai
DeepSeek LLM: Let there be answers · (huggingface) · (mp.weixin.qq)
-
· (huggingface) · (jiqizhixin)
-
XVERSE-13B - xverse-ai
XVERSE-13B: A multilingual large language model developed by XVERSE Technology Inc. · (qbitai) · (huggingface)
-
Qwen - QwenLM
The official repo of Qwen (通义千问) chat & pretrained large language model proposed by Alibaba Cloud.
-
Qwen-7B - QwenLM
The official repo of Qwen-7B (通义千问-7B) chat & pretrained large language model proposed by Alibaba Cloud. · (mp.weixin.qq) · (qbitai)
-
Qwen-72B-Chat-Demo - Qwen 🤗
-
Baichuan 2: Open Large-scale Language Models,
arXiv, 2309.10305
, arxiv, pdf, cication: 16Aiyuan Yang, Bin Xiao, Bingning Wang, Borong Zhang, Ce Bian, Chao Yin, Chenxu Lv, Da Pan, Dian Wang, Dong Yan · (Baichuan2 - baichuan-inc)
· (cdn.baichuan-ai) · (mp.weixin.qq) · (jiqizhixin)
-
Baichuan-13B - baichuan-inc
A 13B large language model developed by Baichuan Intelligent Technology · (mp.weixin.qq)
-
baichuan-7B - baichuan-inc
A large-scale 7B pretraining language model developed by BaiChuan-Inc.
-
ChatGLM3 - THUDM
ChatGLM3 series: Open Bilingual Chat LLMs | 开源双语对话语言模型 · (qbitai)
-
ChatGLM2-6B - THUDM
ChatGLM2-6B: An Open Bilingual Chat LLM | 开源双语对话语言模型 · (qbitai)
-
chatglm.cpp - li-plus
C++ implementation of ChatGLM-6B & ChatGLM2-6B
-
TigerBot - TigerResearch
TigerBot: A multi-language multi-task LLM · (qbitai)
-
Aurora:Activating Chinese chat capability for Mistral-8x7B sparse Mixture-of-Experts through Instruction-Tuning,
arXiv, 2312.14557
, arxiv, pdf, cication: -1Rongsheng Wang, Haoming Chen, Ruizhe Zhou, Yaofei Duan, Kunyan Cai, Han Ma, Jiaxi Cui, Jian Li, Patrick Cheong-Iao Pang, Yapeng Wang
· (Aurora - WangRongsheng)
-
Taiwan-LLaMa - MiuLab
Traditional Mandarin LLMs for Taiwan
-
Chinese-LLaMA-Alpaca-2 - ymcui
中文LLaMA-2 & Alpaca-2大语言模型 (Chinese LLaMA-2 & Alpaca-2 LLMs)
-
TransGPT - DUOMO
· (jiqizhixin)
-
Llama2-Chinese - FlagAlpha
Llama中文社区,最好的中文Llama大模型,完全开源可商用
-
Chinese-Llama-2-7b - LinkSoul-AI
开源社区第一个能下载、能运行的中文 LLaMA2 模型!
-
ChatGLM-Efficient-Tuning - hiyouga
Fine-tuning ChatGLM-6B with PEFT | 基于 PEFT 的高效 ChatGLM 微调
-
BayLing: Bridging Cross-lingual Alignment and Instruction Following through Interactive Translation for Large Language Models,
arXiv, 2306.10968
, arxiv, pdf, cication: -1Shaolei Zhang, Qingkai Fang, Zhuocheng Zhang, Zhengrui Ma, Yan Zhou, Langlin Huang, Mengyu Bu, Shangtong Gui, Yunji Chen, Xilin Chen · (jiqizhixin) · (BayLing - ictnlp)
· (huggingface)
-
LLMs-In-China - wgwang
中国大模型
-
· (mp.weixin.qq)
-
CroissantLLM: A Truly Bilingual French-English Language Model,
arXiv, 2402.00786
, arxiv, pdf, cication: -1Manuel Faysse, Patrick Fernandes, Nuno Guerreiro, António Loison, Duarte Alves, Caio Corro, Nicolas Boizard, João Alves, Ricardo Rei, Pedro Martins
-
MaLA-500: Massive Language Adaptation of Large Language Models,
arXiv, 2401.13303
, arxiv, pdf, cication: -1Peiqin Lin, Shaoxiong Ji, Jörg Tiedemann, André F. T. Martins, Hinrich Schütze · (huggingface)
-
Multilingual Instruction Tuning With Just a Pinch of Multilinguality,
arXiv, 2401.01854
, arxiv, pdf, cication: -1Uri Shaham, Jonathan Herzig, Roee Aharoni, Idan Szpektor, Reut Tsarfaty, Matan Eyal
-
LLaMA Beyond English: An Empirical Study on Language Capability Transfer,
arXiv, 2401.01055
, arxiv, pdf, cication: -1Jun Zhao, Zhihao Zhang, Qi Zhang, Tao Gui, Xuanjing Huang
-
FinGPT: Large Generative Models for a Small Language,
arXiv, 2311.05640
, arxiv, pdf, cication: -1Risto Luukkonen, Ville Komulainen, Jouni Luoma, Anni Eskelinen, Jenna Kanerva, Hanna-Mari Kupari, Filip Ginter, Veronika Laippala, Niklas Muennighoff, Aleksandra Piktus · (turkunlp)
-
LLMZoo - FreedomIntelligence
⚡LLM Zoo is a project that provides data, models, and evaluation benchmark for large language models.⚡
-
open-llms - eugeneyan
📋 A list of open LLMs available for commercial use.
-
List of Open Sourced Fine-Tuned Large Language Models (LLM) | by Sung Kim | Medium
-
Awesome-Chinese-LLM - HqWu-HITCS
整理开源的中文大语言模型,以规模较小、可私有化部署、训练成本较低的模型为主,包括底座模型,垂直领域微调及应用,数据集与教程等。
-
self-llm - datawhalechina
《开源大模型食用指南》基于AutoDL快速部署开源大模型,更适合中国宝宝的部署教程