Skip to content

Commit

Permalink
Apply automatic changes
Browse files Browse the repository at this point in the history
  • Loading branch information
mdingemanse authored and github-actions[bot] committed Apr 19, 2024
1 parent 0983289 commit 4a243ef
Show file tree
Hide file tree
Showing 3 changed files with 51 additions and 51 deletions.
2 changes: 1 addition & 1 deletion docs/df.csv
Original file line number Diff line number Diff line change
Expand Up @@ -15,7 +15,7 @@ https://huggingface.co/lmsys/vicuna-13b-v1.3,Vicuna is a chat assistant trained
hhttps://github.com/ethanyanjiali/minChatGPT,,GPT2,anthropic,GNU General Public License v3.0,ethanyanjiali,https://github.com/ethanyanjiali/minChatGPT,,open,,,open,,,open,,,partial,,,closed,,,open,,,open,,,partial,,,closed,,,closed,,,closed,,,closed,,,closed,,,open,,,/projects/minChatGPT.yaml,7.0
https://github.com/Cerebras,,,Alpaca (synthetic),Apache 2.0,Cerebras + Schramm,https://github.com/Cerebras,,open,,,open,,,open,,,open,,,closed,,,open,,,closed,,,open,,,partial,https://arxiv.org/abs/2304.03208,,closed,,,closed,,,closed,,,closed,,,closed,,,/projects/Cerebras-GPT-111m.yaml,6.5
https://github.com/BlinkDL/ChatRWKV,,RWKV-LM,"alpaca, shareGPT (synthetic)",,BlinkDL/RWKV,https://www.rwkv.com/,,open,https://github.com/BlinkDL/ChatRWKV,Various community-contributed enhancements available,partial,https://pile.eleuther.ai/,Trained on The Pile. Recent versions also build on Red Pajama (https://huggingface.co/datasets/togethercomputer/RedPajama-Data-1T),open,https://huggingface.co/BlinkDL/rwkv-4-world/tree/main,Model weights released across different HuggingFace spaces,closed,,"Instruction tuning data not separately available. Documentation 'These are RWKV-4-Pile 1.5/3/7/14B models finetuned on Alpaca, CodeAlpaca, Guanaco, GPT4All, ShareGPT and more'",closed,,Weights not separately available.,open,https://github.com/BlinkDL/ChatRWKV/blob/main/LICENSE,Apache 2.0,partial,,Code documentation scattered across github repo and HuggingFace spaces,partial,,Architecture described in preprint (LM part) but not all details clearly documented.,partial,https://arxiv.org/abs/2305.13048,"Preprint covers only LLM (RNN based), not instruction fine-tuning, so partial.",closed,,No peer-reviewed paper or published data audit known,closed,https://huggingface.co/BlinkDL/rwkv-4-raven,"No modelcard, HuggingFace spaces only used to share files",closed,https://huggingface.co/BlinkDL/rwkv-4-raven,"No data sheet, HuggingFac spaces only used to share files",open,https://pypi.org/project/rwkv/,Available through pip install rwkv,partial,,API via HuggingFace,/projects/ChatRWKV.yaml,6.5
https://github.com/LianjiaTech/BELLE,,LLaMA & BLOOMZ,"alpaca, shareGPT, Belle (synthetic)",Apache License 2.0,KE Technologies,http://www.ke.com,,open,,,partial,,"Open for variants based on BLOOMZ. Closed for variants based on LLaMA, whose pretraining data is nowhere disclosed or documented.",partial,,LLaMA based but copyright status unclear,partial,https://github.com/LianjiaTech/BELLE/tree/main/data/1.5M,Synthetic BELLE training data in Chinese released in batches,partial,https://github.com/LianjiaTech/BELLE/tree/main/models,"Some models available, most only as delta weights requiring separate access to LLaMA",closed,,Lowest common denominator is non-OSI approved LLaMA licence agreement,partial,https://github.com/LianjiaTech/BELLE/blob/main/README_en.md,"Quite some documentation on Github, though not all well-organized",open,https://github.com/LianjiaTech/BELLE/blob/main/README_en.md,Specified in a fair bit of detail on github,open,https://arxiv.org/abs/2303.14742,,closed,,No peer-reviewed paper found,closed,,No model card found,partial,,No data sheet found,closed,,No dedicated package available,closed,,No API found,/projects/BELLE.yaml,6.0
https://github.com/LianjiaTech/BELLE,,LLaMA & BLOOMZ,"alpaca, shareGPT, Belle (synthetic)",Apache License 2.0,KE Technologies,http://www.ke.com,,open,https://github.com/LianjiaTech/BELLE,Repository contains a fair bit of code,partial,,"Open for variants based on BLOOMZ. Closed for variants based on LLaMA, whose pretraining data is nowhere disclosed or documented.",partial,,LLaMA based but copyright status unclear,partial,https://github.com/LianjiaTech/BELLE/tree/main/data/1.5M,Synthetic BELLE training data in Chinese released in batches,partial,https://github.com/LianjiaTech/BELLE/tree/main/models,"Some models available, most only as delta weights requiring separate access to LLaMA",closed,,Lowest common denominator is non-OSI approved LLaMA licence agreement,partial,https://github.com/LianjiaTech/BELLE/blob/main/README_en.md,"Quite some documentation on Github, though not all well-organized",open,https://github.com/LianjiaTech/BELLE/blob/main/README_en.md,Specified in a fair bit of detail on github,open,https://arxiv.org/abs/2303.14742,,closed,,No peer-reviewed paper found,closed,,No model card found,partial,,No data sheet found,closed,,No dedicated package available,closed,,No API found,/projects/BELLE.yaml,6.0
https://huggingface.co/WizardLM/WizardLM-13B-V1.2,Empowering Large Pre-Trained Language Models to Follow Complex Instructions,LLaMA2-13B,Evol-Instruct (synthetic),CC-BY-NC-4.0,Microsoft & Peking University,https://github.com/nlpxucan,,partial,https://github.com/nlpxucan/WizardLM/tree/main/WizardLM,Fast-evolving repository contains WizardLM code,closed,https://github.com/opening-up-chatgpt/opening-up-chatgpt.github.io/blob/main/projects/llama-2-chat.yaml,"Based on LLaMA2, which is claimed to be public but nowhere exactly documented.",partial,https://ai.meta.com/resources/models-and-libraries/llama-downloads/,"Based on LLaMA2 weights, which are made conditionally available by Meta.",open,https://huggingface.co/datasets/WizardLM/WizardLM_evol_instruct_V2_196k,The Evol-Instruct V2 dataset contains 196k instruction-following sequences generated from Evol-Instruct,open,https://huggingface.co/WizardLM/WizardLM-13B-V1.2,Model weights offered in HuggingFace repository,partial,https://github.com/nlpxucan/WizardLM/blob/main/WizardLM/MODEL_DIFF_LICENSE,"Restricted for academic research purposes only. Code and Model diff release under CC-BY-NC-4.0, software code under Apache 2.0",partial,https://github.com/nlpxucan/WizardLM/tree/main/WizardLM,"Code is only partially documented, not clearly versioned, and appears to be in flux.",open,https://arxiv.org/abs/2304.12244,Architecture described in preprint and partly accessible in code repository,open,https://arxiv.org/abs/2304.12244,Preprint describes method for creating large amounts of LLM-based synthetic RLHF data and fine-tuning WizardLM based on it,closed,,No peer-reviewed paper or data audit found,closed,https://huggingface.co/WizardLM/WizardLM-13B-V1.2,Model card is only a placeholder and generates an error (missing yaml metadata),closed,https://huggingface.co/datasets/WizardLM/WizardLM_evol_instruct_V2_196k,Dataset card for Evol-Instruct generates an error,closed,,No package available,closed,,No API available,/projects/wizardlm-13B.yaml,6.0
https://huggingface.co/jondurbin/airoboros-l2-70b-gpt4-1.4.1,,Llama2,Airoboros (synthetic),Purposely left ambiguous,Jon Durbin,https://github.com/jondurbin,Only active on GitHub since May 2023,partial,https://gist.github.com/jondurbin/87fc040b92a3073125ed516b04bc6e19,Repo exists for RL data but only a gist exists for model training and architecture,closed,,Llama2 training data is nowhere documented or disclosed,partial,,"Llama2, made conditionally available by Meta",open,https://github.com/jondurbin/airoboros,"Airoboros, an implementation of the Self-Instruct paper",open,https://huggingface.co/jondurbin/airoboros-l2-70b-gpt4-1.4.1/tree/main,Made available through HuggingFace,partial,https://huggingface.co/jondurbin/airoboros-l2-70b-gpt4-1.4.1#licence-and-usage-restrictions,Licensing left ambiguous because of murky status of OpenAI-derived Self-Instruct data,partial,,What little code available is not very systematically documented,partial,https://huggingface.co/jondurbin/airoboros-l2-70b-gpt4-1.4.1/discussions/2#64c29e4c617b36543dedac9a,Some info can be gleaned at link but most remains undocumented,closed,,No preprint found,closed,,No peer-reviewed paper found,partial,https://huggingface.co/jondurbin/airoboros-65b-gpt4-1.4,Instructs reader to look up model card for prior 65B Llama1 version,partial,https://huggingface.co/datasets/jondurbin/airoboros-gpt4-1.4.1,Datasheet for RL data only,closed,,No package found,closed,,No API found,/projects/airoboros.yaml,5.5
https://github.com/THUDM/ChatGLM-6B/blob/main/README_en.md,"From the readme, ""ChatGLM-6B uses technology similar to ChatGPT, optimized for Chinese QA and dialogue. The model is trained for about 1T tokens of Chinese and English corpus, supplemented by supervised fine-tuning, feedback bootstrap, and reinforcement learning wit human feedback. With only about 6.2 billion parameters, the model is able to generate answers that are in line with human preference.""",GLM (own),Unspecified,Apache 2.0,THUDM,https://github.com/THUDM,Knowledge Engineering Group (KEG) & Data Mining at Tsinghua University,partial,https://github.com/THUDM/ChatGLM-6B/blob/main/README_en.md#deployment,Some code made available on Github,partial,http://doi.org/10.18653/v1/2022.acl-long.26,"Training data not centrally made available, but described in 2022 ACL paper, appears to be mostly public datasets",open,https://huggingface.co/THUDM/chatglm-6b/tree/main,Model made available through HuggingFace,closed,,"docs mention ""supervised fine-tuning, feedback bootstrap, and reinforcement learning wit human feedback"", but none of the datasets used are clearly specified.",closed,,No weights or checkpoints corresponding to the delta of the LLM vs RLHF provided,open,https://github.com/THUDM/ChatGLM-6B/blob/main/LICENSE,Apache 2.0,partial,https://github.com/THUDM/ChatGLM-6B/blob/main/ptuning/README_en.md,"Some documentation available, but a lot of code is not commented or explained.",partial,,Full details architecture not specified in a single place,closed,,,partial,https://aclanthology.org/2022.acl-long.26/,"ACL 2022 paper describes the training of the GLM base model, but the RLHF portion is more recent (there is also a related ICLR paper for a newer generation https://openreview.net/forum?id=-Aw0rrrPUF)",closed,https://huggingface.co/THUDM/chatglm-6b,No modelcard; the HuggingFace modelcard spot is used just as the homepage for the model.,closed,,No datasheet,closed,,No package,open,https://github.com/THUDM/ChatGLM-6B/blob/main/README_en.md#api-deployment,API provided through fastapi uvicorn,/projects/ChatGLM-6B.yaml,5.5
Expand Down
Loading

0 comments on commit 4a243ef

Please sign in to comment.