mirror of
https://github.com/bentoml/OpenLLM.git
synced 2026-04-22 16:07:24 -04:00
chore: add NousResearch's as non-gated Llama (#152)
This commit is contained in:
@@ -0,0 +1 @@
|
||||
# You can use this file with 'git config blame.ignoreRevsFile .git-blame-ignore-revs'
|
||||
|
||||
@@ -153,6 +153,11 @@ To update the snapshot, do the following:
|
||||
hatch run tests:snapshot-models
|
||||
```
|
||||
|
||||
## Working with Git
|
||||
|
||||
To filter out most of the generated commits for infrastructure, use ``--invert-grep`` in conjunction with ``--grep``
|
||||
to filter out all commits with regex `"[generated]"`
|
||||
|
||||
## Releasing a New Version
|
||||
|
||||
To release a new version, use `./tools/run-release-action`. It requires `gh`,
|
||||
|
||||
18
README.md
18
README.md
@@ -254,12 +254,18 @@ pip install openllm
|
||||
<td><a href=https://huggingface.co/docs/transformers/main/model_doc/llama#transformers.LlamaForCausalLM><code>LlamaForCausalLM</code></a></td>
|
||||
<td>
|
||||
|
||||
<ul><li><a href=https://huggingface.co/meta-llama/llama-2-70b-chat-hf><code>meta-llama/llama-2-70b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/llama-2-13b-chat-hf><code>meta-llama/llama-2-13b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/llama-2-7b-chat-hf><code>meta-llama/llama-2-7b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/llama-2-70b-hf><code>meta-llama/llama-2-70b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/llama-2-13b-hf><code>meta-llama/llama-2-13b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/llama-2-7b-hf><code>meta-llama/llama-2-7b-hf</code></a></li>
|
||||
<ul><li><a href=https://huggingface.co/meta-llama/Llama-2-70b-chat-hf><code>meta-llama/Llama-2-70b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/Llama-2-13b-chat-hf><code>meta-llama/Llama-2-13b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/Llama-2-7b-chat-hf><code>meta-llama/Llama-2-7b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/Llama-2-70b-hf><code>meta-llama/Llama-2-70b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/Llama-2-13b-hf><code>meta-llama/Llama-2-13b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/meta-llama/Llama-2-7b-hf><code>meta-llama/Llama-2-7b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-70b-chat-hf><code>NousResearch/llama-2-70b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-13b-chat-hf><code>NousResearch/llama-2-13b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-7b-chat-hf><code>NousResearch/llama-2-7b-chat-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-70b-hf><code>NousResearch/llama-2-70b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-13b-hf><code>NousResearch/llama-2-13b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/NousResearch/llama-2-7b-hf><code>NousResearch/llama-2-7b-hf</code></a></li>
|
||||
<li><a href=https://huggingface.co/openlm-research/open_llama_7b_v2><code>openlm-research/open_llama_7b_v2</code></a></li>
|
||||
<li><a href=https://huggingface.co/openlm-research/open_llama_3b_v2><code>openlm-research/open_llama_3b_v2</code></a></li>
|
||||
<li><a href=https://huggingface.co/openlm-research/open_llama_13b><code>openlm-research/open_llama_13b</code></a></li>
|
||||
|
||||
@@ -36,12 +36,18 @@ class LlaMAConfig(openllm.LLMConfig):
|
||||
"architecture": "LlamaForCausalLM",
|
||||
"requirements": ["fairscale", "sentencepiece"],
|
||||
"model_ids": [
|
||||
"meta-llama/llama-2-70b-chat-hf",
|
||||
"meta-llama/llama-2-13b-chat-hf",
|
||||
"meta-llama/llama-2-7b-chat-hf",
|
||||
"meta-llama/llama-2-70b-hf",
|
||||
"meta-llama/llama-2-13b-hf",
|
||||
"meta-llama/llama-2-7b-hf",
|
||||
"meta-llama/Llama-2-70b-chat-hf",
|
||||
"meta-llama/Llama-2-13b-chat-hf",
|
||||
"meta-llama/Llama-2-7b-chat-hf",
|
||||
"meta-llama/Llama-2-70b-hf",
|
||||
"meta-llama/Llama-2-13b-hf",
|
||||
"meta-llama/Llama-2-7b-hf",
|
||||
"NousResearch/llama-2-70b-chat-hf",
|
||||
"NousResearch/llama-2-13b-chat-hf",
|
||||
"NousResearch/llama-2-7b-chat-hf",
|
||||
"NousResearch/llama-2-70b-hf",
|
||||
"NousResearch/llama-2-13b-hf",
|
||||
"NousResearch/llama-2-7b-hf",
|
||||
"openlm-research/open_llama_7b_v2",
|
||||
"openlm-research/open_llama_3b_v2",
|
||||
"openlm-research/open_llama_13b",
|
||||
|
||||
Reference in New Issue
Block a user