diff --git a/README.md b/README.md
index af75783e..bb4b5178 100644
--- a/README.md
+++ b/README.md
@@ -261,21 +261,13 @@ pip install "openllm[llama]"
meta-llama/llama-2-70b-hf
meta-llama/llama-2-13b-hf
meta-llama/llama-2-7b-hf
-decapoda-research/llama-65b-hf
-decapoda-research/llama-30b-hf
-decapoda-research/llama-13b-hf
-decapoda-research/llama-7b-hf-int8
-decapoda-research/llama-7b-hf
openlm-research/open_llama_7b_v2
openlm-research/open_llama_3b_v2
openlm-research/open_llama_13b
-openlm-research/open_llama_7b
-openlm-research/open_llama_3b
huggyllama/llama-65b
huggyllama/llama-30b
huggyllama/llama-13b
-huggyllama/llama-7b
-syzymon/long_llama_3b
+huggyllama/llama-7b
diff --git a/src/openllm/models/llama/configuration_llama.py b/src/openllm/models/llama/configuration_llama.py
index d0b01e62..fcd98415 100644
--- a/src/openllm/models/llama/configuration_llama.py
+++ b/src/openllm/models/llama/configuration_llama.py
@@ -45,21 +45,13 @@ class LlaMAConfig(openllm.LLMConfig):
"meta-llama/llama-2-70b-hf",
"meta-llama/llama-2-13b-hf",
"meta-llama/llama-2-7b-hf",
- "decapoda-research/llama-65b-hf",
- "decapoda-research/llama-30b-hf",
- "decapoda-research/llama-13b-hf",
- "decapoda-research/llama-7b-hf-int8",
- "decapoda-research/llama-7b-hf",
"openlm-research/open_llama_7b_v2",
"openlm-research/open_llama_3b_v2",
"openlm-research/open_llama_13b",
- "openlm-research/open_llama_7b",
- "openlm-research/open_llama_3b",
"huggyllama/llama-65b",
"huggyllama/llama-30b",
"huggyllama/llama-13b",
"huggyllama/llama-7b",
- "syzymon/long_llama_3b", # NOTE: use ``openllm.LongLLaMA`` to load this variant. Otherwise it will be limited to context length of 2048
],
"tokenizer_class": "LlamaTokenizerFast",
}