mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-15 08:08:09 +08:00
32 lines
1.1 KiB
Python
32 lines
1.1 KiB
Python
IGNORE_INDEX = -100
|
|
|
|
VALUE_HEAD_FILE_NAME = "value_head.bin"
|
|
|
|
FINETUNING_ARGS_NAME = "finetuning_args.json"
|
|
|
|
LAYERNORM_NAMES = ["norm", "ln_f", "ln_attn", "ln_mlp"] # for LLaMA, BLOOM and Falcon settings
|
|
|
|
METHODS = ["full", "freeze", "lora"]
|
|
|
|
SUPPORTED_MODELS = {
|
|
"LLaMA-7B": "huggyllama/llama-7b",
|
|
"LLaMA-13B": "huggyllama/llama-13b",
|
|
"LLaMA-30B": "huggyllama/llama-30b",
|
|
"LLaMA-65B": "huggyllama/llama-65b",
|
|
"BLOOM-560M": "bigscience/bloom-560m",
|
|
"BLOOM-3B": "bigscience/bloom-3b",
|
|
"BLOOM-7B1": "bigscience/bloom-7b1",
|
|
"BLOOMZ-560M": "bigscience/bloomz-560m",
|
|
"BLOOMZ-3B": "bigscience/bloomz-3b",
|
|
"BLOOMZ-7B1-mt": "bigscience/bloomz-7b1-mt",
|
|
"Falcon-7B-Base": "tiiuae/falcon-7b",
|
|
"Falcon-7B-Chat": "tiiuae/falcon-7b-instruct",
|
|
"Falcon-40B-Base": "tiiuae/falcon-40b",
|
|
"Falcon-40B-Chat": "tiiuae/falcon-40b-instruct",
|
|
"Baichuan-7B": "baichuan-inc/Baichuan-7B",
|
|
"Baichuan-13B-Base": "baichuan-inc/Baichuan-13B-Base",
|
|
"Baichuan-13B-Chat": "baichuan-inc/Baichuan-13B-Chat",
|
|
"InternLM-7B-Base": "internlm/internlm-7b",
|
|
"InternLM-7B-Chat": "internlm/internlm-chat-7b"
|
|
}
|