mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 08:38:09 +08:00
221 lines
13 KiB
Python
221 lines
13 KiB
Python
LOCALES = {
|
||
"lang": {"en": {"label": "Lang"}, "zh": {"label": "语言"}},
|
||
"model_name": {"en": {"label": "Model name"}, "zh": {"label": "模型名称"}},
|
||
"model_path": {
|
||
"en": {"label": "Model path", "info": "Path to pretrained model or model identifier from Hugging Face."},
|
||
"zh": {"label": "模型路径", "info": "本地模型的文件路径或 Hugging Face 的模型标识符。"},
|
||
},
|
||
"finetuning_type": {"en": {"label": "Finetuning method"}, "zh": {"label": "微调方法"}},
|
||
"adapter_path": {"en": {"label": "Adapter path"}, "zh": {"label": "适配器路径"}},
|
||
"refresh_btn": {"en": {"value": "Refresh adapters"}, "zh": {"value": "刷新适配器"}},
|
||
"advanced_tab": {"en": {"label": "Advanced configurations"}, "zh": {"label": "高级设置"}},
|
||
"quantization_bit": {
|
||
"en": {"label": "Quantization bit", "info": "Enable 4/8-bit model quantization (QLoRA)."},
|
||
"zh": {"label": "量化等级", "info": "启用 4/8 比特模型量化(QLoRA)。"},
|
||
},
|
||
"template": {
|
||
"en": {"label": "Prompt template", "info": "The template used in constructing prompts."},
|
||
"zh": {"label": "提示模板", "info": "构建提示词时使用的模板"},
|
||
},
|
||
"rope_scaling": {"en": {"label": "RoPE scaling"}, "zh": {"label": "RoPE 插值方法"}},
|
||
"booster": {"en": {"label": "Booster"}, "zh": {"label": "加速方式"}},
|
||
"training_stage": {
|
||
"en": {"label": "Stage", "info": "The stage to perform in training."},
|
||
"zh": {"label": "训练阶段", "info": "目前采用的训练方式。"},
|
||
},
|
||
"dataset_dir": {
|
||
"en": {"label": "Data dir", "info": "Path to the data directory."},
|
||
"zh": {"label": "数据路径", "info": "数据文件夹的路径。"},
|
||
},
|
||
"dataset": {"en": {"label": "Dataset"}, "zh": {"label": "数据集"}},
|
||
"data_preview_btn": {"en": {"value": "Preview dataset"}, "zh": {"value": "预览数据集"}},
|
||
"preview_count": {"en": {"label": "Count"}, "zh": {"label": "数量"}},
|
||
"page_index": {"en": {"label": "Page"}, "zh": {"label": "页数"}},
|
||
"prev_btn": {"en": {"value": "Prev"}, "zh": {"value": "上一页"}},
|
||
"next_btn": {"en": {"value": "Next"}, "zh": {"value": "下一页"}},
|
||
"close_btn": {"en": {"value": "Close"}, "zh": {"value": "关闭"}},
|
||
"preview_samples": {"en": {"label": "Samples"}, "zh": {"label": "样例"}},
|
||
"cutoff_len": {
|
||
"en": {"label": "Cutoff length", "info": "Max tokens in input sequence."},
|
||
"zh": {"label": "截断长度", "info": "输入序列分词后的最大长度。"},
|
||
},
|
||
"learning_rate": {
|
||
"en": {"label": "Learning rate", "info": "Initial learning rate for AdamW."},
|
||
"zh": {"label": "学习率", "info": "AdamW 优化器的初始学习率。"},
|
||
},
|
||
"num_train_epochs": {
|
||
"en": {"label": "Epochs", "info": "Total number of training epochs to perform."},
|
||
"zh": {"label": "训练轮数", "info": "需要执行的训练总轮数。"},
|
||
},
|
||
"max_samples": {
|
||
"en": {"label": "Max samples", "info": "Maximum samples per dataset."},
|
||
"zh": {"label": "最大样本数", "info": "每个数据集最多使用的样本数。"},
|
||
},
|
||
"compute_type": {
|
||
"en": {"label": "Compute type", "info": "Whether to use fp16 or bf16 mixed precision training."},
|
||
"zh": {"label": "计算类型", "info": "是否启用 FP16 或 BF16 混合精度训练。"},
|
||
},
|
||
"batch_size": {
|
||
"en": {"label": "Batch size", "info": "Number of samples to process per GPU."},
|
||
"zh": {"label": "批处理大小", "info": "每块 GPU 上处理的样本数量。"},
|
||
},
|
||
"gradient_accumulation_steps": {
|
||
"en": {"label": "Gradient accumulation", "info": "Number of gradient accumulation steps."},
|
||
"zh": {"label": "梯度累积", "info": "梯度累积的步数。"},
|
||
},
|
||
"lr_scheduler_type": {
|
||
"en": {
|
||
"label": "LR Scheduler",
|
||
"info": "Name of learning rate scheduler.",
|
||
},
|
||
"zh": {"label": "学习率调节器", "info": "采用的学习率调节器名称。"},
|
||
},
|
||
"max_grad_norm": {
|
||
"en": {"label": "Maximum gradient norm", "info": "Norm for gradient clipping.."},
|
||
"zh": {"label": "最大梯度范数", "info": "用于梯度裁剪的范数。"},
|
||
},
|
||
"val_size": {
|
||
"en": {"label": "Val size", "info": "Proportion of data in the dev set."},
|
||
"zh": {"label": "验证集比例", "info": "验证集占全部样本的百分比。"},
|
||
},
|
||
"extra_tab": {"en": {"label": "Extra configurations"}, "zh": {"label": "其它参数设置"}},
|
||
"logging_steps": {
|
||
"en": {"label": "Logging steps", "info": "Number of steps between two logs."},
|
||
"zh": {"label": "日志间隔", "info": "每两次日志输出间的更新步数。"},
|
||
},
|
||
"save_steps": {
|
||
"en": {"label": "Save steps", "info": "Number of steps between two checkpoints."},
|
||
"zh": {"label": "保存间隔", "info": "每两次断点保存间的更新步数。"},
|
||
},
|
||
"warmup_steps": {
|
||
"en": {"label": "Warmup steps", "info": "Number of steps used for warmup."},
|
||
"zh": {"label": "预热步数", "info": "学习率预热采用的步数。"},
|
||
},
|
||
"neftune_alpha": {
|
||
"en": {"label": "NEFTune Alpha", "info": "Magnitude of noise adding to embedding vectors."},
|
||
"zh": {"label": "NEFTune 噪声参数", "info": "嵌入向量所添加的噪声大小。"},
|
||
},
|
||
"sft_packing": {
|
||
"en": {
|
||
"label": "Pack sequences",
|
||
"info": "Pack sequences into samples of fixed length in supervised fine-tuning.",
|
||
},
|
||
"zh": {"label": "序列打包", "info": "在有监督微调阶段将序列打包为相同长度的样本。"},
|
||
},
|
||
"upcast_layernorm": {
|
||
"en": {"label": "Upcast LayerNorm", "info": "Upcast weights of layernorm in float32."},
|
||
"zh": {"label": "缩放归一化层", "info": "将归一化层权重缩放至 32 位精度。"},
|
||
},
|
||
"lora_tab": {"en": {"label": "LoRA configurations"}, "zh": {"label": "LoRA 参数设置"}},
|
||
"lora_rank": {
|
||
"en": {"label": "LoRA rank", "info": "The rank of LoRA matrices."},
|
||
"zh": {"label": "LoRA 秩", "info": "LoRA 矩阵的秩。"},
|
||
},
|
||
"lora_dropout": {
|
||
"en": {"label": "LoRA Dropout", "info": "Dropout ratio of LoRA weights."},
|
||
"zh": {"label": "LoRA 随机丢弃", "info": "LoRA 权重随机丢弃的概率。"},
|
||
},
|
||
"lora_target": {
|
||
"en": {
|
||
"label": "LoRA modules (optional)",
|
||
"info": "Name(s) of target modules to apply LoRA. Use commas to separate multiple modules.",
|
||
},
|
||
"zh": {"label": "LoRA 作用模块(非必填)", "info": "应用 LoRA 的目标模块名称。使用英文逗号分隔多个名称。"},
|
||
},
|
||
"additional_target": {
|
||
"en": {
|
||
"label": "Additional modules (optional)",
|
||
"info": "Name(s) of modules apart from LoRA layers to be set as trainable. Use commas to separate multiple modules.",
|
||
},
|
||
"zh": {"label": "附加模块(非必填)", "info": "除 LoRA 层以外的可训练模块名称。使用英文逗号分隔多个名称。"},
|
||
},
|
||
"create_new_adapter": {
|
||
"en": {
|
||
"label": "Create new adapter",
|
||
"info": "Whether to create a new adapter with randomly initialized weight or not.",
|
||
},
|
||
"zh": {"label": "新建适配器", "info": "是否创建一个经过随机初始化的新适配器。"},
|
||
},
|
||
"rlhf_tab": {"en": {"label": "RLHF configurations"}, "zh": {"label": "RLHF 参数设置"}},
|
||
"dpo_beta": {
|
||
"en": {"label": "DPO beta", "info": "Value of the beta parameter in the DPO loss."},
|
||
"zh": {"label": "DPO beta 参数", "info": "DPO 损失函数中 beta 超参数大小。"},
|
||
},
|
||
"dpo_ftx": {
|
||
"en": {"label": "DPO-ftx weight", "info": "The weight of SFT loss in the DPO-ftx."},
|
||
"zh": {"label": "DPO-ftx 权重", "info": "DPO-ftx 中 SFT 损失的权重大小。"},
|
||
},
|
||
"reward_model": {
|
||
"en": {
|
||
"label": "Reward model",
|
||
"info": "Adapter of the reward model for PPO training. (Needs to refresh adapters)",
|
||
},
|
||
"zh": {"label": "奖励模型", "info": "PPO 训练中奖励模型的适配器路径。(需要刷新适配器)"},
|
||
},
|
||
"cmd_preview_btn": {"en": {"value": "Preview command"}, "zh": {"value": "预览命令"}},
|
||
"start_btn": {"en": {"value": "Start"}, "zh": {"value": "开始"}},
|
||
"stop_btn": {"en": {"value": "Abort"}, "zh": {"value": "中断"}},
|
||
"output_dir": {
|
||
"en": {"label": "Output dir", "info": "Directory for saving results."},
|
||
"zh": {"label": "输出目录", "info": "保存结果的路径。"},
|
||
},
|
||
"output_box": {"en": {"value": "Ready."}, "zh": {"value": "准备就绪。"}},
|
||
"loss_viewer": {"en": {"label": "Loss"}, "zh": {"label": "损失"}},
|
||
"predict": {"en": {"label": "Save predictions"}, "zh": {"label": "保存预测结果"}},
|
||
"load_btn": {"en": {"value": "Load model"}, "zh": {"value": "加载模型"}},
|
||
"unload_btn": {"en": {"value": "Unload model"}, "zh": {"value": "卸载模型"}},
|
||
"info_box": {"en": {"value": "Model unloaded, please load a model first."}, "zh": {"value": "模型未加载,请先加载模型。"}},
|
||
"system": {"en": {"placeholder": "System prompt (optional)"}, "zh": {"placeholder": "系统提示词(非必填)"}},
|
||
"tools": {"en": {"placeholder": "Tools (optional)"}, "zh": {"placeholder": "工具列表(非必填)"}},
|
||
"query": {"en": {"placeholder": "Input..."}, "zh": {"placeholder": "输入..."}},
|
||
"submit_btn": {"en": {"value": "Submit"}, "zh": {"value": "提交"}},
|
||
"clear_btn": {"en": {"value": "Clear history"}, "zh": {"value": "清空历史"}},
|
||
"max_length": {"en": {"label": "Maximum length"}, "zh": {"label": "最大长度"}},
|
||
"max_new_tokens": {"en": {"label": "Maximum new tokens"}, "zh": {"label": "最大生成长度"}},
|
||
"top_p": {"en": {"label": "Top-p"}, "zh": {"label": "Top-p 采样值"}},
|
||
"temperature": {"en": {"label": "Temperature"}, "zh": {"label": "温度系数"}},
|
||
"max_shard_size": {
|
||
"en": {"label": "Max shard size (GB)", "info": "The maximum size for a model file."},
|
||
"zh": {"label": "最大分块大小(GB)", "info": "单个模型文件的最大大小。"},
|
||
},
|
||
"export_quantization_bit": {
|
||
"en": {"label": "Export quantization bit.", "info": "Quantizing the exported model."},
|
||
"zh": {"label": "导出量化等级", "info": "量化导出模型。"},
|
||
},
|
||
"export_quantization_dataset": {
|
||
"en": {"label": "Export quantization dataset.", "info": "The calibration dataset used for quantization."},
|
||
"zh": {"label": "导出量化数据集", "info": "量化过程中使用的校准数据集。"},
|
||
},
|
||
"export_dir": {
|
||
"en": {"label": "Export dir", "info": "Directory to save exported model."},
|
||
"zh": {"label": "导出目录", "info": "保存导出模型的文件夹路径。"},
|
||
},
|
||
"export_btn": {"en": {"value": "Export"}, "zh": {"value": "开始导出"}},
|
||
}
|
||
|
||
|
||
ALERTS = {
|
||
"err_conflict": {"en": "A process is in running, please abort it firstly.", "zh": "任务已存在,请先中断训练。"},
|
||
"err_exists": {"en": "You have loaded a model, please unload it first.", "zh": "模型已存在,请先卸载模型。"},
|
||
"err_no_model": {"en": "Please select a model.", "zh": "请选择模型。"},
|
||
"err_no_path": {"en": "Model not found.", "zh": "模型未找到。"},
|
||
"err_no_dataset": {"en": "Please choose a dataset.", "zh": "请选择数据集。"},
|
||
"err_no_adapter": {"en": "Please select an adapter.", "zh": "请选择一个适配器。"},
|
||
"err_no_export_dir": {"en": "Please provide export dir.", "zh": "请填写导出目录"},
|
||
"err_failed": {"en": "Failed.", "zh": "训练出错。"},
|
||
"err_demo": {
|
||
"en": "Training is unavailable in demo mode, duplicate the space to a private one first.",
|
||
"zh": "展示模式不支持训练,请先复制到私人空间。",
|
||
},
|
||
"err_device_count": {"en": "Multiple GPUs are not supported yet.", "zh": "尚不支持多 GPU 训练。"},
|
||
"info_aborting": {"en": "Aborted, wait for terminating...", "zh": "训练中断,正在等待线程结束……"},
|
||
"info_aborted": {"en": "Ready.", "zh": "准备就绪。"},
|
||
"info_finished": {"en": "Finished.", "zh": "训练完毕。"},
|
||
"info_loading": {"en": "Loading model...", "zh": "加载中……"},
|
||
"info_unloading": {"en": "Unloading model...", "zh": "卸载中……"},
|
||
"info_loaded": {"en": "Model loaded, now you can chat with your model!", "zh": "模型已加载,可以开始聊天了!"},
|
||
"info_unloaded": {"en": "Model unloaded.", "zh": "模型已卸载。"},
|
||
"info_exporting": {"en": "Exporting model...", "zh": "正在导出模型……"},
|
||
"info_exported": {"en": "Model exported.", "zh": "模型导出完成。"},
|
||
}
|