mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-08-02 19:52:50 +08:00
reduce repetition penalty
Former-commit-id: 926291940de4b59a40489e6a509fdc0135c8616d
This commit is contained in:
parent
4200d5d558
commit
3bf4b20d0b
@ -53,7 +53,7 @@ def main():
|
|||||||
"temperature": 0.95,
|
"temperature": 0.95,
|
||||||
"num_beams": 1,
|
"num_beams": 1,
|
||||||
"max_new_tokens": 256,
|
"max_new_tokens": 256,
|
||||||
"repetition_penalty": 1.5,
|
"repetition_penalty": 1.0,
|
||||||
"logits_processor": get_logits_processor()
|
"logits_processor": get_logits_processor()
|
||||||
}
|
}
|
||||||
with torch.no_grad():
|
with torch.no_grad():
|
||||||
|
@ -105,7 +105,7 @@ def predict(input, chatbot, max_length, top_p, temperature, history):
|
|||||||
"temperature": temperature,
|
"temperature": temperature,
|
||||||
"num_beams": 1,
|
"num_beams": 1,
|
||||||
"max_length": max_length,
|
"max_length": max_length,
|
||||||
"repetition_penalty": 1.5,
|
"repetition_penalty": 1.0,
|
||||||
"logits_processor": get_logits_processor()
|
"logits_processor": get_logits_processor()
|
||||||
}
|
}
|
||||||
with torch.no_grad():
|
with torch.no_grad():
|
||||||
|
Loading…
x
Reference in New Issue
Block a user