mirror of
https://github.com/hiyouga/LLaMA-Factory.git
synced 2025-10-16 00:28:10 +08:00
update gc kwargs
Former-commit-id: 0cb81c156bc8c21a4bbdd3289a491f78dfcaf730
This commit is contained in:
parent
b564b97b7e
commit
f67f781fed
@ -227,7 +227,9 @@ def _prepare_model_for_training(
|
|||||||
if not getattr(model, "supports_gradient_checkpointing", False):
|
if not getattr(model, "supports_gradient_checkpointing", False):
|
||||||
logger.warning("Current model does not support gradient checkpointing.")
|
logger.warning("Current model does not support gradient checkpointing.")
|
||||||
else:
|
else:
|
||||||
model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": False})
|
# use_reentrant=False might increase VRAM usage (have not been empirically verified yet)
|
||||||
|
# According to: https://github.com/huggingface/transformers/issues/28339
|
||||||
|
model.gradient_checkpointing_enable(gradient_checkpointing_kwargs={"use_reentrant": True})
|
||||||
model.enable_input_require_grads()
|
model.enable_input_require_grads()
|
||||||
model.config.use_cache = False # turn off when gradient checkpointing is enabled
|
model.config.use_cache = False # turn off when gradient checkpointing is enabled
|
||||||
logger.info("Gradient checkpointing enabled.")
|
logger.info("Gradient checkpointing enabled.")
|
||||||
|
Loading…
x
Reference in New Issue
Block a user