mirror of
				https://github.com/hiyouga/LLaMA-Factory.git
				synced 2025-11-04 18:02:19 +08:00 
			
		
		
		
	release v0.1.3
Former-commit-id: 62c68bcbf591516e8f90b47810bea6f710fd23f6
This commit is contained in:
		
							parent
							
								
									a1468139a5
								
							
						
					
					
						commit
						dad7ca6633
					
				@ -1,8 +1,8 @@
 | 
			
		||||
torch>=1.13.1
 | 
			
		||||
transformers>=4.29.1
 | 
			
		||||
datasets>=2.12.0
 | 
			
		||||
accelerate>=0.19.0
 | 
			
		||||
peft>=0.3.0
 | 
			
		||||
accelerate>=0.21.0
 | 
			
		||||
peft>=0.4.0
 | 
			
		||||
trl>=0.4.7
 | 
			
		||||
sentencepiece
 | 
			
		||||
jieba
 | 
			
		||||
 | 
			
		||||
@ -1,4 +1,4 @@
 | 
			
		||||
from llmtuner.chat import ChatModel
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
__version__ = "0.1.2"
 | 
			
		||||
__version__ = "0.1.3"
 | 
			
		||||
 | 
			
		||||
@ -27,8 +27,8 @@ logger = get_logger(__name__)
 | 
			
		||||
 | 
			
		||||
check_min_version("4.29.1")
 | 
			
		||||
require_version("datasets>=2.12.0", "To fix: pip install datasets>=2.12.0")
 | 
			
		||||
require_version("accelerate>=0.19.0", "To fix: pip install accelerate>=0.19.0")
 | 
			
		||||
require_version("peft>=0.3.0", "To fix: pip install peft>=0.3.0")
 | 
			
		||||
require_version("accelerate>=0.21.0", "To fix: pip install accelerate>=0.21.0")
 | 
			
		||||
require_version("peft>=0.4.0", "To fix: pip install peft>=0.4.0")
 | 
			
		||||
require_version("trl>=0.4.7", "To fix: pip install trl>=0.4.7")
 | 
			
		||||
 | 
			
		||||
 | 
			
		||||
@ -81,9 +81,6 @@ def load_model_and_tokenizer(
 | 
			
		||||
 | 
			
		||||
        elif model_args.quantization_bit == 4:
 | 
			
		||||
            require_version("bitsandbytes>=0.39.0", "To fix: pip install bitsandbytes>=0.39.0")
 | 
			
		||||
            require_version("transformers>=4.30.1", "To fix: pip install transformers>=4.30.1")
 | 
			
		||||
            require_version("accelerate>=0.20.3", "To fix: pip install accelerate>=0.20.3")
 | 
			
		||||
            require_version("peft>=0.4.0.dev0", "To fix: pip install git+https://github.com/huggingface/peft.git")
 | 
			
		||||
            config_kwargs["load_in_4bit"] = True
 | 
			
		||||
            config_kwargs["quantization_config"] = BitsAndBytesConfig(
 | 
			
		||||
                load_in_4bit=True,
 | 
			
		||||
 | 
			
		||||
@ -84,6 +84,12 @@ class WebChatModel(ChatModel):
 | 
			
		||||
            query, history, prefix, max_new_tokens=max_new_tokens, top_p=top_p, temperature=temperature
 | 
			
		||||
        ):
 | 
			
		||||
            response += new_text
 | 
			
		||||
            response = self.postprocess(response)
 | 
			
		||||
            new_history = history + [(query, response)]
 | 
			
		||||
            chatbot[-1] = [query, response]
 | 
			
		||||
            yield chatbot, new_history
 | 
			
		||||
 | 
			
		||||
    def postprocess(self, response: str) -> str:
 | 
			
		||||
        response = response.replace("<", "<")
 | 
			
		||||
        response = response.replace(">", ">")
 | 
			
		||||
        return response
 | 
			
		||||
 | 
			
		||||
		Loading…
	
	
			
			x
			
			
		
	
		Reference in New Issue
	
	Block a user