git clone https://gitcode.com/tencent_hunyuan/Hunyuan-4B-Instruct-FP8
第二步:安装依赖
pip install "transformers>=4.56.0" torch accelerate
第三步:基础推理示例
from transformers import AutoModelForCausalLM, AutoTokenizer import re
model_name_or_path = "tencent/Hunyuan-4B-Instruct" tokenizer = AutoTokenizer.from_pretrained(model_name_or_path) model = AutoModelForCausalLM.from_pretrained(model_name_or_path, device_map="auto")
快思维模式示例
messages = [{"role": "user", "content": "/no_think 请解释什么是人工智能"}] inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device) outputs = model.generate(inputs, max_new_tokens=2048) print(tokenizer.decode(outputs[0], skip_special_tokens=True))
慢思维模式示例
messages = [{"role": "user", "content": "/think 解方程:3x + 7 = 22"}] inputs = tokenizer.apply_chat_template(messages, return_tensors="pt").to(model.device) outputs = model.generate(inputs, max_new_tokens=2048)
提取思考过程和答案
output_text = tokenizer.decode(outputs[0]) think_pattern = r'(.*?)' think_matches = re.findall(think_pattern, output_text, re.DOTALL) think_content = [match.strip() for match in think_matches][0] answer_content = [match.strip() for match in think_matches][1]
print(f"思考过程: {think_content}\n\n答案: {answer_content}")
创作声明:本文部分内容由AI辅助生成(AIGC),仅供参考



