| --- |
| language: |
| - zh |
| license: apache-2.0 |
| tasks: |
| - text-generation |
| --- |
| |
| <!-- markdownlint-disable first-line-h1 --> |
| <!-- markdownlint-disable html --> |
| <div align="center"> |
| <h1> |
| HuatuoGPT2-34B-4bits |
| </h1> |
| </div> |
|
|
| <div align="center"> |
| <a href="https://github.com/FreedomIntelligence/HuatuoGPT-II" target="_blank">GitHub</a> | <a href="https://arxiv.org/pdf/2311.09774.pdf" target="_blank">Our Paper</a> |
| </div> |
|
|
| # <span id="Start">Quick Start</span> |
|
|
| ```Python |
| import torch |
| from transformers import AutoModelForCausalLM, AutoTokenizer |
| from transformers.generation.utils import GenerationConfig |
| tokenizer = AutoTokenizer.from_pretrained("FreedomIntelligence/HuatuoGPT2-34B-4bits", use_fast=True, trust_remote_code=True) |
| model = AutoModelForCausalLM.from_pretrained("FreedomIntelligence/HuatuoGPT2-34B-4bits", device_map="auto", torch_dtype="auto", trust_remote_code=True) |
| model.generation_config = GenerationConfig.from_pretrained("FreedomIntelligence/HuatuoGPT2-34B-4bits") |
| messages = [] |
| messages.append({"role": "user", "content": "肚子疼怎么办?"}) |
| response = model.HuatuoChat(tokenizer, messages) |
| print(response) |
| ``` |