Comments (2)
model = dict(
type=SupervisedFinetune,
llm=dict(
type=AutoModelForCausalLM.from_pretrained,
pretrained_model_name_or_path=pretrained_model_name_or_path,
trust_remote_code=True,
torch_dtype=torch.float16,
quantization_config=dict(
type=BitsAndBytesConfig,
load_in_4bit=True,
load_in_8bit=False,
llm_int8_threshold=6.0,
llm_int8_has_fp16_weight=False,
bnb_4bit_compute_dtype=torch.float16,
bnb_4bit_use_double_quant=True,
bnb_4bit_quant_type='nf4')
),
# lora=dict(
# type=LoraConfig,
# r=64,
# lora_alpha=16,
# lora_dropout=0.1,
# bias='none',
# task_type='CAUSAL_LM')
)
全参数训练,在进行后续模型转换时,不要加入 quantization_config,不然会导致模型不匹配
from xtuner.
model = dict( type=SupervisedFinetune, llm=dict( type=AutoModelForCausalLM.from_pretrained, pretrained_model_name_or_path=pretrained_model_name_or_path, trust_remote_code=True, torch_dtype=torch.float16, quantization_config=dict( type=BitsAndBytesConfig, load_in_4bit=True, load_in_8bit=False, llm_int8_threshold=6.0, llm_int8_has_fp16_weight=False, bnb_4bit_compute_dtype=torch.float16, bnb_4bit_use_double_quant=True, bnb_4bit_quant_type='nf4') ), # lora=dict( # type=LoraConfig, # r=64, # lora_alpha=16, # lora_dropout=0.1, # bias='none', # task_type='CAUSAL_LM') )
全参数训练,在进行后续模型转换时,不要加入 quantization_config,不然会导致模型不匹配
收到!!!大佬您终于来了哈哈哈哈哈哈
from xtuner.
Related Issues (20)
- EOFError: Ran out of input HOT 1
- 多轮对话训练后测试不能多轮对话。 HOT 8
- XTuner内置几个config文件的log打印疑问 HOT 3
- 怎么在IDE中debug xtuner? HOT 7
- about tokenizer HOT 5
- RuntimeError: Expected all tensors to be on the same device, but found at least two devices, cuda:6 and cuda:7! (when checking argumen t for argument mat2 in method wrapper_CUDA_mm) HOT 4
- xtuner convert merge 报段错误 HOT 3
- xtuner convert merge CUDA out of memory HOT 1
- 认知微调失败 HOT 4
- InternLM2_Chat模型工具调用能力微调是否有相关样例? HOT 2
- KeyError: 'Cache only has 0 layers, attempted to access layer with index 0' HOT 1
- 超大规模预训练 HOT 1
- Inheritance between configuration files HOT 4
- RLHF and DPO
- [Bug] Empty available tools during execute `msagent_react_map_fn` HOT 3
- Do you have plan of adding RWKV finetuning? HOT 2
- 微调interlm20b-chat报错:UnicodeEncodeError: 'ascii' codec can't encode characters in position 28-39: ordinal not in range(128) HOT 2
- 微调llava后,转换模型报错 HOT 2
- Support model sharding for training large models HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from xtuner.