Comments (1)
Hi, thanks for the question.
Note the model here with win rate 40.8% is an SFT model trained on 52k data (a reproduction of the original Alpaca model).
This model is not the base SFT model we use for reward modelling and RLHF, which is the SFT model trained on 10k data.
I have rerun the auto-annotations with the exact models used in our paper. While there's stochasticity in the pooled auto-annotator (due to the assignment of examples to different auto-annotators and randomization in ordering), the difference compared to our paper's results is quite small (see Table 2 of the paper).
Below are the results based on a rerun.
n_draws n_total n_wins n_wins_base standard_error win_rate
GPT4 17.00 805.00 639.00 149.00 1.38 80.43
ChatGPT 9.00 804.00 489.00 306.00 1.71 61.38
rlhf_llama_7b_regen_v7_3ep_v12_ckpt_20 9.00 803.00 370.00 424.00 1.75 46.64
sft_52k 19.00 805.00 325.00 461.00 1.72 41.55
sft_llama_7b_regen_v7_3ep 16.00 804.00 320.00 468.00 1.72 40.80
sft_10k 19.00 802.00 278.00 505.00 1.67 35.85
Davinci001 0.00 805.00 201.00 604.00 1.53 24.97
LLaMA 7B 0.00 786.00 94.00 692.00 1.16 11.96
The sft_52k
and sft_10k
entries are based on reruns.
I will send a patch to clarify this point.
from alpaca_farm.
Related Issues (20)
- [Bug] Error importing is_deepspeed_zero3_enabled HOT 1
- Question about KL term HOT 1
- Reproducibility of pretuned reward model
- Use with Llama-2-70b-hf? HOT 4
- KeyError: 'llama' in /recover_model_weights.py HOT 2
- Confusing detail preference mapping HOT 1
- recover_model_weights.py gives WARNING:root:Your base LLaMA checkpoint is converted with transformers==4.27.0.dev0
- Huge memory demand of recover_model_weights.py?
- RewardModel.from_pretrained() loads redundant weights (incurs extra ~30GB of RAM)
- Why use FSDP instead of Deepspeed? HOT 2
- lower cuda version? HOT 1
- integrity_check error with sft10k
- score with reward model HOT 2
- RecursionError: maximum recursion depth exceeded
- BaseAnnotator.__init__() got an unexpected keyword argument 'other_keys_to_keep' HOT 4
- PairwiseAutoAnnotator always "Annotating 0 examples with gpt4_3" HOT 1
- Repeated Deprecation Error HOT 3
- Error downloading pre-trained weights
- Using pretrained models HOT 3
- Inquiry Regarding Supervised Fine-Tuning with AlpacaFarm Framework for Pythia Models HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from alpaca_farm.