Skip to content

Commit c57f38c

Browse files
authored
Fix changes i forgot to push? (#100)
1 parent 11732cf commit c57f38c

File tree

3 files changed

+120
-3
lines changed

3 files changed

+120
-3
lines changed

chapters/16-evaluation.md

Lines changed: 69 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -22,7 +22,7 @@ Evaluation for RLHF and post-training has gone a few distinct phases in its earl
2222
Beyond this, new domains will evolve.
2323
Throughout this chapter we will include details that map to how these evaluations were implemented and understood.
2424

25-
## Formatting: From Few-shot to Zero-shot Prompting
25+
## Formatting as prompting: From Few-shot to Zero-shot Prompting to CoT
2626

2727
Early language models were only used as intelligent autocomplete.
2828
In order to use these models in an more open ended way, multiple examples were shown to the model and then a prompt that is an incomplete phrase. This was called few-shot or in-context learning [@brown2020language], and at the time instruction tuning or RLHF was not involved.
@@ -107,19 +107,85 @@ With more open-ended usage, generative evaluation became increasingly popular as
107107
In this period through recent years after ChatGPT, some multiple-choice evaluations were still used in RLHF research as a holdback to common practice.
108108

109109
With the rise of reasoning models at the end of 2024 and the beginning of 2025, a major change in model behavior was the addition of a long Chain-of-Thought (CoT [@wei2022chain]) reasoning process before every answer.
110+
111+
For example, for every prompt there can specially designed prompts to help extract behavior from the model.
112+
Tülu 3 details some prompts used for CoT answering on multiple choice questions [@lambert2024t]:
113+
114+
```
115+
Answer the following multiple-choice question by giving the correct answer letter in parentheses. Provide CONCISE reasoning for the answer, and make sure to finish the response with “Therefore, the answer is (ANSWER_LETTER)” where (ANSWER_LETTER) is one of (A), (B), (C), (D), (E), etc.
116+
117+
Question: {question}
118+
(A) {choice_A}
119+
(B) {choice_B}
120+
(C) …
121+
122+
Answer the above question and REMEMBER to finish your response with the exact phrase “Therefore, the answer is (ANSWER_LETTER)” where (ANSWER_LETTER) is one of (A), (B), (C), (D), (E), etc.
123+
```
124+
110125
This, especially when the models use special formatting to separate thinking tokens from answer tokens, necessitated the most recent major update to evaluation regimes.
111126
Evaluation is moving to where the models are tested to respond in a generative manner with a chain of thought prompting.
112127

113128
## Prompting
114129

115130
Prompting, i.e. crafting the correct query for a model, is a crucial portion of using them as the models are evolving rapidly.
116131

117-
## Evaluation
132+
TODO expand
133+
134+
## Tooling
135+
136+
*Needs editing, taken from blog*
137+
138+
Language model evaluations done within companies can only be compared to their peers with large error bars. As evaluation scores have become central components of corporate marketing schemes, their implementations within companies have drifted. There are rumors of major AI labs using “custom prompts” for important evaluations like GSM8k or MATH. At the same time, OpenAI has released open-source code for configurations (and basic prompts) for their models.
139+
140+
It would seem that open-source language model builders should have an advantage in trust and comparability of models, but the open-source AI community also hasn’t converged on a single rubric for evaluating language models. The experience that motivated this article has happened every time — it is looking at an open weights model (such as Llama) and realizing we need to re-run evaluations on our own setup to get a sense of what is actually happening. This is very salient and hard to communicate to the general audience, where open means more reproducible. It should also mean the results are easier to interpret. What tools do we need for this?
141+
142+
In my article on “Big Tech’s LLM evals are just marketing,” I didn’t uncover the deeper reasons as to why can’t fully believe these evaluations. Language model evaluation stacks are perceived as marketing because the evaluations have no hard source of truth. What is happening inside frontier labs is that evaluation suites are being tuned to suit their internal needs. When results are shared, we get output in the form of the numbers a lab got for their models, but not all the inputs to that function. The inputs are very sensitive configurations, and they’re different at all of OpenAI, Meta, Anthropic, and Google. Even fully open evaluation standards are hard to guarantee reproducibility on. Focusing efforts on your own models is the only way to get close to repeatable evaluation techniques. There are good intentions underpinning the marketing, starting with the technical teams.
143+
144+
Evaluation of frontier language models is every bit as much an art today as it is a science.
145+
146+
Different groups choose different evaluations to maintain independence on, i.e. making them a true test set, but no one discloses which ones they choose. For example, popular reasoning evaluations MATH and GSM8k both have training sets with prompts that can easily be used to improve performance. Improving performance with the prompts from the same distribution is very different than generalizing to these tasks by training on general math data.
147+
148+
Labs like OpenAI hillclimb by focusing on a few key evaluations and report scores on the core public set at the end. The key point is that some of their evaluations for tracking progress, such as the datasets for cross-entropy loss predictions in scaling from the GPT-4 report, are often not public.
149+
150+
151+
The post-training evaluations are heavily co-dependent on human evaluation. Human evaluation for generative language models yields Elo rankings (popular in early Anthropic papers, such as Constitutional AI), and human evaluation for reward models shows agreement.
118152

119-
### Formatting and Overview
153+
The limited set of evaluations they choose to focus on forms a close link between evaluation and training. At one point one evaluation of focus was MMLU. Now, GPQA is likely. Labs will change the evaluations to make them better suited to their needs, such as OpenAI releasing SWE-Bench-Verified. There are more internally we don’t know about.
154+
155+
The key “capability” that improving evaluations internally has on downstream training is improving the statistical power when comparing training runs. By changing evaluations, these labs reduce the noise on their prioritized signals in order to make more informed training decisions.
156+
157+
This is compounded by the sophistication of post-training in the modern language model training stacks. Evaluating language models today involves a moderate amount of generating tokens (rather than just looking at log probabilities of answers). It is accepted that small tricks are used by frontier labs to boost performance on many tasks — the most common explanation is one-off prompts for certain evaluations. Does Claude use its special thinking tokens when being evaluated on reasoning benchmarks? I have no way of knowing.
158+
159+
Depending on how your data is formatted in post-training, models will have substantial differences. For example, two popular, open math datasets Numina and MetaMathQA conflict with each other in training due to small differences in how the answers are formatted — training on both can make performance worse than with just one. Meta’s models use a very specific format for MATH, Minerva, while this varies substantially with post-training decisions.
160+
161+
In the end we are left with a few key points on the state of evaluating closed models:
162+
163+
We do not know or necessarily have the key test sets that labs are climbing on, so some evaluations are proxies (or worse, blissfully cheated on by training on text on the public web).
164+
165+
Inference of frontier models is becoming more complicated with special system prompts, special tokens, etc., and we don’t know how it impacts evaluations, and
166+
167+
We do not know all the formats and details used to numerically report the closed evaluations.
168+
169+
There are a lot of asterisks to put on the state of closed evaluations. To get ahead of some comments, it is important to acknowledge that there are some very useful efforts by closed labs to communicate their evaluation efforts.
170+
Some include OpenAI’s prompts in simple-evals, OpenAI releasing entire evals to the community, such as MLE Bench, or Meta’s documentation on reproducing their results with a popular open-source eval tools.
171+
172+
TODO open tools here:
173+
174+
There are many open-sourced evaluation tools for people to choose from. There’s Inspect AI from the UK Safety Institute [@inspectAI2024], HuggingFace’s LightEval [@fourrier2023lighteval] that powers the LLM Leaderboard, Eleuther AI’s evaluation harness [@gao2023evalharness] built on top of the infrastructure from their GPT-Neo-X model (around GPT-3 evaluation config), AI2’s library based on OLMES [@gu2024olmes], Stanford’s Center for Research on Foundation Model’s HELM [@liang2023helm], Mosaic’s (now Databricks’) Eval Gauntlet [@mosaicml2024gauntlet], and surely more I did not mention.
175+
176+
## Contamination
177+
178+
TODO mini section
179+
180+
## Case Studies
181+
182+
Below are a few types of evaluations that are at the core of how RLHF and post-training is evolving today.
120183

121184
### ChatBotArena
122185

186+
*Needs editing, taken from blog*
187+
188+
123189
ChatBotArena is the largest community evaluation tool for language models. The LMSYS team, which emerged early in the post-ChatGPT craze, works with most of the model providers to host all of the relevant models. If you’re looking to get to know how multiple models compare to each other, ChatBotArena is the place to start.
124190

125191
ChatBotArena casts language model evaluation through the wisdom of the crowd. For getting an initial ranking of how models stack up and how the models in the ecosystem are getting better, it has been and will remain crucial.

chapters/bib.bib

Lines changed: 50 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -1482,6 +1482,56 @@ @inproceedings{sanh2022multitask
14821482
booktitle={International Conference on Learning Representations},
14831483
year={2022}
14841484
}
1485+
# eval tools
1486+
@misc{inspectAI2024,
1487+
author = {{UK AI Safety Institute}},
1488+
title = {{Inspect AI: Framework for Large Language Model Evaluations}},
1489+
year = {2024},
1490+
howpublished = {\url{https://github.com/UKGovernmentBEIS/inspect_ai}}
1491+
}
1492+
@misc{fourrier2023lighteval,
1493+
author = {Fourrier, Clementine and Habib, Nathan and Kydlicek, Hynek and Wolf, Thomas and Tunstall, Lewis},
1494+
title = {{LightEval: A lightweight framework for LLM evaluation}},
1495+
year = {2023},
1496+
howpublished = {\url{https://github.com/huggingface/lighteval}}
1497+
}
1498+
@misc{gao2023evalharness,
1499+
author = {Gao, Leo and Tow, Jonathan and Abbasi, Baber and Biderman, Stella and Black, Sid and DiPofi, Anthony and Foster, Charles and Golding, Laurence and Hsu, Jeffrey and Le Noac'h, Alain and Li, Haonan and McDonell, Kyle and Muennighoff, Niklas and Ociepa, Chris and Phang, Jason and Reynolds, Laria and Schoelkopf, Hailey and Skowron, Aviya and Sutawika, Lintang and Tang, Eric and Thite, Anish and Wang, Ben and Wang, Kevin and Zou, Andy},
1500+
title = {{A Framework for Few-Shot Language Model Evaluation}},
1501+
year = {2023},
1502+
publisher = {Zenodo},
1503+
doi = {10.5281/zenodo.10256836},
1504+
url = {https://zenodo.org/record/10256836}
1505+
}
1506+
@article{gu2024olmes,
1507+
author = {Gu, Yuling and Tafjord, Oyvind and Kuehl, Bailey and Haddad, Dany and Dodge, Jesse and Hajishirzi, Hannaneh},
1508+
title = {{OLMES: A Standard for Language Model Evaluations}},
1509+
journal = {arXiv preprint arXiv:2406.08446},
1510+
year = {2024}
1511+
}
1512+
@article{liang2023helm,
1513+
author = {Liang, Percy and Bommasani, Rishi and Lee, Tony and Tsipras, Dimitris and Soylu, Dilara and Yasunaga, Michihiro and Zhang, Yian and Narayanan, Deepak and Wu, Yuhuai and Kumar, Ananya and Newman, Benjamin and Yuan, Binhang and Yan, Bobby and Zhang, Ce and Cosgrove, Christian and Manning, Christopher D. and R\'e, Christopher and Acosta-Navas, Diana and Hudson, Drew A. and Zelikman, Eric and Durmus, Esin and Ladhak, Faisal and Rong, Frieda and Ren, Hongyu and Yao, Huaxiu and Wang, Jue and Santhanam, Keshav and Orr, Laurel J. and Zheng, Lucia and Y\'uksekg\"on\"ul, Mert and Suzgun, Mirac and Kim, Nathan and Guha, Neel and Chatterji, Niladri S. and Khattab, Omar and Henderson, Peter and Huang, Qian and Chi, Ryan and Xie, Sang Michael and Santurkar, Shibani and Ganguli, Surya and Hashimoto, Tatsunori and Icard, Thomas and Zhang, Tianyi and Chaudhary, Vishrav and Wang, William and Li, Xuechen and Mai, Yifan and Zhang, Yuhui and Koreeda, Yuta},
1514+
title = {{Holistic Evaluation of Language Models}},
1515+
journal = {Transactions on Machine Learning Research},
1516+
year = {2023},
1517+
doi = {10.1111/nyas.15007},
1518+
note = {Also available as arXiv:2211.09110}
1519+
}
1520+
@misc{mosaicml2024gauntlet,
1521+
author = {{MosaicML}},
1522+
title = {{Mosaic Eval Gauntlet v0.3.0 \textemdash{} Evaluation Suite}},
1523+
year = {2024},
1524+
howpublished = {\url{https://github.com/mosaicml/llm-foundry/blob/main/scripts/eval/local_data/EVAL_GAUNTLET.md}},
1525+
note = {Accessed 2024-05-29}
1526+
}
1527+
@article{lewkowycz2022solving,
1528+
title={Solving quantitative reasoning problems with language models},
1529+
author={Lewkowycz, Aitor and Andreassen, Anders and Dohan, David and Dyer, Ethan and Michalewski, Henryk and Ramasesh, Vinay and Slone, Ambrose and Anil, Cem and Schlag, Imanol and Gutman-Solo, Theo and others},
1530+
journal={Advances in Neural Information Processing Systems},
1531+
volume={35},
1532+
pages={3843--3857},
1533+
year={2022}
1534+
}
14851535
# Misc Blogs / talks #############################################################################
14861536
@misc{schulman2016klapprox,
14871537
author = {Schulman, John},

templates/html.html

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -86,6 +86,7 @@ <h2>Abstract</h2>
8686
<body>
8787
<section id="changelog" style="padding: 20px; text-align: center;">
8888
<h2>Changelog</h2>
89+
<p><strong>6 Apr. - 12 Apr. 2025.</strong>: Evaluation section </p>
8990
<p><strong>28 Mar. - 5 Apr. 2025.</strong>: Research on RLHF x Product, cleaning, improving website, reasoning section </p>
9091
<p><strong>17 Mar. - 27 Mar 2025.</strong>: Improving policy gradient section, minor changes </p>
9192
<p><strong>6 Mar. - 16 Mar 2025.</strong>: Finish DPO, major cleaning </p>

0 commit comments

Comments
 (0)