Recursive amplification
Subquestions can have subquestions
Now we’d like to generalize the recipe above so that we can run it at different depths:
Depth 0: Just answer the question, no subquestions.
Depth 1: One layer of subquestions.
Depth 2: Use subquestions when answering subquestions.
Etc.
To do this, we add a depth
parameter to answer_by_amplification
and get_subs
and only get subquestions if we’re at depth > 0. This simplifies the amplification recipe to:
Now we have a parameterized recipe that we can run at different depths:
Depth 0
Depth 1
Depth 2
The trace for depth 2, partially expanded:
Exercises
Right now we’re answering subquestions without the context of the question they’re intended to help with. Provide the question (or questions) that are further up in the hierarchy as additional context to the model.
Running subquestions in parallel is nice because it’s fast, but has the disadvantage that the answers to later subquestions can’t inform what question to ask next. Modify the recipe to support sequential choice of questions based on earlier responses.
Now make the recipe from step 1 adaptive: Let the model decide whether to answer or whether to ask another subquestion. If you don’t limit the depth, what is the effective depth that the model ends up with for different questions?
References
Christiano, Paul, Buck Shlegeris, and Dario Amodei. Supervising Strong Learners by Amplifying Weak Experts. October 19, 2018.
Leike, Jan, David Krueger, Tom Everitt, Miljan Martic, Vishal Maini, and Shane Legg. Scalable Agent Alignment via Reward Modeling: A Research Direction. arXiv cs.LG, November 19, 2018.
Wu, Jeff, Long Ouyang, Daniel M. Ziegler, Nisan Stiennon, Ryan Lowe, Jan Leike, and Paul Christiano. Recursively Summarizing Books with Human Feedback. arXiv, September 27, 2021.
Perez, Ethan, Patrick Lewis, Wen-tau Yih, Kyunghyun Cho, and Douwe Kiela. Unsupervised Question Decomposition for Question Answering. In Proceedings of the 2020 Conference on Empirical Methods in Natural Language Processing (EMNLP), 8864–80. Online: Association for Computational Linguistics, 2020.
Last updated