Timezone: »
Sampling diverse programs from a code language model and reranking with model likelihood is a popular method for code generation but it is prone to preferring degenerate solutions. Inspired by collaborative programming, we propose Coder-Reviewer reranking. We augment Coder language models from past work, which generate programs given language instructions, with Reviewer models, which evaluate the likelihood of the instruction given the generated programs. We perform an extensive study across six datasets with eight models from three model families. Experimental results show that Coder-Reviewer reranking leads to consistent and significant improvement (up to 17% absolute accuracy gain) over reranking with the Coder model only. When combined with executability filtering, Coder-Reviewer reranking can often outperform the minimum Bayes risk method. Coder-Reviewer reranking is easy to implement by prompting, can generalize to different programming languages, and works well with off-the-shelf hyperparameters.
Author Information
Tianyi Zhang (Stanford University)
Tao Yu (The University of Hong Kong)
Tatsunori Hashimoto (Stanford)
Mike Lewis (Facebook)
Scott Yih (Meta AI - FAIR)
Daniel Fried (Carnegie Mellon University)
Sida Wang (Meta AI)
More from the Same Authors
-
2023 : Exploiting Programmatic Behavior of LLMs: Dual-Use Through Standard Security Attacks »
Daniel Kang · Xuechen Li · Ion Stoica · Carlos Guestrin · Matei Zaharia · Tatsunori Hashimoto -
2023 : Retrieval-Augmented Multimodal Language Modeling »
Michihiro Yasunaga · Armen Aghajanyan · Weijia Shi · Rich James · Jure Leskovec · Percy Liang · Mike Lewis · Luke Zettlemoyer · Wen-tau Yih -
2023 Poster: LEVER: Learning to Verify Language-to-Code Generation with Execution »
Ansong Ni · Srinivasan Iyer · Dragomir Radev · Veselin Stoyanov · Scott Yih · Sida Wang · Xi Victoria Lin -
2023 Poster: Compositional Exemplars for In-context Learning »
Jiacheng Ye · Zhiyong Wu · Jiangtao Feng · Tao Yu · Lingpeng Kong -
2023 Poster: Data Feedback Loops: Model-driven Amplification of Dataset Biases »
Rohan Taori · Tatsunori Hashimoto -
2023 Poster: Whose Opinions Do Language Models Reflect? »
Shibani Santurkar · Esin Durmus · Faisal Ladhak · Cinoo Lee · Percy Liang · Tatsunori Hashimoto -
2023 Poster: Grounding Language Models to Images for Multimodal Inputs and Outputs »
Jing Yu Koh · Ruslan Salakhutdinov · Daniel Fried -
2023 Oral: Data Feedback Loops: Model-driven Amplification of Dataset Biases »
Rohan Taori · Tatsunori Hashimoto -
2023 Oral: Whose Opinions Do Language Models Reflect? »
Shibani Santurkar · Esin Durmus · Faisal Ladhak · Cinoo Lee · Percy Liang · Tatsunori Hashimoto -
2023 Oral: Evaluating Self-Supervised Learning via Risk Decomposition »
Yann Dubois · Tatsunori Hashimoto · Percy Liang -
2023 Poster: Evaluating Self-Supervised Learning via Risk Decomposition »
Yann Dubois · Tatsunori Hashimoto · Percy Liang -
2023 Poster: Out-of-Domain Robustness via Targeted Augmentations »
Irena Gao · Shiori Sagawa · Pang Wei Koh · Tatsunori Hashimoto · Percy Liang -
2023 Poster: DS-1000: A Natural and Reliable Benchmark for Data Science Code Generation »
Yuhang Lai · Chengxi Li · Yiming Wang · Tianyi Zhang · Ruiqi Zhong · Luke Zettlemoyer · Scott Yih · Daniel Fried · Sida Wang · Tao Yu -
2023 Poster: Retrieval-Augmented Multimodal Language Modeling »
Michihiro Yasunaga · Armen Aghajanyan · Weijia Shi · Richard James · Jure Leskovec · Percy Liang · Mike Lewis · Luke Zettlemoyer · Scott Yih -
2022 Poster: Identifiability Conditions for Domain Adaptation »
Ishaan Gulrajani · Tatsunori Hashimoto -
2022 Spotlight: Identifiability Conditions for Domain Adaptation »
Ishaan Gulrajani · Tatsunori Hashimoto -
2021 Poster: BASE Layers: Simplifying Training of Large, Sparse Models »
Mike Lewis · Shruti Bhosale · Tim Dettmers · Naman Goyal · Luke Zettlemoyer -
2021 Spotlight: BASE Layers: Simplifying Training of Large, Sparse Models »
Mike Lewis · Shruti Bhosale · Tim Dettmers · Naman Goyal · Luke Zettlemoyer -
2019 Poster: SWALP : Stochastic Weight Averaging in Low Precision Training »
Guandao Yang · Tianyi Zhang · Polina Kirichenko · Junwen Bai · Andrew Wilson · Christopher De Sa -
2019 Poster: Simplifying Graph Convolutional Networks »
Felix Wu · Amauri Souza · Tianyi Zhang · Christopher Fifty · Tao Yu · Kilian Weinberger -
2019 Oral: SWALP : Stochastic Weight Averaging in Low Precision Training »
Guandao Yang · Tianyi Zhang · Polina Kirichenko · Junwen Bai · Andrew Wilson · Christopher De Sa -
2019 Oral: Simplifying Graph Convolutional Networks »
Felix Wu · Amauri Souza · Tianyi Zhang · Christopher Fifty · Tao Yu · Kilian Weinberger -
2018 Poster: Hierarchical Text Generation and Planning for Strategic Dialogue »
Denis Yarats · Mike Lewis -
2018 Oral: Hierarchical Text Generation and Planning for Strategic Dialogue »
Denis Yarats · Mike Lewis