Yann Dubois, Chen Xuechen Li, Rohan Taori, Tianyi
Zhang, Ishaan Gulrajani, Jimmy Ba, Carlos Guestrin,
Percy S Liang, and Tatsunori B Hashimoto. 2024.
Alpacafarm: A simulation framework for methods
that learn from human feedback. Advances in Neural
Information Processing Systems, 36.
Alex Fabbri, Prafulla Kumar Choubey, Jesse Vig, Chien-
Sheng Wu, and Caiming Xiong. 2022. Improving fac-
tual consistency in summarization with compression-
based post-editing. In Proceedings of the 2022 Con-
ference on Empirical Methods in Natural Language
Processing, pages 9149–9156, Abu Dhabi, United
Arab Emirates. Association for Computational Lin-
guistics.
Zelalem Gero, Chandan Singh, Hao Cheng, Tristan Nau-
mann, Michel Galley, Jianfeng Gao, and Hoifung
Poon. 2023. Self-verification improves few-shot clin-
ical information extraction. In ICML 3rd Workshop
on Interpretable Machine Learning in Healthcare
(IMLH).
Zhibin Gou, Zhihong Shao, Yeyun Gong, Yelong Shen,
Yujiu Yang, Nan Duan, and Weizhu Chen. 2024.
CRITIC: Large Language Models Can Self-Correct
with Tool-Interactive Critiquing.
Edward J Hu, yelong shen, Phillip Wallis, Zeyuan Allen-
Zhu, Yuanzhi Li, Shean Wang, Lu Wang, and Weizhu
Chen. 2022. LoRA: Low-rank adaptation of large
language models. In International Conference on
Learning Representations.
Dongfu Jiang, Yishan Li, Ge Zhang, Wenhao Huang,
Bill Yuchen Lin, and Wenhu Chen. 2024. TIGER-
Score: Building Explainable Metric for All Text Gen-
eration Task.
Shuyang Jiang, Yuhao Wang, and Yu Wang. 2023. Self-
Evolve: A Code Evolution Framework via Large
Language Models. arXiv preprint arXiv:2306.02907.
Seungone Kim, Jamin Shin, Yejin Cho, Joel Jang,
Shayne Longpre, Hwaran Lee, Sangdoo Yun,
Seongjin Shin, Sungdong Kim, James Thorne, et al.
2023. Prometheus: Inducing evaluation capability
in language models. In NeurIPS 2023 Workshop on
Instruction Tuning and Instruction Following.
Seungone Kim, Juyoung Suk, Shayne Longpre,
Bill Yuchen Lin, Jamin Shin, Sean Welleck, Graham
Neubig, Moontae Lee, Kyungjae Lee, and Minjoon
Seo. 2024a. Prometheus 2: An Open Source Lan-
guage Model Specialized in Evaluating Other Lan-
guage Models. arXiv preprint arXiv:2405.01535.
Tae Soo Kim, Yoonjoo Lee, Jamin Shin, Young-Ho
Kim, and Juho Kim. 2024b. EvalLM: Interactive
Evaluation of Large Language Model Prompts on
User-Defined Criteria. In Proceedings of the CHI
Conference on Human Factors in Computing Systems,
pages 1–21.
Leandro von Werra, Younes Belkada, Lewis Tunstall,
Edward Beeching, Tristan Thrush, Nathan Lambert,
and Shengyi Huang. 2020. TRL: Transformer Rein-
forcement Learning.
Vladimir I Levenshtein et al. 1966. Binary codes capa-
ble of correcting deletions, insertions, and reversals.
In Soviet physics doklady, volume 10, pages 707–710.
Soviet Union.
Sharon Levy, Michael Saxon, and William Yang Wang.
2021. Investigating memorization of conspiracy theo-
ries in text generation. In Findings of the Association
for Computational Linguistics: ACL-IJCNLP 2021,
pages 4718–4729, Online. Association for Computa-
tional Linguistics.
Zhen Li, Xiaohan Xu, Tao Shen, Can Xu, Jia-Chen
Gu, and Chongyang Tao. 2024. Leveraging Large
Language Models for NLG Evaluation: A Survey.
arXiv preprint arXiv:2401.07103.
Stephanie C. Lin, Jacob Hilton, and Owain Evans. 2021.
TruthfulQA: Measuring How Models Mimic Human
Falsehoods. In Annual Meeting of the Association
for Computational Linguistics.
Yinhan Liu, Myle Ott, Naman Goyal, Jingfei Du, Man-
dar Joshi, Danqi Chen, Omer Levy, Mike Lewis,
Luke Zettlemoyer, and Veselin Stoyanov. 2019.
RoBERTa: A Robustly Optimized BERT Pretrain-
ing Approach. arXiv preprint arXiv:1907.11692.
Yixin Liu, Budhaditya Deb, Milagro Teruel, Aaron Hal-
faker, Dragomir Radev, and Ahmed Hassan Awadal-
lah. 2023. On improving summarization factual con-
sistency from natural language feedback. In Proceed-
ings of the 61st Annual Meeting of the Association for
Computational Linguistics (Volume 1: Long Papers),
pages 15144–15161, Toronto, Canada. Association
for Computational Linguistics.
Aman Madaan, Niket Tandon, Prakhar Gupta, Skyler
Hallinan, Luyu Gao, Sarah Wiegreffe, Uri Alon,
Nouha Dziri, Shrimai Prabhumoye, Yiming Yang,
et al. 2023. Self-refine: Iterative refinement with
self-feedback. Advances in Neural Information Pro-
cessing Systems, 36.
Meta. 2024. Introducing Meta Llama 3: The most
capable openly available LLM to date.
Todor Mihaylov, Peter Clark, Tushar Khot, and Ashish
Sabharwal. 2018. Can a suit of armor conduct elec-
tricity? a new dataset for open book question answer-
ing. In Conference on Empirical Methods in Natural
Language Processing.
Abhika Mishra, Akari Asai, Vidhisha Balachandran,
Yizhong Wang, Graham Neubig, Yulia Tsvetkov, and
Hannaneh Hajishirzi. 2024. Fine-grained Halluci-
nation Detection and Editing For Language Models.
arXiv preprint arXiv:2401.06855.
Yixin Nie, Xiang Zhou, and Mohit Bansal. 2020. What
can we learn from collective human opinions on nat-
ural language inference data? In Proceedings of the
2020 Conference on Empirical Methods in Natural
10