arxiv Preprint – Shepherd: A Critic for Language Model Generation


In this episode we discuss Shepherd: A Critic for Language Model Generation
by Tianlu Wang, Ping Yu, Xiaoqing Ellen Tan, Sean O’Brien, Ramakanth Pasunuru, Jane Dwivedi-Yu, Olga Golovneva, Luke Zettlemoyer, Maryam Fazel-Zarandi, Asli Celikyilmaz. The paper introduces Shepherd, a language model trained to critique responses generated by large language models (LLMs) and offer suggestions for improvement. Despite its smaller size, Shepherd’s critiques are on par or preferred over established models like ChatGPT. Evaluation results demonstrate Shepherd’s superior performance and highlight its potential in enhancing the reliability and coherence of LLM outputs.


Posted

in

by

Tags: