Microsoft AI Researchers Open-Source ‘GODEL’: A Large Scale Pre-Trained Language Model For Dialog

Recent developments in machine learning have made virtual assistants reliable for various activities, including restaurant recommendations, bill-paying assistance, and appointment reminders.

A novel work by the Microsoft research team now presents GODEL, a Grounded Open Dialogue Language Model. Their work introduces a new class of pretrained language models that permit both task-oriented and social dialogue and are assessed by the utility of their responses. With GODEL, they aim to help researchers and developers design dialogue agents that are unlimited in the types of queries they may react to and the sources of information they can draw from.

The potential for meaningful, open-ended conversational exchanges is present in modern state-of-the-art models that use massive PLMs. Still, they are resistant to meaningful comparison because there is no agreement on how to evaluate them. Their method overcomes the absence of reliable automated evaluation criteria, which has long been a barrier to general-purpose open-ended discussion models. 

According to the researchers, the general-purpose dialogue model should be fluid and socially engaging. In fact, these inherent communication characteristics make up the majority of SOTA PLM evaluations. Beyond that, though, most machine-human conversations have a purpose and work to help the user achieve one or more objectives. That is, the model must be useful to the user. Therefore, they emphasize that automated evaluation in general-domain models should be this external dimension of functional value.

The technology that underpins these dialogue bots, conversational AI, includes trained language models among its engines. Alternatively, they can engage in chitchat or open-domain conversations with no predetermined goal. GODEL combines these two skills, enabling dialogue agents to respond based on both the conversation’s context and outside data—content that was not included in the dataset used to train the model. This covers unstructured content, such as restaurant reviews, Wikipedia articles, and other publicly accessible content found on the web, and organized content, such as data saved in databases. 

The adaptability that GODEL offers users in designing their model’s grounding—the sources from which their dialogue agents get information. It is one of its primary characteristics. This adaptability contributes to GODEL’s adaptability in many conversational contexts. GODEL would be able to respond to questions even when the information used to train it may not have included that establishment.


Depending on whether the grounding information is null, a portion of a document, a search result (unstructured text), or details taken from a database about the restaurant, the responses would change (structured text). Each comment, though, would be pertinent and helpful. In addition to increasing specificity, grounded generation contributes to model maintenance because the grounded text can include data that wasn’t necessarily present when the model was trained.

The variety of conversational applications offered by GODEL is another key characteristic. GODEL can be used for a range of dialogues, including task-oriented, question-answering, and grounded chitchat, in contrast to its predecessors, which mostly focused on social bots. They also show that a range of query forms, such as general inquiries or requests for particular tasks, can be satisfactorily addressed by GODEL in the same conversation.

The researchers conducted experiments to show the usefulness of GODEL’s responses. They show that goal-directed datasets are more reliably evaluated and that when asked to rank their utility in terms of fulfilling particular goals, people generally agree on which answers are better. With this reliable evaluation setup, they compared their model to several reliable baselines and cutting-edge techniques. Their findings suggest that GODEL is superior in terms of both human and automatic evaluation.

With its cutting-edge dialogue models and lack of considerable GPU resources, the team hopes that GODEL will help various academic research teams advance the conversational AI field. 

This Article is written as a summary article by Marktechpost Staff based on the paper 'GODEL: Large-Scale Pre-Training for Goal-Directed Dialog'. All Credit For This Research Goes To Researchers on This Project. Checkout the paper, github, project and post.

Please Don't Forget To Join Our ML Subreddit
✅ [Featured Tool] Check out Taipy Enterprise Edition