Llama-3-based OpenBioLLM-Llama3-70B and 8B: Outperforming GPT-4, Gemini, Meditron-70B, Med-PaLM-1 and Med-PaLM-2 in Medical-Domain

With the significant development in the rapidly developing field of Artificial Intelligence driven healthcare, a team of researchers has introduced OpenBioLLM-Llama3-70B & 8B models. These state-of-the-art Large Language Models (LLMs) have the potential to completely transform medical natural language processing (NLP) by establishing new standards for functionality and performance in the biomedical field. 

The release of these models marks a substantial advancement in medical-domain LLM technology. Their ability to outperform models such as GPT-4, Gemini, Meditron-70B, Med-PaLM-1, and Med-PaLM-2 in biomedical tasks is a testament to their superiority and represents a significant breakthrough in the usability and effectiveness of freely available medical language models.

OpenBioLLM-70B has proven state-of-the-art performance, exhibiting unmatched capabilities in relation to its size. This model, which outperforms GPT-3.5, Gemini, and Meditron-70B, has demonstrated the revolutionary power of targeted fine-tuning and creative training approaches. 

The team has shared that they intend to improve these models in the upcoming months by adding multimodal capabilities, extended context windows, better benchmarks, and more coverage of the medical sector. This iterative process demonstrates a dedication to continuous improvement and modification to satisfy the changing needs of the medical AI market.

The development process included Direct Preference Optimisation (DPO) and careful fine-tuning using the LLama-3 70B & 8B models as a foundation. With accuracy, dependability, and versatility as top priorities, this methodological rigor guarantees that OpenBioLLM-Llama3-70B & 8B are optimized for practical medical applications. 

A key component of the model’s performance is the large-scale, multi-dimensional training dataset. Over the course of four months, medical specialists were consulted in the curation process to guarantee the quality and usefulness of the data. With more than ten medical subjects and more than 3,000 healthcare topics, the dataset demonstrates a dedication to inclusion and representativeness in the field of medical AI.

OpenBioLLM-70B’s effect has been demonstrated by its exceptional performance on nine different biomedical datasets, outperforming larger models in spite of having fewer parameters. With an 86.06% average score, this model is a prime example of effectiveness and efficiency in medical NLP. 

OpenBioLLM-70B & 8B’s adaptability covers a wide range of vital medical applications, which are as follows.

  1. Extracting important details from intricate clinical narratives, i.e., summarising clinical notes. 
  1. Providing precise answers to a broad range of medical questions.
  1. Biomedical Classification: Disease prediction, sentiment analysis, and medical document classification. 
  1. De-Identification: Removing personally identifiable information (PII) from medical records in order to protect patient privacy.

In conclusion, a new era in medical NLP has been marked by improved performance, accessibility, and practicality in healthcare contexts with the release of OpenBioLLM-Llama3-70B & 8B. With more development and growth ahead of them, these models have the potential to completely transform medical AI and open the door to more effective, precise, and morally sound healthcare solutions.


Check out the Open Medical-LLM Leaderboard, OpenBioLLM-70B project page, and OpenBioLLM-8B project page. All credit for this research goes to the researchers of this project. Also, don’t forget to follow us on Twitter. Join our Telegram Channel, Discord Channel, and LinkedIn Group.

If you like our work, you will love our newsletter..

Don’t Forget to join our 40k+ ML SubReddit

Tanya Malhotra is a final year undergrad from the University of Petroleum & Energy Studies, Dehradun, pursuing BTech in Computer Science Engineering with a specialization in Artificial Intelligence and Machine Learning.
She is a Data Science enthusiast with good analytical and critical thinking, along with an ardent interest in acquiring new skills, leading groups, and managing work in an organized manner.