Skip to content

rajat709/Llama2-7b-Medical-Model

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

6 Commits
 
 
 
 
 
 
 
 

Repository files navigation

title emoji sdk sdk_version app_file
Llama 2 7b Med Model
⚕️
gradio
3.48.0
app.py

App URL: https://rjvishwa-rajat-llama-2-7b-med-model.hf.space/

Model: https://huggingface.co/rjvishwa/Llama-2-7b-Med-Model/tree/main

Dataset Overview

Name Source n n included in training
Medical Flashcards medalpaca/medical_meadow_medical_flashcards 33955 33955
Wikidoc medalpaca/medical_meadow_wikidoc 67704 10000
Wikidoc Patient Information medalpaca/medical_meadow_wikidoc_patient_information 5942 5942
Stackexchange academia medalpaca/medical_meadow_stack_exchange 40865 40865
Stackexchange biology medalpaca/medical_meadow_stack_exchange 27887 27887
Stackexchange fitness medalpaca/medical_meadow_stack_exchange 9833 9833
Stackexchange health medalpaca/medical_meadow_stack_exchange 7721 7721
Stackexchange bioinformatics medalpaca/medical_meadow_stack_exchange 5407 5407
USMLE Self Assessment Step 1 medalpaca/medical_meadow_usmle_self 119 92 (test only)
USMLE Self Assessment Step 2 medalpaca/medical_meadow_usmle_self 120 110 (test only)
USMLE Self Assessment Step 3 medalpaca/medical_meadow_usmle_self 135 122 (test only)
MEDIQA original, preprocessed 2208 2208
CORD-19 original, preprocessed 1056660 50000
MMMLU original, preprocessed 3787 3787
Pubmed Health Advice original, preprocessed 10178 10178
Pubmed Causal original, preprocessed 2446 2446
ChatDoctor original 215000 10000
OpenAssistant original 9209 9209

The Llama Med Model represents an advanced suite of large language models meticulously fine-tuned for the specific purpose of medical question-answering and dialogue applications. Our primary goal is to provide a comprehensive selection of open-source language models, thereby facilitating the streamlined development of medical chatbot solutions.

The Meta Llama2 model, boasting an impressive 7 billion parameters, has been fine-tuned on a meticulously curated medical dataset. This training process was conducted on a robust infrastructure comprising two GPU nodes, each equipped with 15 GB of VRAM. Remarkably, this intricate training endeavor was successfully completed in a mere 8 hours.