Medical Chat Model
Model descrisption
The Medical_chat_model is a large language model (LLM) fine-tuned specifically for conversational tasks in the healthcare domain. It is designed to assist with initial medical inquiries, symptom clarification, and general health information retrieval by simulating doctor-patient interactions.
This model was created by fine-tuning a powerful base language model (e.g., Llama, Mistral, etc.) on specialized medical dialogue data to enhance its knowledge base and conversational fluency within a clinical context.
Model training details
- Model Type --> "Causal Language Model (e.g., Decoder-only Transformer)"
- Task --> "Conversational AI, Text Generation, Medical Chatbot"
- Language --> English
- Fine-tuning Dataset --> all_medtext - medical health conversation data
Model Usage
Requirements
pip install transformers torch accelerate
Model loading and inference
from transformers import AutoModelForCausalLM, AutoTokenizer
import torch
# The model identifier provided by the user
model_id = "avikumart/Medical_chat_model"
# 1. Load the tokenizer
tokenizer = AutoTokenizer.from_pretrained(model_id)
# 2. Load the model
# Using AutoModelForCausalLM for generative chat models
# Setting torch_dtype=torch.bfloat16 for modern LLMs and device_map="auto" for multi-GPU setups
model = AutoModelForCausalLM.from_pretrained(
model_id,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# 3. Define the conversational prompt
# Format the prompt to guide the model to act as a doctor
prompt_template = (
"A doctor and a patient are discussing symptoms. "
"Patient: I've been having a persistent dry cough and low-grade fever for the past three days. "
"Doctor:"
)
inputs = tokenizer(prompt_template, return_tensors="pt").to(model.device)
# 4. Generate the response
# Use model.generate for text generation
output_ids = model.generate(
**inputs,
max_new_tokens=256,
do_sample=True,
temperature=0.7,
pad_token_id=tokenizer.eos_token_id
)
# 5. Decode and print the output
response = tokenizer.decode(output_ids[0], skip_special_tokens=True)
print(response.split("Doctor:")[-1].strip())
Dataset
This model was fine-tuned on the all_medtext - medical health conversation data corpus, a high-quality, domain-specific dataset crucial for developing robust healthcare LLMs.
The dataset features a large collection of structured doctor-patient dialogues and/or medical instructions. It typically covers:
Scale: Over 50,000+ to 100,000+ conversational pairs, ensuring broad coverage.
Specialties: Interactions span a wide array of medical specialties, including internal medicine, cardiology, dermatology, pediatrics, and more.
Content: The conversations focus on symptom elicitation, diagnosis-oriented discussion, treatment suggestions (e.g., medication, home remedies), and medical advice.
The rich, specialized nature of this data ensures that the fine-tuned model is capable of generating clinically relevant, context-aware, and structured medical responses, moving beyond generic LLM capabilities.
- Downloads last month
- 29
Model tree for avikumart/Medical_chat_model
Base model
meta-llama/Llama-3.2-1B