๐Ÿง  OphthaScholarโ€‘1.2B

OphthaScholarโ€‘1.2B is a fine-tuned clinical assistant model focused specifically on ophthalmology. It is built on top of the LiquidAI LFMโ€‘1.2B base model and fine-tuned using ophthalmology-relevant question-answer pairs extracted from the MIRIADโ€‘4.4M dataset.


๐Ÿ“Œ Model Use

  • โœ”๏ธ Good at: Direct factual question answering related to ophthalmology (short-form)
  • โš ๏ธ Not suitable for: Multi-choice QA or reasoning-based multi-step inference
  • โŒ Do not use for: Real clinical decision-making or diagnostic support

โš ๏ธ Important Caution (from MIRIAD Dataset)

This model was trained using a filtered subset of the MIRIADโ€‘4.4M dataset. While the dataset was curated using LLMs and rule-based methods, no licensed medical professionals reviewed it. As noted by the dataset authors:

โ€œDespite quality filtering, hallucinations or factual errors may remain. The dataset should not be used to train clinical decision tools or models intended for real medical use.โ€


๐Ÿ”ง How to Use

from transformers import AutoTokenizer, AutoModelForCausalLM
import torch

model_id = "yasserrmd/OphthaScholar"
tokenizer = AutoTokenizer.from_pretrained(model_id, trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained(model_id, 
             device_map="auto", 
             torch_dtype=torch.float16 if torch.cuda.is_available() else torch.float32,
             trust_remote_code=True).eval()

question = "What is the treatment for neovascular age-related macular degeneration?"
prompt = f"System: You are an ophthalmology expert.\nUser: {question}\nAnswer:"

inputs = tokenizer(prompt, return_tensors="pt").to(model.device)
outputs = model.generate(**inputs, max_new_tokens=64, do_sample=False)
print(tokenizer.decode(outputs[0], skip_special_tokens=True).split("Answer:")[-1].strip())

๐Ÿง  Suggested Applications

  • โœ… Clinical education (ophthalmology)
  • โœ… Dataset generation for supervised fine-tuning
  • โœ… Question answering for medical research chatbots

โŒ Not Intended For

  • โŒ Clinical use
  • โŒ Diagnostic automation
  • โŒ Emergency decision tools

๐Ÿ“– Citation

@misc{ophthascholar2025,
  title={OphthaScholar-1.2B: A factual ophthalmology model fine-tuned on MIRIAD},
  author={Yasserrmd},
  year={2025},
  url={https://huggingface.co/yasserrmd/OphthaScholar}
}

๐Ÿ“Ž License

  • Base: MIT (Liquid LFM-1.2B)
  • Data: MIRIAD-4.4M (filtered use only; not for clinical deployment)

Downloads last month
11
Safetensors
Model size
1.17B params
Tensor type
BF16
ยท
F16
ยท
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support

Model tree for yasserrmd/OphthaScholar-1.2B

Base model

LiquidAI/LFM2-1.2B
Finetuned
(17)
this model