Plaid-1B MedDialog LoRA

LoRA adapter fine-tuned on the OpenMed/MedDialog patient-doctor conversation dataset on top of the Plaid-1B continuous diffusion language model. Final-year-project artifact.

Research demo only. The outputs are not medical advice. Always consult a qualified clinician for real health questions.

What's in this repo

File Purpose
lora_model/adapter_config.json PEFT LoRA config (rank, targets, ...)
lora_model/adapter_model.safetensors LoRA weights (~200 MB)
embedding_matrix.pt Custom token embedding matrix saved at checkpoint time
noise_schedule.pt Learned diffusion noise schedule (if present)
gamma_bounds.pt (gamma_0, gamma_1) bounds (if present)
training_info.json Hyperparameters, step count, dataset split, etc.

The ~5.1 GB Plaid-1B base weights are not included - they are distributed separately by the Plaid authors. At inference time, load the base weights first, then apply this LoRA adapter on top.

Training

  • Base model: Plaid-1B (continuous-latent diffusion transformer, ~1B params)
  • Adapter: LoRA, rank 64, targeting attn_qkv, attn_out, fc1, fc2, and the embedding matrix.
  • Dataset: OpenMed/MedDialog (patient <-> doctor dialogues).
  • Hardware: 1x NVIDIA RTX 3090 (24 GB), float64 precision.
  • Epochs: 10 (this checkpoint = epoch 8, picked for best val loss).

Usage

This adapter is intended to be loaded in the full Plaid training environment (torch, mup, peft, apex, flash_attn). A runnable interactive demo that wires everything together is available in the FYP repo fyp-final-demo (FastAPI backend + Streamlit frontend + cloudflared tunnel).

from huggingface_hub import snapshot_download

path = snapshot_download(repo_id="Senum/plaid-meddialog-lora")
# then load with sample_meddialog.load_model(
#     weights_path="/path/to/plaid1b_weights",
#     checkpoint_path=path,
#     ...
# )

Citation

If you use this artifact in academic work, please also cite the Plaid paper (continuous diffusion language models) and the MedDialog dataset.

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Dataset used to train Senum/plaid-meddialog-lora