CPT (4 generators) — Qwen2.5-14B-Instruct

Qwen2.5-14B-Instruct continually pre-trained on synthetic data from 4 generators (GPT-4.1-mini, GPT-5-nano, GPT-OSS-20B, Qwen3-235B). The CPT-only configuration that already surpasses all frontier models on PCDT-QA.

Test-split accuracy

Benchmark Accuracy
HealthBench-BR 71.1%
PCDT-QA 86.3%

Usage

from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("hugo/protocolos-clinicos-br-cpt-4gen-14b", torch_dtype="auto", device_map="auto")
tok   = AutoTokenizer.from_pretrained("hugo/protocolos-clinicos-br-cpt-4gen-14b")

Intended use & limitations

Research model for studying domain adaptation of LLMs to Brazilian clinical guidelines. Not a certified medical device. Even at the best accuracy reported in the paper, residual errors may involve consequential details (dosages, contraindications). Use only under qualified professional supervision.

Citation

See the paper and code at the project repository:

Code & paper: https://github.com/hugoabonizio/clinical-protocols-br

Downloads last month
26
Safetensors
Model size
15B params
Tensor type
BF16
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for hugo/protocolos-clinicos-br-cpt-4gen-14b

Adapters
1 model

Collection including hugo/protocolos-clinicos-br-cpt-4gen-14b