Qihoo360/360-LLaMA-Factory
slowing
significant_divergence
Selected Choose this fork if sequence parallelism is the requirement and you can tolerate upstream lag and merge overhead. Choose upstream if you want the broadest, freshest LlamaFactory feature set and easier maintenance.
xiezhe-24/ChatTS-Training
active
significant_divergence
Choose this fork if you need a ChatTS-oriented training workflow and can accept divergence from upstream. Stay with upstream if you want the widest model/backend support, fresher fixes, and the full documentation/demo surface.
LUMIA-Group/PonderingLM
stale
significant_divergence
Choose this fork if your goal is pondering/continuous-space research and you want the paper's implementation details. Choose upstream if you need the broader, actively maintained fine-tuning platform with current backend support and fewer missing workflows.
g1f1/LLaMA-Factory
stale
significant_divergence
Choose this fork only if its custom training/data behavior matches your needs and you can tolerate maintenance debt. If you want broad model support and current upstream training backend work, upstream is the safer default.
Choose this fork if you want PEFT-focused fine-tuning plus benchmark/evaluation helpers and can tolerate upstream lag. Choose upstream if you need the latest model/backend support and the broadest maintained feature set.
OpenLLM-Ro/LLaMA-Factory
stale
significant_divergence
Prefer this fork if you are explicitly building Romanian LLM workflows and want the fork's curated data/branding/customization. Prefer upstream if you need the newest training backends, broader docs/examples, or maximum compatibility with the fast-moving LlamaFactory ecosystem.
emrecanacikgoz/Medical-Factory
Choose this fork if you want the added dataset package and are comfortable being behind upstream; choose upstream if you need the newest training backends, fixes, and broader model support.
githisw/LLaMA-Factory
stale
significant_divergence
Prefer this fork if your priority is long-sequence training with Ulysses and you want that capability integrated into LlamaFactory. Prefer upstream if you want current model support, broader feature coverage, and lower maintenance risk.
WGS-note/LLaMA-Factory
stale
significant_divergence
Prefer this fork only if you specifically need the channel-loss work and the bundled experimental assets. For general LlamaFactory adoption, upstream is the better choice because this fork is stale and materially behind on recent fixes and backend support.
littepan/LLaMA-Efficient-Tuning
stale
significant_divergence
Choose this fork only if you specifically want the older, simpler PEFT/QLoRA training workflow and its bundled data/fixes. For anyone starting fresh or needing current model, backend, and multimodal support, upstream LlamaFactory is the better default.