Adapting open-source LLMs to Nepali using parameter‑efficient techniques (LoRA/QLoRA) and carefully scoped distributed training on Tribhuvan University High Performance Computing. Focus: resource‑aware methods that preserve quality while enabling reproducibility and open access.
Most state‑of‑the‑art LLMs are trained on high‑resource languages; Nepali remains under‑served despite tens of millions of speakers. This project adapts an open‑source LLM to the Nepali domain using parameter‑efficient fine‑tuning (LoRA/QLoRA) with quantization and memory‑aware optimizations, enabling single‑GPU training on Tribhuvan University High Performance Computing System while preserving quality. We emphasize reproducibility and open release of models and code as NepaliGPT.
Illustrative placeholder. Replace with your TU‑HPC measurements.
Replace with measured outcomes and ablations.
| ID | Base Model | Tokenizer | Method | Seq Len | Batch | Epochs | PPL | Notes |
|---|---|---|---|---|---|---|---|---|
| E‑FT‑7B‑01 | 7B | SP‑base | Full FT | 1024 | 1× GA8 | 1 | — | Baseline |
| E‑LoRA‑13B‑02 | 13B | SP‑base | LoRA r=8 | 1024 | 2× GA16 | 2 | — | Attn+MLP |
| E‑QLoRA‑33B‑03 | 33B | SP‑ext | QLoRA r=16 | 1024 | 2× GA32 | 1 | — | NF4 + paged opt |
Author: Aatiz Ghimire (MSc Data Science, Tribhuvan University)
Advisor: Dr. Madhav Prasad Ghimire (Associate Professor, Central Department of Physics),
Siman Giri(Center of AI, Herald College Kathmandu)
Infrastructure: Tribhuvan University High Performance Computing