All documents

Model Fine-Tuning

    Summary
    Summary
    Updated on 05 Nov 2025

    The Japan Advanced Institute of Science and Technology (JAIST), a leading national research university, required a robust and scalable infrastructure to build a state-of-the-art Large Language Model (LLM) specifically for the Japanese language. Their goal was to conduct extensive experimentation, from optimal data combination discovery to large-scale continual pre-training, demanding significant computational power and a streamlined MLOps platform.

    Partnering with FPT AI Factory, JAIST leveraged a comprehensive suite of services, including FPT AI Studio and FPT AI Inference, to accelerate their research and development pipeline. This collaboration enabled JAIST to systematically identify the best data mixtures, execute multiple large-scale continual pre-training phases on massive datasets, and efficiently evaluate model performance. By offloading the complexities of GPU infrastructure management to FPT AI Factory, JAIST's research team could focus on their core mission: advancing the frontiers of natural language processing for Japanese.