I am an Incoming PhD Student at Institute of Science Tokyo (formerly Tokyo Tech), starting in April 2026. My research lies at the intersection of HPC and Machine Learning, specifically focusing on distributed training and low-precision training (FP8/NVFP4) for Large Language Models.
I am a core contributor to the Swallow Project, a Japanese LLM development initiative, where I maintain the pre-training library and lead large-scale training experiments.
- [Mar 2026] I will be presenting Swallow LLM at NVIDIA GTC 2026 in San Jose! 🗣️
- [Jan 2026] My paper "Rewriting Pre-Training Data Boosts LLM Performance in Math and Code" has been accepted to ICLR 2026! 🎉
I am actively looking for Research Internship opportunities in the US If you are interested in my work on LLM systems and low-precision training, please reach out!
- 🌐 Website: https://okoge-kaz.github.io/
- 🎓 Google Scholar: Citations Profile
- 💼 LinkedIn: kazuki-fujii
- 🐦 X (Twitter): @okoge_kaz




