Get the latest Science News and Discoveries

Release of “Fugaku-LLM” – a large language model trained on the supercomputer “Fugaku” - EurekAlert


A team of researchers in Japan released Fugaku-LLM, a large language model&nbsp;with enhanced Japanese language capability, using the RIKEN supercomputer Fugaku.&nbsp;&nbsp; <p>To train large language models on Fugaku, the researchers developed distributed training methods, including porting the deep learning framework Megatron-DeepSpeed to Fugaku in order to optimize the performance of Transformers on Fugaku. They accelerated the dense matrix multiplication library for Transformers, and optimized communication performance for Fugaku by combining three types of parallelization techniques and accelerated the collective communication library on the Tofu interconnect D.</p>

None

Get the Android app

Or read this on Eureka Alert

Read more on:

Photo of LLM

LLM

Photo of large language model

large language model

Photo of Fugaku” - EurekAlert

Fugaku” - EurekAlert

Related news:

News photo

It Takes a Supercomputer to Properly Simulate a Neutron Star’s Surface

News photo

Scientists Ignited a Thermonuclear Explosion Inside a Supercomputer

News photo

Scientists Use Supercomputer To Unravel Mysteries of Dark Matter and the Universe’s Evolution