Open-Source GPT Model Trained at Wafer Scale eetimes.com - get the latest breaking news, showbiz & celebrity photos, sport news & rumours, viral videos and top stories from eetimes.com Daily Mail and Mail on Sunday newspapers.
Cerebras Systems has trained and is releasing a series of seven GPT-based large language models (LLM) for open use by the research community, according to the company. This is the first time a company has used non-GPU based AI systems to train LLMs up to 13 billion parameters and is sharing the models, weights, and training recipe via the industry standard Apache 2.0 license. All seven models were trained on the 16 CS-2 systems in the Cerebras Andromeda AI supercomputer.
Silicon Valley-based Cerebras released seven models all trained on its AI supercomputer called Andromeda, including smaller 111 million parameter language models to a larger 13 billion parameter model.
As the Only European Provider of Cerebras Cloud, Green AI Cloud Delivers AI Super Compute in an Easy-to-Use Solution for AI and ML Applications, Data Science and Simulation Workloads Cerebras Systems