The Chinese Academy of Sciences' Institute of Automation in Beijing has recently unveiled its newest SpikingBrain 1.0 large language model. This LLM is claimed to be the world's first "brain-like" LLM, designed to use significantly less energy than conventional AI systems, such as ChatGPT. Rather than relying on Nvidia hardware, it operates entirely on Chinese-made MetaX chips, marking a significant advancement in neuromorphic computing.
The system utilizes "spiking computation" technology, which mimics the human brain's neuron firing patterns. This technology enables only the necessary neurons to activate, rather than the entire network, as in traditional models. This event-driven approach also keeps the system energy-efficient, reducing power consumption. The system can also learn from a minuscule two percent of the training data required compared to conventional systems. Two versions have been developed: one with 7 billion and another with 76 billion parameters.
The researchers report up to a whopping 100x faster performance compared to traditional models in certain tasks, with the smaller model responding to a 4-million-token prompt at least 100 times faster than standard systems. Additionally, there's a 26.5-fold speed improvement over conventional Transformer architectures for first token generation. The new model was trained on approximately 150 billion tokens, just a fraction of the typical requirements for conventional systems. Despite this reduced training data, the system still performs comparably well to other popular open-source alternatives.
The new system holds strategic significance for China, given that the LLM operates entirely within China's homegrown AI ecosystem, utilizing the MetaX chip platform. This becomes particularly important as the U.S. tightens export controls on advanced AI chips. The research also demonstrates that it is possible to train an efficient large model on non-Nvidia platforms. Lead researcher Li Guoqi also highlights optimization for Chinese chip architecture. Potential applications for such systems include legal documents, medical records, and even scientific simulations.
The team has open-sourced the smaller version of the model for public access, with the larger version only available online for public testing via a demo site. The research is also published as a non-peer-reviewed paper on the arXiv repository, so it's best to take all these claims with a grain of salt. Nevertheless, this breakthrough could pave the way for more energy-efficient AI systems and advance brain-inspired computing approaches.
Source(s)
SCMP (in English)






