Programmable AI chip maker Quadrics is making waves in the tech industry with its innovative approach to reducing power consumption in artificial intelligence systems. The company's latest development, the Kolmogorov Arnold Network (KAN), is set to revolutionize the way AI frameworks operate.
The KAN framework eliminates matrix multiply operations, a significant breakthrough that can lead to a drastic reduction in AI power consumption. This approach is reminiscent of a method pioneered at the University of California Santa Cruz, which also focused on removing matrix multiply operations to enhance energy efficiency for large language models.
Researchers from MIT and CalTech recently published a research paper showcasing the KAN framework and its potential to transform machine learning networks. Early analysis indicates that KANs could be significantly smaller in size compared to traditional transformer-based models while delivering comparable results.
One of the key advantages of KAN is its ability to execute a vast number of univariate functions, such as polynomials, with minimal matrix multiplication operations. This streamlined process not only improves efficiency but also reduces the memory requirements for edge AI applications.
Quadrics' Chimera general-purpose NPU is designed to support KANs, along with the necessary matrix multiplication hardware for running conventional neural networks efficiently. The processor boasts a massively parallel array of general-purpose, C++ programmable ALUs, making it versatile for various machine learning models.
Quadrics' CEO, Veerbhan Kheterpal, emphasized the importance of staying ahead of the curve in the rapidly evolving field of machine learning. The company's forward-thinking approach with the Chimera GPNPU architecture ensures that licensees are well-equipped to tackle future challenges in AI development.