Austrian Research Company NXAI releases its new xLSTM 7B model, once again highlighting the efficiency advantage and performance of the xLSTM architecture.
The new pre-trained model plays in the premier league of 7B models even without fine-tuning. It evolved into the best non-transformer large language model as demonstrated by its next word prediction performance and evaluations on standard benchmarks like Massive Multitask Language Understanding (MMLU). However, the main feature of the 7B xLSTM model is its speed. It is much more efficient than other large language models, that is, it generates results with considerably less compute resources and much faster. Consequently, the 7B xLSTM is the champion with respect to speed and energy efficiency.
The xLSTM architecture is available on Github at NXAI and a pre-trained model is available on Huggingface for fine-tuning for the developer communities.