BRIEF

on Nota AI

Nota AI Enhances Memory Efficiency of Solar LLM by 72%

Nota AI, a tech company specializing in AI optimization, introduced an innovative quantization method known as "Nota AI MoE Quantization" that reduces the memory usage of Upstage's Solar LLM by 72%. This advance maintains model accuracy and improves processing speed, leading to reduced inference costs.

The new technology, developed under the "Sovereign AI Foundation Model Project" in South Korea, addresses challenges in Mixture of Experts (MoE) architectures. It selectively preserves precision in crucial areas of the model, achieving substantial memory compression without significant performance loss.

Applying this approach to the Solar 100B model reduced its memory from 191.2GB to 51.9GB. Performance remained stable with a Perplexity (PPL) score close to the original model. Moreover, the technology is poised to facilitate deployment in fields like robotics and automotive systems, even on limited GPU infrastructures.

R. P.

Copyright © 2026 FinanzWire, all reproduction and representation rights reserved.
Disclaimer: although drawn from the best sources, the information and analyzes disseminated by FinanzWire are provided for informational purposes only and in no way constitute an incentive to take a position on the financial markets.

Click here to consult the press release on which this article is based

See all Nota AI news