in

Huawei Zurich Lab’s New Open-Source Tech Lets LLMs Run on Consumer GPUs

Huawei Zurich Lab’s New Open-Source Tech Lets LLMs Run on Consumer GPUs

Huawei’s Zurich Computing Systems Laboratory has released SINQ (Sinkhorn Normalization Quantization), an open-source quantization method that reduces the memory requirements of large language models (LLMs) by up to 70%. The breakthrough allows workloads that once needed enterprise GPUs like Nvidia’s A100 or H100 to run efficiently on consumer-grade cards such as the RTX 4090, cutting both hardware and cloud compute costs.

The Apache 2.0–licensed project is now available on GitHub and Hugging Face for free use and commercialization. Huawei says SINQ achieves accuracy close to data-calibrated approaches while outperforming other calibration-free methods such as RTN and HQQ in both speed and precision. [TechNode reporting]

Report

What do you think?

Newbie

Written by Mr Viral

Leave a Reply

Your email address will not be published. Required fields are marked *

GIPHY App Key not set. Please check settings

17 Best Beauty Advent Calendars Worth Buying Before December 2025

17 Best Beauty Advent Calendars Worth Buying Before December 2025

China’s “Rat-Dried” Plush Toys Go Viral: How Ugly Dolls Became a Cultural Mirror for Gen Z

China’s “Rat-Dried” Plush Toys Go Viral: How Ugly Dolls Became a Cultural Mirror for Gen Z