#Huawei’s Computing Systems Lab introduced #SINQ, an #opensource #quantisationmethod for large language models (#LLMs). SINQ reduces #memoryusage by 60-70% without sacrificing output quality, enabling models to run on less powerful #hardware. The technique, available on GitHub and Hugging Face, uses #dualaxisscaling and #SinkhornKnopp-style #normalisation for improved performance. https://venturebeat.com/ai/huaweis-new-open-source-technique-shrinks-llms-to-make-them-run-on-less?eicker.news #tech #media #news