Neural Compressor: Boosting AI Model Efficiency

In the age of the AI PC, AI-infused applications will become the norm, and developers are increasingly replacing traditional code fragments with AI models. This accelerating trend is unleashing exciting user experiences, enhancing productivity, providing new tools for creators, and enabling seamless and natural collaborative experiences.

To meet the computing demand for these models, AI PCs are providing the foundational computing blocks to enable these AI experiences with the combination of CPU, GPU (Graphics Processing Unit), and NPU (Neural Processing Unit). However, to fully take advantage of an AI PC and each of these computing engines to provide the most optimal user experiences, developers need to compress these AI models, a non-trivial task. To help tackle this problem, Intel is proud to announce we are embracing the open-source community and have made the Neural Compressor utility available under the ONNX project.

Ce contenu a été publié dans Non classé. Vous pouvez le mettre en favoris avec ce permalien.