-
-
Articles récents
- Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
- Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
- KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference
- Scaling AI with Confidence: Lenovo’s Approach to Responsible and Practical Adoption
- Unlocking AI-Driven Media Monetization with Intel® Xeon® CPUs and Broadpeak BannersIn2
-
Neural networks news
Intel NN News
- Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
The vLLM (Virtualized Large Language Model) framework, optimized for CPU inference, is emerging as […]
- Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
The United Nations (UN) has taken a bold step toward digital sovereignty by developing an […]
- KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference
Developed by Intel, KVCrush can improve LLM inference throughput up to 4x with less than 1% […]
- Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
-
Archives de catégorie : Non classé
Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
The United Nations (UN) has taken a bold step toward digital sovereignty by developing an open-source AI infrastructure in collaboration with Intel and the Open Platform for Enterprise AI (OPEA).
Publié dans Non classé
Commentaires fermés sur Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
The vLLM (Virtualized Large Language Model) framework, optimized for CPU inference, is emerging as a powerful solution for efficiently serving large language models (LLMs).
Publié dans Non classé
Commentaires fermés sur Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference
Developed by Intel, KVCrush can improve LLM inference throughput up to 4x with less than 1% accuracy drop.
Publié dans Non classé
Commentaires fermés sur KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference
Scaling AI with Confidence: Lenovo’s Approach to Responsible and Practical Adoption
In the race to operationalize AI, success depends not on flashy pilots, but on turning experimentation into measurable business value. According to David Ellison, Chief Data Scientist and Director of AI Engineering at Lenovo, the most successful AI projects start … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur Scaling AI with Confidence: Lenovo’s Approach to Responsible and Practical Adoption
Unlocking AI-Driven Media Monetization with Intel® Xeon® CPUs and Broadpeak BannersIn2
In this article, we will cover how to deploy high-performance AI inferencing for media data curation and retrieval-augmented generation (RAG) without requiring discrete GPUs.
Publié dans Non classé
Commentaires fermés sur Unlocking AI-Driven Media Monetization with Intel® Xeon® CPUs and Broadpeak BannersIn2
AI at the Edge: Intel’s Vision for Real-World Impact
When it comes to scaling AI, the conversation isn’t only about the cloud—it’s about the edge. According to Matthew Formica, Senior Director and Head of Edge Product Marketing & AI PC/Edge AI Software Developer Relations at Intel, the edge represents … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur AI at the Edge: Intel’s Vision for Real-World Impact
Intel® Xeon® Processors: The Most Preferred CPU for AI Host Nodes
Today’s AI workloads are not purely offloaded to GPU accelerators. Host CPUs such as the Intel® Xeon® 6 processors play a significant role in maximizing the performance of AI-accelerated systems.
Publié dans Non classé
Commentaires fermés sur Intel® Xeon® Processors: The Most Preferred CPU for AI Host Nodes
Building AI With Empathy: Sorenson’s Mission for Accessibility
For Sorenson Senior Director of AI Mariam Rahmani, the future of AI isn’t about building the flashiest models—it’s about creating solutions that close communication gaps and empower people, especially the Deaf and Hard of Hearing community. With a third of … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur Building AI With Empathy: Sorenson’s Mission for Accessibility
Multi-node deployments using Intel® AI for Enterprise RAG
As enterprises scale generative AI across diverse infrastructures, Intel® AI for Enterprise RAG solution delivers a modular, hardware-aware framework for Retrieval-Augmented Generation (RAG) optimized for Kubernetes and Intel platforms. With intelligent scheduling, NUMA-aware resource isolation, and dynamic scaling, it ensures predictable, … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur Multi-node deployments using Intel® AI for Enterprise RAG
Connected Data is the Future: How Neo4j Is Enabling the Next Generation of AI
In the evolving landscape of artificial intelligence, connected data is becoming a core competitive advantage.
Publié dans Non classé
Commentaires fermés sur Connected Data is the Future: How Neo4j Is Enabling the Next Generation of AI