-
-
Articles récents
- Next-Gen AI Inference: Intel® Xeon® Processors Power Vision, NLP, and Recommender Workloads
- Document Summarization: Transforming Enterprise Content with Intel® AI for Enterprise RAG
- AutoRound Meets SGLang: Enabling Quantized Model Inference with AutoRound
- In-production AI Optimization Guide for Xeon: Search and Recommendation Use Case
- Argonne’s Aurora Supercomputer Helps Power Breakthrough Simulations of Quantum Materials
-
Neural networks news
Intel NN News
- Next-Gen AI Inference: Intel® Xeon® Processors Power Vision, NLP, and Recommender Workloads
Intel® Xeon® processors can deliver a CPU-first platform built for modern AI workloads without […]
- Document Summarization: Transforming Enterprise Content with Intel® AI for Enterprise RAG
Transform enterprise documents into insights with Document Summarization, optimized for Intel® […]
- AutoRound Meets SGLang: Enabling Quantized Model Inference with AutoRound
We are thrilled to announce an official collaboration between SGLang and AutoRound, enabling […]
- Next-Gen AI Inference: Intel® Xeon® Processors Power Vision, NLP, and Recommender Workloads
-
Archives mensuelles : septembre 2025
Powering AI-Driven Workflows: Intel’s Vision for the Future of Professional Computing
The race to operationalize AI isn’t about who’s fastest—it’s about who’s smartest. While the tech world obsesses over processing speeds and benchmark scores, enterprise leaders are asking different questions: Can this AI system actually integrate with our existing workflows? Will it … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur Powering AI-Driven Workflows: Intel’s Vision for the Future of Professional Computing
Deploying Deepseek models on Intel® Gaudi® accelerators using vLLM
Deepseek is a model that utilizes Deepseek Mixture of Experts (MoE) and Multi-Head Latent Attention (MLA). Weights are natively stored in FP8 with block quantization scales.It comes in two forms: V3, which is a standard model, and R1, which is … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur Deploying Deepseek models on Intel® Gaudi® accelerators using vLLM
Powering Agentic AI with CPUs: LangChain, MCP, and vLLM on Google Cloud
Google Cloud now offers access to Intel® Xeon® 6 CPUs. These processors are well-suited for a variety of workloads, including agentic AI systems.
Publié dans Non classé
Commentaires fermés sur Powering Agentic AI with CPUs: LangChain, MCP, and vLLM on Google Cloud
How Intel Creates Better AI Video Understanding with Scene Graph Technology
Intel researchers developed EASG-Bench, a new benchmark with more than 1,800 question-answer pairs that tests how well AI models understand what happens in first-person videos by using structured scene graphs instead of narrative video descriptions
Publié dans Non classé
Commentaires fermés sur How Intel Creates Better AI Video Understanding with Scene Graph Technology
Intel® Xeon 6® Processors: Delivering High Throughput and Low Latency with Data Center LLMs
When optimizing LLM workloads, hardware is only one piece of the equation. Intel brings decades of experience working with software developers, building a strong ecosystem to optimize software for Intel Xeon processors.
Publié dans Non classé
Commentaires fermés sur Intel® Xeon 6® Processors: Delivering High Throughput and Low Latency with Data Center LLMs
AI PCs and the Future of Cybersecurity: AI-Powered Protection from Deepfakes
The rise of deepfakes has introduced a new dimension of risk in today’s digital landscape. What started as fringe technology is now a mainstream concern for consumers, businesses, and governments alike. For Leena Elias, Chief Product Officer at Gen Digital, … Continuer la lecture
Publié dans Non classé
Commentaires fermés sur AI PCs and the Future of Cybersecurity: AI-Powered Protection from Deepfakes
Powering Agentic AI with CPUs: LangChain, MCP, and vLLM on Google Cloud
With the launch of the C4 series, Google Cloud now offers access to Intel® Xeon® 6 processor with P-cores which are well-suited for a variety of workloads, including agentic AI systems.
Publié dans Non classé
Commentaires fermés sur Powering Agentic AI with CPUs: LangChain, MCP, and vLLM on Google Cloud
Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
The United Nations (UN) has taken a bold step toward digital sovereignty by developing an open-source AI infrastructure in collaboration with Intel and the Open Platform for Enterprise AI (OPEA).
Publié dans Non classé
Commentaires fermés sur Building a Sovereign GenAI Stack for the United Nations with Intel and OPEA
Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
The vLLM (Virtualized Large Language Model) framework, optimized for CPU inference, is emerging as a powerful solution for efficiently serving large language models (LLMs).
Publié dans Non classé
Commentaires fermés sur Accelerating vLLM Inference: Intel® Xeon® 6 Processor Advantage over AMD EPYC
KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference
Developed by Intel, KVCrush can improve LLM inference throughput up to 4x with less than 1% accuracy drop.
Publié dans Non classé
Commentaires fermés sur KVCrush: Rethinking KV Cache Alternative Representation for Faster LLM Inference