Scalable Vector Search: Deep Dive Series

Vector search is at the core of the AI revolution, and this blog series is here to teach you all about it. Our blog series introduces scalable vector search and dives deeper into sector compression, dimensionality reduction, and Retrieval-Augmented Generation systems.

Publié dans Non classé | Commentaires fermés sur Scalable Vector Search: Deep Dive Series

Leveling Up Your AI Skills in 30 Minutes

 

 

 

Publié dans Non classé | Commentaires fermés sur Leveling Up Your AI Skills in 30 Minutes

Building Agentic AI Foundations: How Intel® Liftoff Startups Are Preparing for the Next GPT Moment

Agentic AI is here: See how Intel® Liftoff startups are building smarter, more autonomous systems that plan, reason, and execute across real-world workflows.

Publié dans Non classé | Commentaires fermés sur Building Agentic AI Foundations: How Intel® Liftoff Startups Are Preparing for the Next GPT Moment

Designing Empathetic AI: The Future of Human-Centered Technology

Ted Shelton, Chief Operating Officer at Inflection AI, discusses how emotionally intelligent AI is transforming business interactions, customer experience, and organizational workflows.

Publié dans Non classé | Commentaires fermés sur Designing Empathetic AI: The Future of Human-Centered Technology

Deploying Llama 4 Scout and Maverick Models on Intel® Gaudi® 3 with vLLM

Learn how to deploy Llama 4 Scout and Maverick models on Intel® Gaudi® 3 using vLLM for efficient, high-performance inference across complex AI tasks.

Publié dans Non classé | Commentaires fermés sur Deploying Llama 4 Scout and Maverick Models on Intel® Gaudi® 3 with vLLM

Intel Labs’ Innovative Low-Rank Model Adaptation Increases Model Accuracy and Compression

Intel Labs’ Neural Low-Rank Adapter Search (NLS) produces accurate models with INT4 weights and is available in OpenVINO’s Neural Network Compression Framework

Publié dans Non classé | Commentaires fermés sur Intel Labs’ Innovative Low-Rank Model Adaptation Increases Model Accuracy and Compression

Running Llama3.3-70B on Intel® Gaudi® 2 with vLLM: A Step-by-Step Inference Guide

Run Llama 3.3-70B efficiently on Intel® Gaudi® 2 using vLLM. Learn setup, configuration, and performance tips for scalable, production-ready inference.

Publié dans Non classé | Commentaires fermés sur Running Llama3.3-70B on Intel® Gaudi® 2 with vLLM: A Step-by-Step Inference Guide

Accelerating Llama 3.3-70B Inference on Intel® Gaudi® 2 via Hugging Face Text Generation Inference

Learn how to deploy Llama 3.3-70B on Intel® Gaudi® 2 AI accelerators using Hugging Face TGI, with practical setup steps and optimization tips.

Publié dans Non classé | Commentaires fermés sur Accelerating Llama 3.3-70B Inference on Intel® Gaudi® 2 via Hugging Face Text Generation Inference

Exploring Vision-Language Models (VLMs) with Text Generation Inference on Intel® Data Center GPU Max

Supercharge VLM deployment with TGI on Intel XPUs. This guide shows how to set up, optimize, and serve blazing-fast models using Intel® Tiber AI Cloud.

Publié dans Non classé | Commentaires fermés sur Exploring Vision-Language Models (VLMs) with Text Generation Inference on Intel® Data Center GPU Max

A Journey Towards Approaching “Why” Question-Answering for Video

Let’s take a super fast journey summarizing the strides taken in an era (2012 to 2025 period) from simple image classification to recent video-LLMs to understand how to proceed with “why” questions in video understanding

Publié dans Non classé | Commentaires fermés sur A Journey Towards Approaching “Why” Question-Answering for Video