Researchers at Intel Labs, in collaboration with Xiamen University, have presented LiSA, the first semantic aware AI framework for highly accurate 3D visual localization. LiSA can leverage pre-trained semantic segmentation models to significantly improve state-of-the-art 3D visual localization accuracy, without introducing computational overhead during inference. LiSA was presented at CVPR 2024 as a highlight paper, an award given to the top 3.6% of conference papers.
-
-
Articles récents
- Accelerating Llama 3.3-70B Inference on Intel® Gaudi® 2 via Hugging Face Text Generation Inference
- Exploring Vision-Language Models (VLMs) with Text Generation Inference on Intel® Data Center GPU Max
- A Journey Towards Approaching “Why” Question-Answering for Video
- From Infrastructure to Impact: How Dell is Scaling AI
- Intel Labs’ Kid Space Conversational AI Facilitates Collaborative Problem-Solving Among Students
-
Neural networks news
Intel NN News
- Accelerating Llama 3.3-70B Inference on Intel® Gaudi® 2 via Hugging Face Text Generation Inference
Learn how to deploy Llama 3.3-70B on Intel® Gaudi® 2 AI accelerators using Hugging Face TGI, with […]
- Exploring Vision-Language Models (VLMs) with Text Generation Inference on Intel® Data Center GPU Max
Supercharge VLM deployment with TGI on Intel XPUs. This guide shows how to set up, optimize, and […]
- Evaluating Trustworthiness of Explanations in Agentic AI Systems
Intel Labs research published at the ACM CHI 2025 Human-Centered Explainable AI Workshop found that […]
- Accelerating Llama 3.3-70B Inference on Intel® Gaudi® 2 via Hugging Face Text Generation Inference
-