Premier AI Development Resources
Explore these deployable AI applications or build on top of the open source code.

Generative AI
DeepSeek* and Small Scale Model Performance
DeepSeek-R1, a new reasoning model, bridges the gap between large-scale performance and local deployment by offering advanced reasoning capabilities in a small language model (SLM).
- Advanced Reasoning: Efficiently transfers complex reasoning to SLMs for local deployment.
- Innovative Techniques: Uses Group Relative Policy Optimization (GRPO) and chain-of-thought (CoT) reasoning.
Custom AI Assistant
Use speech recognition and natural language processing (NLP) in a voice-activated tool that can respond with annotated summaries, actionable recommendations, and insight.
- Natural language processing, speech recognition
- Uses Llama 2 or later and Distil-Whisper
- Use cases include healthcare, retail, and finance
ChatQnA Application
This ChatQnA use case performs RAG using LangChain, Redis VectorDB and Text Generation Inference on Intel Gaudi2 or Intel XEON Scalable Processors. The Intel Gaudi2 accelerator supports both training and inference for deep learning models in particular for LLMs.
Intel Models on Hugging Face*
Computer vision is a field within AI that gives machines a human-like ability to see and understand the world. This cutting-edge technology is transforming lives by enabling a wide range of devices to extract meaningful data from digital images and video.
OpenVINO™ Toolkit Notebooks
A collection of ready-to-run Jupyter notebooks for learning and experimenting with the OpenVINO™ Toolkit. The notebooks provide an introduction to OpenVINO basics and teach developers how to leverage our API for optimized deep learning inference.
Use Cases
GenAI has the potential to revolutionize creative industries, enhance content generation processes, and drive innovation across industries and applications.
Training, Tutorials and Certifications
Explore the possibilities of developing on the Intel® Developer Cloud with our curated list of AI tutorials.
LLMs
With the ability to generate text, summarize and translate content, respond to questions, engage in conversations, and perform complex tasks such as solving math problems or reasoning, LLMs have the potential to benefit society at scale.
Llama* 2 Fine-Tuning with Low-Rank Adaptations (LoRA) on Intel® Gaudi®2 AI Accelerator
Learn how to fine-tune Llama 2 70B with DeepSpeed ZeRO-3 and LoRA* techniques on eight Intel® Gaudi® 2 AI accelerators.
GenAI Essentials: LLM Inference
See LLM inference on the latest Intel Data Center GPU Max 1100 using two models; Camel 5B and Open LLaMA 3B v2l.
Custom AI Assistant
Learn how to run Llama 2 inference on Windows* and Windows Subsystem for Linux* (WSL2) with Intel® Arc™ A-Series GPU.
Comprehensive Strategies for Optimization and Deployment with the OpenVINO™ Toolkit
Use the OpenVINO toolkit to compress LLMs, integrate them into AI-assistant applications, and deploy them for maximum performance, whether on edge devices or in the cloud.
Weight-Only Quantization to Improve LLM Inference on AI PCs
Take advantage of weight-only quantization (WOQ) to improve LLM inference using Intel® Extension for PyTorch* on AI PCs featuring Intel® Core™ Ultra processors (formerly code named Meteor Lake).
Retrieval Augmented Generation (RAG)
RAG supplements pretrained models with up-to-date, proprietary, and confidential data from vector databases during inference. This simplifies customization and updating, and enables attribution of generated information to its source.
RAG Inference Engines with LangChain* on CPUs
Develop a unique vantage point on RAG — understanding its role and potential in scalable inference within production-scale LLM deployments.
Create a RAG system using OpenVINO and LlamaIndex
Build AI applications that can reason about private data or data introduced after a model’s cutoff date by augmenting the knowledge of the model with the specific information it needs. The process of bringing the appropriate information and inserting it into the model prompt is known as Retrieval Augmented Generation (RAG).
Building Blocks of RAG: Harness the Value of Your Data with Domain-specific LLMs
Learn how retrieval augmented generation plus Intel hardware and software can help you yield incredible value and performance for your production GenAI models.
Image, Video, and Audio Generation
Image or video generation creates new content based on descriptive text or image input. Audio generation clones and creates voices, sounds, and music from text prompts or input audio. These techniques typically rely on diffusion or transformer models.
Controllable Music Generation with MusicGen and OpenVINO™ Toolkit
MusicGen is a single-stage auto-regressive Transformer model capable of generating high-quality music samples conditioned on text descriptions or audio prompts.
Stable Diffusion Inference on Intel Arc and Data Center GPUs
Set up and run a stable diffusion model using Intel® Arc™ Discrete Graphics and Pytorch or Tensorflow framework.
Text-to-Video Generation with ZeroScope and OpenVINO Toolkit
The ZeroScope model is a free and open-source text-to-video model that can generate realistic and engaging videos from text descriptions.
Text-to-Image Generation with Stable Diffusion* on Intel® Developer Cloud
These models take text as input and generate an image from that text prompt.
Code Generation
GenAI can create or suggest new code snippets using natural language or text prompts. This technology can also translate code from one programming language to another and efficiently test and debug computer code.
Optimize Code Development with LLMs on the Intel Developer Cloud
Generate code more efficiently with optimized LLMs on Intel® Data Center GPU Max Series
GenAI Software and Ecosystem
Data
Data engineering at scale:
Vector embedding:
Vector databases:
Develop
Pretrained models:
Distributed training and fine-tuning:
RAG and agents:
Optimize & Deploy
Model optimization:
Inference serving:
Monitoring: