Skip to content

Practical guide for performing local inference on LLM on a laptop equipped with an Intel Iris GPU, even if you do not have an Nvidia GPU.

Notifications You must be signed in to change notification settings

coela-oss/devino

Repository files navigation

Project Overview

This repository provides an environment for leveraging OpenVINO and OneAPI to enable LLM inference on Intel devices. It includes scripts for setting up an Ubuntu-based environment, converting models to OpenVINO IR format, and deploying an OpenVINO model server.

Repository Structure

  • setup: Provides scripts for setting up a Pytorch-XPU environment using Ubuntu 22 and Poetry. installation.
  • Model Directories: Named according to Hugging Face model IDs, each containing conversion and inference scripts based on the setup environment.
  • playground: Contains sample scripts tested in an OpenVINO 2025 and Ubuntu 24 environment.

Key Features

  • OpenVINO IR Conversion: Converts Hugging Face models to OpenVINO IR format for optimized inference.
  • OpenVINO Model Server (OVMC): Implements an OpenVINO model server for running converted models.
  • GPU Acceleration: Provides performance improvements for inference using Intel GPUs.

Getting Started

  1. Verify Ubuntu Compatibility: Check the appropriate WSL Ubuntu version using intel-gpu-wsl-advisor.
  • intel-gpu-wsl-advisor repo is a prerequisite repository to determine the appropriate Ubuntu version for WSL
  1. Setup Environment: Use the scripts in setup/ to install dependencies and configure Pytorch-XPU.
  2. Convert Models: Run the provided conversion scripts to transform models into OpenVINO IR format.
  3. Deploy Model Server: Install OpenVINO GenAI's OVMC server and execute converted models. by Makefile

References

This repository is under active development, integrating new features for optimized inference and deployment on Intel hardware.

About

Practical guide for performing local inference on LLM on a laptop equipped with an Intel Iris GPU, even if you do not have an Nvidia GPU.

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published