Web Analytics

ollama-intel-arc

⭐ 241 stars English by eleiton

Run Ollama, Stable Diffusion and Automatic Speech Recognition with your Intel Arc GPU

[Blog]

Effortlessly deploy a Docker-based solution that uses Open WebUI as your user-friendly AI Interface and Ollama for integrating Large Language Models (LLM).

Additionally, you can run ComfyUI or SD.Next docker containers to streamline Stable Diffusion capabilities.

You can also run an optional docker container with OpenAI Whisper to perform Automatic Speech Recognition (ASR) tasks.

All these containers have been optimized for Intel Arc Series GPUs on Linux systems by using Intel® Extension for PyTorch.

screenshot

Services

Setup

Run the following commands to start your Ollama instance with Open WebUI
$ git clone https://github.com/eleiton/ollama-intel-arc.git
$ cd ollama-intel-arc
$ podman compose up

Additionally, if you want to run one or more of the image generation tools, run these commands in a different terminal:

For ComfyUI

$ podman compose -f docker-compose.comfyui.yml up
For SD.Next

$ podman compose -f docker-compose.sdnext.yml up

If you want to run Whisper for automatic speech recognition, run this command in a different terminal:

$ podman compose -f docker-compose.whisper.yml up

Validate

Run the following command to verify your Ollama instance is up and running
$ curl http://localhost:11434/
Ollama is running
When using Open WebUI, you should see this partial output in your console, indicating your arc gpu was detected
[ollama-intel-arc] | Found 1 SYCL devices:
[ollama-intel-arc] | |  |                   |                                       |       |Max    |        |Max  |Global |                     |
[ollama-intel-arc] | |  |                   |                                       |       |compute|Max work|sub  |mem    |                     |
[ollama-intel-arc] | |ID|        Device Type|                                   Name|Version|units  |group   |group|size   |       Driver version|
[ollama-intel-arc] | |--|-------------------|---------------------------------------|-------|-------|--------|-----|-------|---------------------|
[ollama-intel-arc] | | 0| [level_zero:gpu:0]|                     Intel Arc Graphics|  12.71|    128|    1024|   32| 62400M|         1.6.32224+14|

Using Image Generation

screenshot screenshot screenshot

Using Automatic Speech Recognition

  podman exec -it  whisper-ipex whisper https://www.lightbulblanguages.co.uk/resources/ge-audio/hobbies-ge.mp3 --device xpu --model small --language German --task transcribe
  [00:00.000 --> 00:08.000]  Ich habe viele Hobbys. In meiner Freizeit mache ich sehr gerne Sport, wie zum Beispiel Wasserball oder Radfahren.
  [00:08.000 --> 00:13.000]  Außerdem lese ich gerne und lerne auch gerne Fremdsprachen.
  [00:13.000 --> 00:19.000]  Ich gehe gerne ins Kino, höre gerne Musik und treffe mich mit meinen Freunden.
  [00:19.000 --> 00:22.000]  Früher habe ich auch viel Basketball gespielt.
  [00:22.000 --> 00:26.000]  Im Frühling und im Sommer werde ich viele Radtouren machen.
  [00:26.000 --> 00:29.000]  Außerdem werde ich viel schwimmen gehen.
  [00:29.000 --> 00:33.000]  Am liebsten würde ich das natürlich im Meer machen.
  podman exec -it  whisper-ipex whisper https://www.lightbulblanguages.co.uk/resources/ge-audio/hobbies-ge.mp3 --device xpu --model small --language German --task translate
  [00:00.000 --> 00:02.000]  I have a lot of hobbies.
  [00:02.000 --> 00:05.000]  In my free time I like to do sports,
  [00:05.000 --> 00:08.000]  such as water ball or cycling.
  [00:08.000 --> 00:10.000]  Besides, I like to read
  [00:10.000 --> 00:13.000]  and also like to learn foreign languages.
  [00:13.000 --> 00:15.000]  I like to go to the cinema,
  [00:15.000 --> 00:16.000]  like to listen to music
  [00:16.000 --> 00:19.000]  and meet my friends.
  [00:19.000 --> 00:22.000]  I used to play a lot of basketball.
  [00:22.000 --> 00:26.000]  In spring and summer I will do a lot of cycling tours.
  [00:26.000 --> 00:29.000]  Besides, I will go swimming a lot.
  [00:29.000 --> 00:33.000]  Of course, I would prefer to do this in the sea.
  podman exec -it  whisper-ipex whisper YOUR_FILE_NAME.mp3 --device xpu --model small --task translate

Updating the containers

If there are new updates in the ipex-llm-inference-cpp-xpu Docker image or in the Open WebUI Docker image, you may want to update your containers to stay up to date.

Before any updates, be sure to stop your containers

$ podman compose down 
Then just run a pull command to retrieve the latest images.

$ podman compose pull
After that, you can run compose up to start your services again.

$ podman compose up

Manually connecting to your Ollama container

You can connect directly to your Ollama container by running these commands:

$ podman exec -it ollama-intel-arc /bin/bash
$ /llm/ollama/ollama -v

My development environment:

References

--- Tranlated By Open Ai Tx | Last indexed: 2026-02-16 ---