Run with container images
LocalAI provides a variety of images to support different environments. These images are available on quay.io and Docker Hub.
All-in-One images comes with a pre-configured set of models and backends, standard images instead do not have any model pre-configured and installed.
For GPU Acceleration support for Nvidia video graphic cards, use the Nvidia/CUDA images, if you don’t have a GPU, use the CPU images. If you have AMD or Mac Silicon, see the build section.
Available Images Types:
Images ending with
-core
are smaller images without predownload python dependencies. Use these images if you plan to usellama.cpp
,stablediffusion-ncn
,tinydream
orrwkv
backends - if you are not sure which one to use, do not use these images.Images containing the
aio
tag are all-in-one images with all the features enabled, and come with an opinionated set of configuration.FFMpeg is not included in the default images due to its licensing. If you need FFMpeg, use the images ending with
-ffmpeg
. Note thatffmpeg
is needed in case of usingaudio-to-text
LocalAI’s features.If using old and outdated CPUs and no GPUs you might need to set
REBUILD
totrue
as environment variable along with options to disable the flags which your CPU does not support, however note that inference will perform poorly and slow. See also flagset compatibility.
All-in-one images
All-In-One images are images that come pre-configured with a set of models and backends to fully leverage almost all the LocalAI featureset. These images are available for both CPU and GPU environments. The AIO images are designed to be easy to use and requires no configuration. Models configuration can be found here separated by size.
In the AIO images there are models configured with the names of OpenAI models, however, they are really backed by Open Source models. You can find the table below
Category | Model name | Real model (CPU) | Real model (GPU) |
---|---|---|---|
Text Generation | gpt-4 | phi-2 | hermes-2-pro-mistral |
Multimodal Vision | gpt-4-vision-preview | bakllava | llava-1.6-mistral |
Image Generation | stablediffusion | stablediffusion | dreamshaper-8 |
Speech to Text | whisper-1 | whisper with whisper-base model | <= same |
Text to Speech | tts-1 | en-us-amy-low.onnx from rhasspy/piper | <= same |
Embeddings | text-embedding-ada-002 | all-MiniLM-L6-v2 in Q4 | all-MiniLM-L6-v2 |
Usage
Select the image (CPU or GPU) and start the container with Docker:
# CPU example
docker run -p 8080:8080 --name local-ai -ti localai/localai:latest-aio-cpu
LocalAI will automatically download all the required models, and the API will be available at localhost:8080.
Available images
Description | Quay | Docker Hub |
---|---|---|
Latest images for CPU | quay.io/go-skynet/local-ai:latest-aio-cpu | localai/localai:latest-aio-cpu |
Versioned image (e.g. for CPU) | quay.io/go-skynet/local-ai:v2.17.1-aio-cpu | localai/localai:v2.17.1-aio-cpu |
Latest images for Nvidia GPU (CUDA11) | quay.io/go-skynet/local-ai:latest-aio-gpu-nvidia-cuda-11 | localai/localai:latest-aio-gpu-nvidia-cuda-11 |
Latest images for Nvidia GPU (CUDA12) | quay.io/go-skynet/local-ai:latest-aio-gpu-nvidia-cuda-12 | localai/localai:latest-aio-gpu-nvidia-cuda-12 |
Latest images for AMD GPU | quay.io/go-skynet/local-ai:latest-aio-gpu-hipblas | localai/localai:latest-aio-gpu-hipblas |
Latest images for Intel GPU (sycl f16) | quay.io/go-skynet/local-ai:latest-aio-gpu-intel-f16 | localai/localai:latest-aio-gpu-intel-f16 |
Latest images for Intel GPU (sycl f32) | quay.io/go-skynet/local-ai:latest-aio-gpu-intel-f32 | localai/localai:latest-aio-gpu-intel-f32 |
Available environment variables
The AIO Images are inheriting the same environment variables as the base images and the environment of LocalAI (that you can inspect by calling --help
). However, it supports additional environment variables available only from the container image
Variable | Default | Description |
---|---|---|
PROFILE | Auto-detected | The size of the model to use. Available: cpu , gpu-8g |
MODELS | Auto-detected | A list of models YAML Configuration file URI/URL (see also running models) |
Standard container images
Standard container images do not have pre-installed models.
Images are available with and without python dependencies. Note that images with python dependencies are bigger (in order of 17GB).
Images with core
in the tag are smaller and do not contain any python dependencies.
See Also
Last updated 15 Jun 2024, 20:08 UTC .