Skip to content

Conversation

richiejp
Copy link
Collaborator

Description

Set /var/localai/capability so that the SYCL backends are downloaded. Also remove the f16 and f32 builds for the core image because they are now identical, only the GGML based backends have this distinction.

Also some other bits have been removed and cleaned up.

Notes for Reviewers

Signed commits

  • Yes, I signed my commits.

Copy link

netlify bot commented Jul 31, 2025

Deploy Preview for localai ready!

Name Link
🔨 Latest commit 92f2160
🔍 Latest deploy log https://app.netlify.com/projects/localai/deploys/688b92587a77d40008f87009
😎 Deploy Preview https://deploy-preview-5945--localai.netlify.app
📱 Preview on mobile
Toggle QR Code...

QR Code

Use your smartphone camera to open QR code link.

To edit notification comments on pull requests, go to your Netlify project configuration.

@github-actions github-actions bot added kind/documentation Improvements or additions to documentation dependencies labels Jul 31, 2025
@@ -109,24 +109,15 @@ jobs:
skip-drivers: 'false'
makeflags: "--jobs=4 --output-sync=target"
aio: "-aio-gpu-vulkan"
- build-type: 'sycl_f16'
- build-type: 'sycl'
Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

shall we simply put intel here? this is not used anywhere else now, backends have their own build_type

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I dunno, Intel seem to have multiple options and I'm not sure if oneapi would cover all of them.

Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

but it's what Intel suggest and are the basic dependency to get going with an intel GPU (common enough for all the other backends). I'd really go simple here and name it intel. Users don't have to learn internals until they want to.

@@ -166,7 +162,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-ai
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-11

# Intel GPU version
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel-f16
Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think we need to update also the https://github.com/mudler/LocalAI/blob/master/docs/static/install.sh script, and I guess documentation too. Maybe better to grep the codebase

Copy link
Collaborator Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I did, but not for the right thing I think.

I'm a bit confused about the AIO image and how this works now. Does it include the backends or are these downloaded? I can't see how that would happen. If they are included then it would make sense to keep the f16 and f32 versions perhaps, but presently I have removed them.

Copy link
Owner

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

all downloaded during start, there aren't anymore backends in any image 👍

@mudler mudler added breaking-change bug Something isn't working and removed kind/documentation Improvements or additions to documentation dependencies labels Jul 31, 2025
Copy link
Owner

@mudler mudler left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

@mudler mudler merged commit c07bc55 into mudler:master Jul 31, 2025
23 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
breaking-change bug Something isn't working
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants