-
-
Notifications
You must be signed in to change notification settings - Fork 2.8k
fix(intel): Set GPU vendor on Intel images and cleanup #5945
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
✅ Deploy Preview for localai ready!
To edit notification comments on pull requests, go to your Netlify project configuration. |
@@ -109,24 +109,15 @@ jobs: | |||
skip-drivers: 'false' | |||
makeflags: "--jobs=4 --output-sync=target" | |||
aio: "-aio-gpu-vulkan" | |||
- build-type: 'sycl_f16' | |||
- build-type: 'sycl' |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
shall we simply put intel
here? this is not used anywhere else now, backends have their own build_type
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I dunno, Intel seem to have multiple options and I'm not sure if oneapi would cover all of them.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
but it's what Intel suggest and are the basic dependency to get going with an intel GPU (common enough for all the other backends). I'd really go simple here and name it intel
. Users don't have to learn internals until they want to.
@@ -166,7 +162,7 @@ docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-ai | |||
docker run -ti --name local-ai -p 8080:8080 --gpus all localai/localai:latest-aio-gpu-nvidia-cuda-11 | |||
|
|||
# Intel GPU version | |||
docker run -ti --name local-ai -p 8080:8080 localai/localai:latest-aio-gpu-intel-f16 |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I think we need to update also the https://github.com/mudler/LocalAI/blob/master/docs/static/install.sh script, and I guess documentation too. Maybe better to grep the codebase
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I did, but not for the right thing I think.
I'm a bit confused about the AIO image and how this works now. Does it include the backends or are these downloaded? I can't see how that would happen. If they are included then it would make sense to keep the f16 and f32 versions perhaps, but presently I have removed them.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
all downloaded during start, there aren't anymore backends in any image 👍
Signed-off-by: Richard Palethorpe <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thanks!
Description
Set /var/localai/capability so that the SYCL backends are downloaded. Also remove the f16 and f32 builds for the core image because they are now identical, only the GGML based backends have this distinction.
Also some other bits have been removed and cleaned up.
Notes for Reviewers
Signed commits