-
-
Notifications
You must be signed in to change notification settings - Fork 3k
feat: update ROCM and use smaller image #2196
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
✅ Deploy Preview for localai canceled.
|
bfb3b07 to
f043597
Compare
|
I think it should be safe - however I don't have an AMD card to test this on, so we might need to collect feedback from master. Maybe @jtwolfe can give this a try? |
Signed-off-by: Chris Jowett <[email protected]>
f043597 to
3854749
Compare
Signed-off-by: Chris Jowett <[email protected]>
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
We have a new release out, I'd say that's good to test on master - thanks @cryptk for this!
|
Tested latest commit, and was able to get the Docker image built and running successfully. Tested multiple models on my Radeon RX 7900XT successfully. I used the sample prompts from the Getting Started page. Note that I did restart the local-ai container between tests, as I saw some errors in sequential tests due to RAM limitations. ✅ Text Generation: ✅ Image Preview: ✅ Function Calling:
✅ Image Generation: ✅ Text to Speech:
❌ Speech to Text:
|
|
Logged this issue for the hipblas-dev package missing the ldconfig trigger: ROCm/ROCm#3081 |
nice! thanks @linuxtek-canada for testing, just in time! :) |
…5.0@f178386 by renovate (#21846) This PR contains the following updates: | Package | Update | Change | |---|---|---| | [docker.io/localai/localai](https://togithub.com/mudler/LocalAI) | minor | `v2.14.0` -> `v2.15.0` | --- > [!WARNING] > Some dependencies could not be looked up. Check the Dependency Dashboard for more information. --- ### Release Notes <details> <summary>mudler/LocalAI (docker.io/localai/localai)</summary> ### [`v2.15.0`](https://togithub.com/mudler/LocalAI/releases/tag/v2.15.0) [Compare Source](https://togithub.com/mudler/LocalAI/compare/v2.14.0...v2.15.0)  ### 🎉 LocalAI v2.15.0! 🚀 Hey awesome people! I'm happy to announce the release of LocalAI version 2.15.0! This update introduces several significant improvements and features, enhancing usability, functionality, and user experience across the board. Dive into the key highlights below, and don't forget to check out the full changelog for more detailed updates. ##### 🌍 WebUI Upgrades: Turbocharged! ##### 🚀 Vision API Integration The Chat WebUI now seamlessly integrates with the Vision API, making it easier for users to test image processing models directly through the browser interface - this is a very simple and hackable interface in less then 400L of code with Alpine.JS and HTMX!  ##### 💬 System Prompts in Chat System prompts can be set in the WebUI chat, which guide the user through interactions more intuitively, making our chat interface smarter and more responsive.  ##### 🌟 Revamped Welcome Page New to LocalAI or haven't installed any models yet? No worries! The updated welcome page now guides users through the model installation process, ensuring you're set up and ready to go without any hassle. This is a great first step for newcomers - thanks for your precious feedback!  ##### 🔄 Background Operations Indicator Don't get lost with our new background operations indicator on the WebUI, which shows when tasks are running in the background.  ##### 🔍 Filter Models by Tag and Category As our model gallery balloons, you can now effortlessly sift through models by tag and category, making finding what you need a breeze.  ##### 🔧 Single Binary Release LocalAI is expanding into offering single binary releases, simplifying the deployment process and making it easier to get LocalAI up and running on any system. For the moment we have condensed the builds which disables AVX and SSE instructions set. We are also planning to include cuda builds as well. ##### 🧠 Expanded Model Gallery This release introduces several exciting new models to our gallery, such as 'Soliloquy', 'tess', 'moondream2', 'llama3-instruct-coder' and 'aurora', enhancing the diversity and capability of our AI offerings. Our selection of one-click-install models is growing! We pick carefully model from the most trending ones on huggingface, feel free to submit your requests in a github issue, hop to our Discord or contribute by hosting your gallery, or.. even by adding models directly to LocalAI!   Want to share your model configurations and customizations? See the docs: https://localai.io/docs/getting-started/customize-model/ #### 📣 Let's Make Some Noise! A gigantic THANK YOU to everyone who’s contributed—your feedback, bug squashing, and feature suggestions are what make LocalAI shine. To all our heroes out there supporting other users and sharing their expertise, you’re the real MVPs! Remember, LocalAI thrives on community support—not big corporate bucks. If you love what we're building, show some love! A shoutout on social (@​LocalAI_OSS and @​mudler_it on twitter/X), joining our sponsors, or simply starring us on GitHub makes all the difference. Also, if you haven't yet joined our Discord, come on over! Here's the link: https://discord.gg/uJAeKSAGDy Thanks a ton, and.. enjoy this release! *** #### What's Changed ##### Bug fixes 🐛 - fix(webui): correct documentation URL for text2img by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2233](https://togithub.com/mudler/LocalAI/pull/2233) - fix(ux): fix small glitches by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2265](https://togithub.com/mudler/LocalAI/pull/2265) ##### Exciting New Features 🎉 - feat: update ROCM and use smaller image by [@​cryptk](https://togithub.com/cryptk) in [https://github.com/mudler/LocalAI/pull/2196](https://togithub.com/mudler/LocalAI/pull/2196) - feat(llama.cpp): do not specify backends to autoload and add llama.cpp variants by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2232](https://togithub.com/mudler/LocalAI/pull/2232) - fix(webui): display small navbar with smaller screens by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2240](https://togithub.com/mudler/LocalAI/pull/2240) - feat(startup): show CPU/GPU information with --debug by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2241](https://togithub.com/mudler/LocalAI/pull/2241) - feat(single-build): generate single binaries for releases by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2246](https://togithub.com/mudler/LocalAI/pull/2246) - feat(webui): ux improvements by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2247](https://togithub.com/mudler/LocalAI/pull/2247) - fix: OpenVINO winograd always disabled by [@​fakezeta](https://togithub.com/fakezeta) in [https://github.com/mudler/LocalAI/pull/2252](https://togithub.com/mudler/LocalAI/pull/2252) - UI: flag `trust_remote_code` to users // favicon support by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2253](https://togithub.com/mudler/LocalAI/pull/2253) - feat(ui): prompt for chat, support vision, enhancements by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2259](https://togithub.com/mudler/LocalAI/pull/2259) ##### 🧠 Models - fix(gallery): hermes-2-pro-llama3 models checksum changed by [@​Nold360](https://togithub.com/Nold360) in [https://github.com/mudler/LocalAI/pull/2236](https://togithub.com/mudler/LocalAI/pull/2236) - models(gallery): add moondream2 by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2237](https://togithub.com/mudler/LocalAI/pull/2237) - models(gallery): add llama3-llava by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2238](https://togithub.com/mudler/LocalAI/pull/2238) - models(gallery): add llama3-instruct-coder by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2242](https://togithub.com/mudler/LocalAI/pull/2242) - models(gallery): update poppy porpoise by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2243](https://togithub.com/mudler/LocalAI/pull/2243) - models(gallery): add lumimaid by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2244](https://togithub.com/mudler/LocalAI/pull/2244) - models(gallery): add openbiollm by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2245](https://togithub.com/mudler/LocalAI/pull/2245) - gallery: Added some OpenVINO models by [@​fakezeta](https://togithub.com/fakezeta) in [https://github.com/mudler/LocalAI/pull/2249](https://togithub.com/mudler/LocalAI/pull/2249) - models(gallery): Add Soliloquy by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2260](https://togithub.com/mudler/LocalAI/pull/2260) - models(gallery): add tess by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2266](https://togithub.com/mudler/LocalAI/pull/2266) - models(gallery): add lumimaid variant by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2267](https://togithub.com/mudler/LocalAI/pull/2267) - models(gallery): add kunocchini by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2268](https://togithub.com/mudler/LocalAI/pull/2268) - models(gallery): add aurora by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2270](https://togithub.com/mudler/LocalAI/pull/2270) - models(gallery): add tiamat by [@​mudler](https://togithub.com/mudler) in [https://github.com/mudler/LocalAI/pull/2269](https://togithub.com/mudler/LocalAI/pull/2269) ##### 📖 Documentation and examples - docs: updated Transformer parameters description by [@​fakezeta](https://togithub.com/fakezeta) in [https://github.com/mudler/LocalAI/pull/2234](https://togithub.com/mudler/LocalAI/pull/2234) - Update readme: add ShellOracle to community integrations by [@​djcopley](https://togithub.com/djcopley) in [https://github.com/mudler/LocalAI/pull/2254](https://togithub.com/mudler/LocalAI/pull/2254) - Add missing Homebrew dependencies by [@​michaelmior](https://togithub.com/michaelmior) in [https://github.com/mudler/LocalAI/pull/2256](https://togithub.com/mudler/LocalAI/pull/2256) ##### 👒 Dependencies - ⬆️ Update docs version mudler/LocalAI by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2228](https://togithub.com/mudler/LocalAI/pull/2228) - ⬆️ Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2229](https://togithub.com/mudler/LocalAI/pull/2229) - ⬆️ Update ggerganov/whisper.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2230](https://togithub.com/mudler/LocalAI/pull/2230) - build(deps): bump tqdm from 4.65.0 to 4.66.3 in /examples/langchain/langchainpy-localai-example in the pip group across 1 directory by [@​dependabot](https://togithub.com/dependabot) in [https://github.com/mudler/LocalAI/pull/2231](https://togithub.com/mudler/LocalAI/pull/2231) - ⬆️ Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2239](https://togithub.com/mudler/LocalAI/pull/2239) - ⬆️ Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2251](https://togithub.com/mudler/LocalAI/pull/2251) - ⬆️ Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2255](https://togithub.com/mudler/LocalAI/pull/2255) - ⬆️ Update ggerganov/llama.cpp by [@​localai-bot](https://togithub.com/localai-bot) in [https://github.com/mudler/LocalAI/pull/2263](https://togithub.com/mudler/LocalAI/pull/2263) ##### Other Changes - test: check the response URL during image gen in `app_test.go` by [@​dave-gray101](https://togithub.com/dave-gray101) in [https://github.com/mudler/LocalAI/pull/2248](https://togithub.com/mudler/LocalAI/pull/2248) #### New Contributors - [@​Nold360](https://togithub.com/Nold360) made their first contribution in [https://github.com/mudler/LocalAI/pull/2236](https://togithub.com/mudler/LocalAI/pull/2236) - [@​djcopley](https://togithub.com/djcopley) made their first contribution in [https://github.com/mudler/LocalAI/pull/2254](https://togithub.com/mudler/LocalAI/pull/2254) - [@​michaelmior](https://togithub.com/michaelmior) made their first contribution in [https://github.com/mudler/LocalAI/pull/2256](https://togithub.com/mudler/LocalAI/pull/2256) **Full Changelog**: mudler/LocalAI@v2.14.0...v2.15.0 </details> --- ### Configuration 📅 **Schedule**: Branch creation - At any time (no schedule defined), Automerge - At any time (no schedule defined). 🚦 **Automerge**: Enabled. ♻ **Rebasing**: Whenever PR becomes conflicted, or you tick the rebase/retry checkbox. 🔕 **Ignore**: Close this PR and you won't be reminded about this update again. --- - [ ] <!-- rebase-check -->If you want to rebase/retry this PR, check this box --- This PR has been generated by [Renovate Bot](https://togithub.com/renovatebot/renovate). <!--renovate-debug:eyJjcmVhdGVkSW5WZXIiOiIzNy4zNTMuMSIsInVwZGF0ZWRJblZlciI6IjM3LjM1My4xIiwidGFyZ2V0QnJhbmNoIjoibWFzdGVyIiwibGFiZWxzIjpbImF1dG9tZXJnZSIsInVwZGF0ZS9kb2NrZXIvZ2VuZXJhbC9ub24tbWFqb3IiXX0=-->
Description
This is an attempt to not only update ROCM to 6.1, which according to the release notes should be compatible with every card that 6.0 is compatible with, but also to move away from the very very large -complete images and instead use the smaller rocm "base" images, while only pulling in the few individual things that we need from the -complete image.
This should save a couple of GB on the resulting images, as well as considerably speed up builds as we won't need to download an ~4.2GB image, but instead an ~970MB image.
Notes for Reviewers
I did some local testing, and I was able to build images, but I don't have a ROCM compatible GPU to do any in-depth testing/validation with.
Below is the complete list of packages that are installed in the -complete image that are not installed in the "base" image.
Of those, the only ones I found that were required for the build to succeed are hipblas-dev and rocblas-dev which pulls in the following package set including dependencies:
We should do some testing and validation on this change in case there are other packages that are somehow needed at runtime but not needed at build time.
Signed commits