Releases: mudler/LocalAI
v2.5.1
Patch release to create /build/models in the container images.
What's Changed
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1562
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1563
Full Changelog: v2.5.0...v2.5.1
v2.5.0
What's Changed
This release adds more embedded models, and shrink image sizes.
You can run now phi-2 ( see here for the full list ) locally by starting localai with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core phi-2
LocalAI accepts now as argument a list of short-hands models and/or URLs pointing to valid yaml file. A popular way to host those files are Github gists.
For instance, you can run llava, by starting local-ai with:
docker run -ti -p 8080:8080 localai/localai:v2.5.0-ffmpeg-core https://raw.githubusercontent.com/mudler/LocalAI/master/embedded/models/llava.yamlExciting New Features 🎉
👒 Dependencies
- deps(conda): use transformers-env with vllm,exllama(2) by @mudler in #1554
- deps(conda): use transformers environment with autogptq by @mudler in #1555
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1558
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1557
Full Changelog: v2.4.1...v2.5.0
v2.4.1
What's Changed
Exciting New Features 🎉
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1546
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1547
- docs: improve getting started by @mudler in #1553
Full Changelog: v2.4.0...v2.4.1
v2.4.0
What's Changed
Bug fixes 🐛
- fix(entrypoint): cd to backend dir before start by @mudler in #1530
- fix(diffusers): add omegaconf dependency by @mudler in #1540
Exciting New Features 🎉
- feat(preload): prepare models in galleries by @mudler in #1515
- feat: share models by url by @mudler in #1522
- feat(prepare): allow to specify additional files to download by @mudler in #1526
- feat: add trimsuffix by @mudler in #1528
👒 Dependencies
- deps(llama.cpp): update and sync grpc server by @mudler in #1527
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1531
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1517
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1518
- docs(phi-2): add example by @mudler in #1525
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1533
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1544
- ci(dockerhub): push images also to dockerhub by @mudler in #1542
Full Changelog: v2.3.1...v2.4.0
v2.3.1
What's Changed
Bug fixes 🐛
- Fix: Set proper Homebrew install location for x86 Macs by @cnatale in #1510
- fix(download): correctly check for not found error by @mudler in #1514
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1509
Other Changes
- Update version.json (2.3.0) by @lunamidori5 in #1511
New Contributors
Full Changelog: v2.3.0...v2.3.1
v2.3.0
What's Changed
Bug fixes 🐛
- fix: exllama2 backend by @sozercan in #1484
- fix: guidance_scale not work in sd by @hibobmaster in #1488
- fix(image-gen): remove default to stablediffusion by @JZacharie in #1500
Exciting New Features 🎉
- env(conda): use transformers for vall-e-x by @mudler in #1481
- feat: add 🐸 coqui by @mudler in #1489
- feat: partial download by @muka in #1486
- feat: add tiny dream stable diffusion support by @M0Rf30 in #1283
👒 Dependencies
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1479
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1478
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1469
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1482
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1483
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1492
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1495
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1496
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1501
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1503
Other Changes
- ⬆️ Update docs version mudler/LocalAI by @localai-bot in #1477
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1480
- docs: add langchain4j integration by @mudler in #1476
- ci(apple): speedups by @mudler in #1471
New Contributors
- @hibobmaster made their first contribution in #1488
- @muka made their first contribution in #1486
- @JZacharie made their first contribution in #1500
Full Changelog: v2.2.0...v2.3.0
v2.2.0
What's Changed
This release brings updates to the backends and includes a fix for recompilation of LocalAI with go-rwkv (#1473). To note, it also tries to reduce the image size by allowing some backends (transformers-based) to share the same environment.
With this release inline templates and models as URLs are supported, for example:
name: mixtral
parameters:
model: https://huggingface.co/TheBloke/Mixtral-8x7B-v0.1-GGUF/resolve/main/mixtral-8x7b-v0.1.Q2_K.gguf
# or huggingface://TheBloke/Mixtral-8x7B-v0.1-GGUF/mixtral-8x7b-v0.1.Q2_K.gguf@main
template:
completion: |
Complete the following: {{.Input}}Exciting New Features 🎉
- docs: automatically track latest versions by @mudler in #1451
- feat: inline templates and accept URLs in models by @mudler in #1452
- feat(conda): share envs with transformer-based backends by @mudler in #1465
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1450
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1455
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1457
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1462
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1461
- ⬆️ Update donomii/go-rwkv.cpp by @localai-bot in #1470
Other Changes
- How To (Updates and Fixes) by @lunamidori5 in #1456
- Revert ":arrow_up: Update donomii/go-rwkv.cpp" by @mudler in #1474
Full Changelog: v2.1.0...v2.2.0
v2.1.0
What's Changed
Breaking Changes 🛠
Bug fixes 🐛
- fix(piper): pin petals, phonemize and espeak by @mudler in #1393
- update(llama.cpp): update server, correctly propagate LLAMA_VERSION by @mudler in #1440
Exciting New Features 🎉
- Added Check API KEYs file to API.go by @lunamidori5 in #1381
- exllama(v2): fix exllamav1, add exllamav2 by @mudler in #1384
- Fix: API Key / JSON Fast Follow #1 by @dave-gray101 in #1388
- feat: add transformers-musicgen backend by @dave-gray101 in #1387
- feat(diffusers): update, add autopipeline, controlnet by @mudler in #1432
- feat(img2vid,txt2vid): Initial support for img2vid,txt2vid by @mudler in #1442
👒 Dependencies
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1378
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1379
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1430
- ⬆️ Update mudler/go-piper by @localai-bot in #1441
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1434
Other Changes
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1385
- docs: site update fixing old image text / How To update updating GPU and CPU docker pages by @lunamidori5 in #1399
- feat: cuda transformers by @mudler in #1401
- feat(entrypoint): optionally prepare extra endpoints by @mudler in #1405
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1390
- ⬆️ Update mudler/go-piper by @localai-bot in #1400
- tests: add diffusers tests by @mudler in #1419
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1418
- How To Updates / Model Used Switched / Removed "docker-compose" (RIP) by @lunamidori5 in #1417
- fix(transformers*): add sentence-transformers and transformers-musicgen tests, fix musicgen wrapper by @mudler in #1420
- extras: add vllm,bark,vall-e-x tests, bump diffusers by @mudler in #1422
- Documentation for Hipblas by @sfxworks in #1425
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1391
- docs: add aikit to integrations by @sozercan in #1412
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1429
- docs(mixtral): add mixtral example by @mudler in #1449
New Contributors
Full Changelog: v2.0.0...v2.1.0
v2.0.0
What's Changed
Breaking Changes 🛠
- 🔥 add LLaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types by @mudler in #1254
- refactor: rename llama-stable to llama-ggml by @mudler in #1287
Bug fixes 🐛
- fix: respect OpenAI spec for response format by @mudler in #1289
- fix: handle grpc and llama-cpp with REBUILD=true by @mudler in #1328
- fix: propagate CMAKE_ARGS when building grpc by @mudler in #1334
- fix(vall-e-x): correctly install reqs in environment by @mudler in #1377
Exciting New Features 🎉
- feat(certificates): add support for custom CA certificates by @vitorstone in #880
- feat(conda): conda environments by @mudler in #1144
- refactor: move backends into the backends directory by @mudler in #1279
- feat: allow to run parallel requests by @mudler in #1290
- feat(transformers): add embeddings with Automodel by @mudler in #1308
- ci(core): add -core images without python deps by @mudler in #1309
- feat: initial watchdog implementation by @mudler in #1341
- feat: update whisper_cpp with CUBLAS, HIPBLAS, METAL, OPENBLAS, CLBLAST support by @wuxxin in #1302
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1231
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1236
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1285
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1288
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1291
Other Changes
- Update .gitignore for backend/llama.cpp by @dave-gray101 in #1235
- llama index example by @sfxworks in #1237
- chianlit example by @sfxworks in #1238
- Fixes the bug 1196 by @diego-minguzzi in #1232
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1242
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1256
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1265
- deps(go-piper): update to 2023.11.6-3 by @M0Rf30 in #1257
- feat(llama.cpp): support lora with scale and yarn by @mudler in #1277
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1272
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1280
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1282
- feat: queue up requests if not running parallel requests by @mudler in #1296
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1297
- fix(api/config): allow YAML config with .yml by @Papawy in #1299
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1300
- llava.yaml (yaml format standardization) by @lunamidori5 in #1303
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1304
- ⬆️ Update mudler/go-piper by @localai-bot in #1305
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1306
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1310
- fix: ExLlama Backend Context Size & Rope Scaling by @ok2sh in #1311
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1313
- docs: Initial import from localai-website by @mudler in #1312
- fix: move python header comments below shebang in some backends by @B4ckslash in #1321
- Feat: OSX Local Codesigning by @dave-gray101 in #1319
- docs: Add llava, update hot topics by @mudler in #1322
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1323
- docs: Update Features->Embeddings page to reflect backend restructuring by @B4ckslash in #1325
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1330
- fix: rename transformers.py to avoid circular import by @mudler in #1337
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1340
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1345
- feat(petals): add backend by @mudler in #1350
- fix: go-piper add libucd at linking time by @M0Rf30 in #1357
- docs: Add docker instructions, add community projects section in README by @mudler in #1359
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1351
- docs: Update getting started and GPU section by @mudler in #1362
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1363
- ci: limit concurrent jobs by @mudler in #1364
- fix/docs: Python backend dependencies by @B4ckslash in #1360
- ci: split into reusable workflows by @mudler in #1366
- fix: OSX Build Fix Part 1: Metal by @dave-gray101 in #1365
- docs: add fine-tuning example by @mudler in #1374
- docs: site/how-to clean up by @lunamidori5 in #1342
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1375
- ⬆️ Update ggerganov/whisper.cpp by @localai-bot in #1227
New Contributors
- @vitorstone made their first contribution in #880
- @sfxworks made their first contribution in #1237
- @diego-minguzzi made their first contribution in #1232
- @M0Rf30 made their first contribution in #1257
- @Papawy made their first contribution in #1299
- @ok2sh made their first contribution in #1311
- @B4ckslash made their first contribution in #1321
- @wuxxin made their first contribution in #1302
Full Changelog: v1.40.0...v2.0.0
v2.0.0_beta
What's Changed
Breaking Changes 🛠
- 🔥 add LLaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types by @mudler in #1254
- refactor: rename llama-stable to llama-ggml by @mudler in #1287
Bug fixes 🐛
- fix: respect OpenAI spec for response format by @mudler in #1289
- fix: handle grpc and llama-cpp with REBUILD=true by @mudler in #1328
- fix: propagate CMAKE_ARGS when building grpc by @mudler in #1334
Exciting New Features 🎉
- feat(certificates): add support for custom CA certificates by @vitorstone in #880
- feat(conda): conda environments by @mudler in #1144
- refactor: move backends into the backends directory by @mudler in #1279
- feat: allow to run parallel requests by @mudler in #1290
- feat(transformers): add embeddings with Automodel by @mudler in #1308
- ci(core): add -core images without python deps by @mudler in #1309
- feat: initial watchdog implementation by @mudler in #1341
👒 Dependencies
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1231
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1236
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1285
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1288
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1291
Other Changes
- Update .gitignore for backend/llama.cpp by @dave-gray101 in #1235
- llama index example by @sfxworks in #1237
- chianlit example by @sfxworks in #1238
- Fixes the bug 1196 by @diego-minguzzi in #1232
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1242
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1256
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1265
- deps(go-piper): update to 2023.11.6-3 by @M0Rf30 in #1257
- feat(llama.cpp): support lora with scale and yarn by @mudler in #1277
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1272
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1280
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1282
- feat: queue up requests if not running parallel requests by @mudler in #1296
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1297
- fix(api/config): allow YAML config with .yml by @Papawy in #1299
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1300
- llava.yaml (yaml format standardization) by @lunamidori5 in #1303
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1304
- ⬆️ Update mudler/go-piper by @localai-bot in #1305
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1306
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1310
- fix: ExLlama Backend Context Size & Rope Scaling by @ok2sh in #1311
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1313
- docs: Initial import from localai-website by @mudler in #1312
- fix: move python header comments below shebang in some backends by @B4ckslash in #1321
- Feat: OSX Local Codesigning by @dave-gray101 in #1319
- docs: Add llava, update hot topics by @mudler in #1322
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1323
- docs: Update Features->Embeddings page to reflect backend restructuring by @B4ckslash in #1325
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1330
- fix: rename transformers.py to avoid circular import by @mudler in #1337
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1340
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1345
- feat(petals): add backend by @mudler in #1350
- fix: go-piper add libucd at linking time by @M0Rf30 in #1357
- docs: Add docker instructions, add community projects section in README by @mudler in #1359
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1351
- docs: Update getting started and GPU section by @mudler in #1362
- ⬆️ Update ggerganov/llama.cpp by @localai-bot in #1363
- ci: limit concurrent jobs by @mudler in #1364
- fix/docs: Python backend dependencies by @B4ckslash in #1360
- ci: split into reusable workflows by @mudler in #1366
- fix: OSX Build Fix Part 1: Metal by @dave-gray101 in #1365
- docs: add fine-tuning example by @mudler in #1374
- docs: site/how-to clean up by @lunamidori5 in #1342
New Contributors
- @vitorstone made their first contribution in #880
- @sfxworks made their first contribution in #1237
- @diego-minguzzi made their first contribution in #1232
- @M0Rf30 made their first contribution in #1257
- @Papawy made their first contribution in #1299
- @ok2sh made their first contribution in #1311
- @B4ckslash made their first contribution in #1321
Full Changelog: v1.40.0...v2.0.0_beta