Proto commits in mudler/LocalAI

These 58 commits are when the Protocol Buffers files have changed:

Commit:2c9279a
Author:Ettore Di Giacinto
Committer:GitHub

feat(video-gen): add endpoint for video generation (#5247) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

The documentation is generated from this commit.

Commit:61cc76c
Author:Ettore Di Giacinto
Committer:GitHub

chore(autogptq): drop archived backend (#5214) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:a7be2d2
Author:Ettore Di Giacinto
Committer:Ettore Di Giacinto

chore(autogptq): drop archived backend Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:67f7bff
Author:Ettore Di Giacinto
Committer:GitHub

chore(deps): update llama.cpp and sync with upstream changes (#4950) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:6a6e1a0
Author:Brandon Beiler
Committer:GitHub

feat(vllm): Additional vLLM config options (Disable logging, dtype, and Per-Prompt media limits) (#4855) * Adding the following vLLM config options: disable_log_status, dtype, limit_mm_per_prompt Signed-off-by: TheDropZone <brandonbeiler@gmail.com> * using " marks in the config.yaml file Signed-off-by: TheDropZone <brandonbeiler@gmail.com> * adding in missing colon Signed-off-by: TheDropZone <brandonbeiler@gmail.com> --------- Signed-off-by: TheDropZone <brandonbeiler@gmail.com>

Commit:1d6afbd
Author:Ettore Di Giacinto
Committer:GitHub

feat(llama.cpp): Add support to grammar triggers (#4733) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:a1d5462
Author:Ettore Di Giacinto

Stores to chromem (WIP) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:96f8ec0
Author:mintyleaf
Committer:GitHub

feat: add machine tag and inference timings (#4577) * Add machine tag option, add extraUsage option, grpc-server -> proto -> endpoint extraUsage data is broken for now Signed-off-by: mintyleaf <mintyleafdev@gmail.com> * remove redurant timing fields, fix not working timings output Signed-off-by: mintyleaf <mintyleafdev@gmail.com> * use middleware for Machine-Tag only if tag is specified Signed-off-by: mintyleaf <mintyleafdev@gmail.com> --------- Signed-off-by: mintyleaf <mintyleafdev@gmail.com>

Commit:9b6826d
Author:Ettore Di Giacinto
Committer:Ettore Di Giacinto

aujdio Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:ebfe8dd
Author:Ettore Di Giacinto
Committer:Ettore Di Giacinto

gRPC client stubs Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:d4c1746
Author:Ettore Di Giacinto
Committer:GitHub

feat(llama.cpp): expose cache_type_k and cache_type_v for quant of kv cache (#4329) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:44a5dac
Author:Ettore Di Giacinto
Committer:GitHub

feat(backend): add stablediffusion-ggml (#4289) * feat(backend): add stablediffusion-ggml Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(ci): track stablediffusion-ggml Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Use default scheduler and sampler if not specified Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Move cfg scale out of diffusers block Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Make it working Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix: set free_params_immediately to false to call the model in sequence https://github.com/leejet/stable-diffusion.cpp/issues/366 Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:b1ea931
Author:Ettore Di Giacinto
Committer:GitHub

feat(silero): add Silero-vad backend (#4204) * feat(vad): add silero-vad backend (WIP) Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(vad): add API endpoint Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(vad): correctly place the onnxruntime libs Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(vad): hook silero-vad to binary and container builds Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(gRPC): register VAD Server Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(Makefile): consume ONNX_OS consistently Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(Makefile): handle macOS Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

Commit:947224b
Author:Ettore Di Giacinto
Committer:GitHub

feat(diffusers): allow multiple lora adapters (#4081) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:61c964d
Author:Ettore Di Giacinto
Committer:GitHub

fix(grpc): pass by modelpath (#4023) Instead of trying to derive it from the model file. In backends that specify HF url this results in a fragile logic. Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:835932e
Author:Ettore Di Giacinto

feat: update proto file Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:f84b55d
Author:siddimore
Committer:GitHub

feat: Add Get Token Metrics to GRPC server (#3687) * Add Get Token Metrics to GRPC server Signed-off-by: Siddharth More <siddimore@gmail.com> * Expose LocalAI endpoint Signed-off-by: Siddharth More <siddimore@gmail.com> --------- Signed-off-by: Siddharth More <siddimore@gmail.com>

Commit:50a3b54
Author:siddimore
Committer:GitHub

feat(api): add correlationID to Track Chat requests (#3668) * Add CorrelationID to chat request Signed-off-by: Siddharth More <siddimore@gmail.com> * remove get_token_metrics Signed-off-by: Siddharth More <siddimore@gmail.com> * Add CorrelationID to proto Signed-off-by: Siddharth More <siddimore@gmail.com> * fix correlation method name Signed-off-by: Siddharth More <siddimore@gmail.com> * Update core/http/endpoints/openai/chat.go Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Signed-off-by: Siddharth More <siddimore@gmail.com> * Update core/http/endpoints/openai/chat.go Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Signed-off-by: Siddharth More <siddimore@gmail.com> --------- Signed-off-by: Siddharth More <siddimore@gmail.com> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

Commit:191bc2e
Author:Ettore Di Giacinto
Committer:GitHub

feat(api): allow to pass audios to backends (#3603) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:fbb9fac
Author:Ettore Di Giacinto
Committer:GitHub

feat(api): allow to pass videos to backends (#3601) This prepares the API to receive videos as well for video understanding. It works similarly to images, where the request should be in the form: { "type": "video_url", "video_url": { "url": "url or base64 data" } } Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:81ae92f
Author:Dave
Committer:GitHub

feat: elevenlabs `sound-generation` api (#3355) * initial version of elevenlabs compatible soundgeneration api and cli command Signed-off-by: Dave Lee <dave@gray101.com> * minor cleanup Signed-off-by: Dave Lee <dave@gray101.com> * restore TTS, add test Signed-off-by: Dave Lee <dave@gray101.com> * remove stray s Signed-off-by: Dave Lee <dave@gray101.com> * fix Signed-off-by: Dave Lee <dave@gray101.com> --------- Signed-off-by: Dave Lee <dave@gray101.com> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

Commit:03b1cf5
Author:Ettore Di Giacinto
Committer:GitHub

feat(whisper): add translate option (#2649) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:b99182c
Author:Chakib Benziane
Committer:GitHub

TTS API improvements (#2308) * update doc on COQUI_LANGUAGE env variable Signed-off-by: blob42 <contact@blob42.xyz> * return errors from tts gRPC backend Signed-off-by: blob42 <contact@blob42.xyz> * handle speaker_id and language in coqui TTS backend Signed-off-by: blob42 <contact@blob42.xyz> * TTS endpoint: add optional language paramter Signed-off-by: blob42 <contact@blob42.xyz> * tts fix: empty language string breaks non-multilingual models Signed-off-by: blob42 <contact@blob42.xyz> * allow tts param definition in config file - consolidate TTS options under `tts` config entry Signed-off-by: blob42 <contact@blob42.xyz> * tts: update doc Signed-off-by: blob42 <contact@blob42.xyz> --------- Signed-off-by: blob42 <contact@blob42.xyz> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>

Commit:e49ea01
Author:Ettore Di Giacinto
Committer:GitHub

feat(llama.cpp): add `flash_attention` and `no_kv_offloading` (#2310) feat(llama.cpp): add flash_attn and no_kv_offload Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:b664edd
Author:Ettore Di Giacinto
Committer:GitHub

feat(rerankers): Add new backend, support jina rerankers API (#2121) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:03adc1f
Author:Taikono-Himazin
Committer:GitHub

Add tensor_parallel_size setting to vllm setting items (#2085) Signed-off-by: Taikono-Himazin <kazu@po.harenet.ne.jp>

Commit:e843d7d
Author:Ettore Di Giacinto
Committer:GitHub

feat(grpc): return consumed token count and update response accordingly (#2035) Fixes: #1920

Commit:12c0d94
Author:Ludovic Leroux
Committer:GitHub

feat: use tokenizer.apply_chat_template() in vLLM (#1990) Use tokenizer.apply_chat_template() in vLLM Signed-off-by: Ludovic LEROUX <ludovic@inpher.io>

Commit:643d85d
Author:Richard Palethorpe
Committer:GitHub

feat(stores): Vector store backend (#1795) Add simple vector store backend Signed-off-by: Richard Palethorpe <io@richiejp.com>

Commit:20136ca
Author:Ettore Di Giacinto
Committer:GitHub

feat(tts): add Elevenlabs and OpenAI TTS compatibility layer (#1834) * feat(elevenlabs): map elevenlabs API support to TTS This allows elevenlabs Clients to work automatically with LocalAI by supporting the elevenlabs API. The elevenlabs server endpoint is implemented such as it is wired to the TTS endpoints. Fixes: https://github.com/mudler/LocalAI/issues/1809 * feat(openai/tts): compat layer with openai tts Fixes: #1276 * fix: adapt tts CLI

Commit:d2934dd
Author:Ettore Di Giacinto
Committer:Ettore Di Giacinto

feat(elevenlabs): map elevenlabs API support to TTS This allows elevenlabs Clients to work automatically with LocalAI by supporting the elevenlabs API. The elevenlabs server endpoint is implemented such as it is wired to the TTS endpoints. Fixes: https://github.com/mudler/LocalAI/issues/1809

Commit:9394113
Author:Ludovic Leroux
Committer:GitHub

Bump vLLM version + more options when loading models in vLLM (#1782) * Bump vLLM version to 0.3.2 * Add vLLM model loading options * Remove transformers-exllama * Fix install exllama

Commit:cb75127
Author:Ettore Di Giacinto
Committer:GitHub

transformers: correctly load automodels (#1643) * backends(transformers): use AutoModel with LLM types * examples: animagine-xl * Add codellama examples

Commit:9e653d6
Author:Ettore Di Giacinto
Committer:GitHub

feat: 🐍 add mamba support (#1589) feat(mamba): Initial import This is a first iteration of the mamba backend, loosely based on mamba-chat(https://github.com/havenhq/mamba-chat).

Commit:7641f92
Author:Ettore Di Giacinto
Committer:GitHub

feat(diffusers): update, add autopipeline, controlnet (#1432) * feat(diffusers): update, add autopipeline, controlenet * tests with AutoPipeline * simplify logic

Commit:ad0e30b
Author:Ettore Di Giacinto
Committer:GitHub

refactor: move backends into the backends directory (#1279) * refactor: move backends into the backends directory Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactor: move main close to implementation for every backend Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:803a0ac
Author:Ettore Di Giacinto
Committer:GitHub

feat(llama.cpp): support lora with scale and yarn (#1277) * feat(llama.cpp): support lora with scale Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(llama.cpp): support yarn Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:0eae727
Author:Ettore Di Giacinto
Committer:GitHub

:fire: add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types (#1254) * wip * wip * Make it functional Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * wip * Small fixups * do not inject space on role encoding, encode img at beginning of messages Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add examples/config defaults * Add include dir of current source dir * cleanup * fixes Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups * Revert "fixups" This reverts commit f1a4731ccadf7226c6589d6d39131376f0811625. * fixes Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:a28ab18
Author:Ettore Di Giacinto
Committer:GitHub

feat(vllm): Allow to set quantization (#1094) This particularly useful to set AWQ **Description** Follow up of #1015 **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:8ccf5b2
Author:Ettore Di Giacinto
Committer:GitHub

feat(speculative-sampling): allow to specify a draft model in the model config (#1052) **Description** This PR fixes #1013. It adds `draft_model` and `n_draft` to the model YAML config in order to load models with speculative sampling. This should be compatible as well with grammars. example: ```yaml backend: llama context_size: 1024 name: my-model-name parameters: model: foo-bar n_draft: 16 draft_model: model-name ``` --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:dc307a1
Author:Ettore Di Giacinto
Committer:GitHub

feat: add vall-e-x (#1007) **Description** This PR fixes #985 **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:44bc7aa
Author:Ettore Di Giacinto
Committer:GitHub

feat: Allow to load lora adapters for llama.cpp (#955) **Description** This PR fixes # **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:901f070
Author:Dave
Committer:GitHub

Feat: rwkv improvements: (#937)

Commit:8cb1061
Author:Dave
Committer:GitHub

Usage Features (#863)

Commit:2bacd01
Author:Ettore Di Giacinto
Committer:GitHub

feat(diffusers): add img2img and clip_skip, support more kernels schedulers (#906) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:37700f2
Author:Ettore Di Giacinto
Committer:GitHub

feat(diffusers): add DPMSolverMultistepScheduler++, DPMSolverMultistepSchedulerSDE++, guidance_scale (#903) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:a96c3bc
Author:Ettore Di Giacinto
Committer:GitHub

feat(diffusers): various enhancements (#895)

Commit:8c781a6
Author:Ettore Di Giacinto
Committer:GitHub

feat: Add Diffusers (#874) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:3c8fc37
Author:Ettore Di Giacinto

feat: Add UseFastTokenizer Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:a843e64
Author:Ettore Di Giacinto
Committer:Ettore Di Giacinto

feat: add initial AutoGPTQ backend implementation

Commit:5ca21ee
Author:Ettore Di Giacinto
Committer:GitHub

feat: add ngqa and RMSNormEps parameters (#860) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:096d98c
Author:Ettore Di Giacinto
Committer:GitHub

fix: add rope settings during model load, fix CUDA (#821) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:b96e30e
Author:Ettore Di Giacinto
Committer:GitHub

fix: use bytes in gRPC proto instead of strings (#813) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:569c1d1
Author:Ettore Di Giacinto
Committer:GitHub

feat: add rope settings and negative prompt, drop grammar backend (#797) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:ae533ca
Author:Ettore Di Giacinto

feat: move gpt4all to a grpc service Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:b816009
Author:Ettore Di Giacinto

feat: add falcon ggllm via grpc client Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:58f6aab
Author:Ettore Di Giacinto

feat: move llama to a grpc Signed-off-by: Ettore Di Giacinto <mudler@localai.io>

Commit:1d0ed95
Author:Ettore Di Giacinto

feat: move other backends to grpc This finally makes everything more consistent Signed-off-by: Ettore Di Giacinto <mudler@localai.io>