These 58 commits are when the Protocol Buffers files have changed:
Commit: | 2c9279a | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(video-gen): add endpoint for video generation (#5247) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
The documentation is generated from this commit.
Commit: | 61cc76c | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
chore(autogptq): drop archived backend (#5214) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | a7be2d2 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | Ettore Di Giacinto |
chore(autogptq): drop archived backend Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 67f7bff | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
chore(deps): update llama.cpp and sync with upstream changes (#4950) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 6a6e1a0 | |
---|---|---|
Author: | Brandon Beiler | |
Committer: | GitHub |
feat(vllm): Additional vLLM config options (Disable logging, dtype, and Per-Prompt media limits) (#4855) * Adding the following vLLM config options: disable_log_status, dtype, limit_mm_per_prompt Signed-off-by: TheDropZone <brandonbeiler@gmail.com> * using " marks in the config.yaml file Signed-off-by: TheDropZone <brandonbeiler@gmail.com> * adding in missing colon Signed-off-by: TheDropZone <brandonbeiler@gmail.com> --------- Signed-off-by: TheDropZone <brandonbeiler@gmail.com>
Commit: | 1d6afbd | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(llama.cpp): Add support to grammar triggers (#4733) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | a1d5462 | |
---|---|---|
Author: | Ettore Di Giacinto |
Stores to chromem (WIP) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 96f8ec0 | |
---|---|---|
Author: | mintyleaf | |
Committer: | GitHub |
feat: add machine tag and inference timings (#4577) * Add machine tag option, add extraUsage option, grpc-server -> proto -> endpoint extraUsage data is broken for now Signed-off-by: mintyleaf <mintyleafdev@gmail.com> * remove redurant timing fields, fix not working timings output Signed-off-by: mintyleaf <mintyleafdev@gmail.com> * use middleware for Machine-Tag only if tag is specified Signed-off-by: mintyleaf <mintyleafdev@gmail.com> --------- Signed-off-by: mintyleaf <mintyleafdev@gmail.com>
Commit: | 9b6826d | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | Ettore Di Giacinto |
aujdio Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | ebfe8dd | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | Ettore Di Giacinto |
gRPC client stubs Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | d4c1746 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(llama.cpp): expose cache_type_k and cache_type_v for quant of kv cache (#4329) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 44a5dac | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(backend): add stablediffusion-ggml (#4289) * feat(backend): add stablediffusion-ggml Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(ci): track stablediffusion-ggml Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Use default scheduler and sampler if not specified Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Move cfg scale out of diffusers block Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Make it working Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix: set free_params_immediately to false to call the model in sequence https://github.com/leejet/stable-diffusion.cpp/issues/366 Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | b1ea931 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(silero): add Silero-vad backend (#4204) * feat(vad): add silero-vad backend (WIP) Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(vad): add API endpoint Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(vad): correctly place the onnxruntime libs Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * chore(vad): hook silero-vad to binary and container builds Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(gRPC): register VAD Server Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(Makefile): consume ONNX_OS consistently Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fix(Makefile): handle macOS Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Commit: | 947224b | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(diffusers): allow multiple lora adapters (#4081) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 61c964d | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
fix(grpc): pass by modelpath (#4023) Instead of trying to derive it from the model file. In backends that specify HF url this results in a fragile logic. Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 835932e | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: update proto file Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | f84b55d | |
---|---|---|
Author: | siddimore | |
Committer: | GitHub |
feat: Add Get Token Metrics to GRPC server (#3687) * Add Get Token Metrics to GRPC server Signed-off-by: Siddharth More <siddimore@gmail.com> * Expose LocalAI endpoint Signed-off-by: Siddharth More <siddimore@gmail.com> --------- Signed-off-by: Siddharth More <siddimore@gmail.com>
Commit: | 50a3b54 | |
---|---|---|
Author: | siddimore | |
Committer: | GitHub |
feat(api): add correlationID to Track Chat requests (#3668) * Add CorrelationID to chat request Signed-off-by: Siddharth More <siddimore@gmail.com> * remove get_token_metrics Signed-off-by: Siddharth More <siddimore@gmail.com> * Add CorrelationID to proto Signed-off-by: Siddharth More <siddimore@gmail.com> * fix correlation method name Signed-off-by: Siddharth More <siddimore@gmail.com> * Update core/http/endpoints/openai/chat.go Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Signed-off-by: Siddharth More <siddimore@gmail.com> * Update core/http/endpoints/openai/chat.go Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Signed-off-by: Siddharth More <siddimore@gmail.com> --------- Signed-off-by: Siddharth More <siddimore@gmail.com> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Commit: | 191bc2e | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(api): allow to pass audios to backends (#3603) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | fbb9fac | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(api): allow to pass videos to backends (#3601) This prepares the API to receive videos as well for video understanding. It works similarly to images, where the request should be in the form: { "type": "video_url", "video_url": { "url": "url or base64 data" } } Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 81ae92f | |
---|---|---|
Author: | Dave | |
Committer: | GitHub |
feat: elevenlabs `sound-generation` api (#3355) * initial version of elevenlabs compatible soundgeneration api and cli command Signed-off-by: Dave Lee <dave@gray101.com> * minor cleanup Signed-off-by: Dave Lee <dave@gray101.com> * restore TTS, add test Signed-off-by: Dave Lee <dave@gray101.com> * remove stray s Signed-off-by: Dave Lee <dave@gray101.com> * fix Signed-off-by: Dave Lee <dave@gray101.com> --------- Signed-off-by: Dave Lee <dave@gray101.com> Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Commit: | 03b1cf5 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(whisper): add translate option (#2649) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | b99182c | |
---|---|---|
Author: | Chakib Benziane | |
Committer: | GitHub |
TTS API improvements (#2308) * update doc on COQUI_LANGUAGE env variable Signed-off-by: blob42 <contact@blob42.xyz> * return errors from tts gRPC backend Signed-off-by: blob42 <contact@blob42.xyz> * handle speaker_id and language in coqui TTS backend Signed-off-by: blob42 <contact@blob42.xyz> * TTS endpoint: add optional language paramter Signed-off-by: blob42 <contact@blob42.xyz> * tts fix: empty language string breaks non-multilingual models Signed-off-by: blob42 <contact@blob42.xyz> * allow tts param definition in config file - consolidate TTS options under `tts` config entry Signed-off-by: blob42 <contact@blob42.xyz> * tts: update doc Signed-off-by: blob42 <contact@blob42.xyz> --------- Signed-off-by: blob42 <contact@blob42.xyz> Co-authored-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
Commit: | e49ea01 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(llama.cpp): add `flash_attention` and `no_kv_offloading` (#2310) feat(llama.cpp): add flash_attn and no_kv_offload Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | b664edd | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(rerankers): Add new backend, support jina rerankers API (#2121) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 03adc1f | |
---|---|---|
Author: | Taikono-Himazin | |
Committer: | GitHub |
Add tensor_parallel_size setting to vllm setting items (#2085) Signed-off-by: Taikono-Himazin <kazu@po.harenet.ne.jp>
Commit: | e843d7d | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(grpc): return consumed token count and update response accordingly (#2035) Fixes: #1920
Commit: | 12c0d94 | |
---|---|---|
Author: | Ludovic Leroux | |
Committer: | GitHub |
feat: use tokenizer.apply_chat_template() in vLLM (#1990) Use tokenizer.apply_chat_template() in vLLM Signed-off-by: Ludovic LEROUX <ludovic@inpher.io>
Commit: | 643d85d | |
---|---|---|
Author: | Richard Palethorpe | |
Committer: | GitHub |
feat(stores): Vector store backend (#1795) Add simple vector store backend Signed-off-by: Richard Palethorpe <io@richiejp.com>
Commit: | 20136ca | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(tts): add Elevenlabs and OpenAI TTS compatibility layer (#1834) * feat(elevenlabs): map elevenlabs API support to TTS This allows elevenlabs Clients to work automatically with LocalAI by supporting the elevenlabs API. The elevenlabs server endpoint is implemented such as it is wired to the TTS endpoints. Fixes: https://github.com/mudler/LocalAI/issues/1809 * feat(openai/tts): compat layer with openai tts Fixes: #1276 * fix: adapt tts CLI
Commit: | d2934dd | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | Ettore Di Giacinto |
feat(elevenlabs): map elevenlabs API support to TTS This allows elevenlabs Clients to work automatically with LocalAI by supporting the elevenlabs API. The elevenlabs server endpoint is implemented such as it is wired to the TTS endpoints. Fixes: https://github.com/mudler/LocalAI/issues/1809
Commit: | 9394113 | |
---|---|---|
Author: | Ludovic Leroux | |
Committer: | GitHub |
Bump vLLM version + more options when loading models in vLLM (#1782) * Bump vLLM version to 0.3.2 * Add vLLM model loading options * Remove transformers-exllama * Fix install exllama
Commit: | cb75127 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
transformers: correctly load automodels (#1643) * backends(transformers): use AutoModel with LLM types * examples: animagine-xl * Add codellama examples
Commit: | 9e653d6 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: 🐍 add mamba support (#1589) feat(mamba): Initial import This is a first iteration of the mamba backend, loosely based on mamba-chat(https://github.com/havenhq/mamba-chat).
Commit: | 7641f92 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(diffusers): update, add autopipeline, controlnet (#1432) * feat(diffusers): update, add autopipeline, controlenet * tests with AutoPipeline * simplify logic
Commit: | ad0e30b | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
refactor: move backends into the backends directory (#1279) * refactor: move backends into the backends directory Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * refactor: move main close to implementation for every backend Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 803a0ac | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(llama.cpp): support lora with scale and yarn (#1277) * feat(llama.cpp): support lora with scale Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * feat(llama.cpp): support yarn Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 0eae727 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
:fire: add LaVA support and GPT vision API, Multiple requests for llama.cpp, return JSON types (#1254) * wip * wip * Make it functional Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * wip * Small fixups * do not inject space on role encoding, encode img at beginning of messages Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add examples/config defaults * Add include dir of current source dir * cleanup * fixes Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * fixups * Revert "fixups" This reverts commit f1a4731ccadf7226c6589d6d39131376f0811625. * fixes Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | a28ab18 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(vllm): Allow to set quantization (#1094) This particularly useful to set AWQ **Description** Follow up of #1015 **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 8ccf5b2 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(speculative-sampling): allow to specify a draft model in the model config (#1052) **Description** This PR fixes #1013. It adds `draft_model` and `n_draft` to the model YAML config in order to load models with speculative sampling. This should be compatible as well with grammars. example: ```yaml backend: llama context_size: 1024 name: my-model-name parameters: model: foo-bar n_draft: 16 draft_model: model-name ``` --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | dc307a1 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: add vall-e-x (#1007) **Description** This PR fixes #985 **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 44bc7aa | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: Allow to load lora adapters for llama.cpp (#955) **Description** This PR fixes # **Notes for Reviewers** **[Signed commits](../CONTRIBUTING.md#signing-off-on-commits-developer-certificate-of-origin)** - [ ] Yes, I signed my commits. <!-- Thank you for contributing to LocalAI! Contributing Conventions: 1. Include descriptive PR titles with [<component-name>] prepended. 2. Build and test your changes before submitting a PR. 3. Sign your commits By following the community's contribution conventions upfront, the review process will be accelerated and your PR merged more quickly. --> Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 901f070 | |
---|---|---|
Author: | Dave | |
Committer: | GitHub |
Feat: rwkv improvements: (#937)
Commit: | 8cb1061 | |
---|---|---|
Author: | Dave | |
Committer: | GitHub |
Usage Features (#863)
Commit: | 2bacd01 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(diffusers): add img2img and clip_skip, support more kernels schedulers (#906) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 37700f2 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(diffusers): add DPMSolverMultistepScheduler++, DPMSolverMultistepSchedulerSDE++, guidance_scale (#903) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | a96c3bc | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat(diffusers): various enhancements (#895)
Commit: | 8c781a6 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: Add Diffusers (#874) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 3c8fc37 | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: Add UseFastTokenizer Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | a843e64 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | Ettore Di Giacinto |
feat: add initial AutoGPTQ backend implementation
Commit: | 5ca21ee | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: add ngqa and RMSNormEps parameters (#860) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 096d98c | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
fix: add rope settings during model load, fix CUDA (#821) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | b96e30e | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
fix: use bytes in gRPC proto instead of strings (#813) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 569c1d1 | |
---|---|---|
Author: | Ettore Di Giacinto | |
Committer: | GitHub |
feat: add rope settings and negative prompt, drop grammar backend (#797) Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | ae533ca | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: move gpt4all to a grpc service Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | b816009 | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: add falcon ggllm via grpc client Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 58f6aab | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: move llama to a grpc Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
Commit: | 1d0ed95 | |
---|---|---|
Author: | Ettore Di Giacinto |
feat: move other backends to grpc This finally makes everything more consistent Signed-off-by: Ettore Di Giacinto <mudler@localai.io>