Awesome LLMOps
An awesome & curated list of the best LLMOps tools for developers.
Contribute
Contributions are most welcome, please adhere to the contribution guidelines.
Table of Contents
Model
Large Language Model
Project | Details | Repository |
---|
Alpaca | Code and documentation to train Stanford's Alpaca models, and generate the data. | |
BELLE | A 7B Large Language Model fine-tune by 34B Chinese Character Corpus, based on LLaMA and Alpaca. | |
Bloom | BigScience Large Open-science Open-access Multilingual Language Model | |
dolly | Databricks’ Dolly, a large language model trained on the Databricks Machine Learning Platform | |
Falcon 40B | Falcon-40B-Instruct is a 40B parameters causal decoder-only model built by TII based on Falcon-40B and finetuned on a mixture of Baize. It is made available under the Apache 2.0 license. | |
FastChat (Vicuna) | An open platform for training, serving, and evaluating large language models. Release repo for Vicuna and FastChat-T5. | |
Gemma | Gemma is a family of lightweight, open models built from the research and technology that Google used to create the Gemini models. | |
GLM-6B (ChatGLM) | An Open Bilingual Pre-Trained Model, quantization of ChatGLM-130B, can run on consumer-level GPUs. | |
ChatGLM2-6B | ChatGLM2-6B is the second-generation version of the open-source bilingual (Chinese-English) chat model ChatGLM-6B. | |
GLM-130B (ChatGLM) | An Open Bilingual Pre-Trained Model (ICLR 2023) | |
GPT-NeoX | An implementation of model parallel autoregressive transformers on GPUs, based on the DeepSpeed library. | |
Luotuo | A Chinese LLM, Based on LLaMA and fine tune by Stanford Alpaca, Alpaca LoRA, Japanese-Alpaca-LoRA. | |
Mixtral-8x7B-v0.1 | The Mixtral-8x7B Large Language Model (LLM) is a pretrained generative Sparse Mixture of Experts. | |
StableLM | StableLM: Stability AI Language Models | |
⬆ back to ToC
CV Foundation Model
Project | Details | Repository |
---|
disco-diffusion | A frankensteinian amalgamation of notebooks, models and techniques for the generation of AI Art and Animations. | |
midjourney | Midjourney is an independent research lab exploring new mediums of thought and expanding the imaginative powers of the human species. | |
segment-anything (SAM) | produces high quality object masks from input prompts such as points or boxes, and it can be used to generate masks for all objects in an image. | |
stable-diffusion | A latent text-to-image diffusion model | |
stable-diffusion v2 | High-Resolution Image Synthesis with Latent Diffusion Models | |
⬆ back to ToC
Audio Foundation Model
Project | Details | Repository |
---|
bark | Bark is a transformer-based text-to-audio model created by Suno. Bark can generate highly realistic, multilingual speech as well as other audio - including music, background noise and simple sound effects. | |
whisper | Robust Speech Recognition via Large-Scale Weak Supervision | |
Serving
Large Model Serving
Project | Details | Repository |
---|
Alpaca-LoRA-Serve | Alpaca-LoRA as Chatbot service | |
CTranslate2 | fast inference engine for Transformer models in C++ | |
Clip-as-a-service | serving the OpenAI CLIP model | |
DeepSpeed-MII | MII makes low-latency and high-throughput inference possible, powered by DeepSpeed. | |
Faster Whisper | fast inference engine for whisper in C++ using CTranslate2. | |
FlexGen | Running large language models on a single GPU for throughput-oriented scenarios. | |
Flowise | Drag & drop UI to build your customized LLM flow using LangchainJS. | |
llama.cpp | Port of Facebook's LLaMA model in C/C++ | |
Infinity | Rest API server for serving text-embeddings | |
Modelz-LLM | OpenAI compatible API for LLMs and embeddings (LLaMA, Vicuna, ChatGLM and many others) | |
Ollama | Serve Llama 2 and other large language models locally from command line or through a browser interface. | |
TensorRT-LLM | Inference engine for TensorRT on Nvidia GPUs | |
text-generation-inference | Large Language Model Text Generation Inference | |
text-embeddings-inference | Inference for text-embedding models | |
vllm | A high-throughput and memory-efficient inference and serving engine for LLMs. | |
whisper.cpp | Port of OpenAI's Whisper model in C/C++ | |
x-stable-diffusion | Real-time inference for Stable Diffusion - 0.88s latency. Covers AITemplate, nvFuser, TensorRT, FlashAttention. | |
⬆ back to ToC
Frameworks/Servers for Serving
Project | Details | Repository |
---|
BentoML | The Unified Model Serving Framework | |
Jina | Build multimodal AI services via cloud native technologies · Model Serving · Generative AI · Neural Search · Cloud Native | |
Mosec | A machine learning model serving framework with dynamic batching and pipelined stages, provides an easy-to-use Python interface. | |
TFServing | A flexible, high-performance serving system for machine learning models. | |
Torchserve | Serve, optimize and scale PyTorch models in production | |
Triton Server (TRTIS) | The Triton Inference Server provides an optimized cloud and edge inferencing solution. | |
langchain-serve | Serverless LLM apps on Production with Jina AI Cloud | |
lanarky | FastAPI framework to build production-grade LLM applications | |
ray-llm | LLMs on Ray - RayLLM | |
Xinference | Replace OpenAI GPT with another LLM in your app by changing a single line of code. Xinference gives you the freedom to use any LLM you need. With Xinference, you're empowered to run inference with any open-source language models, speech recognition models, and multimodal models, whether in the cloud, on-premises, or even on your laptop. | |
⬆ back to ToC
Security
Frameworks for LLM security
| Project | Details | Repository