LMOS is modular, scalable AI system for deploying full-featured language models, transcription, embeddings, and reranking, with support for Docker and Kubernetes orchestration.
Coming Soon to Apache 2.0
LMOS is modular, scalable AI system for deploying full-featured language models, transcription, embeddings, and reranking, with support for Docker and Kubernetes orchestration.
Coming Soon to Apache 2.0
Contains Pydantic and MsgSpec models derived from OpenAI's OpenAPI spec. Used across the organization.
A high-throughput and memory-efficient inference and serving engine for LLMs
Python SDK, Proxy Server (LLM Gateway) to call 100+ LLM APIs in OpenAI format - [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, Replicate, Groq]
The LMOS-Config repo contains a shared config provider for the entire LMOS system.
A fast inference library for running LLMs locally on modern consumer-class GPUs
This contains a relational database schema for the LMOS system. It is stored in this repo so that it can be included as a dependency inside other systems to eliminate ambiguity and diverging specifications.
This organization has no public members. You must be a member to see who’s a part of this organization.
Loading…
Loading…