🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. 🍊YC W23
-
Updated
Oct 17, 2025 - TypeScript
🪢 Open source LLM engineering platform: LLM Observability, metrics, evals, prompt management, playground, datasets. Integrates with OpenTelemetry, Langchain, OpenAI SDK, LiteLLM, and more. 🍊YC W23
Next-generation AI Agent Optimization Platform: Cozeloop addresses challenges in AI agent development by providing full-lifecycle management capabilities from development, debugging, and evaluation to monitoring.
🧊 Open source LLM observability platform. One line of code to monitor, evaluate, and experiment. YC W23 🍓
The open-source LLMOps platform: prompt playground, prompt management, LLM evaluation, and LLM observability all in one place.
🕹️ Open-source, developer-first LLMOps platform designed to streamline prompt design, version management, instant delivery, collaboration, troubleshooting, observability and more.
Prompt & Conversation Management Middleware for Conversational AI APIs such as OpenAI ChatGPT, Facebook Hugging Face, Anthropic Claude, Google Gemini, Ollama and Jlama. Lean, restful, scalable, and cloud-native. Developed in Java, powered by Quarkus, provided with Docker, and orchestrated with Kubernetes or Openshift.
Markdown for the AI era
An Automated AI-Powered Prompt Optimization Framework
Open-source versioning, tracing, and annotation tooling.
🪢 Langfuse documentation -- Langfuse is the open source LLM Engineering Platform. Observability, evals, prompt management, playground and metrics to debug and improve LLM apps
Model Context Protocol (MCP) Server for Langfuse Prompt Management. This server allows you to access and manage your Langfuse prompts through the Model Context Protocol.
Modular, open source LLMOps stack that separates concerns: LiteLLM unifies LLM APIs, manages routing and cost controls, and ensures high-availability, while Langfuse focuses on detailed observability, prompt versioning, and performance evaluations.
Claude MCP Server - Intelligent Prompt Engineering & Management
Open Source LLM proxy that transparently captures and logs all interactions with LLM API
These guides are designed to help teams and individuals leverage AI tools like GitHub Copilot, OpenAI, and Claude to build software projects efficiently and effectively
Manage, version, and hot‐reload prompts with type‐safe inputs and zero‐deploy updates.
PromptRose 🌹 is your AI prompt companion, blooming at your fingertips.
Managed Prompt Engineering
An easy-to-use structured prompt builder for LLMs in TypeScript.
PromptCode: The VS Code extension that saves the day when your AI code agent (Cursor, etc.) hits a wall. You're in control—select files, craft precise prompts, and effortlessly apply fixes using your favorite LLM.
Add a description, image, and links to the prompt-management topic page so that developers can more easily learn about it.
To associate your repository with the prompt-management topic, visit your repo's landing page and select "manage topics."