Traces, evals, prompt management and metrics to debug and improve your LLM application. Onboard via https://langfuse.com Langfuse helps you build and improve LLM applications across the entire lifecycle: - Develop: Observability, Langfuse UI & Prompt Management - Monitor: Traces, Analytics, Metrics & Evaluations - Test: Experiments, Releases & Datasets We are hiring: https://langfuse.com/careers
Marc has diverse experience across Product, Sales, Business Intelligence and full-stack engineering at companies from large (Google, DHL) to early stage startups. He graduated within the top 1% of his Masters in Management and Computer Science from Technical University Munich. Besides that, he loves to hack on personal project and connect with other builders (something he does not get to too much right now).
Max built trading systems at European 5bn Fintech, Trade Republic. He knows the ins and outs of building reliable, scalable systems to handle our customers’ most critical business processes. While Max started out studying Management in undergrad, he quickly found his love for computer science and transitioned into engineer self-taught and with a graduate degree.
Before starting Langfuse, Clemens worked with the founder-CEOs of German Fintech Unicorn Scalable Capital including a unicorn fundraising, an acquisition and helped scale the org and team from 100 - 400 employees. On another note, he studied Economic History, dropped out of a PhD at Oxford, and was a competitive wine taster.
⭐ Star us on Github & follow us on Twitter
TLDR: Langfuse is building open-source product analytics (think ‘Mixpanel’) for LLM apps. We help companies to track and analyze quality, cost and latency across product releases and use cases.
Hi everyone, we’re Max, Marc and Clemens. We were part of the Winter 23 batch and work on Langfuse, where we help teams make sense of how their LLM applications perform.
LLMs represent a new paradigm in software. Single LLM calls are probabilistic and add substantial latency and cost. Applications use LLMs in new ways via advanced prompting, embedding-based retrieval, chains, and agents with tools. Teams building production-grade LLM applications have new product analytics and monitoring needs:
Langfuse derives actionable insights from production data. Our customers use Langfuse to answer questions such as: ‘How helpful are my LLM app’s outputs? What is my LLM API spend by customer? What do latencies look like across geographies and steps of LLM chains? Did the quality of the application improve in newer versions? What was the impact of switching from zero-shotting GPT4 to using few-shotted Llama calls?’
Metrics
Insights
Integrations
Langfuse can be self-hosted or used with a generous free tier in our managed cloud version.
Based on the ingested data, Langfuse helps developers debug complex LLM apps in production:
Star us on GitHub + follow along on Twitter & LinkedIn.