From fd2e323140c6dc118006ac37e16724c2f117d23d Mon Sep 17 00:00:00 2001 From: Marc Klingen Date: Tue, 17 Dec 2024 13:17:36 +0100 Subject: [PATCH] push --- .../2024-12-langfuse-v3-infrastructure-evolution.mdx | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/pages/blog/2024-12-langfuse-v3-infrastructure-evolution.mdx b/pages/blog/2024-12-langfuse-v3-infrastructure-evolution.mdx index a27fce306..b3d958ffe 100644 --- a/pages/blog/2024-12-langfuse-v3-infrastructure-evolution.mdx +++ b/pages/blog/2024-12-langfuse-v3-infrastructure-evolution.mdx @@ -17,10 +17,16 @@ import Link from "next/link"; authors={["steffenschmitz", "maxdeichmann"]} /> -Langfuse, the open-source LLM observability platform, emerged from the Y Combinator Winter 2023 batch. After building many LLM applications ourselves and realizing that it is hard to go from demo to production, we worked closely with a few of our batchmates to quickly develop a v0 LLM observability platform. We fouced on getting a few core features right: SDKs were asynchronous, based on tracing, as well as open source and easily self-hostable. The first version was written on NextJs, Vercel, and Postgres. Little did we know we would rapidly evolve from an experiment to processing tens of thousands of events per minute. -Our recent V3 release marks a significant milestone in ensuring Langfuse can scale for all of our users. Additionally, it unlocks capabilities for self-hosting users. Features like online evaluations, async ingestion, and cached prompts are now also available in self-hosting. +Langfuse, the open-source LLM observability platform, emerged from the Y Combinator Winter 2023 batch. After building many LLM applications ourselves and realizing that it is hard to go from demo to production, we worked closely with a few of our batchmates to quickly develop a v0 LLM observability platform. + +Initially, we focused on getting a few core features right: SDKs were asynchronous, Langfuse was based on tracing, and all components were open source and easily self-hostable. The first version was written on NextJs, Vercel, and Postgres. Little did we know we would rapidly evolve from an experiment to processing tens of thousands of events per minute. + +Our recent V3 release marks a significant milestone in ensuring Langfuse can scale for all of our users. Additionally, it makes features that we piloted over the recent months available to self-hosting users, such as online evaluations, async/queued ingestion, and cached prompts. + In this post, we will walk you through the scaling challenges we faced while building Langfuse and how our “hypothesis - experiment - feedback” loop helped us arrive at Langfuse v3. If you are interested in solving similar challenges with us - we are hiring in Berlin! +## Context: Architecture Changes + import ArchitectureDiagramV2 from "@/components-mdx/architecture-diagram-v2.mdx"; import ArchitectureDiagramV3 from "@/components-mdx/architecture-diagram-v3.mdx"; import ArchitectureDescriptionV3 from "@/components-mdx/architecture-description-v3.mdx";