DocsObservabilityConcepts

Core Concepts

This page digs into the underlying concepts of how Langfuse structures and captures your data. Understanding these will make debugging and working with traces easier.

Ready to start? Check out the Get Started guide to ingest your first trace.

Observations, Traces, and Sessions

Langfuse organizes an application’s data into three core concepts: observations, traces, and sessions.

Observations

Observations are the individual steps within a trace. Langfuse supports a number of LLM application specific observation types, for example generations, toolcalls, RAG retrieval steps, etc.

Observations can be nested. The example below shows a trace with a nested observation.

Hierarchical structure of observations in Langfuse

Example trace in Langfuse UI

Trace in Langfuse UI

Traces

A trace typically represents a single request or operation. For example, when a user asks a question to a chatbot, that interaction, from the user’s question to the bot’s response, is captured as one trace.

It contains the overall input and output of the function, as well as metadata about the request ( i.e. user, session, tags, etc.).

Sessions

Optionally, traces can be grouped into sessions. Sessions are used to group traces that are part of the same user interaction. A common example is a thread in a chat interface.

Optionally, sessions aggregate traces

Example session in Langfuse UI

Session view

Using sessions is recommended for applications with multi-turn conversations or workflows. Please refer to the Sessions documentation to add sessions to your traces.

Adding Attributes

Once you’ve structured your data into traces and observations, you can enrich them with additional attributes. These attributes act as labels that help you filter, segment, and analyze your traces for specific use cases.

There are different types of attributes you can add:

AttributeDescription
EnvironmentsSeparate data from different deployment contexts like production, staging, or development
TagsFlexible labels to categorize traces by feature, API endpoint, or workflow
UserTrack which end-user triggered each trace
MetadataFlexible key-value store for custom information
Releases & VersionsTrack application versions and component changes

Observation-Centric Data Model

The observation-centric data model is available with Python SDK v4+ and JS/TS SDK v5+. Both data models currently coexist, but Langfuse is gradually transitioning to the observation-centric model. See the changelog post for more details.

To improve query performance at scale, Langfuse introduces an observation-centric data model as part of Langfuse v4 alongside the existing model. In this model, context attributes (user_id, session_id, metadata, tags) that previously lived only on the trace are propagated to every observation. This eliminates expensive joins between trace and observation tables, enabling single-table queries.

Observation-centric data model

What changes

Classic modelObservation-centric model
StorageSeparate mutable Traces and Observations tablesSingle immutable Observations table
Context attributesStored on the trace, joined at query timePropagated to every observation at ingestion time
Trace input/outputSet directly on the traceRemoved, use observation IO instead. (For legacy LLM-as-a-judge evaluators that depend on trace input and output, users can use the deprecated set_trace_io methods in the SDKs.)
MutabilityTraces and observations are mutableObservations are immutable (written once)

How to adopt

Upgrade to the latest SDK versions to start ingesting data in the new format:

The key change is that update_current_trace() / updateActiveTrace() is replaced by propagate_attributes(), which automatically flows attributes to all child observations created within its scope.

How Langfuse Captures Data

Now that you understand the data model, let’s explore how Langfuse actually captures and processes your traces.

Built on OpenTelemetry

Langfuse is built on OpenTelemetry, an open standard for collecting telemetry data from applications.

This means you’re not locked into using only Langfuse-specific SDKs. You can also send your traces to multiple destinations at once, like Langfuse for LLM observability and Datadog for infrastructure monitoring.

See the OpenTelemetry integration guide for detailed documentation on integrating OpenTelemetry with Langfuse.

Instrumentation

Instrumentation is the process of adding code to record its behavior. Once this recording is turned on, Langfuse (through OpenTelemetry) can automatically capture these events and structure them into traces and observations.

The Get Started guide walks you through the process of instrumenting a function in your application.

Background Processing

In order to avoid slowing down your application, Langfuse doesn’t send traces synchronously the moment they’re created. Instead, Langfuse batches traces locally and sends them in the background, keeping your application fast and responsive.

Long-running applications

The approach above works well for long-running applications (like web servers or APIs) because the background exporter continuously runs and has plenty of time to flush batches on its own.

Short-lived applications

For applications that start, execute something, and shut down quickly (short-lived applications), there’s a risk that the application terminates while there are still unsent traces in the queue.

To avoid losing data, short-lived applications must explicitly call flush() before exiting. This forces the exporter to send all buffered traces immediately, so nothing is lost when the process terminates.

Was this page helpful?