Instrument LangChain and LangGraph Applications with OpenTelemetry
Learn how to instrument LangChain and LangGraph applications with OpenTelemetry and export large language model (LLM) telemetry via OTLP into Cribl Search. Once your data is in Cribl Search, you can explore and investigate it to troubleshoot issues, analyze model behavior, and monitor performance.
You’ll complete the following high-level steps:
- Configure an OpenTelemetry (OTel) Source in Cribl Search to receive OTLP data.
- Instrument your LangChain application (Node.js or Python) with OpenTelemetry.
- Explore LLM telemetry use cases in Cribl Search.
Telemetry Captured
LangChain and LangGraph runs produce trace spans with these attributes:
| Category | Data | Example |
|---|---|---|
| Timing | Duration per chain step, tool run, or graph node. | 1.2s for an LLM span, 45ms for a tool. |
| Structure | Chain or graph hierarchy, span names for LLM, retriever, tool, or agent steps. | ChatOpenAI, tool:add_numbers, LangGraph node names. |
| Model usage | Model id and token-related fields when the underlying LLM instrumentation is active. | gpt-4o-mini, input/output tokens. |
| Tools | Tool name, arguments metadata (content may be opt-in). | Tool id, invocation status. |
| Errors | Exceptions or API errors surfaced on the relevant span. | HTTP 429, validation errors. |
| Content | Prompt and completion text. | (opt-in, disabled by default). |
Semantic conventions follow OpenTelemetry GenAI Semantic Conventions where applicable.
Prerequisites
You’ll need:
- Cribl.Cloud Enterprise.
- Search Admin permission or higher.
- An OpenAI API key or other provider credentials with access to the models you want to monitor (examples below use
langchain-openaiandOPENAI_API_KEY). - Node.js 18+ or Python 3.9+ depending on your application language.
Get Data Into Cribl Search
Complete these steps before you instrument your LLM application:
- Add a lakehouse engine. This provides storage and compute for data you’re going to ingest into Cribl Search.
- Add a Cribl Search OpenTelemetry Source to start ingesting the data.
- Set up your Datatype rules to parse and structure the incoming data.
- Set up your Dataset rules to route ingested events into individual Search Datasets. This will let you scope your queries and control retention.
Add a Lakehouse Engine
A lakehouse engine is the storage-and-compute unit in Cribl Search that holds ingested OTLP (and other Source data) until Dataset retention expires. See Lakehouse Engines in Cribl Search to learn how to setup a new lakehouse engine.
Configure the OpenTelemetry (OTel) Source in Cribl Search
To receive OTLP from your LLM application directly in Cribl Search, add an OpenTelemetry Source.
On the Cribl.Cloud top bar, select Products > Search. Under Data, select Add Source > OpenTelemetry.
In the New Source modal, configure the following under General Settings:
- ID: Unique Source ID across your Cribl.Cloud Workspace. Use letters, numbers, underscores, and hyphens.
- Description (optional): Describe the Source (for example, OTLP from LLM-instrumented apps).
- Address: Hostname that your OpenTelemetry collector or agent connects to. You will use this in exporter configuration.
- Port: Network port to listen on. Defaults to
4317for gRPC and4318for HTTP (OTLP standard). Change if you use a custom port. - OTLP version: Version that matches your upstream sender (default
1.3.1). - Protocol: gRPC (default) or HTTP. This must match your OpenTelemetry exporter.
Under Authentication, choose None, Basic, Basic (Credentials Secret), Auth Tokens, or Auth Token (Text Secret) as required. See Set up Authentication.
Under Encrypt, enable TLS and set the minimum TLS version when senders must connect over TLS. See Set Up Encryption.
Select Save to create the Source.
Set Datatype Rules
Next, configure Datatype rules to parse, filter, and normalize your data into structured fields.
On the Cribl.Cloud top bar, select Products > Search > Data > Datatyping (auto). Here, you can:
- Use Auto-Datatyping to parse your data automatically.
- Check for uncategorized data that didn’t match any Datatype rules.
- Handle the uncategorized data by adding custom Datatype rules.
See also:
- Datatypes in Cribl Search
- v2 Datatypes in Cribl Search
- List of Stock v2 Datatypes
- Add a Custom v2 Datatype
Add a Dataset and Set Dataset Rules
Next, create a Dataset and add Dataset rules to route parsed events into it.
Add a Dataset
- On the Cribl.Cloud top bar, select Products > Search > Data > Datasets.
- Select Add Dataset.
- Enter a unique ID.
- Optionally add a Description and Tags.
- Set Dataset Provider set to lakehouse.
- Select the Lakehouse engine that will store the data.
- Set the Retention period.
- Select Save.
Set Dataset Rules
Dataset rules enable you to route ingested events into individual Search Datasets so you can scope your queries and control retention.
See Organize Your Data for details around how to configure your Dataset rules and plan your Search Datasets based on estimated future storage costs.
Instrument Your Application
Pick Node.js or Python below. Both paths emit OTLP spans from LangChain and LangGraph through OpenInference instrumentation into the Source you configured earlier.
Use @arizeai/openinference-instrumentation-langchain with the OpenTelemetry Node SDK. LangChain.js requires manual instrumentation of @langchain/core/callbacks/manager because of how the package loads callbacks.
See the OpenInference LangChain.js documentation for compatibility between
@langchain/coreand instrumentation versions.
SDK-Based Setup
- Install the required packages:
npm install @opentelemetry/sdk-node \
@opentelemetry/api \
@opentelemetry/resources \
@opentelemetry/semantic-conventions \
@opentelemetry/exporter-trace-otlp-grpc \
@arizeai/openinference-instrumentation-langchain \
@langchain/core- Create a
tracing.jsfile and require it before your application code:
// tracing.js: load this BEFORE your app code
const { NodeSDK } = require('@opentelemetry/sdk-node');
const { Resource } = require('@opentelemetry/resources');
const { ATTR_SERVICE_NAME } = require('@opentelemetry/semantic-conventions');
const { OTLPTraceExporter } = require('@opentelemetry/exporter-trace-otlp-grpc');
const { LangChainInstrumentation } = require('@arizeai/openinference-instrumentation-langchain');
const CallbackManagerModule = require('@langchain/core/callbacks/manager');
const sdk = new NodeSDK({
resource: new Resource({
[ATTR_SERVICE_NAME]: process.env.OTEL_SERVICE_NAME || 'my-langchain-app',
}),
traceExporter: new OTLPTraceExporter(),
});
sdk.start();
const lcInstrumentation = new LangChainInstrumentation();
lcInstrumentation.manuallyInstrument(CallbackManagerModule);- Run your application:
node -r ./tracing.js app.jsThe exporter reads its endpoint from OTEL_EXPORTER_OTLP_ENDPOINT automatically.
Use OpenInference LangChainInstrumentor with Python 3.9+. Start with automatic instrumentation for the least code churn, or use code-based setup when you want to own tracer initialization and import order.
Example Applications
The examples below use optional packages such as langgraph, langchain-openai, and opentelemetry-instrumentation-httpx to trace outbound HTTP to model APIs. Set your provider API keys before running.
LangGraph ReAct agent (stable on current PyPI):
from langchain_core.messages import HumanMessage
from langchain_core.tools import tool
from langchain_openai import ChatOpenAI
from langgraph.prebuilt import create_react_agent
@tool
def add_numbers(a: int, b: int) -> int:
"""Add two integers and return the sum."""
return a + b
llm = ChatOpenAI(model="gpt-4o-mini", temperature=0)
agent = create_react_agent(llm, [add_numbers])
result = agent.invoke(
{"messages": [HumanMessage(content="What is 42 + 58? Use the tool.")]}
)
print(result["messages"][-1].content)LCEL chain (minimal):
from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate
from langchain_openai import ChatOpenAI
prompt = ChatPromptTemplate.from_messages(
[("system", "Answer in one short sentence."), ("user", "{input}")]
)
chain = prompt | ChatOpenAI(model="gpt-4o-mini", temperature=0) | StrOutputParser()
print(chain.invoke({"input": "What is OpenTelemetry?"}))Auto-Instrumentation
Install the OpenInference LangChain instrumentor packages and run your app under opentelemetry-instrument so LangChain and LangGraph emit spans with minimal code changes.
- Install the distro and OpenInference LangChain instrumentation packages:
pip install opentelemetry-distro opentelemetry-exporter-otlp \
opentelemetry-instrumentation-httpx \
openinference-instrumentation-langchain \
langchain langchain-openai langgraph- Optionally, bootstrap additional instrumentation packages:
opentelemetry-bootstrap -a install- Run your application with the auto-instrumentor:
opentelemetry-instrument python app.pyThe instrumentor reads its configuration from OTEL_* environment variables.
Code-Based Setup
Initialize OpenTelemetry and LangChainInstrumentor in a dedicated module, and import that module before you build or invoke LangChain or LangGraph chains or agents.
- Create a
tracing.pyfile:
# tracing.py: import this BEFORE your app code
from openinference.instrumentation.langchain import LangChainInstrumentor
from opentelemetry import trace
from opentelemetry.sdk.resources import Resource
from opentelemetry.sdk.trace import TracerProvider
from opentelemetry.sdk.trace.export import BatchSpanProcessor
from opentelemetry.exporter.otlp.proto.grpc.trace_exporter import OTLPSpanExporter
resource = Resource.create({"service.name": "my-langchain-app"})
provider = TracerProvider(resource=resource)
provider.add_span_processor(BatchSpanProcessor(OTLPSpanExporter()))
trace.set_tracer_provider(provider)
LangChainInstrumentor().instrument(tracer_provider=provider)- Import
tracingat the top of your application entry point before any LangChain or LangGraph calls:
import tracing # must be first
# ... then import and run LangChain / LangGraph codeSet Environment Variables
Set these environment variables before running your instrumented application.
Use gRPC (port 4317)
Set environment variables using gRPC. This is the recommended method:
export OTEL_EXPORTER_OTLP_ENDPOINT="http://<cribl-host>:4317"
export OTEL_EXPORTER_OTLP_PROTOCOL="grpc"
export OTEL_SERVICE_NAME="my-langchain-app"
# Only if you enabled auth on the Cribl OTEL Source:
export OTEL_EXPORTER_OTLP_HEADERS="Authorization=Bearer <token>"Use HTTP/protobuf (port 4318)
If your OTLP exporter or network path requires HTTP/protobuf instead of gRPC, use port 4318 and set the protocol as shown.
export OTEL_EXPORTER_OTLP_ENDPOINT="http://<cribl-host>:4318"
export OTEL_EXPORTER_OTLP_PROTOCOL="http/protobuf"
export OTEL_SERVICE_NAME="my-langchain-app"
# Only if you enabled auth on the Cribl OTEL Source:
export OTEL_EXPORTER_OTLP_HEADERS="Authorization=Bearer <token>"Environment Variable Reference
Use this table to set OTLP export options, optional auth headers for the Cribl Source, and provider API keys for the models your LangChain stack calls.
| Variable | Required | Description |
|---|---|---|
OTEL_EXPORTER_OTLP_ENDPOINT | Yes | Cribl Search OpenTelemetry Source URL (http://<cribl-host>:4317 for gRPC, :4318 for HTTP). |
OTEL_EXPORTER_OTLP_PROTOCOL | Yes | grpc or http/protobuf. |
OTEL_SERVICE_NAME | Yes | Logical name for your application (appears in traces). |
OTEL_EXPORTER_OTLP_HEADERS | No | Auth header if the Source requires it (e.g., Authorization=Bearer <token>). |
| Provider API keys | As needed | For example OPENAI_API_KEY, ANTHROPIC_API_KEY, or keys for other model backends your LangChain stack uses. |
Capture Prompt/Completion Content
Enabling content capture may expose sensitive data in your telemetry pipeline.
By default, prompt and completion content is not captured to protect sensitive data. To enable it:
export OTEL_INSTRUMENTATION_GENAI_CAPTURE_MESSAGE_CONTENT="true"Verify Data Flow in Cribl Search
After completing the instrumentation process, confirm OTLP events flow from your instrumented app into Cribl Search.
- On the Cribl.Cloud top bar, select Products > Search > Data > Live Data.
- Select your OpenTelemetry Source and confirm that events appear while you trigger traffic from your LangChain application.
- For full Live Data behavior and troubleshooting, check See Live Data Flow.
Common Search Issues and Fixes
| Symptom | Cause | Fix |
|---|---|---|
| No events returned. | OpenTelemetry Source is not receiving data, or Dataset rules are not routing data correctly. | Verify the Source address, port, protocol, and auth settings. Check Data > Live Data for the OpenTelemetry Source, then confirm your Dataset rule sends matching events to the correct Search Dataset. |
| Empty time range. | No recent traffic, or event timestamps are outside your selected window. | Widen Time range; generate a test LangChain or LangGraph run from your instrumented app. Review timestamp parsing in your Datatype configuration. |
| No spans from LangChain. | Instrumentation initialized after LangChain loaded callbacks. | Run LangChainInstrumentor().instrument() or import tracing before importing modules that construct chains or agents. For Node.js, load tracing.js before other app imports. |
create_agent missing (Python). | Older langchain package. | Upgrade LangChain or use the LangGraph / LCEL examples above. |
| 401 from a model provider. | Missing or invalid API key. | Set the correct provider key (for example OPENAI_API_KEY) in the process environment. |
| Fields not where you expect. | Auto-Datatyping did not classify the events as expected, or your Datatype rules need refinement. | Review the data in Live Data, check for Uncategorized events, and update or add Datatype rules so the data parses into the fields you expect. |
| Slow or expensive searches. | Large partitions scanned. | Narrow time range, increase sampling, or pre-aggregate in Stream. |
| Permission errors. | Insufficient Search role. | Ask an Admin to grant Editor (or higher) for Dataset management, or User for running searches per org policy. |
LLM Telemetry Use Cases
With data flowing into Cribl Search, you can explore and gain visibility towards your LangChain telemetry:
- Explore LLM telemetry: Explore your Search Dataset in Cribl Search and query it with KQL. Filter and slice by service, model, environment, token usage, cost signals, and error or status fields.
- Investigate incidents and performance regressions: During elevated errors or latency, investigate your LLM Search Dataset. Use aggregations such as
summarizeortimestatsto break down failures by model, deployment, feature, tenant, or region. Usejoinor related patterns with infrastructure, gateway, or security Datasets for end-to-end context. - Build dashboards for LLM usage and cost: Use Cribl Search Dashboards to track requests over time by model or app, token usage and derived cost, error rates by environment or tenant, and latency percentiles (for example P95/P99) plus token distributions and traffic share.
- Automate checks with scheduled searches and notifications: Turn important queries (for example, when estimated LLM cost or error rate crosses a threshold over a time window) into Scheduled Searches and attach Notifications to email, Slack, SNS, webhooks, and other targets.
Learn how to instrument LangChain and LangGraph applications with OpenTelemetry, export large language model (LLM) telemetry via OTLP, and send it into Cribl Stream. Once data is in Cribl Stream, you can route, mask, and enrich it before it reaches your observability backends.
You’ll complete the following high-level steps:
- Configure an OpenTelemetry (OTel) Source in Cribl Stream to receive OTLP data.
- Instrument your LangChain application (Node.js or Python) with OpenTelemetry.
- Explore LLM telemetry use cases for Cribl Stream.
Telemetry Captured
LangChain and LangGraph runs produce trace spans with these attributes:
| Category | Data | Example |
|---|---|---|
| Timing | Duration per chain step, tool run, or graph node. | 1.2s for an LLM span, 45ms for a tool. |
| Structure | Chain or graph hierarchy, span names for LLM, retriever, tool, or agent steps. | ChatOpenAI, tool:add_numbers, LangGraph node names. |
| Model usage | Model id and token-related fields when the underlying LLM instrumentation is active. | gpt-4o-mini, input/output tokens. |
| Tools | Tool name, arguments metadata (content may be opt-in). | Tool id, invocation status. |
| Errors | Exceptions or API errors on the relevant span. | HTTP 429, validation errors. |
| Content | Prompt and completion text. | (opt-in, disabled by default). |
Semantic conventions follow OpenTelemetry GenAI Semantic Conventions where applicable.
Prerequisites
- A Cribl Stream instance (v4.x+) with an OpenTelemetry (OTEL) Source enabled.
- An OpenAI API key or other provider credentials as your LangChain stack requires (examples below use
langchain-openaiandOPENAI_API_KEY). - Node.js 18+ or Python 3.9+ depending on your application language.
Configure the OpenTelemetry (OTel) Source in Cribl Stream
Before instrumenting your LangChain app, configure Cribl Stream to receive OTLP data.
Configure an OTel Source
To receive OTLP from your LLM application, add or edit an OpenTelemetry Source on your Worker Group as follows.
On the top bar, select Products, and then select Cribl Stream. Under Worker Groups, select a Worker Group. Next, you have two options:
- To configure via QuickConnect, navigate to Routing > QuickConnect (Stream) or Collect (Edge). Select Add Source and select the Source you want from the list, choosing either Select Existing or Add New.
- To configure via the Routes, select Data > Sources (Stream) or More > Sources (Edge). Select the Source you want. Next, select Add Source.
In the New Source modal, configure the following under General Settings:
- Input ID: Unique ID for this Source. For example,
OTel042. - Description: Optionally, enter a description.
- OTLP version: The drop-down offers
0.10.0and1.3.1(default). - Protocol: Use the drop-down to choose the protocol matching the data you will ingest:
gRPC(default), orHTTP. - Address: Enter the hostname/IP to listen on. Defaults to
0.0.0.0(all addresses, IPv4 format). - Port: By default, OTel applications send output to port
4317when using the gRPC protocol, and port4318when using HTTP. This setting defaults to4317- you must change it if you set Protocol (below) toHTTP, or you want Cribl Stream to collect data from an OTel application that is using a different port.
Port
4318is not available on Cribl-managed Worker Groups in Cribl.Cloud.- The Extract spans, Extract metrics, and Extract logs settings are specific to the OpenTelemetry Source. By default, these options are toggled off, allowing Cribl Stream to act as a pass-through that generates a single event for each incoming OTel payload. This is useful when you want to forward complete OTel events to downstream systems, such as persistent storage, without breaking them apart. You can enable these settings to extract and process individual records from within OTel events:
- Extract spans: Generates an individual event for each span in a trace. Traces typically contain multiple spans.
- Extract metrics: Generates an individual event for each data point in a metric event. OTel metrics often contain multiple data points per event.
- Extract logs: Available only when OTLP version is set to
1.3.1. Generates an individual event for each log record. Cribl recommends enabling this option to simplify log transformation and manipulation.
- Tags: Optionally, add tags to help filter and group Sources within Cribl Stream’s UI. Tags are not included in the event data. Separate tag names using a tab or hard return.
- Input ID: Unique ID for this Source. For example,
Optionally, you can adjust the Authentication, TLS, Persistent Queue Settings, Processing, and Advanced settings, or Connected Destinations.
Select Save, then Commit & Deploy.
Instrument Your Application
Pick Node.js or Python below. Both paths emit OTLP spans from LangChain and LangGraph through OpenInference instrumentation into the Source you configured earlier.
Use @arizeai/openinference-instrumentation-langchain with the OpenTelemetry Node SDK. LangChain.js requires manual instrumentation of @langchain/core/callbacks/manager because of how the package loads callbacks.
See the OpenInference LangChain.js documentation for compatibility between
@langchain/coreand instrumentation versions.
SDK-Based Setup
- Install the required packages:
npm install @opentelemetry/sdk-node \
@opentelemetry/api \
@opentelemetry/resources \
@opentelemetry/semantic-conventions \
@opentelemetry/exporter-trace-otlp-grpc \
@arizeai/openinference-instrumentation-langchain \
@langchain/core- Create a
tracing.jsfile and require it before your application code:
// tracing.js: load this BEFORE your app code
const { NodeSDK } = require('@opentelemetry/sdk-node');
const { Resource } = require('@opentelemetry/resources');
const { ATTR_SERVICE_NAME } = require('@opentelemetry/semantic-conventions');
const { OTLPTraceExporter } = require('@opentelemetry/exporter-trace-otlp-grpc');
const { LangChainInstrumentation } = require('@arizeai/openinference-instrumentation-langchain');
const CallbackManagerModule = require('@langchain/core/callbacks/manager');
const sdk = new NodeSDK({
resource: new Resource({
[ATTR_SERVICE_NAME]: process.env.OTEL_SERVICE_NAME || 'my-langchain-app',
}),
traceExporter: new OTLPTraceExporter(),
});
sdk.start();
// Essential for LangChain.js: Manual hook into the callback manager
const lcInstrumentation = new LangChainInstrumentation();
lcInstrumentation.manuallyInstrument(CallbackManagerModule);- Run your application:
node -r ./tracing.js app.jsThe exporter reads its endpoint from OTEL_EXPORTER_OTLP_ENDPOINT automatically.
Use OpenInference LangChainInstrumentor with Python 3.9+. Start with automatic instrumentation for the least code churn, or use code-based setup when you want to own tracer initialization and import order.
Example Applications
The examples below use optional packages such as langgraph, langchain-openai, and opentelemetry-instrumentation-httpx to trace outbound HTTP to model APIs. Set your provider API keys before running.
LangGraph ReAct agent (stable on current PyPI):
from langchain_core.messages import HumanMessage
from langchain_core.tools import tool
from langchain_openai import ChatOpenAI
from langgraph.prebuilt import create_react_agent
@tool
def add_numbers(a: int, b: int) -> int:
"""Add two integers and return the sum."""
return a + b
llm = ChatOpenAI(model="gpt-4o-mini", temperature=0)
agent = create_react_agent(llm, [add_numbers])
result = agent.invoke(
{"messages": [HumanMessage(content="What is 42 + 58? Use the tool.")]}
)
print(result["messages"][-1].content)LCEL chain (minimal):
from langchain_core.output_parsers import StrOutputParser
from langchain_core.prompts import ChatPromptTemplate
from langchain_openai import ChatOpenAI
prompt = ChatPromptTemplate.from_messages(
[("system", "Answer in one short sentence."), ("user", "{input}")]
)
chain = prompt | ChatOpenAI(model="gpt-4o-mini", temperature=0) | StrOutputParser()
print(chain.invoke({"input": "What is OpenTelemetry?"}))If langchain.agents.create_agent is not available in your environment, use LangGraph or LCEL as above.
Auto-Instrumentation
Install the OpenInference LangChain instrumentor packages and run your app under opentelemetry-instrument so LangChain and LangGraph emit spans with minimal code changes.
- Install the distro and OpenInference LangChain instrumentation packages:
pip install opentelemetry-distro opentelemetry-exporter-otlp \
opentelemetry-instrumentation-httpx \
openinference-instrumentation-langchain \
langchain langchain-openai langgraph- Optionally, bootstrap additional instrumentation packages:
opentelemetry-bootstrap -a install- Run your application with the auto-instrumentor:
opentelemetry-instrument python app.pyThe instrumentor reads its configuration from OTEL_* environment variables.
Code-Based Setup
Initialize OpenTelemetry and LangChainInstrumentor in a dedicated module, and import that module before you build or invoke LangChain or LangGraph chains or agents.
- Create a
tracing.pyfile:
# tracing.py: import this BEFORE your app code
from openinference.instrumentation.langchain import LangChainInstrumentor
from opentelemetry import trace
from opentelemetry.sdk.resources import Resource
from opentelemetry.sdk.trace import TracerProvider
from opentelemetry.sdk.trace.export import BatchSpanProcessor
from opentelemetry.exporter.otlp.proto.grpc.trace_exporter import OTLPSpanExporter
resource = Resource.create({"service.name": "my-langchain-app"})
provider = TracerProvider(resource=resource)
provider.add_span_processor(BatchSpanProcessor(OTLPSpanExporter()))
trace.set_tracer_provider(provider)
LangChainInstrumentor().instrument(tracer_provider=provider)- Import
tracingat the top of your application entry point before any LangChain or LangGraph calls:
import tracing # must be first
# ... then import and run LangChain / LangGraph codeSet Environment Variables
Set these environment variables before running your instrumented application:
Use gRPC (port 4317)
Set environment variables using gRPC. This is the recommended method:
export OTEL_EXPORTER_OTLP_ENDPOINT="http://<cribl-host>:4317"
export OTEL_EXPORTER_OTLP_PROTOCOL="grpc"
export OTEL_SERVICE_NAME="my-langchain-app"
# Only if you enabled auth on the Cribl OTEL Source:
export OTEL_EXPORTER_OTLP_HEADERS="Authorization=Bearer <token>"Use HTTP/protobuf (port 4318)
If your OTLP exporter or network path requires HTTP/protobuf instead of gRPC, use port 4318 and set the protocol as shown.
export OTEL_EXPORTER_OTLP_ENDPOINT="http://<cribl-host>:4318"
export OTEL_EXPORTER_OTLP_PROTOCOL="http/protobuf"
export OTEL_SERVICE_NAME="my-langchain-app"
# Only if you enabled auth on the Cribl OTEL Source:
export OTEL_EXPORTER_OTLP_HEADERS="Authorization=Bearer <token>"Environment Variable Reference
Use this table to set OTLP export options, optional auth headers for the Cribl Source, and provider API keys for the models your LangChain stack calls.
| Variable | Required | Description |
|---|---|---|
OTEL_EXPORTER_OTLP_ENDPOINT | Yes | Cribl Stream OTEL Source URL (http://<cribl-host>:4317 for gRPC, :4318 for HTTP). |
OTEL_EXPORTER_OTLP_PROTOCOL | Yes | grpc or http/protobuf. |
OTEL_SERVICE_NAME | Yes | Logical name for your application (appears in traces). |
OTEL_EXPORTER_OTLP_HEADERS | No | Auth header if Cribl Source requires it (e.g., Authorization=Bearer <token>). |
| Provider API keys | As needed | For example OPENAI_API_KEY or other keys your LangChain stack requires. |
Capture Prompt/Completion Content
Enabling content capture may expose sensitive data in your telemetry pipeline. Use Cribl Stream’s masking and redaction functions to sanitize data before routing to downstream destinations.
By default, prompt and completion content is not captured to protect sensitive data. To enable it:
export OTEL_INSTRUMENTATION_GENAI_CAPTURE_MESSAGE_CONTENT="true"Verify Data Flow in Cribl Stream
After completing the instrumentation process, confirm OTLP events flow from your instrumented app into Cribl Stream.
- In Cribl Stream, go to Monitoring > Sources and select your OTel Source.
- Open the Live Data tab.
- Select Start Capture.
- Trigger a request in your LangChain application.
- Confirm that events appear in Live Data for your OTel Source.
Route Data to Cribl Search
After LangChain OTLP data reaches Cribl Stream, add a Cribl Search Destination so you can explore and investigate LLM telemetry to troubleshoot issues, analyze model behavior, and monitor performance. See the Cribl Search Destination guide for setup instructions.
Common Stream Issues and Fixes
If spans are missing or the exporter cannot reach Cribl Stream, use this table to narrow down endpoint, authentication, TLS, library versions, and how instrumentation is loaded.
| Symptom | Cause | Fix |
|---|---|---|
| No data in Cribl Stream. | Wrong endpoint or port. | Verify OTEL_EXPORTER_OTLP_ENDPOINT matches the Cribl OTEL Source address and port. |
| Connection refused. | Source not running or firewall blocking. | Ensure the OTEL Source is enabled and the port is open. |
| 401 Unauthorized. | Auth mismatch. | Check OTEL_EXPORTER_OTLP_HEADERS matches the auth config on the Cribl Source. |
| Missing spans. | Instrumentation not loaded. | Ensure tracing.js is required first (Node.js) or opentelemetry-instrument is used (Python). For LangChain, also run LangChainInstrumentor().instrument() or manuallyInstrument(CallbackManagerModule) before building or invoking chains or agents. |
create_agent not importable (Python). | Using legacy or version-specific LangChain helper functions. | Upgrade langchain or use the LangGraph / LCEL examples above. |
| 401 from a model provider. | Invalid or missing API key. | Set the correct provider environment variables (for example OPENAI_API_KEY). |
| TLS errors. | TLS mismatch. | Use https:// in the endpoint if TLS is enabled on the Source, or disable TLS on the Source for local testing. |
LLM Telemetry Use Cases
With data flowing into Cribl Stream, you can route LangChain telemetry to any supported Destination. For example:
- Cribl Search: Search and analyze full-fidelity LLM traces and logs interactively.
- Cribl Lake: Store and search traces natively with partitioning optimized for Cribl Search, so you can query LangChain traces at scale without re-ingesting them into a separate system.
- Amazon S3 or other object storage: Archive traces for compliance or long-term analysis using partitioning schemes aligned with Cribl Search, so you can query data in place without moving it.
- OpenTelemetry: Forward LLM traces in standard OTLP format to downstream systems.
To set up routing:
- Add Routes in Routing > Data Routes to match OTel data and direct it to your chosen Destinations.
- Go to Processing > Pipelines and create a Pipeline to process data flowing down your Route.
- Use Functions to enrich, filter, sample, or redact data before delivery. For example:
- Mask prompt/completion content containing sensitive information.
- Sample high-volume traces to reduce cost.
- Aggregate token usage metrics by model or service.
For more detailed LLM telemetry use cases, see LLM Telemetry Use Cases in Cribl.