mirror of
https://github.com/katanemo/plano.git
synced 2026-04-25 00:36:34 +02:00
* Rename all arch references to plano across the codebase
Complete rebrand from "Arch"/"archgw" to "Plano" including:
- Config files: arch_config_schema.yaml, workflow, demo configs
- Environment variables: ARCH_CONFIG_* → PLANO_CONFIG_*
- Python CLI: variables, functions, file paths, docker mounts
- Rust crates: config paths, log messages, metadata keys
- Docker/build: Dockerfile, supervisord, .dockerignore, .gitignore
- Docker Compose: volume mounts and env vars across all demos/tests
- GitHub workflows: job/step names
- Shell scripts: log messages
- Demos: Python code, READMEs, VS Code configs, Grafana dashboard
- Docs: RST includes, code comments, config references
- Package metadata: package.json, pyproject.toml, uv.lock
External URLs (docs.archgw.com, github.com/katanemo/archgw) left as-is.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
* Update remaining arch references in docs
- Rename RST cross-reference labels: arch_access_logging, arch_overview_tracing, arch_overview_threading → plano_*
- Update label references in request_lifecycle.rst
- Rename arch_config_state_storage_example.yaml → plano_config_state_storage_example.yaml
- Update config YAML comments: "Arch creates/uses" → "Plano creates/uses"
- Update "the Arch gateway" → "the Plano gateway" in configuration_reference.rst
- Update arch_config_schema.yaml reference in provider_models.py
- Rename arch_agent_router → plano_agent_router in config example
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
* Fix remaining arch references found in second pass
- config/docker-compose.dev.yaml: ARCH_CONFIG_FILE → PLANO_CONFIG_FILE,
arch_config.yaml → plano_config.yaml, archgw_logs → plano_logs
- config/test_passthrough.yaml: container mount path
- tests/e2e/docker-compose.yaml: source file path (was still arch_config.yaml)
- cli/planoai/core.py: comment and log message
- crates/brightstaff/src/tracing/constants.rs: doc comment
- tests/{e2e,archgw}/common.py: get_arch_messages → get_plano_messages,
arch_state/arch_messages variables renamed
- tests/{e2e,archgw}/test_prompt_gateway.py: updated imports and usages
- demos/shared/test_runner/{common,test_demos}.py: same renames
- tests/e2e/test_model_alias_routing.py: docstring
- .dockerignore: archgw_modelserver → plano_modelserver
- demos/use_cases/claude_code_router/pretty_model_resolution.sh: container name
Note: x-arch-* HTTP header values and Rust constant names intentionally
preserved for backwards compatibility with existing deployments.
Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
---------
Co-authored-by: Claude Opus 4.6 <noreply@anthropic.com>
560 lines
19 KiB
ReStructuredText
560 lines
19 KiB
ReStructuredText
.. _plano_overview_tracing:
|
|
|
|
Tracing
|
|
=======
|
|
|
|
Overview
|
|
--------
|
|
|
|
`OpenTelemetry <https://opentelemetry.io/>`_ is an open-source observability framework providing APIs
|
|
and instrumentation for generating, collecting, processing, and exporting telemetry data, such as traces,
|
|
metrics, and logs. Its flexible design supports a wide range of backends and seamlessly integrates with
|
|
modern application tools. A key feature of OpenTelemetry is its commitment to standards like the
|
|
`W3C Trace Context <https://www.w3.org/TR/trace-context/>`_
|
|
|
|
**Tracing** is a critical tool that allows developers to visualize and understand the flow of
|
|
requests in an AI application. With tracing, you can capture a detailed view of how requests propagate
|
|
through various services and components, which is crucial for **debugging**, **performance optimization**,
|
|
and understanding complex AI agent architectures like Co-pilots.
|
|
|
|
**Plano** propagates trace context using the W3C Trace Context standard, specifically through the
|
|
``traceparent`` header. This allows each component in the system to record its part of the request
|
|
flow, enabling **end-to-end tracing** across the entire application. By using OpenTelemetry, Plano ensures
|
|
that developers can capture this trace data consistently and in a format compatible with various observability
|
|
tools.
|
|
|
|
.. image:: /_static/img/tracing.png
|
|
:width: 100%
|
|
:align: center
|
|
|
|
|
|
Understanding Plano Traces
|
|
--------------------------
|
|
|
|
Plano creates structured traces that capture the complete flow of requests through your AI system. Each trace consists of multiple spans representing different stages of processing.
|
|
|
|
Inbound Request Handling
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
When a request enters Plano, it creates an **inbound span** (``plano(inbound)``) that represents the initial request reception and processing. This span captures:
|
|
|
|
- HTTP request details (method, path, headers)
|
|
- Request payload size
|
|
- Initial validation and authentication
|
|
|
|
Orchestration & Routing
|
|
~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
For agent systems, Plano performs intelligent routing through orchestration spans:
|
|
|
|
- **Agent Orchestration** (``plano(orchestrator)``): When multiple agents are available, Plano uses an LLM to analyze the user's intent and select the most appropriate agent. This span captures the orchestration decision-making process.
|
|
|
|
- **LLM Routing** (``plano(routing)``): For direct LLM requests, Plano determines the optimal endpoint based on your routing strategy (round-robin, least-latency, cost-optimized). This span includes:
|
|
|
|
- Routing strategy used
|
|
- Selected upstream endpoint
|
|
- Route determination time
|
|
- Fallback indicators (if applicable)
|
|
|
|
Agent Processing
|
|
~~~~~~~~~~~~~~~~
|
|
|
|
When requests are routed to agents, Plano creates spans for agent execution:
|
|
|
|
- **Agent Filter Chains** (``plano(filter)``): If filters are configured (guardrails, context enrichment, query rewriting), each filter execution is captured in its own span, showing the transformation pipeline.
|
|
|
|
- **Agent Execution** (``plano(agent)``): The main agent processing span that captures the agent's work, including any tools invoked and intermediate reasoning steps.
|
|
|
|
Outbound LLM Calls
|
|
~~~~~~~~~~~~~~~~~~
|
|
|
|
All LLM calls—whether from Plano's routing layer or from agents—are traced with **LLM spans** (``plano(llm)``) that capture:
|
|
|
|
- Model name and provider (e.g., ``gpt-4``, ``claude-3-sonnet``)
|
|
- Request parameters (temperature, max_tokens, top_p)
|
|
- Token usage (prompt_tokens, completion_tokens)
|
|
- Streaming indicators and time-to-first-token
|
|
- Response metadata
|
|
|
|
**Example Span Attributes**::
|
|
|
|
# LLM call span
|
|
llm.model = "gpt-4"
|
|
llm.provider = "openai"
|
|
llm.usage.prompt_tokens = 150
|
|
llm.usage.completion_tokens = 75
|
|
llm.duration_ms = 1250
|
|
llm.time_to_first_token = 320
|
|
|
|
Handoff to Upstream Services
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
When Plano forwards requests to upstream services (agents, APIs, or LLM providers), it creates **handoff spans** (``plano(handoff)``) that capture:
|
|
|
|
- Upstream endpoint URL
|
|
- Request/response sizes
|
|
- HTTP status codes
|
|
- Upstream response times
|
|
|
|
This creates a complete end-to-end trace showing the full request lifecycle through all system components.
|
|
|
|
Behavioral Signals in Traces
|
|
----------------------------
|
|
|
|
Plano automatically enriches OpenTelemetry traces with :doc:`../../concepts/signals` — behavioral quality indicators computed from conversation patterns. These signals are attached as span attributes, providing immediate visibility into interaction quality.
|
|
|
|
**What Signals Provide**
|
|
|
|
Signals act as early warning indicators embedded in your traces:
|
|
|
|
- **Quality Assessment**: Overall interaction quality (Excellent/Good/Neutral/Poor/Severe)
|
|
- **Efficiency Metrics**: Turn count, efficiency scores, repair frequency
|
|
- **User Sentiment**: Frustration indicators, positive feedback, escalation requests
|
|
- **Agent Behavior**: Repetition detection, looping patterns
|
|
|
|
**Visual Flag Markers**
|
|
|
|
When concerning signals are detected (frustration, looping, escalation, or poor/severe quality), Plano automatically appends a flag marker **🚩** to the span's operation name. This makes problematic traces immediately visible in your tracing UI without requiring additional queries.
|
|
|
|
**Example Span with Signals**::
|
|
|
|
# Span name: "POST /v1/chat/completions gpt-4 🚩"
|
|
# Standard LLM attributes:
|
|
llm.model = "gpt-4"
|
|
llm.usage.total_tokens = 225
|
|
|
|
# Behavioral signal attributes:
|
|
signals.quality = "Severe"
|
|
signals.turn_count = 15
|
|
signals.efficiency_score = 0.234
|
|
signals.frustration.severity = 3
|
|
signals.escalation.requested = "true"
|
|
|
|
**Querying Signal Data**
|
|
|
|
In your observability platform (Jaeger, Grafana Tempo, Datadog, etc.), filter traces by signal attributes:
|
|
|
|
- Find severe interactions: ``signals.quality = "Severe"``
|
|
- Find frustrated users: ``signals.frustration.severity >= 2``
|
|
- Find inefficient flows: ``signals.efficiency_score < 0.5``
|
|
- Find escalations: ``signals.escalation.requested = "true"``
|
|
|
|
For complete details on all available signals, detection methods, and best practices, see the :doc:`../../concepts/signals` guide.
|
|
|
|
|
|
Benefits of Using ``Traceparent`` Headers
|
|
-----------------------------------------
|
|
|
|
- **Standardization**: The W3C Trace Context standard ensures compatibility across ecosystem tools, allowing
|
|
traces to be propagated uniformly through different layers of the system.
|
|
- **Ease of Integration**: OpenTelemetry's design allows developers to easily integrate tracing with minimal
|
|
changes to their codebase, enabling quick adoption of end-to-end observability.
|
|
- **Interoperability**: Works seamlessly with popular tracing tools like AWS X-Ray, Datadog, Jaeger, and many others,
|
|
making it easy to visualize traces in the tools you're already usi
|
|
|
|
How to Initiate A Trace
|
|
-----------------------
|
|
|
|
1. **Enable Tracing Configuration**: Simply add the ``random_sampling`` in ``tracing`` section to 100`` flag to in the :ref:`listener <plano_overview_listeners>` config
|
|
|
|
2. **Trace Context Propagation**: Plano automatically propagates the ``traceparent`` header. When a request is received, Plano will:
|
|
|
|
- Generate a new ``traceparent`` header if one is not present.
|
|
- Extract the trace context from the ``traceparent`` header if it exists.
|
|
- Start a new span representing its processing of the request.
|
|
- Forward the ``traceparent`` header to downstream services.
|
|
|
|
3. **Sampling Policy**: The 100 in ``random_sampling: 100`` means that all the requests as sampled for tracing.
|
|
You can adjust this value from 0-100.
|
|
|
|
|
|
Tracing with the CLI
|
|
--------------------
|
|
|
|
The Plano CLI ships with a local OTLP/gRPC listener and a trace viewer so you can inspect spans without wiring a full observability backend. This is ideal for development, debugging, and quick QA.
|
|
|
|
Quick Start
|
|
~~~~~~~~~~~
|
|
|
|
You can enable tracing in either of these ways:
|
|
|
|
1. Start the local listener explicitly:
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace listen
|
|
|
|
2. Or start Plano with tracing enabled (auto-starts the local OTLP listener):
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai up --with-tracing
|
|
|
|
# Optional: choose a different listener port
|
|
$ planoai up --with-tracing --tracing-port 4318
|
|
|
|
3. Send requests through Plano as usual. The listener accepts OTLP/gRPC on:
|
|
|
|
- ``0.0.0.0:4317`` (default)
|
|
|
|
4. View the most recent trace:
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace
|
|
|
|
Inspect and Filter Traces
|
|
~~~~~~~~~~~~~~~~~~~~~~~~~
|
|
|
|
List available trace IDs:
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace --list
|
|
|
|
Open a specific trace (full or short trace ID):
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace 7f4e9a1c
|
|
$ planoai trace 7f4e9a1c0d9d4a0bb9bf5a8a7d13f62a
|
|
|
|
Filter by attributes and time window:
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace --where llm.model=gpt-4o-mini --since 30m
|
|
$ planoai trace --filter "http.*" --limit 5
|
|
|
|
Return JSON for automation:
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace --json
|
|
$ planoai trace --list --json
|
|
|
|
Show full span attributes (disable default compact view):
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace --verbose
|
|
$ planoai trace -v
|
|
|
|
Point the CLI at a different local listener port:
|
|
|
|
.. code-block:: console
|
|
|
|
$ export PLANO_TRACE_PORT=50051
|
|
$ planoai trace --list
|
|
|
|
Notes
|
|
~~~~~
|
|
|
|
- ``--where`` accepts repeatable ``key=value`` filters and uses AND semantics.
|
|
- ``--filter`` supports wildcards (``*``) to limit displayed attributes.
|
|
- ``--no-interactive`` disables prompts when listing traces.
|
|
- By default, inbound/outbound spans use a compact attribute view.
|
|
|
|
|
|
Trace Propagation
|
|
-----------------
|
|
|
|
Plano uses the W3C Trace Context standard for trace propagation, which relies on the ``traceparent`` header.
|
|
This header carries tracing information in a standardized format, enabling interoperability between different
|
|
tracing systems.
|
|
|
|
Header Format
|
|
~~~~~~~~~~~~~
|
|
|
|
The ``traceparent`` header has the following format::
|
|
|
|
traceparent: {version}-{trace-id}-{parent-id}-{trace-flags}
|
|
|
|
- ``{version}``: The version of the Trace Context specification (e.g., ``00``).
|
|
- ``{trace-id}``: A 16-byte (32-character hexadecimal) unique identifier for the trace.
|
|
- ``{parent-id}``: An 8-byte (16-character hexadecimal) identifier for the parent span.
|
|
- ``{trace-flags}``: Flags indicating trace options (e.g., sampling).
|
|
|
|
Instrumentation
|
|
~~~~~~~~~~~~~~~
|
|
|
|
To integrate AI tracing, your application needs to follow a few simple steps. The steps
|
|
below are very common practice, and not unique to Plano, when you reading tracing headers and export
|
|
`spans <https://docs.lightstep.com/docs/understand-distributed-tracing>`_ for distributed tracing.
|
|
|
|
- Read the ``traceparent`` header from incoming requests.
|
|
- Start new spans as children of the extracted context.
|
|
- Include the ``traceparent`` header in outbound requests to propagate trace context.
|
|
- Send tracing data to a collector or tracing backend to export spans
|
|
|
|
Example with OpenTelemetry in Python
|
|
************************************
|
|
|
|
Install OpenTelemetry packages:
|
|
|
|
.. code-block:: console
|
|
|
|
$ pip install opentelemetry-api opentelemetry-sdk opentelemetry-exporter-otlp
|
|
$ pip install opentelemetry-instrumentation-requests
|
|
|
|
Set up the tracer and exporter:
|
|
|
|
.. code-block:: python
|
|
|
|
from opentelemetry import trace
|
|
from opentelemetry.exporter.otlp.proto.grpc.trace_exporter import OTLPSpanExporter
|
|
from opentelemetry.instrumentation.requests import RequestsInstrumentor
|
|
from opentelemetry.sdk.resources import Resource
|
|
from opentelemetry.sdk.trace import TracerProvider
|
|
from opentelemetry.sdk.trace.export import BatchSpanProcessor
|
|
|
|
# Define the service name
|
|
resource = Resource(attributes={
|
|
"service.name": "customer-support-agent"
|
|
})
|
|
|
|
# Set up the tracer provider and exporter
|
|
tracer_provider = TracerProvider(resource=resource)
|
|
otlp_exporter = OTLPSpanExporter(endpoint="otel-collector:4317", insecure=True)
|
|
span_processor = BatchSpanProcessor(otlp_exporter)
|
|
tracer_provider.add_span_processor(span_processor)
|
|
trace.set_tracer_provider(tracer_provider)
|
|
|
|
# Instrument HTTP requests
|
|
RequestsInstrumentor().instrument()
|
|
|
|
Handle incoming requests:
|
|
|
|
.. code-block:: python
|
|
|
|
from opentelemetry import trace
|
|
from opentelemetry.propagate import extract, inject
|
|
import requests
|
|
|
|
def handle_request(request):
|
|
# Extract the trace context
|
|
context = extract(request.headers)
|
|
tracer = trace.get_tracer(__name__)
|
|
|
|
with tracer.start_as_current_span("process_customer_request", context=context):
|
|
# Example of processing a customer request
|
|
print("Processing customer request...")
|
|
|
|
# Prepare headers for outgoing request to payment service
|
|
headers = {}
|
|
inject(headers)
|
|
|
|
# Make outgoing request to external service (e.g., payment gateway)
|
|
response = requests.get("http://payment-service/api", headers=headers)
|
|
|
|
print(f"Payment service response: {response.content}")
|
|
|
|
|
|
Integrating with Tracing Tools
|
|
------------------------------
|
|
|
|
AWS X-Ray
|
|
~~~~~~~~~
|
|
|
|
To send tracing data to `AWS X-Ray <https://aws.amazon.com/xray/>`_ :
|
|
|
|
1. **Configure OpenTelemetry Collector**: Set up the collector to export traces to AWS X-Ray.
|
|
|
|
Collector configuration (``otel-collector-config.yaml``):
|
|
|
|
.. code-block:: yaml
|
|
|
|
receivers:
|
|
otlp:
|
|
protocols:
|
|
grpc:
|
|
|
|
processors:
|
|
batch:
|
|
|
|
exporters:
|
|
awsxray:
|
|
region: <Your-Aws-Region>
|
|
|
|
service:
|
|
pipelines:
|
|
traces:
|
|
receivers: [otlp]
|
|
processors: [batch]
|
|
exporters: [awsxray]
|
|
|
|
2. **Deploy the Collector**: Run the collector as a Docker container, Kubernetes pod, or standalone service.
|
|
3. **Ensure AWS Credentials**: Provide AWS credentials to the collector, preferably via IAM roles.
|
|
4. **Verify Traces**: Access the AWS X-Ray console to view your traces.
|
|
|
|
Datadog
|
|
~~~~~~~
|
|
|
|
Datadog
|
|
|
|
To send tracing data to `Datadog <https://docs.datadoghq.com/getting_started/tracing/>`_:
|
|
|
|
1. **Configure OpenTelemetry Collector**: Set up the collector to export traces to Datadog.
|
|
|
|
Collector configuration (``otel-collector-config.yaml``):
|
|
|
|
.. code-block:: yaml
|
|
|
|
receivers:
|
|
otlp:
|
|
protocols:
|
|
grpc:
|
|
|
|
processors:
|
|
batch:
|
|
|
|
exporters:
|
|
datadog:
|
|
api:
|
|
key: "${<Your-Datadog-Api-Key>}"
|
|
site: "${DD_SITE}"
|
|
|
|
service:
|
|
pipelines:
|
|
traces:
|
|
receivers: [otlp]
|
|
processors: [batch]
|
|
exporters: [datadog]
|
|
|
|
2. **Set Environment Variables**: Provide your Datadog API key and site.
|
|
|
|
.. code-block:: console
|
|
|
|
$ export <Your-Datadog-Api-Key>=<Your-Datadog-Api-Key>
|
|
$ export DD_SITE=datadoghq.com # Or datadoghq.eu
|
|
|
|
3. **Deploy the Collector**: Run the collector in your environment.
|
|
4. **Verify Traces**: Access the Datadog APM dashboard to view your traces.
|
|
|
|
Langtrace
|
|
~~~~~~~~~
|
|
|
|
Langtrace is an observability tool designed specifically for large language models (LLMs). It helps you capture, analyze, and understand how LLMs are used in your applications including those built using Plano.
|
|
|
|
To send tracing data to `Langtrace <https://docs.langtrace.ai/supported-integrations/llm-tools/arch>`_:
|
|
|
|
1. **Configure Plano**: Make sure Plano is installed and setup correctly. For more information, refer to the `installation guide <https://github.com/katanemo/archgw?tab=readme-ov-file#prerequisites>`_.
|
|
|
|
2. **Install Langtrace**: Install the Langtrace SDK.:
|
|
|
|
.. code-block:: console
|
|
|
|
$ pip install langtrace-python-sdk
|
|
|
|
3. **Set Environment Variables**: Provide your Langtrace API key.
|
|
|
|
.. code-block:: console
|
|
|
|
$ export LANGTRACE_API_KEY=<Your-Langtrace-Api-Key>
|
|
|
|
4. **Trace Requests**: Once you have Langtrace set up, you can start tracing requests.
|
|
|
|
Here's an example of how to trace a request using the Langtrace Python SDK:
|
|
|
|
.. code-block:: python
|
|
|
|
import os
|
|
from langtrace_python_sdk import langtrace # Must precede any llm module imports
|
|
from openai import OpenAI
|
|
|
|
langtrace.init(api_key=os.environ['LANGTRACE_API_KEY'])
|
|
|
|
client = OpenAI(api_key=os.environ['OPENAI_API_KEY'], base_url="http://localhost:12000/v1")
|
|
|
|
response = client.chat.completions.create(
|
|
model="gpt-4o-mini",
|
|
messages=[
|
|
{"role": "system", "content": "You are a helpful assistant"},
|
|
{"role": "user", "content": "Hello"},
|
|
]
|
|
)
|
|
|
|
print(chat_completion.choices[0].message.content)
|
|
|
|
5. **Verify Traces**: Access the Langtrace dashboard to view your traces.
|
|
|
|
|
|
Best Practices
|
|
--------------
|
|
|
|
- **Consistent Instrumentation**: Ensure all services propagate the ``traceparent`` header.
|
|
- **Secure Configuration**: Protect sensitive data and secure communication between services.
|
|
- **Performance Monitoring**: Be mindful of the performance impact and adjust sampling rates accordingly.
|
|
- **Error Handling**: Implement proper error handling to prevent tracing issues from affecting your application.
|
|
|
|
Summary
|
|
----------
|
|
|
|
By leveraging the ``traceparent`` header for trace context propagation, Plano enables developers to implement
|
|
tracing efficiently. This approach simplifies the process of collecting and analyzing tracing data in common
|
|
tools like AWS X-Ray and Datadog, enhancing observability and facilitating faster debugging and optimization.
|
|
|
|
Additional Resources
|
|
--------------------
|
|
|
|
CLI Reference
|
|
~~~~~~~~~~~~~
|
|
|
|
``planoai trace``
|
|
Trace requests captured by the local OTLP listener.
|
|
|
|
**Synopsis**
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace [TARGET] [OPTIONS]
|
|
|
|
**Targets**
|
|
|
|
- ``last`` (default): show the most recent trace.
|
|
- ``any``: allow interactive selection when available.
|
|
- ``<trace-id>``: full 32-hex trace ID.
|
|
- ``<short-id>``: first 8 hex characters.
|
|
|
|
**Options**
|
|
|
|
- ``--filter <pattern>``: limit displayed attributes to matching keys (supports ``*``).
|
|
- ``--where <key=value>``: match traces containing a specific attribute (repeatable, AND).
|
|
- ``--list``: list trace IDs only.
|
|
- ``--no-interactive``: disable interactive prompts/selections.
|
|
- ``--limit <n>``: limit the number of traces returned.
|
|
- ``--since <window>``: look back window (``5m``, ``2h``, ``1d``).
|
|
- ``--json``: output raw JSON instead of formatted output.
|
|
- ``--verbose, -v``: show all span attributes. By default, inbound/outbound
|
|
spans are displayed in a compact view.
|
|
|
|
**Environment**
|
|
|
|
- ``PLANO_TRACE_PORT``: gRPC port used by ``planoai trace`` to query traces
|
|
(defaults to ``4317``).
|
|
|
|
``planoai trace listen``
|
|
Start a local OTLP/gRPC listener.
|
|
|
|
**Synopsis**
|
|
|
|
.. code-block:: console
|
|
|
|
$ planoai trace listen [OPTIONS]
|
|
|
|
**Options**
|
|
|
|
- ``--host <host>``: bind address (default: ``0.0.0.0``).
|
|
- ``--port <port>``: gRPC listener port (default: ``4317``).
|
|
|
|
External References
|
|
~~~~~~~~~~~~~~~~~~~
|
|
|
|
- `OpenTelemetry Documentation <https://opentelemetry.io/docs/>`_
|
|
- `W3C Trace Context Specification <https://www.w3.org/TR/trace-context/>`_
|
|
- `AWS X-Ray Exporter <https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/exporter/awsxrayexporter>`_
|
|
- `Datadog Exporter <https://github.com/open-telemetry/opentelemetry-collector-contrib/tree/main/exporter/datadogexporter>`_
|
|
- `Langtrace Documentation <https://docs.langtrace.ai/introduction>`_
|
|
|
|
.. Note::
|
|
Replace placeholders such as ``<Your-Aws-Region>`` and ``<Your-Datadog-Api-Key>`` with your actual configurations.
|