The Logs Page
Navigate to Logs in the sidebar to access the log explorer.Log Entry Information
Each log entry shows:| Field | Description |
|---|---|
| Timestamp | When the request occurred |
| Endpoint | Which endpoint was called |
| Prompt | Which prompt was used |
| Model | LLM model used |
| Status | Success or failure |
| Latency | Time to complete (ms) |
| Tokens | Input + output tokens |
| Cost | Estimated cost |
Filtering Logs
Use filters to find specific requests:Time Range
Last hour, 24 hours, 7 days, or custom range
Endpoint
Filter by specific endpoint
Status
Success, failed, or all
Model
Filter by LLM model used
Search
Search logs by:- Request ID
- Input content (partial match)
- Output content (partial match)
Log Details
Click any log entry to see full details:Request Tab
Response Tab
Rendered Prompt Tab
See the actual text sent to the LLM after template rendering:Metadata Tab
| Field | Value |
|---|---|
| Request ID | req_xxxxxxxxxxxx |
| Endpoint | /api/v1/summarize |
| Prompt | summarizer-v2 |
| Model | gpt-4o |
| Temperature | 0.3 |
| Input Tokens | 256 |
| Output Tokens | 89 |
| Latency | 1,234ms |
| Cache Hit | No |
| Cost | $0.0012 |
Replaying Requests
Replay any logged request to:- Debug issues with the same input
- Compare results after prompt changes
- Test with different models or settings
Monitoring Metrics
Endpoint Metrics
On each endpoint’s dashboard:| Metric | Description |
|---|---|
| Requests | Total requests over time |
| Success Rate | Percentage of successful requests |
| Avg Latency | Average response time |
| Token Usage | Total tokens consumed |
| Cost | Total estimated cost |
Charts
- Request Volume — Requests over time
- Latency Distribution — P50, P95, P99 latencies
- Error Rate — Failures over time
- Model Usage — Breakdown by model
Setting Up Alerts
Custom alerting is available on Pro and Enterprise plans.
- Error rate exceeds threshold
- Latency exceeds threshold
- Daily cost exceeds budget
- Unusual traffic patterns
Debugging Common Issues
High Latency
- Check the model — Some models are slower
- Check input size — Large inputs take longer
- Check time of day — Provider congestion varies
- Check max tokens — Higher limits may cause longer generation
High Error Rate
- Check validation errors — Are inputs malformed?
- Check provider status — Is OpenAI/Anthropic down?
- Check rate limits — Are you exceeding limits?
- Check prompt — Is it producing valid JSON?
Unexpected Outputs
- View rendered prompt — Is the template rendering correctly?
- Check temperature — Too high may cause variations
- Review recent changes — Did the prompt change?
- Replay request — Reproduce the issue
Log Retention
| Plan | Retention |
|---|---|
| Free | 7 days |
| Pro | 30 days |
| Enterprise | 90 days (or custom) |
Exporting Logs
Export logs for external analysis:- Apply filters to select the logs you need
- Click Export
- Choose format (CSV or JSON)
- Download the file
API Access to Logs
Logs API is available on Enterprise plans.

