Osaurus includes built-in developer tools for debugging, monitoring, and testing your integration. Access them via the Management window (⌘ Shift M).
The Insights tab provides real-time monitoring of all API requests flowing through Osaurus.
- Open the Management window (
⌘ Shift M) - Click Insights in the sidebar
Every API request is logged with:
| Field | Description |
|---|---|
| Time | Request timestamp |
| Source | Origin: Chat UI or HTTP API |
| Method | HTTP method (GET/POST) |
| Path | Request endpoint |
| Status | HTTP status code |
| Duration | Total response time |
Click any row to expand and see full request/response details.
Filter requests to find what you need:
| Filter | Options |
|---|---|
| Search | Filter by path or model name |
| Method | All, GET only, POST only |
| Source | All, Chat UI, HTTP API |
The stats bar shows real-time metrics:
| Stat | Description |
|---|---|
| Requests | Total request count |
| Success | Success rate percentage |
| Avg Time | Average response duration |
| Errors | Total error count |
| Inferences | Chat completion requests (if any) |
| Avg Speed | Average tokens/second (for inference) |
Expand a request row to see:
Request Panel:
- Full request body (formatted JSON)
- Copy to clipboard
Response Panel:
- Full response body (formatted JSON)
- Status indicator (green for success, red for error)
- Response duration
- Copy to clipboard
Inference Details (for chat completions):
- Model used
- Token counts (input → output)
- Generation speed (tok/s)
- Temperature
- Max tokens
- Finish reason
Tool Calls (if applicable):
- Tool name
- Arguments
- Duration
- Success/error status
- Debugging API integration — See exactly what's being sent and received
- Performance monitoring — Track latency and throughput
- Tool call inspection — Debug tool calling behavior
- Error investigation — Understand why requests fail
The Server tab provides an interactive API reference and testing interface.
- Open the Management window (
⌘ Shift M) - Click Server in the sidebar
View current server state:
| Info | Description |
|---|---|
| Server URL | Base URL for API requests |
| Status | Running, Stopped, Starting, etc. |
Copy the server URL with one click for use in your applications.
Browse all available endpoints, organized by category:
| Category | Endpoints |
|---|---|
| Core | /, /health, /models, /tags |
| Chat | /chat/completions, /chat, /messages, /responses |
| Audio | /audio/transcriptions |
| MCP | /mcp/health, /mcp/tools, /mcp/call |
Each endpoint shows:
- HTTP method (GET/POST)
- Path
- Compatibility badge (OpenAI, Ollama, Anthropic, Open Responses, MCP)
- Description
Test any endpoint directly:
- Click an endpoint row to expand it
- For POST requests, edit the JSON payload
- Click Send Request
- View the formatted response
Request Panel (left):
- Editable JSON payload for POST requests
- Request preview for GET requests
- Reset button to restore default payload
- Send Request button
Response Panel (right):
- Formatted response body
- Status code badge
- Response duration
- Copy button
- Clear button
Quick access to the full documentation at docs.osaurus.ai.
- API exploration — Discover available endpoints
- Quick testing — Test endpoints without external tools
- Payload experimentation — Try different request formats
- Response inspection — See formatted API responses
- Open Insights
- Send a request from your application
- Find the request in the log (filter by path if needed)
- Expand to see request/response details
- Check for errors in the response
- If using tools, inspect tool call details
- Open Server Explorer
- Expand
/chat/completions - Modify the payload to include tools:
{
"model": "foundation",
"messages": [{ "role": "user", "content": "What time is it?" }],
"tools": [
{
"type": "function",
"function": {
"name": "current_time",
"description": "Get the current time"
}
}
]
}- Click Send Request
- Observe the tool call in the response
- Check Insights for the full request flow
- Open Insights
- Run your test workload
- Observe:
- Avg Time (should be consistent)
- Success rate (should be high)
- Avg Speed for inference (tok/s)
- Expand slow requests to investigate
- Open Server Explorer
- Expand
GET /mcp/tools - Click Send Request
- Verify your expected tools are listed
- Test a specific tool with
POST /mcp/call
The Insights log grows over time. Use the Clear button to reset when debugging a specific issue.
Filter by source to distinguish between:
- Chat — Requests from the built-in chat UI
- HTTP — Requests from external applications
Use the copy button to quickly grab response payloads for debugging in other tools.
The Server Explorer requires the server to be running. If endpoints show as disabled, start the server first.
- OpenAI API Guide — API usage and examples
- FEATURES.md — Feature inventory
- README — Quick start guide