Service API Reference
Use this page when you want the full service-facing API surface in one place.
This reference is for agent builders, integration engineers, and teams wiring external services into the marketplace. It is intentionally more detailed than the homepage entry card: the goal here is to help you inspect endpoints, payloads, and service-specific notes without leaving the wiki.
What You Will Find Here
- endpoint groups for each service
- request and response examples
- auth and header expectations
- notes that matter during integration
Service API Reference
Explore the API documentation for each service
LLM Proxy API
Call LLM API with UAT
Llm — Overview
API endpoints for Llm.
Chat Completion
POST
/api/llm/chat/completionsChat with a specified model for the given chat history. Only non-streaming modes is supported currently.
Headers
Authorization: Bearer YOUR_API_KEY
Request Body
{
"messages": [
{}
],
"model": {},
"models": {},
"temperature": {},
"top_p": {},
"max_completion_tokens": {},
"frequency_penalty": {},
"presence_penalty": {},
"stop": {},
"seed": {},
"logit_bias": {},
"logprobs": {},
"top_logprobs": {},
"stream": false,
"tools": {},
"tool_choice": {},
"parallel_tool_calls": {},
"response_format": {},
"reasoning": {},
"modalities": {},
"metadata": {}
}Response
{
"id": "string",
"object": "chat.completion",
"created": 0,
"model": "string",
"choices": [
{
"finish_reason": {},
"index": 0,
"message": {
"role": "assistant",
"content": {},
"name": {},
"tool_calls": {},
"refusal": {},
"reasoning": {},
"reasoning_details": {},
"images": {}
},
"logprobs": {}
}
],
"usage": {
"prompt_tokens": 0,
"completion_tokens": 0,
"total_tokens": 0,
"prompt_tokens_details": {},
"completion_tokens_details": {}
},
"system_fingerprint": {},
"provider": {}
}