Chat inference via API Gateway (buffered responses) with multimodal support
  POST /api/v3/organizations/{organisation}/ai/chat    
 Sends requests to the AI API Gateway endpoint which buffers responses. Supports text, images, videos, and documents via base64 encoding. * * Multimodal Support: * - Text: Simple string content * - Images: Base64-encoded PNG, JPEG, GIF, WebP (up to 25MB) * - Videos: Base64-encoded MP4, MOV, WebM, etc. (up to 25MB) * - Documents: Base64-encoded PDF, DOCX, CSV, etc. (up to 25MB) * * Supported Models: * - Amazon Nova Lite, Micro, Pro (all support multimodal) * - Claude models (text only) * * Usage Tips: * - Use base64 encoding for images/videos < 5-10MB * - Place media before text prompts for best results * - Label multiple media files (e.g., ‘Image 1:’, ‘Image 2:’) * - Maximum 25MB total payload size
Authorizations
Parameters
Path Parameters
The organisation ID
Request Body required
Chat request with optional multimodal content blocks
object
Array of chat messages. Content can be a simple string or an array of content blocks for multimodal input.
object
Simple text message
Multimodal content blocks (text, image, video, document)
object
What's in this image?object
object
object
Base64-encoded image data
object
object
object
Base64-encoded video data
object
object
report.pdfobject
Base64-encoded document data
Model ID. Use Nova models for multimodal support.
amazon.nova-lite-v1:0Ignored in buffered mode, always returns complete response
Optional custom system prompt. When tools are enabled, this is prepended with tool usage guidance.
Custom stop sequences
Structured JSON output (Claude 3.5 Sonnet v1/v2, Nova Pro)
object
JSON Schema defining expected structure
object
Function calling configuration (Claude 3+, Nova Pro)
object
object
object
object
JSON Schema for function parameters
object
Responses
200
Chat inference completed
object
Generated response text or JSON
Model used for generation
amazon.nova-pro-v1:0Token usage information
object
Number of input tokens
25Number of output tokens
150Total tokens consumed
175500
Failed to perform chat inference
