Models are grouped by what they produce (text, audio, video, image, or 3D assets) so you can jump straight to the modality you care about. Each category page explains typical technical topics and lists every model in that group with links to detailed pages.Documentation Index
Fetch the complete documentation index at: https://docs.gmicloud.ai/llms.txt
Use this file to discover all available pages before exploring further.
Browse by modality
| Modality | What it covers | Start here |
|---|---|---|
| Language | Chat, code, reasoning, OCR / vision-language | LLM models |
| Audio | TTS, voice cloning, music | Audio models |
| Video | Text-to-video, image-to-video, editing, avatars | Video models |
| Image | Generation, editing, batch inference | Image models |
Model API (serving, pricing, limits)
Individual model pages describe that model’s inputs and examples. For how to call the platform end-to-end (marketplace, serverless vs dedicated deployments, LLM and video API references, SDKs, rate limits, billing, tasks, and artifacts), use Model API in the sidebar.Choosing a model
- Match modality — Pick LLM vs audio vs video vs image vs 3D first; hybrid needs may use multiple APIs.
- Read constraints on the model page — Context length, resolution, duration, and rate limits vary.
- Start with defaults — Official examples on each page reflect supported parameters today.