# AnyFast > AnyFast API documentation — unified access to AI models from Anthropic, ByteDance, OpenAI, and more. ## Docs - [Create Message](https://docs.anyfast.ai/api-reference/endpoints/anthropic.md): Send a structured list of input messages with text and/or image content, and the model will generate the next message in the conversation. - [Chat Completions](https://docs.anyfast.ai/api-reference/endpoints/doubao.md): Create a chat completion using Doubao models. The `content` field accepts a plain string for text, or an array of content parts for multimodal input. - [Create Response](https://docs.anyfast.ai/api-reference/endpoints/doubao-responses.md): Creates a model response using Doubao models. The `input` field accepts a plain string for simple text, or an array of message objects for multimodal content. - [Chat Completions](https://docs.anyfast.ai/api-reference/endpoints/gemini-chat.md): Creates a model response for Gemini models using OpenAI-compatible format. - [Generate Content](https://docs.anyfast.ai/api-reference/endpoints/gemini-native.md): Generates a model response given an input GenerateContentRequest. - [Chat Completions](https://docs.anyfast.ai/api-reference/endpoints/openai.md): Creates a model response for the given chat conversation. - [Introduction](https://docs.anyfast.ai/api-reference/introduction.md): Base URL, authentication, request format, and error handling for the AnyFast API. - [HappyHorse 1.0 I2V](https://docs.anyfast.ai/api-reference/model-api/alibaba/happyhorse-1.0-i2v.md): Submit an image-to-video generation task using a single first-frame image. Returns a `task_id` immediately. Poll `GET /v1/video/generations/{task_id}` for the result. - [HappyHorse 1.0 R2V](https://docs.anyfast.ai/api-reference/model-api/alibaba/happyhorse-1.0-r2v.md): Submit a video generation task with 1–3 reference images. Returns a `task_id` immediately. Poll `GET /v1/video/generations/{task_id}` for the result. - [HappyHorse 1.0 T2V](https://docs.anyfast.ai/api-reference/model-api/alibaba/happyhorse-1.0-t2v.md): Submit a text-to-video generation task. Returns a `task_id` immediately. Poll `GET /v1/video/generations/{task_id}` to check status and retrieve the video URL. - [HappyHorse 1.0 Video Edit](https://docs.anyfast.ai/api-reference/model-api/alibaba/happyhorse-1.0-video-edit.md): Submit a video editing task that combines an input video and a reference image. Must use the **native** request format with `input.media[]` array — flat top-level / unified-framework simplifications are not supported. Returns a `task_id` immediately. - [HappyHorse Task Query](https://docs.anyfast.ai/api-reference/model-api/alibaba/happyhorse-task-query.md): Retrieve the status and result of a HappyHorse video generation task by its ID. - [Qwen Image Plus](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen-image-plus.md): Generate an image from a text prompt. Returns the result synchronously. - [Qwen2-VL-72B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2-vl-72b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2-VL-7B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2-vl-7b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-32B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-32b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-72B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-72b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-7B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-7b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-VL-32B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-vl-32b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-VL-72B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-vl-72b-instruct.md): Creates a model response for the given chat conversation. - [Qwen2.5-VL-7B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen2.5-vl-7b-instruct.md): Creates a model response for the given chat conversation. - [Qwen3-14B](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-14b.md): Creates a model response for the given chat conversation. - [Qwen3-235B-A22B](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-235b-a22b.md): Creates a model response for the given chat conversation. - [Qwen3-235B-A22B-Instruct-2507](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-235b-a22b-instruct-2507.md): Creates a model response for the given chat conversation. - [Qwen3-235B-A22B-Thinking-2507](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-235b-a22b-thinking-2507.md): Creates a model response for the given chat conversation. - [Qwen3-30B-A3B-Instruct-2507](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-30b-a3b-instruct-2507.md): Creates a model response for the given chat conversation. - [Qwen3-30B-A3B-Thinking-2507](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-30b-a3b-thinking-2507.md): Creates a model response for the given chat conversation. - [Qwen3-32B](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-32b.md): Creates a model response for the given chat conversation. - [Qwen3-Coder](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-coder.md): Creates a model response for the given chat conversation. - [Qwen3-Coder-Next](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-coder-next.md): Creates a model response for the given chat conversation. - [Qwen3-Next-80B-A3B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-next-80b-a3b-instruct.md): Creates a model response for the given chat conversation. - [Qwen3-Next-80B-A3B-Thinking](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-next-80b-a3b-thinking.md): Creates a model response for the given chat conversation. - [Qwen3-VL-235B-A22B-Instruct](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-vl-235b-a22b-instruct.md): Creates a model response for the given chat conversation. - [Qwen3-VL-235B-A22B-Thinking](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3-vl-235b-a22b-thinking.md): Creates a model response for the given chat conversation. - [Qwen3.5-397B-A17B](https://docs.anyfast.ai/api-reference/model-api/alibaba/qwen3.5-397b-a17b.md): Creates a model response for the given chat conversation. - [Wan2 Task Query](https://docs.anyfast.ai/api-reference/model-api/alibaba/wan2-task-query.md): Retrieve the status and result of a Wan2 video generation task. - [Wan2.2 I2V A14B](https://docs.anyfast.ai/api-reference/model-api/alibaba/wan2.2-i2v-a14b.md): Submit a video generation task. Returns a `task_id` immediately. Poll `GET /v1/video/generations/{task_id}` to check status and retrieve the video URL. - [Wan2.2 T2V A14B](https://docs.anyfast.ai/api-reference/model-api/alibaba/wan2.2-t2v-a14b.md): Submit a video generation task. Returns a `task_id` immediately. Poll `GET /v1/video/generations/{task_id}` to check status and retrieve the video URL. - [Wan2.6 T2I](https://docs.anyfast.ai/api-reference/model-api/alibaba/wan2.6-t2i.md): Generate an image from a text prompt. Returns the result synchronously. - [claude-3-5-sonnet-20241022](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-3-5-sonnet-20241022.md): Creates a model response for the given chat conversation. - [claude-3-7-sonnet-20250219](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-3-7-sonnet-20250219.md): Creates a model response for the given chat conversation. - [claude-3-7-sonnet-20250219-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-3-7-sonnet-20250219-thinking.md): Creates a model response for the given chat conversation. - [claude-haiku-4-5-20251001](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-haiku-4-5-20251001.md): Creates a model response for the given chat conversation. - [claude-haiku-4-5-20251001-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-haiku-4-5-20251001-thinking.md): Creates a model response for the given chat conversation. - [claude-opus-4-1-20250805](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-1-20250805.md): Creates a model response for the given chat conversation. - [claude-opus-4-1-20250805-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-1-20250805-thinking.md): Creates a model response for the given chat conversation. - [claude-opus-4-20250514](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-20250514.md): Creates a model response for the given chat conversation. - [claude-opus-4-20250514-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-20250514-thinking.md): Creates a model response for the given chat conversation. - [claude-opus-4-5-20251101](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-5-20251101.md): Creates a model response for the given chat conversation. - [claude-opus-4-5-20251101-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-5-20251101-thinking.md): Creates a model response for the given chat conversation. - [claude-opus-4-6](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-6.md): Creates a model response for the given chat conversation. - [claude-opus-4-6-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-6-thinking.md): Creates a model response for the given chat conversation. - [claude-opus-4-7](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-7.md): Creates a model response for the given chat conversation. - [claude-opus-4-7-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-opus-4-7-thinking.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-20250514](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-20250514.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-20250514-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-20250514-thinking.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-5](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-5.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-5-20250929](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-5-20250929.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-5-20250929-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-5-20250929-thinking.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-6](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-6.md): Creates a model response for the given chat conversation. - [claude-sonnet-4-6-thinking](https://docs.anyfast.ai/api-reference/model-api/anthropic/claude-sonnet-4-6-thinking.md): Creates a model response for the given chat conversation. - [flux-1-1-pro](https://docs.anyfast.ai/api-reference/model-api/blackforestlabs/flux-1-1-pro.md): Generate an image from a text prompt using FLUX.1.1 [pro]. Returns the result synchronously as base64-encoded JSON. - [flux-1-kontext-pro](https://docs.anyfast.ai/api-reference/model-api/blackforestlabs/flux-1-kontext-pro.md): Edit an existing image or generate a new one using FLUX.1 Kontext [pro]. Pass `input_image` for editing, or omit it for text-to-image generation. Returns the result synchronously as base64-encoded JSON. - [flux-2-flex](https://docs.anyfast.ai/api-reference/model-api/blackforestlabs/flux-2-flex.md): Generate an image from a text prompt using FLUX.2 [flex]. Offers fine-grained control over inference steps and guidance scale. Returns the result synchronously as base64-encoded JSON. - [flux-2-pro](https://docs.anyfast.ai/api-reference/model-api/blackforestlabs/flux-2-pro.md): Generate an image from a text prompt using FLUX.2 [pro]. Supports up to 4MP resolution and multi-reference consistency. Returns the result synchronously as base64-encoded JSON. - [Doubao Seed 1.6](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-1-6-251015.md): Creates a model response for the given chat conversation. - [Doubao Seed 1.6 Flash](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-1-6-flash-250828.md): Creates a model response for the given chat conversation. - [Doubao Seed 1.6 Lite](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-1-6-lite-251015.md): Creates a model response for the given chat conversation. - [Doubao Seed 1.6 Vision](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-1-6-vision-250815.md): Creates a model response for the given chat conversation. - [Doubao Seed 1.8](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-1-8-251228.md): Creates a model response for the given chat conversation. - [Doubao Seed 2.0 Code](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-2.0-code.md): Creates a model response for the given chat conversation. - [Doubao Seed 2.0 Lite](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-2.0-lite.md): Creates a model response for the given chat conversation. - [Doubao Seed 2.0 Mini](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-2.0-mini.md): Creates a model response for the given chat conversation. - [Doubao Seed 2.0 Pro](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seed-2.0-pro.md): Creates a model response for the given chat conversation. - [Seedance 1.0 Lite i2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-lite-i2v-250428.md): Generate a video using the doubao-seedance-1-0-lite-i2v-250428 model. - [Seedance 1.0 Lite t2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-lite-t2v-250428.md): Generate a video using the doubao-seedance-1-0-lite-t2v-250428 model. - [Seedance 1.0 Pro i2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-pro-250528-i2v.md): Generate a video from reference images using the doubao-seedance-1-0-pro-250528 model. - [Seedance 1.0 Pro t2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-pro-250528-t2v.md): Generate a video from text using the doubao-seedance-1-0-pro-250528 model. - [Seedance 1.0 Pro Fast i2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-pro-fast-251015-i2v.md): Generate a video from reference images using the doubao-seedance-1-0-pro-fast-251015 model. - [Seedance 1.0 Pro Fast t2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-0-pro-fast-251015-t2v.md): Generate a video from text using the doubao-seedance-1-0-pro-fast-251015 model. - [Seedance 1.5 Pro i2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-5-pro-251215-i2v.md): Generate a video from reference images using the doubao-seedance-1-5-pro-251215 model. - [Seedance 1.5 Pro t2v](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedance-1-5-pro-251215-t2v.md): Generate a video from text using the doubao-seedance-1-5-pro-251215 model. - [Seedream 3.0 t2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-3-0-t2i-250415.md): Generate an image from text using the doubao-seedream-3-0-t2i-250415 model. - [Seedream 4.0 i2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-4-0-250828-i2i.md): Generate an image from text and reference images using the doubao-seedream-4-0-250828 model. - [Seedream 4.0 t2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-4-0-250828-t2i.md): Generate an image from text using the doubao-seedream-4-0-250828 model. - [Seedream 4.5 i2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-4-5-251128-i2i.md): Generate an image from text and reference images using the doubao-seedream-4-5-251128 model. - [Seedream 4.5 t2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-4-5-251128-t2i.md): Generate an image from text using the doubao-seedream-4-5-251128 model. - [Seedream 5.0 i2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-5-0-260128-i2i.md): Generate an image from text and reference images using the doubao-seedream-5-0-260128 model. - [Seedream 5.0 t2i](https://docs.anyfast.ai/api-reference/model-api/bytedance/doubao-seedream-5-0-260128-t2i.md): Generate an image from text using the doubao-seedream-5-0-260128 model. - [Seedance 2.0](https://docs.anyfast.ai/api-reference/model-api/bytedance/seedance-2-0.md): Create an asynchronous video generation task. Supports multiple input modes: - Text-to-video - Image-to-video (first frame / first+last frame) - Multimodal reference (image + video + audio combinations) - Video editing - Video extension - Web search enhanced generation - [Seedance 2.0 Fast](https://docs.anyfast.ai/api-reference/model-api/bytedance/seedance-2-0-fast.md): Create an asynchronous video generation task. Supports multiple input modes: - Text-to-video - Image-to-video (first frame / first+last frame) - Multimodal reference (image + video + audio combinations) - Video editing - Video extension - Web search enhanced generation - [Seedance 2.0 Ultra](https://docs.anyfast.ai/api-reference/model-api/bytedance/seedance-2-0-ultra.md): Create an asynchronous video generation task. Supports multiple input modes: - Text-to-video - Image-to-video (first frame / first+last frame) - Multimodal reference (image + video + audio combinations) - Video editing - Video extension - Web search enhanced generation - [Seedance Task Query](https://docs.anyfast.ai/api-reference/model-api/bytedance/seedance-task-query.md): Retrieve the status and result of a Seedance 2.0 video generation task by its ID. - [Seedance Asset — Create Asset](https://docs.anyfast.ai/api-reference/model-api/bytedance/volc-asset-create-assets.md): Upload an asset to an existing asset group. Supports Image, Video, and Audio types. Supports JSON (URL / Base64 / Data URI) and multipart file upload. The returned Asset ID can be used as `Asset://` in Seedance 2.0 requests. - [Seedance Asset — Create Group](https://docs.anyfast.ai/api-reference/model-api/bytedance/volc-asset-create-group.md): Create a new asset group to organize media resources (images, videos, audio) for use with Seedance 2.0 video generation. - [Seedance Asset — List Assets](https://docs.anyfast.ai/api-reference/model-api/bytedance/volc-asset-list-assets.md): Query assets with optional filters. This is a read-only operation and is not billed. - [Seedance Asset — List Groups](https://docs.anyfast.ai/api-reference/model-api/bytedance/volc-asset-list-groups.md): Query asset groups with optional filters. This is a read-only operation and is not billed. - [deepseek-v3](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3.md): Creates a model response for the given chat conversation. - [deepseek-v3-1](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-1.md): Creates a model response for the given chat conversation. - [deepseek-v3-1-terminus](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-1-terminus.md): Creates a model response for the given chat conversation. - [deepseek-v3-2](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-2.md): Creates a model response for the given chat conversation. - [deepseek-v3-2-251201](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-2-251201.md): Creates a model response for the given chat conversation. - [deepseek-v3-2-exp](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-2-exp.md): Creates a model response for the given chat conversation. - [deepseek-v3-2-speciale](https://docs.anyfast.ai/api-reference/model-api/deepseek/deepseek-v3-2-speciale.md): Creates a model response for the given chat conversation. - [gemini-2.0-flash](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2-0-flash.md): Edit images using Gemini model. - [gemini-2.5-flash](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2-5-flash.md): Generate response using Gemini model with Google Search grounding. - [gemini-2.5-flash-image](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2-5-flash-image.md): Generate images using Gemini model. - [gemini-2.5-pro](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2-5-pro.md): Generate model response based on conversation content. - [gemini-2.5-pro (Stream)](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2-5-pro-stream.md): Generate model response in streaming mode via SSE. - [gemini-2.5-flash-lite](https://docs.anyfast.ai/api-reference/model-api/google/gemini-2.5-flash-lite.md): Generate model response based on conversation content. - [gemini-3-flash-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3-flash-preview.md): Generate model response based on conversation content. - [gemini-3-pro-image-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3-pro-image-preview.md): Generate images using Gemini model. - [gemini-3-pro-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3-pro-preview.md): Generate model response based on conversation content. - [gemini-3-pro-preview (Stream)](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3-pro-preview-stream.md): Generate model response in streaming mode via SSE. - [gemini-3.1-flash-image-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3.1-flash-image-preview.md): Generate images using Gemini model. - [gemini-3.1-flash-lite-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3.1-flash-lite-preview.md): Generate model response based on conversation content. - [gemini-3.1-pro-preview](https://docs.anyfast.ai/api-reference/model-api/google/gemini-3.1-pro-preview.md): Generate model response based on conversation content. - [kling-element-create](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-element-create.md): Create a custom element by uploading reference images. The element can be a character, animal, prop, outfit, scene, or special effect. Returns a task_id for polling status via the query endpoint. - [kling-element-delete](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-element-delete.md): Delete a custom element by its element_id. Only custom elements (not Kling presets) can be deleted. - [kling-element-list](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-element-list.md): Retrieve a paginated list of all custom elements created by the current user. - [kling-element-preset-list](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-element-preset-list.md): Retrieve a paginated list of official preset elements provided by Kling. These elements have `owned_by` set to `"kling"`. - [kling-element-query](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-element-query.md): Query the status and result of a custom element creation task by task_id. - [kling-identify-face](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-identify-face.md): Analyze a video and detect all human faces in it. Returns a `session_id` and a `face_list` that are required for the subsequent lip sync generation step. - [kling-lip-sync](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-lip-sync.md): Create an asynchronous lip sync video generation task. The facial movements in the source video are re-animated to match the provided speech. - [kling-task-query](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-task-query.md): Query the status and result of a Kling generation task. Use the task_id returned from the create task endpoint. - [Kling 1.5 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-5-i2v.md): Generate a video from reference images using the kling-v1-5 model. - [Kling 1.6 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-6-i2v.md): Generate a video from reference images using the kling-v1-6 model. - [Kling 1.6 mi2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-6-mi2v.md): Generate a video from multiple reference images (up to 4) using the kling-v1-6 model. - [Kling 1.6 t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-6-t2v.md): Generate a video from text prompt using the kling-v1-6 model. - [Kling 1.0 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-i2v.md): Generate a video from reference images using the kling-v1 model. - [Kling 1.0 t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v1-t2v.md): Generate a video from text prompt using the kling-v1 model. - [Kling 2.1 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-1-i2v.md): Generate a video from reference images using the kling-v2-1 model. - [Kling 2.1 Master i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-1-master-i2v.md): Generate a video from reference images using the kling-v2-1-master model. - [Kling 2.1 Master t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-1-master-t2v.md): Generate a video from text prompt using the kling-v2-1-master model. - [Kling 2.5 Turbo i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-5-turbo-i2v.md): Generate a video from reference images using the kling-v2-5-turbo model. - [Kling 2.5 Turbo t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-5-turbo-t2v.md): Generate a video from text prompt using the kling-v2-5-turbo model. - [Kling 2.6 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-6-i2v.md): Generate a video from reference images using the kling-v2-6 model. - [Kling 2.6 t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-6-t2v.md): Generate a video from text prompt using the kling-v2-6 model. - [Kling 2.0 Master i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-master-i2v.md): Generate a video from reference images using the kling-v2-master model. - [Kling 2.0 Master t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v2-master-t2v.md): Generate a video from text prompt using the kling-v2-master model. - [Kling 3.0 i2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v3-i2v.md): Generate a video from reference images using the kling-v3 model. - [kling-v3-omni](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v3-omni.md): Create an Omni video generation task. The Omni model can achieve various capabilities through Prompt with elements, images, videos, and other content. - [Kling 3.0 t2v](https://docs.anyfast.ai/api-reference/model-api/kuaishou/kling-v3-t2v.md): Generate a video from text prompt using the kling-v3 model. - [MiniMax M2.1](https://docs.anyfast.ai/api-reference/model-api/minimax/minimax-m2.1.md): Creates a model response for the given chat conversation. - [MiniMax M2.5](https://docs.anyfast.ai/api-reference/model-api/minimax/minimax-m2.5.md): Creates a model response for the given chat conversation. - [Kimi-K2.5](https://docs.anyfast.ai/api-reference/model-api/moonshot/kimi-k2-5.md): Creates a model response for the given chat conversation. - [gpt-4](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-4.md): Creates a model response for the given chat conversation. - [gpt-4.1](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-4-1.md): Creates a model response for the given chat conversation. - [gpt-4o](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-4o.md): Creates a model response for the given chat conversation. - [gpt-4o-mini](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-4o-mini.md): Creates a model response for the given chat conversation. - [gpt-5](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5.md): Creates a model response for the given chat conversation. - [gpt-5.1](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-1.md): Creates a model response for the given chat conversation. - [gpt-5.1-chat](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-1-chat.md): Creates a model response for the given chat conversation. - [gpt-5.1-chat-latest](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-1-chat-latest.md): Creates a model response for the given chat conversation. - [gpt-5.1-codex-mini](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-1-codex-mini.md): Creates a model response for the given chat conversation. - [gpt-5.2](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-2.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5.2-chat](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-2-chat.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5.2-chat-latest](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-2-chat-latest.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5.2-codex](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-2-codex.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5.3-codex](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-3-codex.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5.4](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-4.md): Creates a model response for the given input. Some OpenAI models only support the Responses format. - [gpt-5-chat-latest](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-chat-latest.md): Creates a model response for the given chat conversation. - [gpt-5-codex](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-codex.md): Creates a model response for the given chat conversation. - [gpt-5-codex-high](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-codex-high.md): Creates a model response for the given chat conversation. - [gpt-5-codex-low](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-codex-low.md): Creates a model response for the given chat conversation. - [gpt-5-mini](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-mini.md): Creates a model response for the given chat conversation. - [gpt-5-nano](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-nano.md): Creates a model response for the given chat conversation. - [gpt-5-pro](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-5-pro.md): Creates a model response for the given chat conversation. - [GPT Image 1](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-image-1.md): Generate images from text prompts using the GPT Image 1 model. - [GPT Image 1.5](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-image-1-5.md): Generate images from text prompts using the GPT Image 1.5 model. - [GPT Image 2](https://docs.anyfast.ai/api-reference/model-api/openai/gpt-image-2.md): Generate images from text prompts using the GPT Image 2 model. - [Sora 2 Create Video](https://docs.anyfast.ai/api-reference/model-api/openai/sora-2.md): Create a video generation task using the Sora 2 model. Supports text-to-video and image-to-video (via input_reference). The workflow is: create task -> query status -> download video. - [Sora 2 Download Video](https://docs.anyfast.ai/api-reference/model-api/openai/sora-2-download.md): Download the generated video after the task is completed. Returns an error if the task is still in progress. - [Sora 2 Query Task](https://docs.anyfast.ai/api-reference/model-api/openai/sora-2-query.md): Query the status and details of a Sora 2 video generation task. - [grok-3](https://docs.anyfast.ai/api-reference/model-api/xai/grok-3.md): Creates a model response for the given chat conversation. - [grok-3-mini](https://docs.anyfast.ai/api-reference/model-api/xai/grok-3-mini.md): Creates a model response for the given chat conversation. - [grok-4-1-fast-non-reasoning](https://docs.anyfast.ai/api-reference/model-api/xai/grok-4-1-fast-non-reasoning.md): Creates a model response for the given chat conversation. - [grok-4-1-fast-reasoning](https://docs.anyfast.ai/api-reference/model-api/xai/grok-4-1-fast-reasoning.md): Creates a model response for the given chat conversation. - [grok-4-fast-non-reasoning](https://docs.anyfast.ai/api-reference/model-api/xai/grok-4-fast-non-reasoning.md): Creates a model response for the given chat conversation. - [grok-4-fast-reasoning](https://docs.anyfast.ai/api-reference/model-api/xai/grok-4-fast-reasoning.md): Creates a model response for the given chat conversation. - [glm-4.5](https://docs.anyfast.ai/api-reference/model-api/zhipu/glm-4.5.md): Creates a model response for the given chat conversation. - [GLM-4-7](https://docs.anyfast.ai/api-reference/model-api/zhipu/glm-4.7.md): Creates a model response for the given chat conversation. - [glm-5](https://docs.anyfast.ai/api-reference/model-api/zhipu/glm-5.md): Creates a model response for the given chat conversation. - [glm-5.1](https://docs.anyfast.ai/api-reference/model-api/zhipu/glm-5.1.md): Creates a model response for the given chat conversation. - [Invite & Earn](https://docs.anyfast.ai/changelog/invite.md): Invite friends to AnyFast and earn balance rewards when they top up. - [Anthropic Claude Compatible Endpoint](https://docs.anyfast.ai/guides/endpoints/anthropic.md): Use the official Anthropic SDK to access Claude models on AnyFast through the native Messages API. - [ByteDance Doubao Compatible Endpoint](https://docs.anyfast.ai/guides/endpoints/doubao.md): Use the Volcengine Ark SDK to access Doubao models on AnyFast through /v1/chat/completions and /v1/responses endpoints. - [Chat Compatible Format](https://docs.anyfast.ai/guides/endpoints/gemini-chat.md): Access Gemini models on AnyFast using the OpenAI-compatible chat completions format. - [Native Format](https://docs.anyfast.ai/guides/endpoints/gemini-native.md): Use the official Google Generative AI SDK to access Gemini models on AnyFast through the native generateContent endpoint. - [ChatGPT Compatible Endpoint](https://docs.anyfast.ai/guides/endpoints/openai.md): Use the OpenAI SDK to access all chat models on AnyFast through a single /v1/chat/completions endpoint. - [HappyHorse 1.0 I2V](https://docs.anyfast.ai/guides/model-api/alibaba/happyhorse-1.0-i2v.md): Alibaba DashScope HappyHorse 1.0 image-to-video model. Animate a single first-frame image via async task API. - [HappyHorse 1.0 R2V](https://docs.anyfast.ai/guides/model-api/alibaba/happyhorse-1.0-r2v.md): Alibaba DashScope HappyHorse 1.0 reference-image-to-video model. Generate video from 1–3 reference images via async task API. - [HappyHorse 1.0 T2V](https://docs.anyfast.ai/guides/model-api/alibaba/happyhorse-1.0-t2v.md): Alibaba DashScope HappyHorse 1.0 text-to-video model. Generate videos from text prompts via async task API. - [HappyHorse 1.0 Video Edit](https://docs.anyfast.ai/guides/model-api/alibaba/happyhorse-1.0-video-edit.md): Alibaba DashScope HappyHorse 1.0 video editing model. Edit an input video using a reference image. - [HappyHorse Task Query](https://docs.anyfast.ai/guides/model-api/alibaba/happyhorse-task-query.md): Query the status and result of HappyHorse 1.0 video generation tasks. - [Qwen Image Plus](https://docs.anyfast.ai/guides/model-api/alibaba/qwen-image-plus.md): Qwen Image Plus text-to-image model via Alibaba via AnyFast API. - [Qwen2-VL-72B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2-vl-72b-instruct.md): Qwen2-VL-72B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2-VL-7B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2-vl-7b-instruct.md): Qwen2-VL-7B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-32B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-32b-instruct.md): Qwen2.5-32B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-72B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-72b-instruct.md): Qwen2.5-72B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-7B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-7b-instruct.md): Qwen2.5-7B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-VL-32B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-vl-32b-instruct.md): Qwen2.5-VL-32B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-VL-72B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-vl-72b-instruct.md): Qwen2.5-VL-72B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen2.5-VL-7B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen2.5-vl-7b-instruct.md): Qwen2.5-VL-7B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-14B](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-14b.md): Qwen3-14B chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-235B-A22B](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-235b-a22b.md): Qwen3-235B-A22B chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-235B-A22B-Instruct-2507](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-235b-a22b-instruct-2507.md): Qwen3-235B-A22B-Instruct-2507 chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-235B-A22B-Thinking-2507](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-235b-a22b-thinking-2507.md): Qwen3-235B-A22B-Thinking-2507 chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-30B-A3B-Instruct-2507](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-30b-a3b-instruct-2507.md): Qwen3-30B-A3B-Instruct-2507 chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-30B-A3B-Thinking-2507](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-30b-a3b-thinking-2507.md): Qwen3-30B-A3B-Thinking-2507 chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-32B](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-32b.md): Qwen3-32B chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-Coder](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-coder.md): Qwen3-Coder chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-Coder-Next](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-coder-next.md): Qwen3-Coder-Next chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-Next-80B-A3B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-next-80b-a3b-instruct.md): Qwen3-Next-80B-A3B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-Next-80B-A3B-Thinking](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-next-80b-a3b-thinking.md): Qwen3-Next-80B-A3B-Thinking chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-VL-235B-A22B-Instruct](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-vl-235b-a22b-instruct.md): Qwen3-VL-235B-A22B-Instruct chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3-VL-235B-A22B-Thinking](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3-vl-235b-a22b-thinking.md): Qwen3-VL-235B-A22B-Thinking chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Qwen3.5-397B-A17B](https://docs.anyfast.ai/guides/model-api/alibaba/qwen3.5-397b-a17b.md): Qwen3.5-397B-A17B chat model via Alibaba via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Wan2 Task Query](https://docs.anyfast.ai/guides/model-api/alibaba/wan2-task-query.md): Query the status and result of a Wan2 video generation task. - [Wan2.2 I2V A14B](https://docs.anyfast.ai/guides/model-api/alibaba/wan2.2-i2v-a14b.md): Wan2.2 I2V A14B video generation model via Alibaba via AnyFast API. - [Wan2.2 T2V A14B](https://docs.anyfast.ai/guides/model-api/alibaba/wan2.2-t2v-a14b.md): Wan2.2 T2V A14B video generation model via Alibaba via AnyFast API. - [Wan2.6 T2I](https://docs.anyfast.ai/guides/model-api/alibaba/wan2.6-t2i.md): Wan2.6 T2I text-to-image model via Alibaba via AnyFast API. - [Claude 3.5 Sonnet](https://docs.anyfast.ai/guides/model-api/anthropic/claude-3-5-sonnet.md): Anthropic's Claude 3.5 Sonnet via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Claude 3.7 Sonnet](https://docs.anyfast.ai/guides/model-api/anthropic/claude-3-7-sonnet.md): Anthropic's Claude 3.7 Sonnet via OpenAI-compatible API. Advanced reasoning with extended thinking support. - [Claude 3.7 Sonnet Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-3-7-sonnet-thinking.md): Anthropic's Claude 3.7 Sonnet with extended thinking enabled. Deep reasoning for math, coding, and complex analysis. - [Claude Haiku 4.5](https://docs.anyfast.ai/guides/model-api/anthropic/claude-haiku-4-5.md): Anthropic's Claude Haiku 4.5 via OpenAI-compatible API. Fast, lightweight, and cost-effective. - [Claude Haiku 4.5 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-haiku-4-5-thinking.md): Anthropic's Claude Haiku 4.5 with extended thinking enabled. Fast reasoning for math, coding, and analysis at low cost. - [Claude Opus 4](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4.md): Anthropic's Claude Opus 4 via OpenAI-compatible API. The latest and most capable Claude model for complex reasoning, coding, and research. - [Claude Opus 4.1](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-1.md): Anthropic's Claude Opus 4.1 via OpenAI-compatible API. The most powerful Claude model for complex reasoning and research. - [Claude Opus 4.1 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-1-thinking.md): Anthropic's Claude Opus 4.1 with extended thinking enabled. The ultimate reasoning model for the most complex problems. - [Claude Opus 4.5](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-5.md): Anthropic's Claude Opus 4.5 via OpenAI-compatible API. The most intelligent Claude model with best-in-class reasoning and creativity. - [Claude Opus 4.5 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-5-thinking.md): Anthropic's Claude Opus 4.5 with extended thinking enabled. The ultimate reasoning model for the most complex problems. - [Claude Opus 4.6](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-6.md): Anthropic's Claude Opus 4.6 via OpenAI-compatible API. The latest and most intelligent Claude model with enhanced reasoning and coding capabilities. - [Claude Opus 4.6 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-6-thinking.md): Anthropic's Claude Opus 4.6 with extended thinking enabled. The latest and most capable reasoning model. - [Claude Opus 4.7](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-7.md): Anthropic's Claude Opus 4.7 via OpenAI-compatible API. The latest and most intelligent Claude model with enhanced reasoning and coding capabilities. - [Claude Opus 4.7 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-7-thinking.md): Anthropic's Claude Opus 4.7 with extended thinking enabled. The latest and most capable reasoning model. - [Claude Opus 4 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-opus-4-thinking.md): Anthropic's Claude Opus 4 with extended thinking enabled. Deep reasoning for complex research, math, and coding. - [Claude Sonnet 4](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4.md): Anthropic's Claude Sonnet 4 via OpenAI-compatible API. The best balance of intelligence and speed for everyday coding and complex tasks. - [Claude Sonnet 4.5](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4-5.md): Anthropic's Claude Sonnet 4.5 via OpenAI-compatible API. The best coding model with excellent speed-to-performance ratio. - [Claude Sonnet 4.5 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4-5-thinking.md): Anthropic's Claude Sonnet 4.5 with extended thinking enabled. Enhanced reasoning with the best coding model's speed and capability. - [Claude Sonnet 4.6](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4-6.md): Anthropic's Claude Sonnet 4.6 via OpenAI-compatible API. The latest and most capable Sonnet model with state-of-the-art coding and reasoning. - [Claude Sonnet 4.6 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4-6-thinking.md): Anthropic's Claude Sonnet 4.6 with extended thinking enabled. Enhanced reasoning with the latest Sonnet's speed and capability. - [Claude Sonnet 4 Thinking](https://docs.anyfast.ai/guides/model-api/anthropic/claude-sonnet-4-thinking.md): Anthropic's Claude Sonnet 4 with extended thinking enabled. Enhanced reasoning at Sonnet-level speed and cost. - [FLUX.1.1 [pro]](https://docs.anyfast.ai/guides/model-api/blackforestlabs/flux-1-1-pro.md): FLUX.1.1 [pro] text-to-image model by Black Forest Labs via AnyFast API. - [FLUX.1 Kontext [pro]](https://docs.anyfast.ai/guides/model-api/blackforestlabs/flux-1-kontext-pro.md): FLUX.1 Kontext [pro] image editing model by Black Forest Labs via AnyFast API. - [FLUX.2 [flex]](https://docs.anyfast.ai/guides/model-api/blackforestlabs/flux-2-flex.md): FLUX.2 [flex] text-to-image model by Black Forest Labs via AnyFast API. - [FLUX.2 [pro]](https://docs.anyfast.ai/guides/model-api/blackforestlabs/flux-2-pro.md): FLUX.2 [pro] text-to-image model by Black Forest Labs via AnyFast API. - [Doubao Seed 1.6](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-1-6.md): ByteDance's balanced performance chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Doubao Seed 1.6 Flash](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-1-6-flash.md): ByteDance's fast and efficient chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Doubao Seed 1.6 Lite](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-1-6-lite.md): ByteDance's lightweight chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Doubao Seed 1.6 Vision](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-1-6-vision.md): ByteDance's vision-capable chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Doubao Seed 1.8](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-1-8.md): ByteDance's advanced Seed 1.8 chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [Doubao Seed 2.0 Code](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-2.0-code.md): ByteDance's Seed 2.0 Code model optimized for code generation and understanding. - [Doubao Seed 2.0 Lite](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-2.0-lite.md): ByteDance's lightweight Seed 2.0 Lite model, cost-effective for everyday tasks. - [Doubao Seed 2.0 Mini](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-2.0-mini.md): ByteDance's compact Seed 2.0 Mini model, designed for high-throughput scenarios. - [Doubao Seed 2.0 Pro](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seed-2.0-pro.md): ByteDance's flagship Seed 2.0 Pro model with the strongest reasoning capabilities. - [Seedance 1.0 Lite i2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-lite-i2v.md): ByteDance's Seedance 1.0 Lite i2v model. Generate videos from text and images with flexible duration, ratios, and resolutions. - [Seedance 1.0 Lite t2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-lite-t2v.md): ByteDance's Seedance 1.0 Lite t2v model. Generate videos from text and images with flexible duration, ratios, and resolutions. - [Seedance 1.0 Pro Fast i2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-pro-fast-i2v.md): ByteDance's Seedance 1.0 Pro Fast image-to-video model. Generate videos from reference images with first frame and last frame control. - [Seedance 1.0 Pro Fast t2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-pro-fast-t2v.md): ByteDance's Seedance 1.0 Pro Fast text-to-video model. Generate videos from text prompts. - [Seedance 1.0 Pro i2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-pro-i2v.md): ByteDance's Seedance 1.0 Pro image-to-video model. Generate videos from reference images with first frame and last frame control. - [Seedance 1.0 Pro t2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-0-pro-t2v.md): ByteDance's Seedance 1.0 Pro text-to-video model. Generate videos from text prompts. - [Seedance 1.5 Pro i2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-5-pro-i2v.md): ByteDance's Seedance 1.5 Pro image-to-video model. Generate videos from reference images with first frame and last frame control. - [Seedance 1.5 Pro t2v](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedance-1-5-pro-t2v.md): ByteDance's Seedance 1.5 Pro text-to-video model. Generate videos from text prompts. - [Seedream 3.0 t2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-3-0-t2i.md): ByteDance Seedream 3.0 t2i model. Generate images from text prompts with flexible sizes. - [Seedream 4.0 i2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-4-0-i2i.md): ByteDance Seedream 4.0 image-to-image model. Generate images from reference images with text prompts. - [Seedream 4.0 t2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-4-0-t2i.md): ByteDance Seedream 4.0 text-to-image model. Generate images from text prompts. - [Seedream 4.5 i2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-4-5-i2i.md): ByteDance Seedream 4.5 image-to-image model. Generate images from reference images with text prompts. - [Seedream 4.5 t2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-4-5-t2i.md): ByteDance Seedream 4.5 text-to-image model. Generate images from text prompts. - [Seedream 5.0 i2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-5-0-i2i.md): ByteDance Seedream 5.0 image-to-image model. Generate images from reference images with text prompts. - [Seedream 5.0 t2i](https://docs.anyfast.ai/guides/model-api/bytedance/doubao-seedream-5-0-t2i.md): ByteDance Seedream 5.0 text-to-image model. Generate images from text prompts. - [Seedance 2.0](https://docs.anyfast.ai/guides/model-api/bytedance/seedance-2-0.md): ByteDance's Seedance 2.0 video generation model with multimodal reference, video editing, extension, audio generation, and web search. - [Seedance 2.0 Fast](https://docs.anyfast.ai/guides/model-api/bytedance/seedance-2-0-fast.md): ByteDance's Seedance 2.0 Fast video generation model with multimodal reference, video editing, extension, audio generation, and web search. - [Seedance 2.0 Ultra](https://docs.anyfast.ai/guides/model-api/bytedance/seedance-2-0-ultra.md): ByteDance's Seedance 2.0 Ultra video generation model with multimodal reference, video editing, extension, audio generation, and web search. - [Seedance 2.0 Task Query](https://docs.anyfast.ai/guides/model-api/bytedance/seedance-task-query.md): Query the status and result of ByteDance Seedance video generation tasks. - [Seedance 2.0 Asset Management](https://docs.anyfast.ai/guides/model-api/bytedance/volc-asset.md): Manage image, video, and audio assets for Seedance 2.0 video generation via AnyFast API. - [DeepSeek V3](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3.md): DeepSeek V3 via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.1](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-1.md): DeepSeek V3.1 via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.1 Terminus](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-1-terminus.md): DeepSeek V3.1 Terminus via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.2](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-2.md): DeepSeek V3.2 via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.2](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-2-251201.md): DeepSeek V3.2 chat model via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.2 Exp](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-2-exp.md): DeepSeek V3.2 Exp via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [DeepSeek V3.2 Speciale](https://docs.anyfast.ai/guides/model-api/deepseek/deepseek-v3-2-speciale.md): DeepSeek V3.2 Speciale via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Gemini 2.0 Flash](https://docs.anyfast.ai/guides/model-api/google/gemini-2-0-flash.md): Google's Gemini 2.0 Flash for AI image editing via Gemini API. - [Gemini 2.5 Flash](https://docs.anyfast.ai/guides/model-api/google/gemini-2-5-flash.md): Google's Gemini 2.5 Flash with Google Search grounding via Gemini API. - [Gemini 2.5 Flash Image](https://docs.anyfast.ai/guides/model-api/google/gemini-2-5-flash-image.md): Google's Gemini 2.5 Flash Image for AI image generation and editing via Gemini API. - [Gemini 2.5 Pro](https://docs.anyfast.ai/guides/model-api/google/gemini-2-5-pro.md): Google's Gemini 2.5 Pro via Gemini API. Advanced reasoning with thinking mode. - [Gemini 2.5 Pro (Stream)](https://docs.anyfast.ai/guides/model-api/google/gemini-2-5-pro-stream.md): Google's Gemini 2.5 Pro streaming via Gemini API. Real-time token streaming with SSE. - [Gemini 2.5 Flash Lite](https://docs.anyfast.ai/guides/model-api/google/gemini-2.5-flash-lite.md): Google's Gemini 2.5 Flash Lite via Gemini API. Lightweight and cost-efficient text generation. - [Gemini 3 Flash Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3-flash-preview.md): Google's Gemini 3 Flash Preview via Gemini API. Fast and capable text generation. - [Gemini 3 Pro Image Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3-pro-image-preview.md): Google's Gemini 3 Pro Image Preview for AI image generation and editing via Gemini API. - [Gemini 3 Pro Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3-pro-preview.md): Google's Gemini 3 Pro Preview via Gemini API. Advanced reasoning with thinking mode. - [Gemini 3 Pro Preview (Stream)](https://docs.anyfast.ai/guides/model-api/google/gemini-3-pro-preview-stream.md): Google's Gemini 3 Pro Preview streaming via Gemini API. Real-time token streaming with SSE. - [Gemini 3.1 Flash Image Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3.1-flash-image-preview.md): Google's Gemini 3.1 Flash Image Preview for AI image generation and editing via Gemini API. - [Gemini 3.1 Flash Lite Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3.1-flash-lite-preview.md): Google's Gemini 3.1 Flash Lite Preview via Gemini API. Fast and efficient text generation. - [Gemini 3.1 Pro Preview](https://docs.anyfast.ai/guides/model-api/google/gemini-3.1-pro-preview.md): Google's Gemini 3.1 Pro Preview via Gemini API. Advanced reasoning with thinking mode. - [Kling Element Create](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-element-create.md): Create a custom element (character, scene, prop, outfit, etc.) for use in Kling video generation. - [Kling Delete Element](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-element-delete.md): Delete a custom element by its element_id. - [Kling Element List](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-element-list.md): List all custom elements created by the current user with pagination. - [Kling Preset Element List](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-element-preset-list.md): List official preset elements provided by Kling, available to all users. - [Kling Element Query](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-element-query.md): Query the status and result of a custom element creation task by task_id. - [Kling Identify Face](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-identify-face.md): Detect faces in a source video to set up a session for lip sync generation. - [Kling Advanced Lip Sync](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-lip-sync.md): Re-animate the facial movements of a source video to match text (TTS) or an existing audio file. - [Kling Task Query](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-task-query.md): Query the status and result of any Kling generation task — video, image, lip-sync, and virtual try-on. - [Kling 1.5 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-5-i2v.md): Kuaishou's Kling 1.5 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 1.6 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-6-i2v.md): Kuaishou's Kling 1.6 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 1.6 mi2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-6-mi2v.md): Kuaishou's Kling 1.6 multi-image-to-video model. Generate videos from up to 4 reference images with a single prompt. - [Kling 1.6 t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-6-t2v.md): Kuaishou's Kling 1.6 t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 1.0 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-i2v.md): Kuaishou's Kling 1.0 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 1.0 t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v1-t2v.md): Kuaishou's Kling 1.0 t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 2.1 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-1-i2v.md): Kuaishou's Kling 2.1 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 2.1 Master i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-1-master-i2v.md): Kuaishou's Kling 2.1 Master i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 2.1 Master t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-1-master-t2v.md): Kuaishou's Kling 2.1 Master t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 2.5 Turbo i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-5-turbo-i2v.md): Kuaishou's Kling 2.5 Turbo i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 2.5 Turbo t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-5-turbo-t2v.md): Kuaishou's Kling 2.5 Turbo t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 2.6 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-6-i2v.md): Kuaishou's Kling 2.6 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 2.6 t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-6-t2v.md): Kuaishou's Kling 2.6 t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 2.0 Master i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-master-i2v.md): Kuaishou's Kling 2.0 Master i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling 2.0 Master t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v2-master-t2v.md): Kuaishou's Kling 2.0 Master t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [Kling 3.0 i2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v3-i2v.md): Kuaishou's Kling 3.0 i2v model. Generate videos from reference images with camera control, motion brush, and multi-shot storyboarding. - [Kling V3 Omni](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v3-omni.md): Kuaishou's Kling V3 Omni video generation model. Create videos from text, images, and video references with multi-shot storyboarding. - [Kling 3.0 t2v](https://docs.anyfast.ai/guides/model-api/kuaishou/kling-v3-t2v.md): Kuaishou's Kling 3.0 t2v model. Generate videos from text prompts with camera control, multi-shot storyboarding, and sound generation. - [MiniMax M2.1](https://docs.anyfast.ai/guides/model-api/minimax/minimax-m2.1.md): MiniMax M2.1 chat model via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [MiniMax M2.5](https://docs.anyfast.ai/guides/model-api/minimax/minimax-m2.5.md): MiniMax M2.5 chat model via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [Kimi-K2.5](https://docs.anyfast.ai/guides/model-api/moonshot/kimi-k2-5.md): Kimi-K2.5 chat model via MoonShot via AnyFast OpenAI-compatible API. Fast, capable, and easy to integrate. - [GPT-4](https://docs.anyfast.ai/guides/model-api/openai/gpt-4.md): OpenAI's GPT-4 via OpenAI-compatible API. Highly capable and reliable for complex tasks. - [GPT-4.1](https://docs.anyfast.ai/guides/model-api/openai/gpt-4-1.md): OpenAI's GPT-4.1 — fast, capable multimodal model with improved instruction following. - [GPT-4o](https://docs.anyfast.ai/guides/model-api/openai/gpt-4o.md): OpenAI's GPT-4o — fast, multimodal flagship model with text, vision, and audio capabilities. - [GPT-4o Mini](https://docs.anyfast.ai/guides/model-api/openai/gpt-4o-mini.md): OpenAI's GPT-4o Mini — small, affordable, and capable multimodal model. - [GPT-5](https://docs.anyfast.ai/guides/model-api/openai/gpt-5.md): OpenAI's GPT-5 — next-generation multimodal model with advanced reasoning. - [GPT-5.1](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-1.md): OpenAI's GPT-5.1 — improved multimodal model building on GPT-5. - [GPT-5.1 Chat](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-1-chat.md): OpenAI's GPT-5.1 Chat — chat-optimized GPT-5.1 variant. - [GPT-5.1 Chat Latest](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-1-chat-latest.md): OpenAI's GPT-5.1 Chat Latest — latest chat-optimized GPT-5.1. - [GPT-5.1 Codex Mini](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-1-codex-mini.md): OpenAI's GPT-5.1 Codex Mini — compact code-optimized GPT-5.1. - [GPT-5.2](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-2.md): OpenAI's GPT-5.2 — latest multimodal model with enhanced capabilities. - [GPT-5.2 Chat](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-2-chat.md): OpenAI's GPT-5.2 Chat — chat-optimized GPT-5.2 variant. - [GPT-5.2 Chat Latest](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-2-chat-latest.md): OpenAI's GPT-5.2 Chat Latest — latest chat-optimized GPT-5.2. - [GPT-5.2 Codex](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-2-codex.md): OpenAI's GPT-5.2 Codex — code-optimized GPT-5.2 variant. - [GPT-5.3 Codex](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-3-codex.md): OpenAI's GPT-5.3 Codex — latest code-optimized model. - [GPT-5.4](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-4.md): OpenAI's GPT-5.4 — newest generation multimodal model. - [GPT-5 Chat Latest](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-chat-latest.md): OpenAI's GPT-5 Chat Latest — latest chat-optimized variant of GPT-5. - [GPT-5 Codex](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-codex.md): OpenAI's GPT-5 Codex — code-optimized variant of GPT-5. - [GPT-5 Codex High](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-codex-high.md): OpenAI's GPT-5 Codex High — high-effort code generation with GPT-5. - [GPT-5 Codex Low](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-codex-low.md): OpenAI's GPT-5 Codex Low — efficient code generation with GPT-5. - [GPT-5 Mini](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-mini.md): OpenAI's GPT-5 Mini — compact, cost-effective variant of GPT-5. - [GPT-5 Nano](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-nano.md): OpenAI's GPT-5 Nano — ultra-lightweight, lowest-cost GPT-5 variant. - [GPT-5 Pro](https://docs.anyfast.ai/guides/model-api/openai/gpt-5-pro.md): OpenAI's GPT-5 Pro — highest-capability multimodal GPT-5 variant. - [GPT Image 1](https://docs.anyfast.ai/guides/model-api/openai/gpt-image-1.md): OpenAI GPT Image 1 image generation model. Generate and edit images from text prompts. - [GPT Image 1.5](https://docs.anyfast.ai/guides/model-api/openai/gpt-image-1-5.md): OpenAI GPT Image 1.5 image generation model. Generate and edit images with enhanced speed and precision. - [GPT Image 2](https://docs.anyfast.ai/guides/model-api/openai/gpt-image-2.md): OpenAI GPT Image 2 image generation model. Next-generation autoregressive + reasoning hybrid model with flexible resolution up to 4K and near-perfect text rendering. - [Sora 2](https://docs.anyfast.ai/guides/model-api/openai/sora-2.md): OpenAI Sora 2 video generation model. Generate videos from text prompts and reference images. - [Grok 3](https://docs.anyfast.ai/guides/model-api/xai/grok-3.md): Grok 3 by xAI via AnyFast OpenAI-compatible API. xAI most capable model. - [Grok 3 Mini](https://docs.anyfast.ai/guides/model-api/xai/grok-3-mini.md): Grok 3 Mini by xAI via AnyFast OpenAI-compatible API. Fast and lightweight. - [Grok 4.1 Fast Non-Reasoning](https://docs.anyfast.ai/guides/model-api/xai/grok-4-1-fast-non-reasoning.md): Grok 4.1 Fast Non-Reasoning by xAI via AnyFast OpenAI-compatible API. Grok 4.1 without reasoning. - [Grok 4.1 Fast Reasoning](https://docs.anyfast.ai/guides/model-api/xai/grok-4-1-fast-reasoning.md): Grok 4.1 Fast Reasoning by xAI via AnyFast OpenAI-compatible API. Grok 4.1 with reasoning. - [Grok 4 Fast Non-Reasoning](https://docs.anyfast.ai/guides/model-api/xai/grok-4-fast-non-reasoning.md): Grok 4 Fast Non-Reasoning by xAI via AnyFast OpenAI-compatible API. Grok 4 without reasoning. - [Grok 4 Fast Reasoning](https://docs.anyfast.ai/guides/model-api/xai/grok-4-fast-reasoning.md): Grok 4 Fast Reasoning by xAI via AnyFast OpenAI-compatible API. Grok 4 with reasoning. - [GLM-4.5](https://docs.anyfast.ai/guides/model-api/zhipu/glm-4.5.md): Zhipu AI's GLM-4.5 chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [GLM-4-7](https://docs.anyfast.ai/guides/model-api/zhipu/glm-4.7.md): Zhipu AI's GLM-4-7 chat model via OpenAI-compatible API. Fast, capable, and easy to integrate. - [GLM-5](https://docs.anyfast.ai/guides/model-api/zhipu/glm-5.md): Zhipu AI's GLM-5 chat model via OpenAI-compatible API. A powerful next-generation GLM model. - [GLM-5.1](https://docs.anyfast.ai/guides/model-api/zhipu/glm-5.1.md): Zhipu AI's GLM-5.1 chat model via OpenAI-compatible API. The latest and most capable GLM model. - [FAQ](https://docs.anyfast.ai/guides/resources/faq.md): Frequently asked questions about AnyFast. - [Payment](https://docs.anyfast.ai/guides/resources/payment.md): Top up your AnyFast account to start making API calls. - [Introduction](https://docs.anyfast.ai/index.md): AnyFast is a unified API platform for accessing AI models across multiple providers. - [Quickstart](https://docs.anyfast.ai/quickstart.md): Start making API calls to AnyFast in minutes. ## OpenAPI Specs - [openapi](https://docs.anyfast.ai/api-reference/endpoints/openapi/openai/openapi.yaml) ## Optional - [Documentation](https://docs.anyfast.ai)