Documentation Index
Fetch the complete documentation index at: https://docs.mavera.io/llms.txt
Use this file to discover all available pages before exploring further.
Scenario
Before sending video content to Mavera’s analysis pipeline, use GPT-4.1’s vision capability to describe keyframes. Extract frames from a video (one per 10 seconds), send each to GPT-4.1 vision for a description, then aggregate those descriptions and pass them to Mavera for marketing analysis — scene composition, brand alignment, emotional tone, and recommendations. Flow: Extract keyframes → OpenAI GPT-4.1 vision (describe frames) → aggregate → MaveraPOST /mave/chat → Marketing analysis
Code
Example Output
Error Handling
Image token limits
Image token limits
Each image consumes tokens based on detail level. Use
detail: "low" (fixed 85 tokens) for keyframe descriptions. Switch to "high" only for frames requiring fine-grained text extraction.ffmpeg not installed
ffmpeg not installed
Install with
brew install ffmpeg (macOS), apt install ffmpeg (Ubuntu), or choco install ffmpeg (Windows). Alternatively, use opencv-python to extract frames programmatically.Large frame counts
Large frame counts
A 10-minute video at 1 frame/10s yields 60 frames. Cap at 20 and increase the interval for longer videos. Adjust the ffmpeg filter to
fps=1/30 for 30-minute+ content.