GPT-4o (pronounced 'four-oh', for 'omni') was released in May 2024 as OpenAI's flagship multimodal model capable of processing text, audio, and images natively.
Key Capabilities
- Real-time voice conversation with human-like latency
- Image understanding and analysis
- Emotional tone detection in voice
- Faster and cheaper than GPT-4 Turbo
- Available to free users with limits
Reference: