VLMs (Vision-Language Models)
VLMs are multimodal tools that bridge the gap between text and visuals. Unlike standard Vision AI (which sees) or LLMs (which read), VLMs can reason across both formats simultaneously. They allow users to input images and ask questions about them, or input text to guide visual tasks. Examples Visual Question Answering (VQA): Uploading a photo of a broken appliance and asking, "How do I fix this?" Image Captioning: Automatically generating descriptive alt-text for images for SEO or accessibility. Document Understanding: Analyzing complex PDFs that contain both charts and text to extract insights. Video Search: Searching through video archives using natural language queries like "Find the moment the dog jumps."
6 tools
Anthropic is an AI safety and research company building reliable, interpretable, and steerable AI systems.