Tiny OSS vision language model
2B parameter vision-language model optimized to run on edge devices and single GPUs. Supports image captioning, visual QA, and object detection. Runs via Ollama or directly with Python.
Vision-language models for image understanding, captioning, visual QA, and document parsing
AIchitect's Genome scanner detects Moondream in your project via these signals:
moondreamAdd to your GitHub README
[](https://aichitect.dev/tool/moondream)Explore the full AI landscape
See how Moondream fits into the bigger picture — browse all 207 tools and their relationships.