Find AI ListFind AI List
HomeBrowseAI NewsMatch Me 🪄
Submit ToolSubmitLogin

Find AI List

Discover, compare, and keep up with the latest AI tools, models, and news.

Explore

  • Home
  • Discover Stacks
  • AI News
  • Compare

Contribute

  • Submit a Tool
  • Edit your Tool
  • Request a Tool

Newsletter

Get concise updates. Unsubscribe any time.

© 2026 Find AI List. All rights reserved.

PrivacyTermsRefund PolicyAbout
Home
Data & Analytics
VGG Image Annotator (VIA)
VGG Image Annotator (VIA) logo
Data & Analytics

VGG Image Annotator (VIA)

The VGG Image Annotator (VIA) is a standalone, open-source software tool developed by the Visual Geometry Group (VGG) at the University of Oxford for manually annotating images, audio, and video. It is a lightweight, web-based application that runs directly in a browser without requiring any installation or setup, making it highly accessible. VIA is designed for creating ground-truth data for computer vision and machine learning projects, supporting a wide range of annotation tasks including object detection (bounding boxes), image segmentation (polygons, circles, ellipses), and classification. It is widely used by researchers, students, and practitioners in academia and industry for tasks like building datasets for object recognition, facial landmark detection, and medical image analysis. The tool saves annotations in a simple JSON file format, promoting easy integration with other data processing pipelines. Its focus is on simplicity, privacy (as data never leaves the user's computer), and flexibility for custom annotation schemas.

Visit Website

📊 At a Glance

Pricing
Free
Reviews
No reviews
Traffic
N/A
Engagement
0🔥
0👁️
Categories
Data & Analytics
Computer Vision

Key Features

Browser-Based & Offline-Capable

VIA runs entirely within a modern web browser without requiring server-side processing or installation. Once the page is loaded, it can function offline, processing data locally on the user's machine.

Multi-Format Media Support

The tool supports annotation of not just images, but also audio and video files. For video, it allows frame-by-frame annotation and temporal segment labeling.

Flexible Annotation Shapes & Attributes

Users can annotate using multiple shape types (rectangles, polygons, circles, points, polylines) and define custom attributes (text, numbers, checkboxes, radio buttons) for both regions and entire files.

Simple JSON Project File

All annotations, project metadata, and file references are saved in a single, human-readable JSON file. This file can be easily version-controlled, shared, and processed by scripts.

Collaborative Annotation Features

While primarily a desktop tool, VIA supports collaboration by allowing project JSON files to be shared and merged. Multiple annotators can work on subsets of data and their work can be combined.

Pricing

Free

$0
  • ✓Unlimited access to all annotation tools (shapes: rectangle, polygon, circle, ellipse, point, polyline).
  • ✓Unlimited projects, files, and annotations.
  • ✓Support for images, audio, and video annotations.
  • ✓Custom attribute definition for regions and files.
  • ✓Project saving/loading via local JSON files.
  • ✓Export to JSON and CSV formats.
  • ✓Runs entirely client-side in the browser; data never uploaded to external servers.

Use Cases

1

Academic Computer Vision Research

Researchers and PhD students use VIA to create ground-truth datasets for novel computer vision tasks, such as segmenting rare biological specimens or annotating historical documents. Its flexibility in defining custom attributes allows them to capture the precise labels needed for their experiments. The free, open-source nature aligns perfectly with academic budgets and the need for transparent, reproducible methodology.

2

Building Custom Object Detection Datasets

Small companies or indie developers working on niche ML applications (e.g., detecting defects in custom machinery, identifying specific retail products) use VIA to manually label their proprietary image collections. They draw bounding boxes or polygons around objects of interest and assign class labels. The local processing ensures their confidential product images remain secure, and the JSON output can be directly fed into training scripts for models like YOLO or Faster R-CNN.

3

Medical Image Annotation for AI Diagnostics

Medical researchers and radiologists utilize VIA to annotate regions of interest in X-rays, MRI scans, or histopathology slides, marking tumors, lesions, or anatomical structures. The ability to use polygons for precise segmentation is crucial. Data privacy is paramount, and VIA's client-side operation ensures sensitive patient data is never uploaded to the cloud, complying with strict data handling protocols common in healthcare research.

4

Video Analysis for Behavioral Studies

Ethologists or social scientists studying animal or human behavior in video footage use VIA to label temporal events and spatial locations frame-by-frame. They can annotate when a specific action starts and ends and mark the actor's position. This creates structured datasets for training models to automatically recognize behaviors, saving countless hours of manual observation and enabling large-scale quantitative analysis.

5

Education and Teaching Machine Learning

Instructors teaching courses on machine learning or computer vision use VIA as a hands-on tool for students to understand dataset creation. Students learn the fundamentals of annotation, label consistency, and the importance of high-quality ground truth by building a small dataset for a class project. The zero-cost and zero-setup barrier makes it ideal for classroom environments with diverse student hardware.

How to Use

  1. Step 1: Access the tool by navigating to the official VIA website (https://www.robots.ox.ac.uk/~vgg/software/via) in a modern web browser. No account creation, sign-up, or installation is required.
  2. Step 2: Load your media files (images, audio, video) into the application by clicking the 'Add Files' button or by dragging and dropping files directly into the browser window.
  3. Step 3: Define your annotation attributes by creating 'Region Attributes' (like object class, color) and 'File Attributes' (like overall image label) in the 'Attributes' panel to structure your annotation schema.
  4. Step 4: Select an annotation shape tool (rectangle, polygon, circle, ellipse, point, polyline) from the toolbar and draw regions directly on the loaded media to annotate objects or areas of interest.
  5. Step 5: Populate the defined attributes for each annotated region by selecting the region and filling in the values (e.g., typing 'cat' for a class attribute) in the side panel.
  6. Step 6: Navigate between files using the thumbnail strip or keyboard shortcuts to annotate multiple items in a batch.
  7. Step 7: Save your annotation project frequently. VIA saves all annotations and project state into a single JSON file via the 'Save Project' option. This file can be reloaded later to continue work.
  8. Step 8: Export your annotations for use in machine learning pipelines. VIA allows exporting the annotations in its native JSON format or in other common formats like CSV, which can then be converted for frameworks like TensorFlow or PyTorch.

Reviews & Ratings

No reviews yet

Sign in to leave a review

Alternatives

15Five logo

15Five

15Five operates in the people analytics and employee experience space, where platforms aggregate HR and feedback data to give organizations insight into their workforce. These tools typically support engagement surveys, performance or goal tracking, and dashboards that help leaders interpret trends. They are intended to augment HR and management decisions, not to replace professional judgment or context. For specific information about 15Five's metrics, integrations, and privacy safeguards, you should refer to the vendor resources published at https://www.15five.com.

0
0
Data & Analytics
Data Analysis Tools
See Pricing
View Details
20-20 Technologies logo

20-20 Technologies

20-20 Technologies is a comprehensive interior design and space planning software platform primarily serving kitchen and bath designers, furniture retailers, and interior design professionals. The company provides specialized tools for creating detailed 3D visualizations, generating accurate quotes, managing projects, and streamlining the entire design-to-sales workflow. Their software enables designers to create photorealistic renderings, produce precise floor plans, and automatically generate material lists and pricing. The platform integrates with manufacturer catalogs, allowing users to access up-to-date product information and specifications. 20-20 Technologies focuses on bridging the gap between design creativity and practical business needs, helping professionals present compelling visual proposals while maintaining accurate costing and project management. The software is particularly strong in the kitchen and bath industry, where precision measurements and material specifications are critical. Users range from independent designers to large retail chains and manufacturing companies seeking to improve their design presentation capabilities and sales processes.

0
0
Data & Analytics
Computer Vision
Paid
View Details
3D Generative Adversarial Network logo

3D Generative Adversarial Network

3D Generative Adversarial Network (3D-GAN) is a pioneering research project and framework for generating three-dimensional objects using Generative Adversarial Networks. Developed primarily in academia, it represents a significant advancement in unsupervised learning for 3D data synthesis. The tool learns to create volumetric 3D models from 2D image datasets, enabling the generation of novel, realistic 3D shapes such as furniture, vehicles, and basic structures without explicit 3D supervision. It is used by researchers, computer vision scientists, and developers exploring 3D content creation, synthetic data generation for robotics and autonomous systems, and advancements in geometric deep learning. The project demonstrates how adversarial training can be applied to 3D convolutional networks, producing high-quality voxel-based outputs. It serves as a foundational reference implementation for subsequent work in 3D generative AI, often cited in papers exploring 3D shape completion, single-view reconstruction, and neural scene representation. While not a commercial product with a polished UI, it provides code and models for the research community to build upon.

0
0
Data & Analytics
Computer Vision
Paid
View Details
Visit Website

At a Glance

Pricing Model
Free
Visit Website