Find AI ListFind AI List
HomeBrowseAI NewsMatch Me 🪄
Submit ToolSubmitLogin

Find AI List

Discover, compare, and keep up with the latest AI tools, models, and news.

Explore

  • Home
  • Discover Stacks
  • AI News
  • Compare

Contribute

  • Submit a Tool
  • Edit your Tool
  • Request a Tool

Newsletter

Get concise updates. Unsubscribe any time.

© 2026 Find AI List. All rights reserved.

PrivacyTermsRefund PolicyAbout
Home
Data & Analytics
YOLACT
YOLACT logo
Data & Analytics

YOLACT

YOLACT (You Only Look At CoefficienTs) is an open-source, real-time instance segmentation model developed by Daniel Bolya and colleagues. It is a deep learning framework designed to perform pixel-level object detection and segmentation in images and video streams at high speeds, making it suitable for applications requiring immediate feedback. Unlike slower two-stage methods like Mask R-CNN, YOLACT employs a single-stage architecture that generates prototype masks and prediction coefficients in parallel, which are then combined to produce final instance masks. This approach achieves a favorable balance between speed and accuracy, enabling real-time performance on standard GPUs. It is primarily used by researchers, developers, and engineers in fields such as robotics, autonomous vehicles, video surveillance, and augmented reality, where quick and precise object delineation is crucial. The model is implemented in PyTorch and is celebrated for its simplicity, efficiency, and strong performance on benchmarks like COCO. YOLACT addresses the problem of computationally expensive instance segmentation, providing a practical solution for deploying advanced computer vision capabilities in resource-constrained or latency-sensitive environments.

Visit Website

📊 At a Glance

Pricing
Paid
Reviews
No reviews
Traffic
N/A
Engagement
0🔥
0👁️
Categories
Data & Analytics
Computer Vision

Key Features

Real-time Instance Segmentation

Performs pixel-accurate segmentation of multiple object instances in images and video at speeds exceeding 30 FPS on a single GPU. It outputs both bounding boxes and masks for each detected object.

Prototype Mask Generation

Generates a set of non-local prototype masks across the entire image and predicts per-instance mask coefficients. The final masks are produced via a linear combination, decoupling mask resolution from detection.

Fast NMS

Implements a lightweight, GPU-optimized Non-Maximum Suppression (NMS) algorithm that efficiently filters overlapping detections post-inference.

Multi-GPU Training Support

The training script supports data-parallel distributed training across multiple GPUs, accelerating the model training process on large datasets.

Easy Custom Dataset Integration

Provides straightforward utilities and configuration files to train YOLACT on custom datasets formatted in the standard COCO annotation style.

Pricing

Open Source

$0
  • ✓Full access to source code on GitHub
  • ✓Pre-trained models on COCO dataset
  • ✓Freedom to modify, distribute, and use commercially under MIT License
  • ✓Community support via GitHub Issues
  • ✓No user or seat limits

Use Cases

1

Autonomous Vehicle Perception

Developers in autonomous driving use YOLACT to process real-time video feeds from vehicle cameras. It segments and identifies pedestrians, vehicles, and road obstacles at high frame rates. This precise, instantaneous understanding of the environment is crucial for path planning and collision avoidance systems, enhancing safety and decision-making.

2

Robotic Vision and Manipulation

Robotics engineers integrate YOLACT into robotic arms or mobile robots to enable object picking and manipulation. By segmenting objects from a cluttered scene, the robot can accurately determine the shape and location of items. This allows for more reliable grasping and sorting in warehouses, manufacturing, or domestic assistance tasks.

3

Video Surveillance and Analytics

Security system developers deploy YOLACT to analyze live surveillance footage. It can track and segment individuals, vehicles, or abandoned objects across video frames in real time. This enables advanced analytics like crowd counting, intrusion detection, and behavior analysis without the latency of cloud processing.

4

Augmented and Virtual Reality

AR/VR creators use YOLACT for real-time scene understanding to overlay digital content accurately onto the physical world. By segmenting users and objects in the camera feed, it enables realistic occlusion and interaction. This improves immersion in applications ranging from gaming to remote assistance and virtual try-on.

5

Medical Image Analysis

Researchers in bioinformatics and radiology fine-tune YOLACT on medical datasets to segment cells, tumors, or anatomical structures from microscopy or MRI images. The model's ability to delineate multiple instances helps in quantitative analysis, such as cell counting or lesion measurement, aiding in diagnosis and research.

How to Use

  1. Step 1: Clone the official YOLACT repository from GitHub using `git clone https://github.com/dbolya/yolact.git` and navigate into the project directory.
  2. Step 2: Set up the Python environment by installing dependencies, preferably using a virtual environment, with `pip install -r requirements.txt`. Ensure PyTorch and torchvision are installed compatible with your CUDA version for GPU acceleration.
  3. Step 3: Download pre-trained model weights from the repository's release page or provided links (e.g., yolact_base_54_800000.pth) and place them in the `weights/` directory.
  4. Step 4: Run inference on an image using the provided script, e.g., `python eval.py --trained_model=weights/yolact_base_54_800000.pth --score_threshold=0.15 --top_k=15 --image=my_image.jpg`. This will generate an output image with segmented instances.
  5. Step 5: For video or webcam input, use the `--video` or `--webcam` flags respectively, adjusting parameters like `--display_masks` and `--display_bboxes` to control visualization.
  6. Step 6: To train YOLACT on a custom dataset, prepare annotations in COCO format, modify the dataset configuration in `data/config.py`, and execute the training script with appropriate hyperparameters.
  7. Step 7: Integrate YOLACT into a larger application by importing its modules and using the API to load the model and perform inference programmatically within Python code.
  8. Step 8: For deployment, consider optimizing the model with tools like TorchScript or ONNX for production environments, and set up a serving pipeline for batch or real-time processing.

Reviews & Ratings

No reviews yet

Sign in to leave a review

Alternatives

15Five logo

15Five

15Five operates in the people analytics and employee experience space, where platforms aggregate HR and feedback data to give organizations insight into their workforce. These tools typically support engagement surveys, performance or goal tracking, and dashboards that help leaders interpret trends. They are intended to augment HR and management decisions, not to replace professional judgment or context. For specific information about 15Five's metrics, integrations, and privacy safeguards, you should refer to the vendor resources published at https://www.15five.com.

0
0
Data & Analytics
Data Analysis Tools
See Pricing
View Details
20-20 Technologies logo

20-20 Technologies

20-20 Technologies is a comprehensive interior design and space planning software platform primarily serving kitchen and bath designers, furniture retailers, and interior design professionals. The company provides specialized tools for creating detailed 3D visualizations, generating accurate quotes, managing projects, and streamlining the entire design-to-sales workflow. Their software enables designers to create photorealistic renderings, produce precise floor plans, and automatically generate material lists and pricing. The platform integrates with manufacturer catalogs, allowing users to access up-to-date product information and specifications. 20-20 Technologies focuses on bridging the gap between design creativity and practical business needs, helping professionals present compelling visual proposals while maintaining accurate costing and project management. The software is particularly strong in the kitchen and bath industry, where precision measurements and material specifications are critical. Users range from independent designers to large retail chains and manufacturing companies seeking to improve their design presentation capabilities and sales processes.

0
0
Data & Analytics
Computer Vision
Paid
View Details
3D Generative Adversarial Network logo

3D Generative Adversarial Network

3D Generative Adversarial Network (3D-GAN) is a pioneering research project and framework for generating three-dimensional objects using Generative Adversarial Networks. Developed primarily in academia, it represents a significant advancement in unsupervised learning for 3D data synthesis. The tool learns to create volumetric 3D models from 2D image datasets, enabling the generation of novel, realistic 3D shapes such as furniture, vehicles, and basic structures without explicit 3D supervision. It is used by researchers, computer vision scientists, and developers exploring 3D content creation, synthetic data generation for robotics and autonomous systems, and advancements in geometric deep learning. The project demonstrates how adversarial training can be applied to 3D convolutional networks, producing high-quality voxel-based outputs. It serves as a foundational reference implementation for subsequent work in 3D generative AI, often cited in papers exploring 3D shape completion, single-view reconstruction, and neural scene representation. While not a commercial product with a polished UI, it provides code and models for the research community to build upon.

0
0
Data & Analytics
Computer Vision
Paid
View Details
Visit Website

At a Glance

Pricing Model
Paid
Visit Website