Browse

Featured

Datasets, Pre-Trained Models, and APIs for Object Detection, Classification

Overview

The Numberplate Dataset is a collection of Licence Plates that can easily be used for Automatic Number Plate Detection.

Example Footage!

Licese Plate Detection

Training and Deployment

The Number Plate model has been trained in Roboflow, and available for inference on the Dataset tab.
One could also build a Automatic Number Plate Recognition [ANPR] App using YOLOR and EasyOCR. This is achieved using the Roboflow Platform which you can deploy the model for robust and real-time ANPR.

About Augmented Startups

We are at the forefront of Artificial Intelligence in computer vision. With over 92k subscribers on YouTube, we embark on fun and innovative projects in this field and create videos and courses so that everyone can be an expert in this field. Our vision is to create a world full of inventors that can turn their dreams into reality

Apple Vision

The Apple Vision annotated data set contains over 350 images of naturally growing apples on an apple tree. Unlike other existing sets, this set attempted to capture apples growing on trees with different exposures of natural light during the daytime.

The training data was comprised of 77 photos taken of Peter Bloch’s home apple tree. These images were shot between July and September of 2021 on an iPhone 11 camera. After the photos were taken, they were sliced into multiple smaller images with a resolution of 360 × 640 pixels per image. This number was selected as the lowest natural resolution for a CV camera later used in this project.

This set was originally created for the ECE 31 Capstone project at Oregon State University.

Pill Classification

Background Information

This dataset was curated and annotated by Mohamed Attia.

The original dataset (v1) is composed of 451 images of various pills that are present on a large variety of surfaces and objects.
Example of an Image from the Dataset

The dataset is available under the Public License.

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 1 (v1) - 496 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 2 (v2) - 1,190 images

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416), all classes remapped (Modify Classes) to "pill"

  • Augmentations:
    Outputs per training example: 3
    90° Rotate: Clockwise, Counter-Clockwise, Upside Down
    Shear: ±5° Horizontal, ±5° Vertical
    Hue: Between -25° and +25°
    Saturation: Between -10% and +10%
    Brightness: Between -10% and +10%
    Exposure: Between -10% and +10%
    Noise: Up to 2% of pixels
    Cutout: 5 boxes with 5% size each

  • Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow

NOTE:

The Isolate Objects preprocessing step was added to convert the original object detection project into a suitable format for export in OpenAI's CLIP annotation format so that it could be used as a classifcation model in this project.

Mohamed Attia - LinkedIn

Bike Helmet Detection

Background Information

This dataset was curated and annotated by Syed Salman Reza. A custom dataset composed of two classes (With Helmet, Without Helmet). Main objetive is to identify if a Biker wearing Helmet or not.

The original custom dataset (v1) is composed of 1,371 images of people with and without bike helmets.

The dataset is available under the Public License.

Example of an Annotated Image from the Dataset

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 1 (v1) - 1,371 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: Augmentations applied prior to import - Bounding Box Blur (up to 10px)
  • Training Metrics: Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow
    • mAP = 74.4%, precision = 54.0%, recall = 77.0%

Version 2 (v2) - 3,735 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: Augmentations applied prior to import - Bounding Box Blur.
    • New augmentations:
      Outputs per training example: 3
      Rotation: Between -30° and +30°
      Shear: ±15° Horizontal, ±15° Vertical
      Blur: Up to 1.5px
      Mosaic: Applied
  • Training Metrics: Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow
    • mAP = 91.5%, precision = 65.1%, recall = 92.8%

Syed Salman Reza - Github

Aerial Docks and Boats

Overview

Drone Example

This dataset contains 74 images of aerial maritime photographs taken with via a Mavic Air 2 drone and 1,151 bounding boxes, consisting of docks, boats, lifts, jetskis, and cars. This is a multi class problem. This is an aerial object detection dataset. This is a maritime object detection dataset.

The drone was flown at 400 ft. No drones were harmed in the making of this dataset.

This dataset was collected and annotated by the Roboflow team, released with MIT license.

Image example

Use Cases

  • Identify number of boats on the water over a lake via quadcopter.
  • Boat object detection dataset
  • Aerial Object Detection proof of concept
  • Identify if boat lifts have been taken out via a drone
  • Identify cars with a UAV drone
  • Find which lakes are inhabited and to which degree.
  • Identify if visitors are visiting the lake house via quad copter.
  • Proof of concept for UAV imagery project
  • Proof of concept for maritime project
  • Etc.

This dataset is a great starter dataset for building an aerial object detection model with your drone.

Getting Started

Fork or download this dataset and follow our How to train state of the art object detector YOLOv4 for more. Stay tuned for particular tutorials on how to teach your UAV drone how to see and comprable airplane imagery and airplane footage.

Annotation Guide

See here for how to use the CVAT annotation tool that was used to create this dataset.

About Roboflow

Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.
:fa-spacer:
Developers reduce 50% of their boilerplate code when using Roboflow's workflow, save training time, and increase model reproducibility.
:fa-spacer:

Roboflow Wordmark

Overview

The Drowsiness dataset is a collection of images of a person in a vehicle (Ritesh Kanjee, of Augmented Startups) simulating "drowsy" and "awake" facial postures. This dataset can easily be used as a benchmark for a "driver alertness" or "driver safety" computer vision model.

Example Footage!

Distracted Driver Model - Example Footage

Training and Deployment

The Drowsiness model has been trained with Roboflow Train, and available for inference on the Dataset tab. We have also trained a YOLOR model for robust detection and tracking of a fatigued driver. You can learn more here: https://augmentedstartups.info/YOLOR-Get-Started

About Augmented Startups

We are at the forefront of Artificial Intelligence in computer vision. With over 94k subscribers on YouTube, we embark on fun and innovative projects in this field and create videos and courses so that everyone can be an expert in this field. Our vision is to create a world full of inventors that can turn their dreams into reality.

Football Player Detection

Overview

This project started over 3 years ago, where I wanted to make something that would draw out football plays automatically. Last year I hit a break through in my python development where I could track players individually. Roboflow has allowed me to track players by position groups.

Classes

Some of them are straight forward like Center, QB (quarterback), db (defensive back), lb (linebacker), but the rest are identified as skill. That means an offensive player like Runningback, Fullback, Tightend, H-back, Wide Reciever.

The project in action

I haven't made a video with myself using roboflow but I will shortly. You can see the project on my linkedin and how it's grown and will continue to grow.
My LinkedIn

Drone Control

Overview

The Drone Gesture Control Dataset is an object detection dataset that mimicks DJI's air gesture capability. This dataset consists of hand and body gesture commands that you can command your drone to either ,'take-off', 'land' and'follow'.

Example Footage

Drone Control

Model Training and Inference

The model for this dataset has been trained on Roboflow the Dataset tab, with exports to the OpenCV AI Kit, which is running on the drone in this example.

One could also build a model using MobileNet SSD using the Roboflow Platform deploy it to the OpenCV AI Kit. Watch the full tutorial here: https://augmentedstartups.info/AI-Drone-Tutorial

Using this Dataset

Use the fork button to copy this dataset to your own Roboflow account and export it with new preprocessing settings, or additional augmentations to make your model generalize better.

About Augmented Startups

We are at the forefront of Artificial Intelligence in computer vision. We embark on fun and innovative projects in this field and create videos and courses so that everyone can be an expert in this field. Our vision is to create a world full of inventors that can turn their dreams into reality.

Ocean Dataset

About Scubotics

Scubotics created https://www.namethatfish.com/. We are a startup dedicated to helping people better understand the ocean, one fish at a time.

About this dataset

The Ocean dataset contains images of ocean imagery depicting a few different species of fish.

Example Footage

Models trained on images like this dataset empower fish identification like the following:

Scubotics

Chicken Detection and Tracking

Background Information

This dataset was curated and annotated by Mohamed Traore from the Roboflow Team. A custom dataset composed of one class (chicken). The main objective is to identify chicken(s) and perform object-tracking on chicken(s) using Roboflow's "zero shot object tracking."

The original video is from Wendy Thomas (Description: "Definitive proof that the chicken crossed the road to get to the other side.")

The original custom dataset (v1) is composed of 106 images of chickens and their surrounding environment.

The dataset is available under the Public License.

Zero Shot Object Tracking

Example - Zero Shot Object Tracking

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 1 (v1) - 106 images

  • Preprocessing: Auto-Orient
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 2 (v2) - 106 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 3 (v3), "v1-augmented-COCO-transferLearning" - 254 images

Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow

  • 3x image generation

Version 11 (v11), "v1-augmented-trainFromScratch" - 463 images

Trained from the Version 3 training checkpoint.

  • Modify Classes was applied to remap the "chickens" class to "rooster" (meaning "rooster" will show up for the bounding boxes when running inference).
  • 3x image generation

Version 12 (v12) - 185 images

  • Preprocessing: Auto-Orient, Modify Classes (remap the "chickens" class to "rooster")
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Mohamed Traore - LinkedIn

Face Detection

Background Information

This dataset was curated and annotated by Mohamed Traore and Justin Brady after forking the raw images from the Roboflow Universe "Mask Wearing" dataset and remapping the "mask" and "no-mask" classes to "face".

Example Image from the Dataset

The main objective is to identify human faces in images or video. However, this model could be used for privacy purposes with changing the output of the bounding boxes to blur the detected face or fill it with a black box.

The original custom dataset (v1) is composed of 867 unaugmented (raw) images of people in various environments. 55 of the images are marked as "null" to help with feature extraction and reducing false detections.

Version 2 (v2) includes the augmented and trained version of the model. This version is trained from the COCO model checkpoint to take advantage of transfer learning and improve initial model training results.

Model Updates:

After a few trainings, and running tests with Roboflow's webcam model and Roboflow's video inference repo, it was clear that edge cases like hands sometimes recognized as faces was an issue. I grabbed some images from Alex Wong's "Hand Signs" dataset (96 images from the dataset) and added them to the project. I uploaded the images, without the annotation files, labeled all the faces, and retrained the model (version 5).

The dataset is available under the CC BY 4.0 license.

Surfer Spotting

Overview

The Surfline Surfer Spotting dataset contains images with surfers floating on the coast. Each image contains one classification called "surfer" but may contain multiple surfers.

Example Footage

Surfers

Using this Dataset

There are several deployment options available, including inferring via API, webcam, and curl command.

Here is a code snippet for to hit the hosted inference API you can use. Here are code snippets for more languages

const axios = require("axios");
                            const fs = require("fs");
                            
                            const image = fs.readFileSync("YOUR_IMAGE.jpg", {
                                encoding: "base64"
                            });
                            
                            axios({
                                method: "POST",
                                url: "https://detect.roboflow.com/surfer-spotting/2",
                                params: {
                                    api_key: "YOUR_KEY"
                                },
                                data: image,
                                headers: {
                                    "Content-Type": "application/x-www-form-urlencoded"
                                }
                            })
                            .then(function(response) {
                                console.log(response.data);
                            })
                            .catch(function(error) {
                                console.log(error.message);
                            });
                            

Download Dataset

On the versions tab you can select the version you like, and choose to download in 26 annotation formats.

Car_Dent_Scratch_Detection(1)

  • Sindhu
  • Damage-Detection Dataset
  • 3072 images

OptiScan

  • New Workspace
  • barcode-qrcode-1205-combined-110 Dataset
  • 2499 images

QR Code and Bar Code Dataset

Dataset (downloadable)

  • 2.5k Images
  • ~50/50 distribution of QR Code and Bar Code Classes

Training Object Detection Model

In progress using Roboflow train (click to train using YOLOV5)

Aerial Sheep

  • Riis
  • sheep Dataset
  • 1727 images

Overview

The Aerial Sheep dataset contains images taken from a birds-eye view with instances of sheep in them. Images do not differentiate between gender or breed of sheep, instead grouping them into a single class named "sheep".

Example Footage

Aerial Sheep

See RIIS's sheep counter application for additional use case examples.
Link - https://riis.com/blog/counting-sheep-using-drones-and-ai/

About RIIS

https://riis.com/about/

COCO 128

COCO 128 is a subset of 128 images of the larger COCO dataset. It reuses the training set for both validation and testing, with the purpose of proving that your training pipeline is working properly and can overfit this small dataset.

COCO 128 is a great dataset to use the first time you are testing out a new model.

wiki art

Art Classification Dataset

This classification dataset contains artistic movement art images that ranges from Abstract Expressionism to Pop Art.

Classes

  • Abstract_Expressionism
  • Action_painting
  • Analytical_Cubism
  • Art_Nouveau_Modern
  • Baroque
  • Color_Field_Painting
  • Contemporary_Realism
  • Cubism
  • Early_Renaissance
  • Expressionism
  • Fauvism
  • High_Renaissance
  • Impressionism
  • Mannerism_Late_Renaissance
  • Minimalism
  • Naive_Art_Primitivism
  • New_Realism
  • Northern_Renaissance
  • Pointillism
  • Pop_Art
  • Post_Impressionism
  • Realism
  • Rococo
  • Romanticism
  • Symbolism
  • Synthetic_Cubism
  • Ukiyo_e

Using this computer vision model

The computer vision model that's been trained for this dataset can be used to help identify art from the artistic movement classes above.

OnePetri

Background Information

This dataset was created by Michael Shamash and contains the images used to train the OnePetri plaque detection model (plaque detection model v1.0).

In microbiology, a plaque is defined as a “clear area on an otherwise opaque field of bacteria that indicates the inhibition or dissolution of the bacterial cells by some agent, either a virus or an antibiotic. Plaques are a sensitive laboratory indicator of the presence of some anti-bacterial factor.”
When working with bacteriophages (phages), viruses which can only infect and kill bacteria, scientists often need to perform the time-intensive monotonous task of counting plaques on Petri dishes. To help solve this problem I developed OnePetri, a set of machine learning models and a mobile phone application (currently iOS-only) that accelerates common microbiological Petri dish assays using AI.

A task that once took microbiologists several minutes to do per Petri dish (adds up quickly considering there are often tens of Petri dishes to analyze at a time!) could now be mostly automated thanks to computer vision, and completed in a matter of seconds.

App in Action

Video Clip

Petri Dish

Example Image

Plaque Detection

A total of 43 source images were used in this dataset with the following split: 29 training, 9 validation, 5 testing (2505 images after preprocessing and augmentations are applied).

OnePetri is a mobile phone application (currently iOS-only) which accelerates common microbiological Petri dish assays using AI. OnePetri's YOLOv5s plaque detection model was trained on a diverse set of images from the HHMI's SEA-PHAGES program, many of which are included in this dataset. This project wouldn't be possible without their support!

The following pre-processing options were applied:

  1. Auto-orient
  2. Tile image into 5 rows x 5 columns
  3. Resize tiles to 416px x 416px

The following augmentation options were applied:

  1. Grayscale (35% of images)
  2. Hue shift (-45deg to +45deg)
  3. Blur up to 2px
  4. Mosaic

OnePetri App In Action

For more information and to download OnePetri please visit: https://onepetri.ai/.

GARBAGE CLASSIFICATION 3

Garbage Object-Detection to Identify Disposal Class

This dataset detects various kinds of waste, labeling with a class that indentifies how it should be disposed

Never Gonna

Introducing RICK: Saving the Internet from Rickroll

We get it. Rickrolling happens.

RICK Introduction

But what if, using the latest state of the art machine learning, you could build ways to detect and prevent you and your loved ones from ever being Rickrolled again?

We're thrilled to introduce RICK: Real-time Intrusion Checker Kernel, a state of the art advancement and foundation model in internet safety. RICK is capable of detecting the presence of Rick Astley in images and video, so that applications can be built to shield you from Rick Astley content (or amplify said content, should you choose 🙃).

Conceived on April 1, 2022 at 9:41 AM ET and hacked together in about 30min during Roboflow's Friday team lunch, RICK has undergone thorough development to address a pressing need in online safety.

RICK is trained on 1200 images of Rick Astley and non-Rick Astley content.

Read more about how and why we built RICK here.

Building with RICK

As an example of RICK's utility, the Roboflow team has built Rickblocker, an open source application that automatically mutes your computer and places a black box on Rick Astley's face whenever he is present in a YouTube video you may accidentally click.

Rickrollprevention

Try RICK Yourself

RICK is fully hosted and free to use as an API or in-browser model. You can even confirm if you are not-rick by trying with your webcam.

face-features-test

A simple dataset for benchmarking CreateML object detection models. The images are sampled from COCO dataset with eyes and nose bounding boxes added. It’s not meant to be serious or useful in a real application. The purpose is to look at how long it takes to train CreateML models with varying dataset and batch sizes.

Training performance is affected by model configuration, dataset size and batch configuration. Larger models and batches require more memory. I used CreateML object detection project to compare the performance.

Hardware

M1 Macbook Air

  • 8 GPU
  • 4/4 CPU
  • 16G memory
  • 512G SSD

M1 Max Macbook Pro

  • 24 GPU
  • 2/8 CPU
  • 32G memory
  • 2T SSD

Small Dataset
Train: 144
Valid: 16
Test: 8

Results

batch M1 ET M1Max ET peak mem G
16 16 11 1.5
32 29 17 2.8
64 56 30 5.4
128 170 57 12

Larger Dataset
Train: 301
Valid: 29
Test: 18

Results

batch M1 ET M1Max ET peak mem G
16 21 10 1.5
32 42 17 3.5
64 85 30 8.4
128 281 54 16.5

CreateML Settings

For all tests, training was set to Full Network. I closed CreateML between each run to make sure memory issues didn't cause a slow down. There is a bug with Monterey as of 11/2021 that leads to memory leak. I kept an eye on the memory usage. If it looked like there was a memory leak, I restarted MacOS.

Observations

In general, more GPU and memory with MBP reduces the training time. Having more memory lets you train with larger datasets. On M1 Macbook Air, the practical limit is 12G before memory pressure impacts performance. On M1 Max MBP, the practical limit is 26G before memory pressure impacts performance. To work around memory pressure, use smaller batch sizes.

On the larger dataset with batch size 128, the M1Max is 5x faster than Macbook Air. Keep in mind a real dataset should have thousands of samples like Coco or Pascal. Ideally, you want a dataset with 100K images for experimentation and millions for the real training. The new M1 Max Macbooks is a cost effective alternative to building a Windows/Linux workstation with RTX 3090 24G. For most of 2021, the price of RTX 3090 with 24G is around $3,000.00. That means an equivalent windows workstation would cost the same as the M1Max Macbook pro I used to run the benchmarks.

Full Network vs Transfer Learning

As of CreateML 3, training with full network doesn't fully utilize the GPU. I don't know why it works that way. You have to select transfer learning to fully use the GPU. The results of transfer learning with the larger dataset. In general, the training time is faster and loss is better.

batch ET min Train Acc Val Acc Test Acc Top IU Train Top IU Valid Top IU Test Peak mem G loss
16 4 75 19 12 78 23 13 1.5 0.41
32 8 75 21 10 78 26 11 2.76 0.02
64 13 75 23 8 78 24 9 5.3 0.017
128 25 75 22 13 78 25 14 8.4 0.012

Github Project

The source code and full results are up on Github https://github.com/woolfel/createmlbench

Overhead Plane Detector

SkyBot

This is the dataset powering http://skybot.cam, an app that captures planes flying over top of my house.

Skycam Tweet

Upon the project gaining popularity on Hacker News from the above tweet, I thought I'd share the dataset and an example model to make it easier for others to build a plane spotting app, too.

About this Project

I built a system to take photos of all of the airplanes that fly over my house. Most of these planes are passing by at more than 30,000 feet! It uses ADS-B to track where the aircraft are relative to the camera, points the camera in the right direction and snaps a photo. I then run a few serverless functions that are running to detect where the aircraft is in the image and make a thumbnail. Much of the services are hosted on Azure. There's more details on the overall project here! http://skybot.cam/about. The project is open source as a part of my work from IQT as well.

Skybot Infrastructure

About the Dataset

The dataset is of airfract that was captured as they flew overhead. It includes a mix of large and small passenger jets and an assortment of business jets. There are also a images with buildings and contrails, where there is not aircraft present.

Use Cases

This dataset should allow for a plane dectector model to be built like for plane spotting and plane detection.

About Me

I'm Luke Berndt, I work on Azure products at Microsoft. You can learn more about me here: http://lukeberndt.com/

Sub-transmission Asset

  • TIPQC
  • STA Dataset
  • 775 images

The dataset includes actual pictures of Transmission and Sub-transmission Electrical Structure captured during the official inspection of assets being amortized by the 17 Electric Cooperative of the Philippines to the National Transmission Corporation (TransCo). The pictures are captured in years 2018, 2019 and 2020. Below are the following Electric Cooperatives:

  1. Bukidnon Sub-transmission Corporation (BSTC)
  2. Northern Negros Electric Cooperative, Inc. (NONECO)
  3. South Cotabato 1 Electric Cooperative, Inc. (SOCOTECO 1)
  4. Cebu 2 Electric Cooperative, Inc. (CEBECO 2)
  5. Peninsula Electric Cooperative, Inc. (PENELCO)
  6. Misamis Oriental 2 Electric Cooperative, Inc. (MORESCO 2)
  7. Davao del Sur Electric Cooperative, Inc. (DASURECO)
  8. Camiguin Electric Cooperative, Inc. (CAMELCO)
  9. Iloilo 2 Electric Cooperative, Inc. (ILECO 2)
  10. Misamis Oriental 1 Electric Cooperative, Inc. (MORESCO 1)
  11. Davao Oriental Electric Cooperative, Inc. (DORECO)
  12. Isabela 1 Electric Cooperative, Inc. (ISELCO 1)
  13. Aklan Electric Cooperative, Inc. (AKELCO)
  14. Sultan Kudarat Electric Cooperative, Inc. (SUKELCO)
  15. Zamboanga City Electric Cooperative, Inc. (ZAMCELCO)
  16. South Cotabato 2 Electric Cooperative, Inc. (SOCOTECO 2)
  17. Camarines Norte Electric Cooperative, Inc. (CANORECO)

Banana Ripening Process

Banana Ripening Process Dataset and Model

This dataset contains images of the classes below:

  • freshripe
  • freshunripe
  • overripe
  • ripe
  • rotten
  • unripe

Usage

This is an object detection model that can be used to possibly identify where in the Fruit Ripening Process fruit at stores are and when to take them off the shelves and put them in composting.

Hard Hat Universe

Overview

The Hard Hat dataset is an object detection dataset of workers in workplace settings that require a hard hat. Annotations also include examples of just "person" and "head," for when an individual may be present without a hard hart.

Example Image:
Example Image

Use Cases

One could use this dataset to, for example, build a classifier of workers that are abiding safety code within a workplace versus those that may not be. It is also a good general dataset for practice.

Using this Dataset

Use the fork button to copy this dataset to your own Roboflow account and export it with new preprocessing settings (perhaps resized for your model's desired format or converted to grayscale), or additional augmentations to make your model generalize better. This particular dataset would be very well suited for Roboflow's new advanced Bounding Box Only Augmentations.

About Roboflow

Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

Developers reduce 50% of their code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

Roboflow Workmark

License Plate Recognition

This is a U.S. license plate dataset + model using object detection. The images for this dataset were collected from Google images and around Central Florida parks. If you see your license plate in this dataset and you wish to remove it, please contact friends@roboflow.com

white-car-with-license-plate-bounding-box-detected

Try it out on this example web app or deploy to Luxonis Oak .

CSGO TRAIN YOLO V5

CSGO AIMBOT

Go Win

Trained on 5.9k Images

aicook

Background Information

This dataset was curated and annotated by - Karel Cornelis.

The original dataset (v1) is composed of 516 images of various ingredients inside a fridge. The project was created as part of a groupwork for a postgraduate applied AI at Erasmus Brussels - we made an object detection model to identify ingredients in a fridge.

From the recipe dataset we used (which is a subset of the recipe1M dataset) we distilled the top50 ingredients and used 30 of those to randomly fill our fridge.

Read this blog post to learn more about the model production process: How I Used Computer Vision to Make Sense of My Fridge

Watch this video to see the model in action: AICook

The dataset is available under the MIT License.

Getting Started

You can download this dataset for use within your own project, fork it into a workspace on Roboflow to create your own model, or test one of the trained versions within the app.

Dataset Versions

Version 1 (v1) - 516 images (original-images)

  • Preprocessing: Auto-Orient
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 2 (v2) - 3,050 images (aicook-augmented-trainFromCOCO)

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416)
  • Augmentations:
    • Outputs per training example: 8
      Rotation: Between -3° and +3°
      Exposure: Between -20% and +20%
      Blur: Up to 3px
      Noise: Up to 5% of pixels
      Cutout: 12 boxes with 10% size each
  • Training Metrics: Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow
    • mAP = 97.6%, precision = 86.9%, recall = 98.5%

Version 3 (v3) - 3,050 images (aicook-augmented-trainFromScratch)

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416)
  • Augmentations:
    • Outputs per training example: 8
      Rotation: Between -3° and +3°
      Exposure: Between -20% and +20%
      Blur: Up to 3px
      Noise: Up to 5% of pixels
      Cutout: 12 boxes with 10% size each
  • Training Metrics: Trained from "scratch" (no transfer learning employed) on Roboflow
    • mAP = 97.9%, precision = 79.6%, recall = 98.6%

Version 4 (v4) - 3,050 images images (aicook-augmented)

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416)
  • Augmentations:
    • Outputs per training example: 8
      Rotation: Between -3° and +3°
      Exposure: Between -20% and +20%
      Blur: Up to 3px
      Noise: Up to 5% of pixels
      Cutout: 12 boxes with 10% size each
  • Training Metrics: This version of the dataset was not trained

Karel Cornelis - LinkedIn

Weeds

Overview

The Weeds dataset is a collection of garden weeds that can easily confuse object detection models due to similiarity of the weeds compared to its surroundings. This dataset was used with YOLOR for object detection to detect weeds in complex backgrounds.

Example Footage!

Weeds Detection

Training and Deployment

The weeds model has been trained in Roboflow, available for inference on the Dataset tab.

One could also build a Weeds Detector using YOLOR. This is achieved using the Roboflow Platform which you can deploy the model for robust and real-time detections. You can learn more here: https://augmentedstartups.info/YOLOR-Get-Started

About Augmented Startups

We are at the forefront of Artificial Intelligence in computer vision. With over 92k subscribers on YouTube, we embark on fun and innovative projects in this field and create videos and courses so that everyone can be an expert in this field. Our vision is to create a world full of inventors that can turn their dreams into reality

Forklift

About this Dataset

This dataset was created by exporting images from images.cv and labeling them as an object detection dataset. The dataset contains 421 raw images (v1 - raw-images) and labeled classes include:

  • forklift
  • person

Example annotated image from the dataset from the dataset

Hyper-Kvasir

  • Simula
  • GI-tract Dataset
  • 10662 images

Overview
This is the largest Gastrointestinal dataset generously provided by Simula Research Laboratory in Norway

You can read their research paper here in Nature

In total, the dataset contains 10,662 labeled images stored using the JPEG format. The images can be found in the images folder. The classes, which each of the images belong to, correspond to the folder they are stored in (e.g., the ’polyp’ folder contains all polyp images, the ’barretts’ folder contains all images of Barrett’s esophagus, etc.). Each class-folder is located in a subfolder describing the type of finding, which again is located in a folder describing wheter it is a lower GI or upper GI finding. The number of images per class are not balanced, which is a general challenge in the medical field due to the fact that some findings occur more often than others. This adds an additional challenge for researchers, since methods applied to the data should also be able to learn from a small amount of training data. The labeled images represent 23 different classes of findings.

The data is collected during real gastro- and colonoscopy examinations at a Hospital in Norway and partly labeled by experienced gastrointestinal endoscopists.

Use Cases

"Artificial intelligence is currently a hot topic in medicine. The fact that medical data is often sparse and hard to obtain due to legal restrictions and lack of medical personnel to perform the cumbersome and tedious labeling of the data, leads to technical limitations. In this respect, we share the Hyper-Kvasir dataset, which is the largest image and video dataset from the gastrointestinal tract available today."

"We have used the labeled data to research the classification and segmentation of GI findings using both computer vision and ML approaches to potentially be used in live and post-analysis of patient examinations. Areas of potential utilization are analysis, classification, segmentation, and retrieval of images and videos with particular findings or particular properties from the computer science area. The labeled data can also be used for teaching and training in medical education. Having expert gastroenterologists providing the ground truths over various findings, HyperKvasir provides a unique and diverse learning set for future clinicians. Moreover, the unlabeled data is well suited for semi-supervised and unsupervised methods, and, if even more ground truth data is needed, the users of the data can use their own local medical experts to provide the needed labels. Finally, the videos can in addition be used to simulate live endoscopies feeding the video into the system like it is captured directly from the endoscopes enable developers to do image classification."

Borgli, H., Thambawita, V., Smedsrud, P.H. et al. HyperKvasir, a comprehensive multi-class image and video dataset for gastrointestinal endoscopy. Sci Data 7, 283 (2020). https://doi.org/10.1038/s41597-020-00622-y

Using this Dataset

Hyper-Kvasir is licensed under a Creative Commons Attribution 4.0 International (CC BY 4.0) License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source. This means that in all documents and papers that use or refer to the Hyper-Kvasir dataset or report experimental results based on the dataset, a reference to the related article needs to be added: PREPRINT: https://osf.io/mkzcq/. Additionally, one should provide a link to the Creative Commons license, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons license, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons license and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this license, visit http://creativecommons.org/licenses/by/4.0/

About Roboflow

Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

Developers reduce 50% of their boilerplate code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

Blood Cell Detection

Overview

This is a dataset of blood cells photos, originally open sourced by cosmicad and akshaylambda.

There are 364 images across three classes: WBC (white blood cells), RBC (red blood cells), and Platelets. There are 4888 labels across 3 classes (and 0 null examples).

Here's a class count from Roboflow's Dataset Health Check:

BCCD health

And here's an example image:

Blood Cell Example

Fork this dataset (upper right hand corner) to receive the raw images, or (to save space) grab the 500x500 export.

Use Cases

This is a small scale object detection dataset, commonly used to assess model performance. It's a first example of medical imaging capabilities.

Using this Dataset

We're releasing the data as public domain. Feel free to use it for any purpose.

It's not required to provide attribution, but it'd be nice! :)

About Roboflow

Roboflow makes managing, preprocessing, augmenting, and versioning datasets for computer vision seamless.

Developers reduce 50% of their boilerplate code when using Roboflow's workflow, automate annotation quality assurance, save training time, and increase model reproducibility.

Roboflow Workmark

Recyclable Items

  • Recycle
  • recyclables Dataset
  • 9802 images

Recyclable Items Dataset and Model

This is an object detection dataset that contains the classes below:

  • Plastic
  • Glass
  • Metal

Usage

This model can be potentially used to detect the objects above in an effort to sort them in a recycling center or an Automated River Cleaning System that uses computer vision.

halo infinite angel aim

Halo Infinite: Spartan Dataset

Classifications

There are four classifications:

  1. Enemy
  2. Enemy Head
  3. Friendly
  4. Friendly Head

Image Settings

Images are 320p by 320p centered on the targeting reticule

Game Settings

Images were gathered on low settings. Enemies are color:pineapple and allies are default blue.

Attribution and License

This dataset was created and annotated by Graham Doerksen and is available under CC BY 4.0 license

Apple Sorting

This project was created by Arfiani Nur Sayidah and is for sorting "apples" from "damaged apples."

The classes are "apple" and "damaged_apples"
Original Class Balance:

  1. apple: 2,152
  2. damaged_apple: 708

Rock Paper Scissors Presentation

Use this to make a scoring app of rock paper scissors (Cheaters beware)!

This dataset contains images of the rock, paper, scissors hand gestures that are detected in the model and can be used in gaming. Repo with counter (Human vs Computer) here

This dataset needs additional data on:

  • diverse representaion
  • null images

Valorant Smokes and Flashes

Real life Valorant Gameplay Expirience

This Dataset was used to create an AI for Valorant that would Smoke and Flash me In Real Life.

Full Video: https://youtu.be/aopXw22iL1M

My Game Pics

Annotated pictures of animals from trail cameras in East Texas.

yoloapex

Apex Enemy Detection

Use this PRE-TRAINED MODEL to create an aimbot and identify enemies

Use your webcam to infer, or use the hosted inference API

More deployment options are available

VOT2015

VOT2015 Dataset

The dataset comprises 60 short sequences showing various objects in challenging backgrounds. The sequences were chosen from a large pool of sequences including the ALOV dataset, OTB2 dataset, non-tracking datasets, Computer Vision Online, Professor Bob Fisher’s Image Database, Videezy, Center for Research in Computer Vision, University of Central Florida, USA, NYU Center for Genomics and Systems Biology, Data Wrangling, Open Access Directory and Learning and Recognition in Vision Group, INRIA, France. The VOT sequence selection protocol was applied to obtain a representative set of challenging sequences. The dataset is automatically downloaded by the evaluation kit when needed, there is no need to separately download the sequences for the challenge.

Annotations
The sequences were annotated by the VOT committee using rotated bounding boxes in order to provide highly accurate ground truth values for comparing results. The annotations are stored in a text file with the format:

frameN: X1, Y1, X2, Y2, X3, Y3, X4, Y4
where Xi and Yi are the coordinates of corner i of the bounding box in frame N, the N-th row in the text file.

The bounding box was be placed on target such that at most ~30% of pixels within the bounding box corresponded to the background pixels, while containing most of the target. For example, in annotating a person with extended arms, the bounding box was placed such that the arms were not included. Note that in some sequences parts of objects rather than entire objects have been annotated. A rotated bounding box was used to address non-axis alignment of the target. The annotation guidelines have been applied at the judgement of the annotators.

Some targets were partially occluded or were partially out of the image frame. In these cases the bounding box were “inferred” by the annotator to fully contain the object, including the occluded part. For example, if a person’s legs were occluded, the bounding box should also include the non-visible legs.

The annotations have been conducted by three groups of annotators. Each annotator group annotated one third of the dataset and these annotations have been cross-checked by two other groups. The final annotations were checked by the coordinator of the annotation process. The final bounding box annotations have been automatically rectified by replacing a rotated bounding box by an axis-aligned if the ratio of the shortest and longest bounding-box side exceeded 0.95.

Annotators:

Gustavo Fernandez (coordinator)
Jingjing Xiao
Georg Nebehay
Roman Pflugfelder
Koray Aytac

https://www.votchallenge.net/vot2015/dataset.html

futbol players

Background Information

This dataset was curated and annotated by Ilyes Talbi, Head of La revue IA, a French publication focused on stories of machine learning applications.

Main objetive is to identify if soccer (futbol) players, the referree and the soccer ball (futbol).

The original custom dataset (v1) is composed of 163 images.

  • Class 0 = players
  • Class 1 = referree
  • Class 2 = soccer ball (or futbol)

The dataset is available under the Public License.

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 7 (v7) - 163 images (raw images)

  • Preprocessing: Auto-Orient, Modify Classes: 3 remapped, 0 dropped
    • Modified Classes: Class 0 = players, Class 1 = referree, Class 2 = futbol
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 2 (v2) - 163 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 3 (v3) - 391 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416), Modify Classes: 3 remapped, 0 dropped
    • Modified Classes: Class 0 = players, Class 1 = referree, Class 2 = futbol
  • Augmentations:
    • Outputs per training example: 3
    • Rotation: Between -25° and +25°
    • Shear: ±15° Horizontal, ±15° Vertical
    • Brightness: Between -25% and +25%
    • Blur: Up to 0.75px
    • Noise: Up to 1% of pixels
    • Bounding Box: Blur: Up to 0.5px
  • Training Metrics: 86.4%mAP, 51.8% precision, 90.4% recall

Version 4 (v4) - 391 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416), Modify Classes: 3 remapped, 0 dropped
    • Modified Classes: Class 0 = players, Class 1 = referree, Class 2 = futbol
  • Augmentations:
    • Outputs per training example: 3
    • Rotation: Between -25° and +25°
    • Shear: ±15° Horizontal, ±15° Vertical
    • Brightness: Between -25% and +25%
    • Blur: Up to 0.75px
    • Noise: Up to 1% of pixels
    • Bounding Box: Blur: Up to 0.5px
  • Training Metrics: 84.6% mAP, 52.3% precision, 85.3% recall

Version 5 (v5) - 391 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416), Modify Classes: 3 remapped, 2 dropped
    • Modified Classes: Class 0 = players, Class 1 = referree, Class 2 = futbol
      • Only Class 0, which was remapped to players was included in this version
  • Augmentations:
    • Outputs per training example: 3
    • Rotation: Between -25° and +25°
    • Shear: ±15° Horizontal, ±15° Vertical
    • Brightness: Between -25% and +25%
    • Blur: Up to 0.75px
    • Noise: Up to 1% of pixels
    • Bounding Box: Blur: Up to 0.5px
  • Training Metrics: Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow
    • 98.8%mAP, 76.3% precision, 99.2% recall

Version 6 (v6) - 391 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416), Modify Classes: 3 remapped, 2 dropped
    • Modified Classes: Class 0 = players, Class 1 = referree, Class 2 = futbol
      • Only Class 0, which was remapped to players was included in this version
  • Augmentations:
    • Outputs per training example: 3
    • Rotation: Between -25° and +25°
    • Shear: ±15° Horizontal, ±15° Vertical
    • Brightness: Between -25% and +25%
    • Blur: Up to 0.75px
    • Noise: Up to 1% of pixels
    • Bounding Box: Blur: Up to 0.5px
  • Training Metrics: Trained from Scratch (no transfer learning employed)
    • 95.5%mAP, 67.8% precision, 95.5% recall

Ilyes Talbi - LinkedIn | La revue IA

FLIR data set

Self-Driving Thermal Object-Detection

Overview

This model detects potentially moving objects (cars, bicycles, people, and dogs), to aid in self-driving and autonomous vehicles.

Dataset

The dataset is comprised of over twelve thousand thermal images, largely annotating cars.

Kernels counter

  • Public
  • Soybeans-kernels Dataset
  • 840 images

Soybeans kernels counter

Pill Detection

Background Information

This dataset was curated and annotated by Mohamed Attia.

The original dataset (v1) is composed of 451 images of various pills that are present on a large variety of surfaces and objects.
Example of an Annotated Image from the Dataset

The dataset is available under the Public License.

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 1 (v1) - 451 images

  • Preprocessing: Auto-Orient and Resize (Stretch to 416x416)
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 2 (v2) - 1,083 images

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416), all classes remapped (Modify Classes) to "pill"
  • Augmentations:
    90° Rotate: Clockwise, Counter-Clockwise, Upside Down
    Crop: 0% Minimum Zoom, 77% Maximum Zoom
    Rotation: Between -45° and +45°
    Shear: ±15° Horizontal, ±15° Vertical
    Hue: Between -22° and +22°
    Saturation: Between -27% and +27%
    Brightness: Between -33% and +33%
    Exposure: Between -25% and +25%
    Blur: Up to 3px
    Noise: Up to 5% of pixels
    Cutout: 3 boxes with 10% size each
    Mosaic: Applied
    Bounding Box: Brightness: Between -25% and +25%
  • Training Metrics: Trained from the COCO Checkpoint in Public Models ("transfer learning") on Roboflow
    • mAP = 91.4%, precision = 61.1%, recall = 93.9%

Version 3 (v3) - 1,083 images

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416), all classes remapped (Modify Classes) to "pill"
  • Augmentations:
    90° Rotate: Clockwise, Counter-Clockwise, Upside Down
    Crop: 0% Minimum Zoom, 77% Maximum Zoom
    Rotation: Between -45° and +45°
    Shear: ±15° Horizontal, ±15° Vertical
    Hue: Between -22° and +22°
    Saturation: Between -27% and +27%
    Brightness: Between -33% and +33%
    Exposure: Between -25% and +25%
    Blur: Up to 3px
    Noise: Up to 5% of pixels
    Cutout: 3 boxes with 10% size each
    Mosaic: Applied
    Bounding Box: Brightness: Between -25% and +25%
  • Training Metrics: Trained from "scratch" (no transfer learning employed) on Roboflow
    • mAP = 84.3%, precision = 53.2%, recall = 86.7%

Version 4 (v4) - 451 images

  • Preprocessing: Auto-Orient, Resize (Stretch to 416x416), all classes remapped (Modify Classes) to "pill"
  • Augmentations: No augmentations applied
  • Training Metrics: This version of the dataset was not trained

Version 5 (v5) - 496 images

Mohamed Attia - LinkedIn

mos

FinalGolf

Cats

About this Dataset

This dataset was created by exporting the Oxford Pets dataset from Roboflow Universe, generating a version with Modify Classes to drop all of the classes for the labeled dog breeds and consolidating all cat breeds under the label, "cat." The bounding boxes were also modified to incude the entirety of the cats within the images, rather than only their faces/heads.

Annotated image of a cat from the dataset

Oxford Pets

  • The Oxford Pets dataset (also known as the "dogs vs cats" dataset) is a collection of images and annotations labeling various breeds of dogs and cats. There are approximately 100 examples of each of the 37 breeds. This dataset contains the object detection portion of the original dataset with bounding boxes around the animals' heads.

  • Origin: This dataset was collected by the Visual Geometry Group (VGG) at the University of Oxford.

Valentines Chocolates

Overview

This is an attempt to make a computer vision model that can identify chocolates in a box of chocolates.

I trained this model on images of a See’s 1lb Classic Red Heart - Assorted Chocolates box. There are 22 different classes, one for each type of chocolate in this box.

This dataset contains 87 original images, and 1,697 annotations.

Learn More

Read more about this project on the Roboflow blog.

Rock Paper Scissors

Kaylee from Team Roboflow demos how to train a rock paper scissors object detector with this dataset.

Try it out on your Webcam! And remember, if it doesn't work so well, to make it better, you can upload and annotate new images of yourself doing rock paper scissors to futher educate the model.

Excavators

This project is trying to create an efficient computer or machine vision model to detect different kinds of construction equipment in construction sites and we are starting with three classes which are excavators, trucks, and wheel loaders.

The dataset is provided by Mohamed Sabek, a Spring 2022 Master of Science graduate from Arizona State University in Construction Management and Technology.

The raw images (v1) contains:

  1. 1,532 annotated examples of "excavators"
  2. 1,269 annotated examples of "dump truck"
  3. 1,080 annotated examples of "wheel loader"

Note: versions 2 and 3 (v2 and v3) contain the raw images resized at 416 by 416 (stretch to) and 640 by 640 (stretch to) without any augmentations.

PPEs

Personal Protective Equipment Dataset and Model

This dataset is a collection of images that contains annotations for the classes below:

  • goggles
  • helmet
  • mask
  • no-suit
  • no_goggles
  • no_helmet
  • no_mask
  • no_shoes
  • shoes
  • suit
  • no_glove
  • glove

Usage

Most of these classes are underrepresented and would need to be balanced for better detection. An improved model can be utilized for use cases that'll detect the classes above in order to minimize exposure to hazards that cause serious workplace injuries.

Table Extraction PDF

The dataset comes from Devashish Prasad, Ayan Gadpal, Kshitij Kapadni, Manish Visave, and Kavita Sultanpure - creators of CascadeTabNet.

Depending on the dataset version downloaded, the images will include annotations for 'borderless' tables, 'bordered' tables', and 'cells'. Borderless tables are those in which every cell in the table does not have a border. Bordered tables are those in which every cell in the table has a border, and the table is bordered. Cells are the individual data points within the table.

A subset of the full dataset, the ICDAR Table Cells Dataset, was extracted and imported to Roboflow to create this hosted version of the Cascade TabNet project. All the additional dataset components used in the full project are available here: All Files.

Versions:

  1. Version 1, raw-images : 342 raw images of tables. No augmentations, preprocessing step of auto-orient was all that was added.
  2. Version 2, tableBordersOnly-rawImages : 342 raw images of tables. This dataset version contains the same images as version 1, but with the caveat of Modify Classes being applied to omit the 'cell' class from all images (rendering these images to be apt for creating a model to detect 'borderless' tables and 'bordered' tables.

For the versions below: Preprocessing step of Resize (416by416 Fit within-white edges) was added along with more augmentations to increase the size of the training set and to make our images more uniform. Preprocessing applies to all images whereas augmentations only apply to training set images.
3. Version 3, augmented-FAST-model : 818 raw images of tables. Trained from Scratch (no transfer learning) with the "Fast" model from Roboflow Train. 3X augmentation (generated images).
4. Version 4, augmented-ACCURATE-model : 818 raw images of tables. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation.
5. Version 5, tableBordersOnly-augmented-FAST-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Fast" model from Roboflow Train. 3X augmentation.
6. Version 6, tableBordersOnly-augmented-ACCURATE-model : 818 raw images of tables. 'Cell' class ommitted with Modify Classes. Trained from Scratch with the "Accurate" model from Roboflow Train. 3X augmentation.

Example Image from the DatasetExample Image from the Dataset

Cascade TabNet in ActionCascade TabNet in Action
CascadeTabNet is an automatic table recognition method for interpretation of tabular data in document images. We present an improved deep learning-based end to end approach for solving both problems of table detection and structure recognition using a single Convolution Neural Network (CNN) model. CascadeTabNet is a Cascade mask Region-based CNN High-Resolution Network (Cascade mask R-CNN HRNet) based model that detects the regions of tables and recognizes the structural body cells from the detected tables at the same time. We evaluate our results on ICDAR 2013, ICDAR 2019 and TableBank public datasets. We achieved 3rd rank in ICDAR 2019 post-competition results for table detection while attaining the best accuracy results for the ICDAR 2013 and TableBank dataset. We also attain the highest accuracy results on the ICDAR 2019 table structure recognition dataset.

From the Original Authors:

If you find this work useful for your research, please cite our paper:
@misc{ cascadetabnet2020,
title={CascadeTabNet: An approach for end to end table detection and structure recognition from image-based documents},
author={Devashish Prasad and Ayan Gadpal and Kshitij Kapadni and Manish Visave and Kavita Sultanpure},
year={2020},
eprint={2004.12629},
archivePrefix={arXiv},
primaryClass={cs.CV}
}

American Mushrooms

This Dataset contains images of popular North American mushrooms, Chicken of the Woods and Chanterelle, differentiating between the two species.

This dataset is an example of an object detection task that is possible via custom training with Roboflow.

Two versions are listed. "416x416" is a 416 resolution version that contains the base images in the dataset. "416x416augmented" contains the same images with various image augmentations applied to build a more robust model.

Rapid React Balls

This dataset was prepared for First Robotics by the 2914 Robotics Team of Wilson High School.

The dataset contains labeled blue and red balls. The original dataset contains 987 blue annotated examples, and 731 red annotated examples of the balls.

The raw images are contained in version 10 (raw-images)

CARLA

Dataset collected from CARLA.

10 classes:
- Good for traffic light detection by color
- Good for traffic sign detection by speed
- Cars,trucks,... have been simplified to "vechicles"
- Bikes, motobikes and persons

More about me

You can find out more about me on my linkedin

The Dreidel Project

  • Salo Levy
  • Hebrew-Letters Dataset
  • 575 images

When learning to play Dreidel, I would sometimes forget what the names of each character are and what action they correspond to in the game. I thought it’d be fun to create a computer vision model that could understand what each symbol on a Dreidel is, making it easier to learn to play the game.

This model tracks the dreidel as it spins and detects the letters that are on the four sided dreidel.

How to Play Dreidel

Rules:
1. The players are dealt gelt (chocolate wrapped in gold paper made to look like a coin)
2. Each player takes a turn at spinning the Dreidel
3. The Dreidel has four sides that each prompt an action to take by the spinner
If נ‎ (nun) is facing up, the player does nothing.
If ג‎ (gimel) is facing up, the player gets everything in the pot.
If ה‎ (hay) is facing up, the player gets half of the pieces in the pool.
If ש‎ (shin) the player adds one of their gelt to the pot
4. The winner, of course, gets to eat all the gelt

Hopefully, with this application, one can create an application that teaches someone how to play dreidel.

golfBall

Golf Ball Object Detection

Usage

This model will perform best on images + videos that are taken on a golf course (similar to photo in thumbnail and dataset).

It's a great model for sports broadcasting and other apps to have automated ball tracking, scoring, lost ball finding and more!

This dataset was found on Kaggle's AFO - Aerial Dataset of floating objects by Jan Gąsienica-Józkowy, and was used to build an object detection model for the "How to Train Computer Vision Models on Aerial Imagery" technical blog.

Future iterations may include the detection of people, as well as the use of Project_2 and Project_3 of the original dataset.

Smoke100

Smoke Detection Dataset

This computer vision smoke detection dataset contains images of synthsized smoke in both indoor and outdoor settings. Check out the source link below for more information on this dataset.

source:

Smoke100k dataset
https://bigmms.github.io/cheng_gcce19_smoke100k/

Use Cases

  • Identifying smoke indoors
  • Identifying smoke outdoors (but not with aerial imagery)
  • Identifying smoke-like object (eg: mist/steam from humidifiers)

Testing

You can test this model by using the Roboflow Inference Widget found above. The action hits the model inference API, which in turn produces the color coded bounding boxes on the objects the model was trained to detect, along with its labels, and confidence for each prediction. The feature also produces the JSON output provided by the API.

Boxpunch Detector

Boxpunch Detector

Onboarding project for Roboflow

This project captures punch types thrown during boxing training

UFC Champion

Ultimate Fighting Champion

This data set and object detection model detects recent UFC champions.

Clash of Clans

Background Information

This dataset was curated and annotated by Find This Base. A custom dataset composed of 16 classes from the popular mobile game, Clash of Clans.

  • Classes: Canon, WizzTower, Xbow, AD, Mortar, Inferno, Scattershot, AirSweeper, BombTower, ClanCastle, Eagle, KingPad, QueenPad, RcPad, TH13 and WardenPad.

Find This Base

How to Use Find This Base
How to Use Find This Base

The original custom dataset (v1) is composed of 125 annotated images.

The dataset is available under the CC BY 4.0 license.

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 1 (v1) - 125 images

  • Preprocessing - Auto-Orient and Resize: Fit (black edges) to 640x640
  • Augmentations - No augmentations applied
  • Training Metrics - Trained from Scratch (no checkpoint used) on Roboflow
    • mAP = 83.1%, precision = 43.0%, recall = 99.1%

Version 4 (v4) - 301 images

  • Preprocessing - Auto-Orient and Resize: Fit (black edges) to 640x640
  • Augmentations - Mosaic
  • Generated Images - Outputs per training example: 3
  • Training Metrics - Trained from Scratch (no checkpoint used) on Roboflow
    • mAP = %, precision = %, recall = %

Find This Base: Official Website | How to Use Find This Base | Discord | Patreon

Chest X-Rays

This classification dataset is from Kaggle and was uploaded to Kaggle by Paul Mooney.

It contains over 5,000 images of chest x-rays in two categories: "PNEUMONIA" and "NORMAL."

  • Version 1 contains the raw images, and only has the pre-processing feature of "Auto-Orient" applied to strip out EXIF data, and ensure all images are "right side up."
  • Version 2 contains the raw images with pre-processing features of "Auto-Orient" and Resize of 640 by 640 applied
  • Version 3 was trained with Roboflow's model architecture for classification datasets and contains the raw images with pre-processing features of "Auto-Orient" and Resize of 640 by 640 applied + augmentations:
    • Outputs per training example: 3
    • Shear: ±3° Horizontal, ±2° Vertical
    • Saturation: Between -5% and +5%
    • Brightness: Between -5% and +5%
    • Exposure: Between -5% and +5%

Below you will find the description provided on Kaggle:

Context

http://www.cell.com/cell/fulltext/S0092-8674(18)30154-5
Figure S6
Figure S6. Illustrative Examples of Chest X-Rays in Patients with Pneumonia, Related to Figure 6
The normal chest X-ray (left panel) depicts clear lungs without any areas of abnormal opacification in the image. Bacterial pneumonia (middle) typically exhibits a focal lobar consolidation, in this case in the right upper lobe (white arrows), whereas viral pneumonia (right) manifests with a more diffuse ‘‘interstitial’’ pattern in both lungs.
http://www.cell.com/cell/fulltext/S0092-8674(18)30154-5

Content

The dataset is organized into 3 folders (train, test, val) and contains subfolders for each image category (Pneumonia/Normal). There are 5,863 X-Ray images (JPEG) and 2 categories (Pneumonia/Normal).

Chest X-ray images (anterior-posterior) were selected from retrospective cohorts of pediatric patients of one to five years old from Guangzhou Women and Children’s Medical Center, Guangzhou. All chest X-ray imaging was performed as part of patients’ routine clinical care.

For the analysis of chest x-ray images, all chest radiographs were initially screened for quality control by removing all low quality or unreadable scans. The diagnoses for the images were then graded by two expert physicians before being cleared for training the AI system. In order to account for any grading errors, the evaluation set was also checked by a third expert.

Acknowledgements

Data: https://data.mendeley.com/datasets/rscbjbr9sj/2

License: CC BY 4.0

Citation: http://www.cell.com/cell/fulltext/S0092-8674(18)30154-5
citation - latest version (Kaggle)

Inspiration

Automated methods to detect and classify human diseases from medical images.

Draughts Board

This dataset was created by Harry Field and contains the labelled images for capturing the game state of a draughts/checkers 8x8 board.

This was a fun project to develop a mobile draughts applciation enabling users to interact with draughts-based software via their mobile device's camera.

The data captured consists of:

  • White Pieces
  • White Kings
  • Black Pieces
  • Black Kings
  • Bottom left corner square
  • Top left corner square
  • Top right corner square
  • Bottom right corner square

Corner squares are captured so the board locations of the detected pieces can be estimated.

Results of Yolov5 model after training with this dataset

From this data, the locations of other squares can be estimated and game state can be captured. The image below shows the data of a different board configuration being captured. Blue circles refer to squares, numbers refer to square index and the coloured circles refer to pieces.

Once game state is captured, integration with other software becomes possible. In this example, I created a simple move suggestion mobile applciation seen working here.

The developed application is a proof of concept and is not available to the public. Further development is required in training the model accross multiple draughts boards and implementing features to add vlaue to the physical draughts game.

The dataset consists of 759 images and was trained using Yolov5 with a 70/20/10 split.

The output of Yolov5 was parsed and filtered to correct for duplicated/overlapping detections before game state could be determined.

I hope you find this dataset useful and if you have any questions feel free to drop me a message on LinkedIn as per the link above.

QR Code

Overview

This dataset contains images taken of QR Codes in variable lighting conditions at different angles.

Trained Model with Roboflow Train

High Performance

  • 99.5% mAP
  • 100.0% precision
  • 99.2% recall

Testing The Model

You can test this trained model by dropping an image on this page or via curl command

base64 YOUR_IMAGE.jpg | curl -d @- "https://detect.roboflow.com/qr-code-oerhe/1?api_key=YOUR_API_KEY"

Garage

Object tracking for cars in my garage for use in home automation.

https://github.com/brianegge/garbage_bin

Honey_bees_dataset

Background Information

This dataset was curated and annotated by Ahmed Elmogtaba Abdelaziz.

The original dataset (v6) is composed of 204 images of honeybees present in a wide variety of scenes.
Example of an Annotated Image from the Dataset

The dataset is available under a Public License.

Getting Started

You can download this dataset for use within your own projects, or fork it into a workspace on Roboflow to create your own model.

Dataset Versions

Version 5 - 490 images

  • Preprocessing: Resize, 416 by 416
  • Augmentations:
  • 90° Rotate: Clockwise, Counter-Clockwise
    Rotation: Between -15° and +15°
    Saturation: Between -10% and +10%
    Brightness: Between -10% and +10%
    Blur: Up to 0.25px
    Mosaic: Applied
  • Output: 3x image generation