Skip to content

Latest commit

 

History

History
50 lines (38 loc) · 1.66 KB

README.md

File metadata and controls

50 lines (38 loc) · 1.66 KB

Tracking Olympiad

Goal

This is a project preseneted by Prof. Dr. Andreas Kist at FAU. The aim of this project is to detect the precise location of all Hexbugs in videos. We are provided with 100 videos, containing ~ 101 frame each, alognside the annotations of Hexbugs' heads' location at each frame.

Approach

Pre-processing

Given data are in mp4 format. We need to extract every single frame and save it in specific directories.

Step I

We are trying to detect the Hexbugs in the each frame using Yolo v5 algorithm. By doing so, we are left with cropped images containing the Hexbugs. By doing so, we got rid of void disturbing areas in the image.

Step II

After obtaining the cropped images, we use Resnet50 model for feature extraction, and then we add Dense layers at the top of the model to perform regression. Now, we have created a model that is capable of estimating the head's coordination.

Post-processing

After computing the head's position, we need to reverse the process to retrieve head's original coordination in the original frame.

How to run

  1. Clone this git repository.
git clone https://github.com/FarzamTP/TRACO-HexBug.git
  1. Enter the directory
cd TRACO-HexBug
  1. [Pre-processing frames]
jupyter notebook frame_prepration.ipynb
jupyter notebook hexbug_object_detection_YOLO_torch.ipynb
  1. [Training]
jupyter notebook hexbug_head_detection_resnet.ipynb
  1. [Load Model]
jupyter notebook load_model.ipynb

Note

To download the annotated dataset for Yolo, you need an api_key. Please use your Roboflow account to export the dataset using your key.