Skip to content

maxencecarrel/strudl

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

STRUDL: Surveillance Tracking Using Deep Learning

STRUDL is an open-source and free framework for tracking objects in videos filmed by static surveillance cameras. It uses a deep learining object detector, camera calibration and tracking to create trajectories of e.g. road users, in world coordinates. It was designed to faciliate traffic safety analysis, using modern computer vision and deep learning, rather than the traditional methods commonly used despite their many flaws. By creating trajectories in world coordinates, truly meaningful metrics and safety measures can be computed. The paper behind this research is available here. Note that the development of STRUDL has continued after the publication; most notably, STRUDL now does tracking in world coordinates rather than pixel coordinates.

STRUDL was developed as a part of the InDeV project.

InDeV Logo

It provides a Web UI that attempts to make it easy to use, even without too much knowledge in computer vision and deep learning.

It uses Docker for sandboxing and handling dependencies, to make installation easier. The code is designed to be modular, so that new features can be added fairly easily without destroying everything else. It uses Swagger and Connexion to provide a REST API and parts of the Web UI. Because all the functionality can be used via a REST API, users can create their own interfaces if they prefer, to simplify some actions and automate tasks.

Some more information about how STRUDL works internally can be found here.

Got any issues with this software? Feel free to open an issue, if there isn't one already for your problem!

Workflow

  1. Import videos alongside log files specifying the times for each frame in the videos
  2. Annotate images
  3. Train an object detector
  4. Provide camera calibration
  5. Perform tracking in world coordinates
  6. Download the tracks, and analyze them with whatever tools you like

Some features

  1. Take advantage of modern deep learning without extensive knowledge of the technology
  2. Easily make object detection annotations via custom web interface
  3. Semi-automatic annotation process, where a detector can be trained on the currently available training data, and those detections can then be used as a starting point for annotations, speeding up the process once a critical number of images has been annotated.
  4. When importing videos, typically from a USB drive, videos can be recoded to arbitrarily long/short clips, and different resolutions.
  5. Tracking in world coordinates, which can optimized if the user provides ground truth tracks made by T-Analyst. In our experience, this is not necessary and the default parameters work OK for most scenarios.
  6. Full trajectories can be obtained as text files (e.g. csv), which the user can then filter, sort and present, using any tools they prefer. This allows e.g. complex traffic analysis.
  7. Visualize each step in the process in summary video files, to better understand what goes wrong when something goes wrong.

Data assumptions

  1. Data should be in video files, filmed from static (non-moving) surveillance cameras.
  2. A TSAI-calibration should exist, allowing translation between pixel coordinates and world coordinates
  3. For each video file, there should be a log file specifying the time for each frame. This allows more accurate measures of time which is useful in tracking. If your videos do not have this information, simply generate such text files using some approximate method of your choice. The log files should be text files in the .log format, where each line looks like this: 00158 2017 05 16 01:28:52.642, where the first number (with five characters) is the frame number, followed by the year, month, day, hour, minute, second and milliseconds.

Requirements

  1. A Linux computer with a powerful, modern NVidia GPU
  2. NVidia CUDA, STRUDL is made for CUDA 8.0, but there is also a CUDA 10.0 branch
  3. docker
  4. nvidia-docker

Installation

The following terminal commands should work on Ubuntu, assuming the requirements are installed correctly. The first thing to do is to run the testsuit to ensure your requirements is set up properly. This will also download strudl if needed:

nvidia-docker run -ti ahrnbom/strudl py.test -v

A data folder has to be created. It can be put anywhere you want, like on a different disk. This could be useful as the data folder can get quite large if large amounts of video are to be processed.

mkdir -p /path/of/your/choosing/strudl/data

Download the starter script from here and place it next to the data directory and run it to start strudl

/path/of/your/choosing/strudl/start_strudl.sh

If you want to upgrade to the latest version before starting add the -u option:

/path/of/your/choosing/strudl/start_strudl.sh -u

It is also possible to run a specific version, say for example version 0.3, using

/path/of/your/choosing/strudl/start_strudl.sh -v 0.3

If you have a recent NVidia GPU, there is also CUDA 10 versions, which can be started with for example,

/path/of/your/choosing/strudl/start_strudl.sh -v 0.3-CUDA10

A list of all availble version can be found at https://cloud.docker.com/repository/docker/ahrnbom/strudl/tags

Visit the host computer via a web browser to see the Web UI and interact with it. For example, if you're using the web browser on the same computer, visit localhost in a web browser like Firefox.

Local installation

If for some reason, you do not want to download STRUDL from Docker Hub and prefer to build the container yourself, you can do so yourself locally. In addition to the previously mentioned requirements, you also need to install git. Clone the repo, enter the folder, make a data folder, and then build and run the container with the following commands

git clone https://github.com/ahrnbom/strudl.git
cd strudl
mkdir data
sudo ./run_docker.sh

If everything works as intended, you will be inside a bash session inside the docker container. From there you can test STRUDL with py.test -v or start the STRUDL server with python server.py

To leave the container, press Ctrl + D. To run start STRUDL's docker container again, only the command sudo ./run_docker.sh needs to be used (inside the right folder of course).

Security notice

This software has not been designed with maximum security in mind. It is recommended to run it in a local network behind a firewall. While docker does provide some sandboxing, this code is not "battle tested" and it should not be assumed to be safe. Leaving the port open to the internet could compromise your computer. One possible security flaw is that your computer's /media folder is being made available in the docker container, to simplify importing videos via e.g. USB. This can be changed by modifying start_strudl.sh (or run_docker.sh if running on a local docker image).

Future work

There's always more to do! On our to-do list contains, among other things, the following:

  1. Simplification of the Web UI, to make it easier to understand what has been done and what to do next
  2. SSD code is currently based on this port by Rykov8. It might be a good idea to change to this one instead, by Pierluigiferrari, which is more nicely documented and runs NMS on the GPU.
  3. Code should become more readable and better commented (this started as, and in many ways still is, experimental research code)
  4. Different tracking algorithms should be examined, possibly replacing the simplistic one currently used
  5. More easter eggs, jokes and memes

About

STRUDL: Surveillance Tracking Using Deep Learning

Resources

License

Stars

Watchers

Forks

Packages

No packages published

Languages

  • Python 89.6%
  • HTML 9.5%
  • Other 0.9%