Skip to content

A CPU based video summarization and classification tool. Extracts every movement out of a given video, aggregates the extracted changes into movement layers, which can be exported or further analysed and classified..

Notifications You must be signed in to change notification settings

Askill/Video-Summary

Repository files navigation

Video Summary and Classification

Example:

    usage:
    main.py input_video.mp4 output_dir ?config_file.json

docs/demo.gif
What you see above is a 15 second excerpt of a 2 minute overlayed synopsis of a 2.5h video from an on campus web cam.
The synopsis took 40 minutes from start to finish on a 8 core machine and used a maximum of 6Gb of RAM.

However since the contour extraction could be performed on a video stream, the benchmark results show that a single core would be enough to process a video faster than real time.

Heatmap

Benchmark

Below you can find the benchmark results for a 10 minutes clip, with the stacked time per component on the x-axis.
The tests were done on a machine with a Ryzen 3700X with 8 cores 16 threads and 32 Gb of RAM.
On my configuration 1 minutes of of the original Video can be processed in about 20 seconds, the expected processing time is about 1/3 of the orignial video length.

  • CE = Contour Extractor
  • LE = LayerFactory
  • LM = LayerManager
  • EX = Exporter

docs/demo.gif

Configuration

./Application/Config.py

    "min_area": 100,            min area in pixels, of a single contour, smaller is ignored
    "max_area": 9000000,        max area in pixels, of a single contour, larger is ignored
    "threshold": 6,            luminance difference threshold, sensitivity of movement detection
    "resizeWidth": 600,         video is scaled down internally
    "inputPath": None,          overwritten in main.py
    "outputPath": None,         overwritten in main.py
    "maxLayerLength": 5000,     max length of Layer in frames
    "minLayerLength": 10,       min length of Layer in frames
    "tolerance": 100,           max distance (in pixels) between contours to be aggragated into layer
    "maxLength": None,          
    "ttolerance": 60,           number of frames movement can be apart until a new layer is created
    "videoBufferLength": 100,   Buffer Length of Video Reader Componenent
    "LayersPerContour": 2,      number of layers a single contour can belong to
    "avgNum": 10,               number of images that should be averaged before calculating the difference 
                                (computationally expensive, needed in outdoor scenarios due to clouds, leaves moving in the wind ...)

notes:

optional:

install tensorflow==1.15.0 and tensorflow-gpu==1.15.0, cuda 10.2 and 10.0, copy missing files from 10.0 to 10.2, restart computer, set maximum vram

About

A CPU based video summarization and classification tool. Extracts every movement out of a given video, aggregates the extracted changes into movement layers, which can be exported or further analysed and classified..

Topics

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

 
 
 

Languages