Pytorch version of - https://github.com/WaqasSultani/AnomalyDetectionCVPR2018
In this section, I list the future improvements I intend to add to this repository. Please feel free to recommend new features. I also happily accept PR's! ๐
- I3D feature extraction
- MFNET feature extraction
- AUC is not exactly as reported in the paper (0.70 vs 0.75) - might be affected by the weights of C3D
conda env create -f environment.yml
conda activate adCVPR18
C3D Weights I couldn't upload here the weights for the C3D model because the file is too big, but it can be found here: https://github.com/DavideA/c3d-pytorch
Can be downloaded from: https://drive.google.com/drive/folders/1rhOuAdUqyJU4hXIhToUnh5XVvYjQiN50?usp=sharing
Check out exps/models for for trained models on the pre-computed features
The loss graph during training is shown here:
Download the dataset from: https://github.com/WaqasSultani/AnomalyDetectionCVPR2018 Arguments:
- dataset_path - path to the directory containing videos to extract features for (the dataset is available for download above)
- model_type - which type of model to use for feature extraction (necessary in order to choose the correct pre-processing)
- pretrained_3d - path to the 3D model to use for feature extraction
python feature_extractor.py --dataset_path "path-to-dataset" --model_type "fe-model-eg-c3d" --pretrained_3d "path-to-pretrained-fe"
- to extract features with stochastic augmentations use
feature_extractor_augs.py
script
Arguments:
- features_path - path to the directory containing the extracted features (pre-computed features are available for download above, or supply your own features extracted from the previous stage)
- annotation_path - path to the annotations file (Available in this repository as
Train_annotations.txt
python TrainingAnomalyDetector_public.py --features_path "path-to-dataset" --annotation_path "path-to-train-annos"
- to train with our implementation of triplet loss and sampling scheme 2 you have to add 2 flags:
--network_name TripletAnomalyDetector --objective_name triplet_objective
- to train with pytorch metrics learning library, sampling scheme 1, you need to use flags:
--network_name TripletAnomalyDetector
--objective_name PytorchMetricLearningObjectiveWithSampling
--loss_name [TripletMarginLoss, CircleLoss, ArcFaceLoss]
--miner_name [MultiSimilarityMiner, TripletMarginMiner, BatchHardMiner]
- to train with SimCLR approach use
TrainingAnomalyDetector_SimCLR.py
script you have to add 4 flags (features1 and features2 - paths to precomputed augmentation of the dataset):--network_name TripletAnomalyDetector --objective_name SimCLRLoss --features_path_1 features1 --features_path features2
- Model parameters available:
--no_use_last_bn or --use_last_bn
if we want BN after the last layer--no_norm_out_to_unit or --norm_out_to_unit
if we want to normalize the embedding to unit norm
- Optimizer parameters:
--optimizer [adam, adadelta]
--lr_base 0.001
Arguments:
- features_path - path to the directory containing the extracted features (pre-computed features are available for download above, or supply your own features extracted from the previous stage)
- annotation_path - path to the annotations file (Available in this repository as
Test_annotations.txt
- model_path - path to the trained anomaly detection model
python generate_ROC.py --features_path "path-to-dataset" --annotation_path "path-to-annos" --model_path "path-to-model"
- to calculate ROC for the representation learning mode in the 1st frame is always normal scheme you have add
--calc_mode triplet --train_features_path path_to_train_features
-
- to calculate ROC for the representation learning mode in centroid scheme you have add
--calc_mode triplet --train_features_path path_to_train_features --train_annotation_path path_to_train_annotations --use_centroid
- to calculate ROC for the representation learning mode in centroid scheme you have add
Using my pre-trained model after 40K iterations, I achieve this following performance on the test-set. I'm aware that the current model doesn't achieve AUC of 0.75 as reported in the original paper. This can be caused by different weights of the C3D model.
Arguments:
- feature_extractor - path to the 3D model to use for feature extraction
- feature_method - which type of model to use for feature extraction (necessary in order to choose the correct pre-processing)
- ad_model - path to the trained anomaly detection model
- n_segments - the number of segments to chunk the video to (the original paper uses 32 segments)
python video_demo.py --feature_extractor "path-to-pretrained-fe" --feature_method "fe-method" --ad_model "path-to-pretrained-ad-model" --n_segments "number-of-segments"
The GUI lets you load a video and run the Anomaly Detection code (including feature extraction) and output a video with a graph of the Anomaly Detection prediction below.
python annotation_methods.py --path_list LIST_OF_VIDEO_PATH --dir_list LIST_OF_LIST_WITH_PATH_AND_VIDEO_NAME --normal_or_not LIST_TRUE_FALUE
This is currently just for demo but will allow training with nex videos
*Contrbuted by Peter Overbury of Sussex Universty IISP Group
@misc{anomaly18cvpr-pytorch,
author = "Eitan Kosman",
title = "Pytorch implementation of Real-World Anomaly Detection in Surveillance Videos",
howpublished = "\url{https://github.com/ekosman/AnomalyDetectionCVPR2018-Pytorch}",
note = "Accessed: 20xx-xx-xx"
}
Q: video_demo doesn't show videos
A: Downlaod and install LAVFilters: http://forum.doom9.org/showthread.php?t=156191