The paper “Convolutional Networks for semantic Heads Segmentation using Top-View Depth Data in Crowded Environment” [1] introduces an approach to track and detect people in cases of heavy occlusions based on CNNs for semantic segmentation using top-view RGB-D visual data. The purpose is the design of a novel U-Net architecture, U-Net 3, that has been modified compared to the previous ones at the end of each layer. In order to evaluate this new architecture a comparison has been made with other networks in the literature used for semantic segmentation. The implementation is in Python code using Keras API with Tensorflow library. The input data consist of depth frames, from Asus Xtion Pro Live OpenNI recordings (.oni). The dataset used for training and testing of the networks has been manually labeled and it is freely available as well as the source code. The aforementioned networks have their stand-alone Python script implementation for training and testing. A Python script for the on-line prediction in OpenNI recordings (.oni) is also provided. Evaluation of the networks has been made with different metrics implementations (precision, recall, F1 Score, Sørensen-Dice coefficient), included in the networks scripts.
CNN Implementation for Semantic Heads Segmentation Using Top-View Depth Data in Crowded Environment
Paolanti M.;Frontoni E.;
2019-01-01
Abstract
The paper “Convolutional Networks for semantic Heads Segmentation using Top-View Depth Data in Crowded Environment” [1] introduces an approach to track and detect people in cases of heavy occlusions based on CNNs for semantic segmentation using top-view RGB-D visual data. The purpose is the design of a novel U-Net architecture, U-Net 3, that has been modified compared to the previous ones at the end of each layer. In order to evaluate this new architecture a comparison has been made with other networks in the literature used for semantic segmentation. The implementation is in Python code using Keras API with Tensorflow library. The input data consist of depth frames, from Asus Xtion Pro Live OpenNI recordings (.oni). The dataset used for training and testing of the networks has been manually labeled and it is freely available as well as the source code. The aforementioned networks have their stand-alone Python script implementation for training and testing. A Python script for the on-line prediction in OpenNI recordings (.oni) is also provided. Evaluation of the networks has been made with different metrics implementations (precision, recall, F1 Score, Sørensen-Dice coefficient), included in the networks scripts.File | Dimensione | Formato | |
---|---|---|---|
CNN_Implementation_for_Semantic_Heads_Segmentation_using_Top_View_Depth_Data_in_Crowded_Environment.pdf
solo utenti autorizzati
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati
Dimensione
689.44 kB
Formato
Adobe PDF
|
689.44 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
CNN_Implementation_for_Semantic_Heads_Segmentation_using_Top_View_Depth_Data_in_Crowded_Environment.pdf
solo utenti autorizzati
Tipologia:
Versione editoriale (versione pubblicata con il layout dell'editore)
Licenza:
Tutti i diritti riservati
Dimensione
689.44 kB
Formato
Adobe PDF
|
689.44 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.