Full metadata record
DC poleHodnotaJazyk
dc.contributor.authorAl-Akam, Rawya
dc.contributor.authorPaulus, Dietrich
dc.contributor.authorGharabaghi, Darius
dc.contributor.editorSkala, Václav
dc.date.accessioned2019-05-10T10:15:24Z-
dc.date.available2019-05-10T10:15:24Z-
dc.date.issued2018
dc.identifier.citationWSCG 2018: poster papers proceedings: 26th International Conference in Central Europe on Computer Graphics, Visualization and Computer Visionin co-operation with EUROGRAPHICS Association, p. 18-26.en
dc.identifier.isbn978-80-86943-42-8
dc.identifier.issn2464-4617
dc.identifier.uriwscg.zcu.cz/WSCG2018/!!_CSRN-2803.pdf
dc.identifier.urihttp://hdl.handle.net/11025/34633
dc.description.abstractHuman action recognition with color and depth sensors has received increasing attention in image processing and computer vision. This paper target is to develop a novel deep model for recognizing human action from the fusion of RGB-D videos based on a Convolutional Neural Network. This work is proposed a novel 3D Convolutional Neural Network architecture that implicitly captures motion information between adjacent frames, which are represented in two main steps: As a First, the optical flow is used to extract motion information from spatio-temporal domains of the different RGB-D video actions. This information is used to compute the features vector values from deep 3D CNN model. Secondly, train and evaluate a 3D CNN from three channels of the input video sequences (i.e. RGB, depth and combining information from both channels (RGB-D)) to obtain a feature representation for a 3D CNN model. For evaluating the accuracy results, a Convolutional Neural Network based on different data channels are trained and additionally the possibilities of feature extraction from 3D Convolutional Neural Network and the features are examined by support vector machine to improve and recognize human actions. From this methods, we demonstrate that the test results from RGB-D channels better than the results from each channel trained separately by baseline Convolutional Neural Network and outperform the state of the art on the same public datasets.en
dc.format9 s.cs
dc.format.mimetypeapplication/pdf
dc.language.isoenen
dc.publisherVáclav Skala - UNION Agencyen
dc.relation.ispartofseriesWSCG 2018: poster papers proceedingsen
dc.rights© Václav Skala - Union Agencycs
dc.subjectrozpoznání akcecs
dc.subjectRGBD videacs
dc.subjectoptický tokcs
dc.subject3D konvoluční neuronová síťcs
dc.subjectpodpora vektorového strojecs
dc.titleHuman action recognition based on 3D convolution neural networks from RGBD videosen
dc.typekonferenční příspěvekcs
dc.typeconferenceObjecten
dc.rights.accessopenAccessen
dc.type.versionpublishedVersionen
dc.subject.translatedaction recognitionen
dc.subject.translatedRGBD videosen
dc.subject.translatedoptical flowen
dc.subject.translated3D convolutional neural networken
dc.subject.translatedsupport vector machinesen
dc.identifier.doihttps://doi.org/10.24132/CSRN.2018.2803.3
dc.type.statusPeer-revieweden
Vyskytuje se v kolekcích:WSCG 2018: Poster Papers Proceedings

Soubory připojené k záznamu:
Soubor Popis VelikostFormát 
Al-Akam.pdfPlný text1,3 MBAdobe PDFZobrazit/otevřít


Použijte tento identifikátor k citaci nebo jako odkaz na tento záznam: http://hdl.handle.net/11025/34633

Všechny záznamy v DSpace jsou chráněny autorskými právy, všechna práva vyhrazena.