People with deafness or hearing disabilities who aim to use computer based systems rely on state-of-art video classification and human action recognition techniques that combine traditional movement pat-tern recognition and deep learning techniques. In this work we present a pipeline for semi-automatic video annotation applied to a non-annotated Peru-vian Signs Language (PSL) corpus along with a novel method for a progressive detection of PSL elements (nSDm). We produced a set of video annotations in-dicating signs appearances for a small set of nouns and numbers along with a labeled PSL dataset (PSL dataset). A model obtained after ensemble a 2D CNN trained with movement patterns extracted from the PSL dataset using Lucas Kanade Opticalflow, and a RNN with LSTM cells trained with raw RGB frames extracted from the PSL dataset reporting state-of-art results over the PSL dataset on signs classification tasks in terms of AUC, Precision and Recall. / Trabajo de investigación
Identifer | oai:union.ndltd.org:PUCP/oai:tesis.pucp.edu.pe:20.500.12404/16906 |
Date | 31 August 2020 |
Creators | Huiza Pereyra, Eric Raphael |
Contributors | Olivares Poggi, Cesar Augusto |
Publisher | Pontificia Universidad Católica del Perú, PE |
Source Sets | Pontificia Universidad Católica del Perú |
Language | English |
Detected Language | English |
Type | info:eu-repo/semantics/masterThesis |
Format | application/pdf |
Rights | Atribución 2.5 Perú, info:eu-repo/semantics/openAccess, http://creativecommons.org/licenses/by/2.5/pe/ |
Page generated in 0.0187 seconds