21 |
Real time Myanmar Sign Language Recognition System using PCA and SVM ...
|
|
|
|
BASE
|
|
Show details
|
|
22 |
Real time Myanmar Sign Language Recognition System using PCA and SVM ...
|
|
|
|
BASE
|
|
Show details
|
|
23 |
Data from: Koe: Web-based software to classify acoustic units and analyse sequence structure in animal vocalisations ...
|
|
|
|
BASE
|
|
Show details
|
|
24 |
Automatic Identification and Extraction of Yellow-Ipe Individuals in Images Obtained by UAV ; Identificação e Extração Automática de Indivíduos de Ipê-Amarelo em Imagens Obtidas por VANT
|
|
|
|
In: Revista Brasileira de Cartografia; v. 71 n. 4 (2019): Outubro/Dezembro; 1114-1134 ; Revista Brasileira de Cartografia; Vol 71 No 4 (2019): Outubro/Dezembro; 1114-1134 ; 1808-0936 ; 0560-4613 (2019)
|
|
BASE
|
|
Show details
|
|
25 |
Computer vision-based tracking and feature extraction for lingual ultrasound
|
|
|
|
Abstract:
Lingual ultrasound is emerging as an important tool for providing visual feedback to second language learners. In this study, ultrasound videos were recorded in sagittal plane as it provides an image for the full tongue surface in one scan, unlike the transverse plane which provides an information for small portion of the tongue in a single scan. The data were collected from five Arabic speakers as they pronounced fourteen Arabic sounds in three different vowel contexts. The sounds were repeated three times to form 630 ultrasound videos. The thesis algorithm was characterized by four steps. First: denoising the ultrasound image by using the combined curvelet transform and shock filter. Second: automatic selection of the tongue contour area. Third: tongue contour approximation and missing data estimation. Fourth: tongue contour transformation from image space to full concatenated signal and features extraction. The automatic tongue tracking results were validated by measuring the mean sum of distances between automatic and manual tongue contour tracking to give an accuracy of 0.9558mm. The validation for the feature extraction showed that the average mean squared error between the extracted tongue signature for different sound repetitions was 0.000858mm, which means that the algorithm could extract a unique signature for each sound and across different vowel contexts with a high degree of similarity. Unlike other related works, the algorithm showed an efficient and robust approach that could extract the tongue contour and the significant feature for the dynamic tongue movement on the full video frames, not just on the significant single and static video frame as used in the conventional method. The algorithm did not need any training data and had no limitation for the video size or the frame number. The algorithm did not fail during tongue extraction and did not need any manual re-initialization. Even when the ultrasound image recordings missed some tongue contour information, the thesis approach could estimate the missing data with a high degree of accuracy. The usefulness of the thesis approach as it can help the linguistic researchers to replace the manual tongue tracking by an automated tracking to save the time, then extracts the dynamics features for the full speech behavior to give better understanding of the tongue movement during the speech to develop a language learning tool for the second language learners. ; Graduate
|
|
Keyword:
computer vision; feature extraction; lingual ultrasound; tongue; tracking
|
|
URL: http://hdl.handle.net/1828/10812
|
|
BASE
|
|
Hide details
|
|
26 |
Extracting features from text to improve statistical machine translation
|
|
|
|
In: Journal of Applied Linguistics and Lexicography, Vol 1, Iss 1 (2019) (2019)
|
|
BASE
|
|
Show details
|
|
27 |
Comparing of Term Clustering Frameworks for Modular Ontology Learning
|
|
|
|
In: 10th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management ; https://hal.archives-ouvertes.fr/hal-01982912 ; 10th International Joint Conference on Knowledge Discovery, Knowledge Engineering and Knowledge Management, Sep 2018, Seville, Spain. pp.128-135, ⟨10.5220/0006960401280135⟩ (2018)
|
|
BASE
|
|
Show details
|
|
28 |
Sharing Learned Models Between Heterogeneous Robots: An Image-Driven Interpretation ...
|
|
|
|
BASE
|
|
Show details
|
|
29 |
Classification of Overt and Covert Speech for Near-Infrared Spectroscopy-Based Brain Computer Interface
|
|
|
|
In: Sensors ; Volume 18 ; Issue 9 (2018)
|
|
BASE
|
|
Show details
|
|
30 |
A Review Paper On Recognition Of Human Emotions Using Speech Signals ...
|
|
|
|
BASE
|
|
Show details
|
|
31 |
A Review Paper On Recognition Of Human Emotions Using Speech Signals ...
|
|
|
|
BASE
|
|
Show details
|
|
32 |
Review of constraints on vision-based gesture recognition for human–computer interaction
|
|
|
|
In: Author (2018)
|
|
BASE
|
|
Show details
|
|
33 |
Lexical rIchness Feature Extraction method (LIFE) for Multilingual and Cross-lingual Authorship Attribution
|
|
|
|
In: Doctoral (2018)
|
|
BASE
|
|
Show details
|
|
34 |
A Hierarchical Feature Extraction Model for Multi-Label Mechanical Patent Classification
|
|
|
|
In: Faculty Publications (2018)
|
|
BASE
|
|
Show details
|
|
35 |
Deep learning approaches to feature extraction, modelling and compensation for short duration language identification
|
|
|
|
BASE
|
|
Show details
|
|
36 |
Distant supervision for neural relation extraction integrated with word attention and property features
|
|
|
|
BASE
|
|
Show details
|
|
37 |
Recognition Of Pre-Segmented Characters In Printed Bilingual Gujarati-English Text ...
|
|
|
|
BASE
|
|
Show details
|
|
38 |
Recognition Of Pre-Segmented Characters In Printed Bilingual Gujarati-English Text ...
|
|
|
|
BASE
|
|
Show details
|
|
39 |
Using linguistic and topic analysis to classify sub-groups of online depression communities
|
|
|
|
BASE
|
|
Show details
|
|
40 |
Learning Spatial-Semantic Context with Fully Convolutional Recurrent Network for Online Handwritten Chinese Text Recognition
|
|
|
|
BASE
|
|
Show details
|
|
|
|