Topics for thesis. Automatic Speech-based Emotion Recognition

Similar documents
FACE RECOGNITION USING INDEPENDENT COMPONENT

Gender Classification

Facial Expression Detection Using Implemented (PCA) Algorithm

Eye Detection by Haar wavelets and cascaded Support Vector Machine

Facial Expression Recognition Using Non-negative Matrix Factorization

Speech Driven Synthesis of Talking Head Sequences

Facial Emotion Recognition using Eye

Automatic Countenance Recognition Using DCT-PCA Technique of Facial Patches with High Detection Rate

Categorization by Learning and Combining Object Parts

Multi-Modal Human- Computer Interaction

International Journal of Emerging Technologies in Computational and Applied Sciences (IJETCAS) Human Face Detection By YCbCr Technique

Recognition of facial expressions in presence of partial occlusion

Generic Face Alignment Using an Improved Active Shape Model

LEARNING TO GENERATE CHAIRS WITH CONVOLUTIONAL NEURAL NETWORKS

Color Local Texture Features Based Face Recognition

Getting Started with Crazy Talk 6

Spoofing Face Recognition Using Neural Network with 3D Mask

Pupil Localization Algorithm based on Hough Transform and Harris Corner Detection

Webpage: Volume 3, Issue VII, July 2015 ISSN

Facial expression recognition using shape and texture information

Latest development in image feature representation and extraction

Presented at the FIG Congress 2018, May 6-11, 2018 in Istanbul, Turkey

A Hybrid Face Detection System using combination of Appearance-based and Feature-based methods

Selection of Location, Frequency and Orientation Parameters of 2D Gabor Wavelets for Face Recognition

LOCAL APPEARANCE BASED FACE RECOGNITION USING DISCRETE COSINE TRANSFORM

Subject-Oriented Image Classification based on Face Detection and Recognition

CHAPTER 4 FACE RECOGNITION DESIGN AND ANALYSIS

TEXT SEGMENTATION ON PHOTOREALISTIC IMAGES

FACE ANALYSIS AND SYNTHESIS FOR INTERACTIVE ENTERTAINMENT

Car tracking in tunnels

Color-based Face Detection using Combination of Modified Local Binary Patterns and embedded Hidden Markov Models

Computer Vision with MATLAB MATLAB Expo 2012 Steve Kuznicki

Recognizing Deformable Shapes. Salvador Ruiz Correa Ph.D. UW EE

Eye and Mouth Openness Estimation in Sign Language and News Broadcast Videos

Defect Detection of Regular Patterned Fabric by Spectral Estimation Technique and Rough Set Classifier

Data-Driven Face Modeling and Animation

22 October, 2012 MVA ENS Cachan. Lecture 5: Introduction to generative models Iasonas Kokkinos

PROBLEM FORMULATION AND RESEARCH METHODOLOGY

Dynamic Human Shape Description and Characterization

Facial Keypoint Detection

REAL-TIME FACE SWAPPING IN VIDEO SEQUENCES: MAGIC MIRROR

Medical images, segmentation and analysis

FACIAL MOVEMENT BASED PERSON AUTHENTICATION

Model-Based Face Computation

automatic digitization. In the context of ever increasing population worldwide and thereby

Component-based Face Recognition with 3D Morphable Models

Facial Feature Extraction Based On FPD and GLCM Algorithms

Shape Context Matching For Efficient OCR

Significant Jet Point for Facial Image Representation and Recognition

CrazyTalk 5 Basic Face Fitting and Techniques

Biometrics Technology: Multi-modal (Part 2)

The Analysis of Faces in Brains and Machines

3D Attention-Driven Depth Acquisition for Object Identification

Tracking facial features using low resolution and low fps cameras under variable light conditions

Real time facial expression recognition from image sequences using Support Vector Machines

Recognizing Micro-Expressions & Spontaneous Expressions

A THREE LAYERED MODEL TO PERFORM CHARACTER RECOGNITION FOR NOISY IMAGES

Recognizing Deformable Shapes. Salvador Ruiz Correa UW Ph.D. Graduate Researcher at Children s Hospital

SYSTEM AND METHOD FOR SPEECH RECOGNITION

Tri-modal Human Body Segmentation

Adaptive Skin Color Classifier for Face Outline Models

AAM Based Facial Feature Tracking with Kinect

Measurement of pinna flare angle and its effect on individualized head-related transfer functions

World Journal of Engineering Research and Technology WJERT

Face analysis : identity vs. expressions

DTU Technical Report: ARTTS

Multi-modal Translation and Evaluation of Lip-synchronization using Noise Added Voice

3D Interaction Techniques for Virtual Environments: Selection and Manipulation. Doug A. Bowman

Countermeasure for the Protection of Face Recognition Systems Against Mask Attacks

coding of various parts showing different features, the possibility of rotation or of hiding covering parts of the object's surface to gain an insight

CS 231. Deformation simulation (and faces)

Visual Perception for Robots

Video Processing for Judicial Applications

Convolutional Neural Networks for Facial Expression Recognition

Automatic Detecting Neutral Face for Face Authentication and Facial Expression Analysis

Facial Expression Recognition Using Improved Artificial Bee Colony Algorithm

A Full-Range of 3D Body Scanning Solutions

Face Recognition Technology Based On Image Processing Chen Xin, Yajuan Li, Zhimin Tian

Recognizing Deformable Shapes. Salvador Ruiz Correa (CSE/EE576 Computer Vision I)

IMPROVED FACE RECOGNITION USING ICP TECHNIQUES INCAMERA SURVEILLANCE SYSTEMS. Kirthiga, M.E-Communication system, PREC, Thanjavur

EFFECTIVE METHODOLOGY FOR DETECTING AND PREVENTING FACE SPOOFING ATTACKS

Facial Expression Analysis

Supplemental Material: Detailed, accurate, human shape estimation from clothed 3D scan sequences

CS 231. Deformation simulation (and faces)

Human Face Classification using Genetic Algorithm

Research on Dynamic Facial Expressions Recognition

Data Transfer Using a Camera and a Three-Dimensional Code

Automatic human face detection in color images

DETC D FACE RECOGNITION UNDER ISOMETRIC EXPRESSION DEFORMATIONS

Face Detection Using Convolutional Neural Networks and Gabor Filters

Gesture Recognition Technique:A Review

Landmark Detection on 3D Face Scans by Facial Model Registration

A New Robust Face Detection Method Based on Corner Points

Fast and Efficient Automated Iris Segmentation by Region Growing

Hybrid Biometric Person Authentication Using Face and Voice Features

DEFORMABLE MATCHING OF HAND SHAPES FOR USER VERIFICATION. Ani1 K. Jain and Nicolae Duta

Glasses Detection for Face Recognition Using Bayes Rules

Global Gabor features for rotation invariant object classification

CS Human Computer Interaction

Multiple Kernel Learning for Emotion Recognition in the Wild

Transcription:

Topics for thesis Bachelor: Automatic Speech-based Emotion Recognition Emotion recognition is an important part of Human-Computer Interaction (HCI). It has various applications in industrial and commercial sectors. In the context of emotion classification, HCI has been divided into three major modalities: vocal-, facial-, and gesture-based recognition. This title focuses solely on audio based input. Humans use intrinsic skills to infer characteristics denoting the emotional patterns from the voice. However, automation of such a process might prove challenging. In order to identify the emotional state of a given utterance one has to derive relevant features, which carry the emotional information. Finding the most discriminative criterion describing emotional states is sought extensively. Proper selection of features is vital for vocal-based emotion recognition systems, as well as realistic and efficient interaction between human and computer. In the paralinguistic approach, many vocal-based emotion recognition systems use high-level spectral and prosodic features, since both of them contain key data about the emotions present in audio signal. The rest of the thesis is organized as follows. First, we will extract the paralinguistic features and select an efficient combination between them. Second, we will investigate category of classifiers such as Adaboost, Random Forest and Neural Networks approaches as our methodology of the proposed method is described. Then it is applied to the databases, for the sake of verifying its efficiency.

Bachelor: Speech-based Paralinguistic Features Investigation Emotion recognition is an important part of Human-Computer Interaction (HCI). It has various applications in industrial and commercial sectors. Humans use intrinsic skills to infer characteristics denoting the emotional patterns from the voice. However, automation of such a process might prove challenging. In order to identify the emotional state of a given utterance one has to derive relevant features, which carry the emotional information. Finding the most discriminative criterion describing emotional states is sought extensively. Proper selection of features is vital for vocal-based emotion recognition systems, as well as realistic and efficient interaction between human and computer. In the paralinguistic approach, many vocal-based emotion recognition systems use high-level spectral and prosodic features, since both of them contain key data about the emotions present in audio signal. The rest of the thesis is organized as follows. First, we will extract different types of paralinguistic features. Second, we will apply the classifiers such as Random Forest or Neural Networks approaches on the combinations of features. Then we select an efficient combination between them. This process can repeat on the different databases to find the robustness of selection.

Bachelor: Estimation of Human Body Measurements according to Image Taken by Mobile Phone Even the most advanced 3D scanners still cannot provide reliable, comprehensive measurement data. More precisely, for any reason, including illumination interferences, etc., the measurements offered by the existing 3D scanners, in the most cases, are neither complete nor accurate. As a result, there is a need to estimate the body measurements based on 2D static visual data obtained by mobile phones, which is also more convenient to implement compared to 3D laser scanners. There are a variety of ratios and measurement values of human body, which are almost the same in all instances. As the final goal of the project, we will exploit the latter list to estimate other measurements of the body based on the ones being already provided in the image taken by a mobile phone.

Master: Automatic Facial Emotion Recognition for Specific Human Races by using the CNN algorithm The majority of human communications, i.e. 93%, are performed through nonverbal means, which mainly involve facial expressions. Thus, detecting facial expressions and acting accordingly is necessary for Human-Computer Interaction (HCI) purposes. Computerized Facial Expression Recognition (FER), which is the main subject of this project, consists in face detection and tracking, feature extraction and emotion recognition. First, the face is detected and tracked throughout a chain of images constituting a video sequence. As facial expressions are dependent on the translation, scaling and rotation of the head, both motionbased and model-based representations are considered, which requires geometric normalization and segmentation. The next step is to extract information from the detected face, which can help distinguish the intended emotion. The main two categories of facial features are geometric and appearance ones. Geometric features consist of the shapes of specific parts of the face, such as eyes, eyebrows and mouth, and the locations of facial points, e.g. the corners of the eyes and mouth. Appearance features are based on the whole face or a certain region. Finally, as one of the most powerful tools in the context of affective computing, a convolutional neural network will be utilized for classification.

Master: Virtual Garment Fitting Most of the previously proposed methods have failed to accurately model the human body in real-time, especially, when it comes to 3D modeling of the whole body, requiring large amount of data and calculations. Besides, the foregoing techniques, mostly, do not render realistic movement representations in real-time. This project intends to overcome the aforementioned shortcomings. The proposed methodology consists in scanning and performing analysis on both the garment and the body, modelling and assigning reference points on them, and finally, superimposing and depicting the resulting garment model on the user's body.

Master: Data Acquisition Using Depth Segmentation and Classification This project concentrates on data acquisition by segmenting the acquired 3D data into separate parts, such as, trousers and shirts, based on previous information about the colors, to model a Gaussian Mixture Model of the body regions, which will be optimized in order to obtain segmented 3D models of the body and the garment. The key points should be defined in a way that can both be detected under various configurations and avoid self-occlusions, which will be fulfilled upon employing anatomical key points with local references through considering pair-wise constraints and solving them by means of a graph matching approach. The common landmarks will be roughly estimated, and according to the key-point information and depth and RGB data, the exact boundaries will be extracted.