Page Not Found
Page not found. Your pixels are in another canvas.
A list of all the posts and pages found on the site. For you robots out there is an XML version available for digesting as well.
Page not found. Your pixels are in another canvas.
This is a page not in th emain menu
Published:
This post will show up by default. To disable scheduling of future posts, edit config.yml
and set future: false
.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
Published:
This is a sample blog post. Lorem ipsum I can’t remember the rest of lorem ipsum and don’t have an internet connection right now. Testing testing testing this blog post. Blog posts are cool.
,
It has been shown that emotions are learned in a cultural way and expressions are often used to help convey these emotional states. Considering this, in this work, we investigate multimodal cultural behavior differences across 6 different cultures. More specifically, we investigate head pose, action units, and facial landmarks in British, Chinese, German, Greek, Hungarian, and Serbian cultures. Along with this, we also investigate the differences along valence and arousal dimensions for these cultures. To conduct this investigation, we evaluate the SEWA multimodal and multi-cultural dataset. We find varying differences exist that are impacted by culture, context, and modality. Based on these findings, we also perform context classification that takes into account these differences in culture. We show that incorporating culture into our pipeline improves classification performance.
,
Expression recognition is an important and growing field in AI. It has applications in fields including, but not limited to, medicine, security, and entertainment. A large portion of research, in this area, has focused on recognizing expressions of young and middle-age adults with less focus on children and elderly subjects. This focus can lead to unintentional bias across age, resulting in less accurate models. Considering this, we investigate the impact of age on expression recognition. To facilitate this investigation, we evaluate two state-of-the-art datasets, that focus on different age ranges (children and elderly), namely EmoReact and ElderReact. We propose a Siamese-network based approach that learns the semantic similarity of expressions relative to each age. We show that the proposed approach, to expression recognition, is able to generalize across age. We show the proposed approach is comparable to or outperforms current state-of-the-art on the EmoReact and ElderReact datasets.
,
People with autism spectrum disorder (ASD) display impairments in social interaction and communication skills, as well as restricted interests and repetitive behaviors, which greatly affect daily life functioning. Current identification of ASD involves a lengthy process that requires an experienced clinician to assess multiple domains of functioning. Considering this, we propose a method for classifying multiple levels of risk of ASD using eye gaze and demographic feature descriptors such as a subject’s age and gennder. We construct feature descriptors that incorporate the subject’s age and gender, as well as features based on eye gaze patterns. We also present an analysis of eye gaze patterns validating the use of the selected hand-crafted features. We assess the efficacy of our descriptors to classify ASD on a National Database for Autism Research dataset, using multiple classifiers including a random forest, C4.5 decision tree, PART, and a deep feedforward neural network.
,
Facial expression Recognition is a growing and important field that has applications in fields such as medicine, security, education, and entertainment. While there have been encouraging approaches that have shown accurate results on a wide variety of datasets, in many cases it is still a difficult problem to explain the results. Considering this, in this paper, we propose to model flow-based latent representations of facial expressions, which allows us to further analyze the features and grants us more granular control over which features are produced for recognition. Our work is focused on posed facial expressions with a tractable density of the latent space. We investigate the behaviour of these tractable latent space features in the case of subject dependent and independent expression recognition. We employ a flow-based generative approach with minimal supervision introduced during training and observe that traditional metrics give encouraging results. When subject independent expressions are evaluated, a shift towards a stochastic nature, in the probability space, is observed. We evaluate our flow-based representation on the BU-EEG dataset showing our approach provides good separation of classes, resulting in more explainable results.
,
The quantification of visual affect data (e.g. face images) is essential to build and monitor automated affect modeling systems efficiently. Considering this, this work proposes quantified facial Temporal-expressiveness Dynamics (TED) to quantify the expressiveness of human faces. The proposed algorithm leverages multimodal facial features by incorporating static and dynamic information to enable accurate measurements of facial expressiveness. We show that TED can be used for high-level tasks such as summarization of unstructured visual data, and expectation from and interpretation of automated affect recognition models. To evaluate the positive impact of using TED, a case study was conducted on spontaneous pain using the UNBC-McMaster spontaneous shoulder pain dataset. Experimental results show the efficacy of using TED for quantified affect analysis.
Video of children performing RITA-T tasks for Autism Spectrum Disorder Screening
, University of South Florida
Research into age-aware continuous authentication. This work was supported by the NSF.
, University of South Florida
Subject identifcation using 3D facial landmarks.
, University of South Florida
Video and physiologial signals for continuous authentication
Datasets, University of South Florida
Videos of children with risk for ASD
Datasets, University of South Florida
Videos and physiological signals.
, University of South Florida
Gesture Recognition for sign language recognition.
News, Affective Computing and Intelligent Interaction, 2024
Multimodal Behavior Analysis and Impact of Culture on Affect.
News, European Conference on Computer Vision
FlowCon: Out-of-Distribution Detection using Flow-Based Contrastive Learning paper accepted to ECCV 2024.
News, University of South Florida
Dr. Taufeeq Uddin graduated with his PhD in Spring 2024. He is now conducting research at Meta.
News, Face and Gesture Rcognition, 2025
I am a general co-chair of FG 2025
Research, University of South Florida
Research into expression and emotion recognition.
Research, University of South Florida
Research into subject identification and age-aware continuous user authentication.
Research, University of South Florida
Research into HCI for automatic sign language recognition.
Published:
This is a description of your talk, which is a markdown files that can be all markdown-ified like any other post. Yay markdown!
PhD Student, University of South Florida
PhD student investigating latent space representations of emotion.
PhD Student, University of South Florida
PhD student investigating ASD screening in children.
PhD Student, University of South Florida
PhD student investigating bias in expression recognition.
Undergraduate Student, University of South Florida
Undergraduate student investigating human behavior during stressful situations.
PhD Student, University of South Florida
PhD student investigating impact of culture on Affect
Graduate course, University of South Florida, Fall, 2024
Analysis techniques for algorithms. Characterizing algorithms in terms of recurrence relations, solutions of recurrence relations, upper and lower bounds. Graph problems, parallel, algorithms, NP completeness and approximation algorithms, with relationship to practical problems.
Undergraduate and graduate course, University of South Florida, Spring, 2025
The study of systems that can recognize, interpret, process, and simulate human affect. Topics may include, but are not limited to, physiology of emotion, lie detection, wearable devices, music, and ethical concerns associated with affective computing.