THESIS
2021
1 online resource (87 pages) : illustrations (some color)
Abstract
With the advancement in Computer Vision, many new applications of mobile sensing platforms have emerged that can satisfy the unique needs of specific population groups and enhance their daily lives. However, there are still some unsolved challenges. This thesis focuses on two population groups and presents our explorations in resolving some of their crucial needs.
For the deaf-mute, there are significant barriers to communicate with the hearing population. However, given the complexity of data collection remains a challenge in realizing a realistic smartwatch-based sign language recognition system. We describe our exploration in transferring online sign language videos into training data for smartwatch-based sign language recognition. We tested pipelines of different human pose/mesh 3D...[
Read more ]
With the advancement in Computer Vision, many new applications of mobile sensing platforms have emerged that can satisfy the unique needs of specific population groups and enhance their daily lives. However, there are still some unsolved challenges. This thesis focuses on two population groups and presents our explorations in resolving some of their crucial needs.
For the deaf-mute, there are significant barriers to communicate with the hearing population. However, given the complexity of data collection remains a challenge in realizing a realistic smartwatch-based sign language recognition system. We describe our exploration in transferring online sign language videos into training data for smartwatch-based sign language recognition. We tested pipelines of different human pose/mesh 3D reconstruction methods from imagery and data processing techniques to eliminate the need to collect actual smartwatch-collected data.
Dysphagia is a common disease for many older people, making them subject to choking risk if the viscosity of liquid intake is not well controlled. We present ViscoCam, the first fluid viscosity classification system for dysphagia patients or caregivers, requiring only a smartphone. It is easy to operate, widely deployable, and robust for daily use. ViscoCam classifies visually indistinguishable liquid of various viscosity levels by exploiting the sloshing motion of viscous liquid decays faster than thin liquid. To perform a measurement, the user shakes a cup of liquid and their smartphone to induce the liquid sloshing motion. Then, ViscoCam senses the cup's movement using the smartphone's built-in accelerometer or microphone and infers liquid viscosity from the fluid surface motion captured by the flashlight camera. To combat changes in camera position, lighting conditions, and liquid sloshing motion, a 3D convolutional neural network is trained to extract reliable motion features for classification. We present the evaluation of ViscoCam's performance in classifying three levels against the IDDSI standard, which is the most up-to-date and internationally adopted one for dysphagia patients.
Post a Comment