Cross-Dataset Generalization for Human Behavior Modeling

Overview; Data; Code

Overview of The Contributions of This Work. We systematically evaluate cross-dataset generalizability of 19 algorithms: 9 prior behavior modeling algorithm for depression detection, 8 recent domain generalization algorithms, and 2 two new algorithms proposed in this paper. Our open-source platform GLOBEM consolidates these 19 algorithms and support using, developing, evaluating various algorithms.

There is a growing body of research revealing that longitudinal passive sensing data from smartphones and wearable devices can capture daily behavior signals for human behavior modeling, such as depression detection. Most prior studies build and evaluate machine learning models using data collected from a single population. However, to ensure that a behavior model can work for a larger group of users, its generalizability needs to be verified on multiple datasets from different populations. We present the first work evaluating cross-dataset generalizability of longitudinal behavior models, using depression detection as an application. We collect multiple longitudinal passive mobile sensing datasets with over 500 users from two institutes over a two-year span, leading to four institute-year datasets. Using the datasets, we closely re-implement and evaluated nine prior depression detection algorithms. Our experiment reveals the lack of model generalizability of these methods. We also implement eight recently popular domain generalization algorithms from the machine learning community. Our results indicate that these methods also do not generalize well on our datasets, with barely any advantage over the naive baseline of guessing the majority. We then present two new algorithms with better generalizability. Our new algorithm, Reorder, significantly and consistently outperforms existing methods on most cross-dataset generalization setups. However, the overall advantage is incremental and still has great room for improvement. Our analysis reveals that the individual differences (both within and between populations) may play the most important role in the cross-dataset generalization challenge. Finally, we provide an open-source benchmark platform GLOBEM – short for Generalization of LOngitudinal BEhavior Modeling – to consolidate all 19 algorithms. GLOBEM can support researchers in using, developing, and evaluating different longitudinal behavior modeling methods. We call for researchers’ attention to model generalizability evaluation for future longitudinal human behavior modeling studies.

Xuhai Xu, Xin Liu, Han Zhang, Weichen Wang, Subigya Nepal, Yasaman S. Sefidgar, Woosuk Seo, Kevin S. Kuehn, Jeremy F. Huckins, Margaret E. Morris, Paula S. Nurius, Eve A. Riskin, Shwetak N. Patel, Tim Althoff, Andrew Campbell, Anind K. Dey, and Jennifer Mankoff. GlOBEM: Cross-Dataset Generalization of Longitudinal Human Behavior Modeling. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 6(4): 190:1-190:34 (2022).

Xuhai XuHan ZhangYasaman S. SefidgarYiyi RenXin LiuWoosuk SeoJennifer BrownKevin S. KuehnMike A. MerrillPaula S. NuriusShwetak N. PatelTim AlthoffMargaret MorrisEve A. Riskin, Jennifer Mankoff, Anind K. Dey:
GLOBEM Dataset: Multi-Year Datasets for Longitudinal Human Behavior Modeling Generalization. NeurIPS 2022

Saidhruv Chittamuri

Dhruv is a freshman at the University of Washington studying Computer Science. In the past, he has began to build a coding expertise in full-stack development and data analytics. He hopes to delve more in the latter through strengthening his Machine Learning skills and utilizing them in meaningful real-world applications, with an emphasis on tackling accessibility. Outside of Computer Science, he maintains an active lifestyle – enjoying weight lifting, playing soccer, and running. Currently, he is working on the Academic Performance Prediction project here at the Make4All Lab.

Srihari Krishnaswamy

Srihari Krishnaswamy is a first-year undergraduate student studying Computer Science at the University of Washington. His programming experience in the past consists of work on VR simulations and cross-platform apps. He’s looking to learn more about Machine Learning, AI and Signal Processing in the near future, and looks forward to working on more projects. In his free time, he enjoys music production and playing sports. 

Emma Chen

An Asian woman smiling at the camera.

Hello! My name is Emma and I’m a third year computer science major at UW. My hobbies include crocheting and trying out new foods. I love talking to other people so feel free to reach out anytime!

Yusuf Shahpurwala

Yusuf Shahpurwala is a freshman studying computer science at the University of Washington. Most of his coding experience so far has been through making web applications. He is hoping to learn more about human-computer interaction, machine learning, and embedded systems over the next couple of years. To relax he enjoys playing sports and reading. He is excited to currently be working on the Accessible Knitting project with Make4all.

Zelin Yang

Zelin is a first-year Master’s student majoring in Mechanical Engineering with Data Science at the University of Washington. He completed his bachelor’s degree in Mechanical Engineering at Colorado State University where he also engaged in research in the field of manufacturing high-performance polymers by 3D printing and robotics.

His current research, led by Momona, focuses on modeling and enhancing biosignals-based human-machine interaction to support accessibility and health. He presently works on data collection and analysis of biosignals.

PSST: Enabling Blind or Visually Impaired Developers to Author Sonifications of Streaming Sensor Data

Venkatesh Potluri, John Thompson, James Devine, Bongshin Lee, Nora Morsi, Peli De Halleux, Steve Hodges, and Jennifer Mankoff. 2022. PSST: Enabling Blind or Visually Impaired Developers to Author Sonifications of Streaming Sensor Data. In Proceedings of the 35th Annual ACM Symposium on User Interface Software and Technology (UIST ’22). Association for Computing Machinery, New York, NY, USA, Article 46, 1–13. https://doi.org/10.1145/3526113.3545700

We present the first toolkit that equips blind and visually impaired (BVI) developers with the tools to create accessible data displays. Called PSST (Physical Computing Streaming Sensor data Toolkit), it enables BVI developers to understand the data generated by sensors from a mouse to a micro: bit physical computing platform. By assuming visual abilities, earlier efforts to make physical computing accessible fail to address the need for BVI developers to access sensor data. PSST enables BVI developers to understand real-time, real-world sensor data by providing control over what should be displayed, as well as when to display and how to display sensor data. PSST supports filtering based on raw or calculated values, highlighting, and transformation of data. Output formats include tonal sonification, nonspeech audio files, speech, and SVGs for laser cutting. We validate PSST through a series of demonstrations and a user study with BVI developers.

The demo video can be found here: https://youtu.be/UDIl9krawxg.

Evan Zhao

Evan is an undergraduate at University of Washington, majoring in Computer Science. He is passionate about computer graphics and the huge potential of combining graphical programming techniques with fabrication such as 3D printing, machine embroidery, and so on. In the meantime, he is also a member of the UW Reality Lab. He learned advanced knowledge on how to design interactive, efficient, and accessible applications that run in virtual reality, but he also wants to make them physically touchable and perceivable and bring those models to real life. Since started discovering the vast potential in computer fabrication, he has decided to become a part of the pioneers in this field and contribute to the goal of making designs for everybody.

Rapid Convergence: The Outcomes of Making PPE during a Healthcare Crisis

Kelly Avery MackMegan HofmannUdaya LakshmiJerry CaoNayha AuradkarRosa I. ArriagaScott E. HudsonJennifer Mankoff. Rapid Convergence: The Outcomes of Making PPE During a Healthcare Crisis. [Link to the paper]

The U.S. National Institute of Health (NIH) 3D Print Exchange is a public, open-source repository for 3D printable medical device designs with contributions from clinicians, expert-amateur makers, and people from industry and academia. In response to the COVID-19 pandemic, the NIH formed a collection to foster submissions of low-cost, locally-manufacturable personal protective equipment (PPE). We evaluated the 623 submissions in this collection to understand: what makers contributed, how they were made, who made them, and key characteristics of their designs. We found an immediate design convergence to manufacturing-focused remixes of a few initial designs affiliated with NIH partners and major for-profit groups. The NIH worked to review safe, effective designs but was overloaded by manufacturing-focused design adaptations. Our work contributes insights into: the outcomes of distributed, community-based medical making; the features that the community accepted as “safe” making; and how platforms can support regulated maker activities in high-risk domains.

Christina Zhang

Christina Zhang is a senior at University of Washington, majoring in Computer Science and Informatics, her research interests are mainly HCI, mHealth, behavioral health, accessibility and social computing.
Her current work involves supporting early identification of mental health issues in adolescents, and software-based solutions to accessible communication in higher education.

In the past, she has worked on a research project that studies how online tests could be leveraged to bridge the gap in the support system of people with cognitive and mental disabilities, the paper she co-authored won the Best Paper Award on ASSETS 2021.

https://www.linkedin.com/in/christina-zhang-03215824b/