Last week I visited Grinnell College talking about the work that my research groups has been doing on mobile sensing applications. I had a wonderful time. The Computer Science Department is top-notch. Lots of interesting discussions about ongoing work at Grinnell in systems and programming languages. Below is the abstract of my talk and my slides:
Mobile sensing applications are an emerging class of mobile applications that take advantage of the increasing sensing, computational, storage, and networking capabilities of mobile devices. Chipara’s research focuses on the systems, networking, and software engineering aspects of developing mobile health (mHealth) systems that continuously monitor and infer the health status of patients. His work combines the design of communication protocols, middleware, and programming tools with large-scale real-world deployments of working systems.
In this talk, Chipara will describe AudioSense – a novel mobile sensing application that allows audiologists to assess the performance of the hearing aids in the real-world. A key limitation of traditional laboratory and survey methods employed by audiologists is that they fail to predict when a hearing aid user will be dissatisfied with its performance in the real-world. In contrast with these techniques, AudioSense jointly characterizes both the user’s auditory context and the performance of the hearing aid in that context. The second part of the talk will cover some of the tools his team has created to simplify the development of mobile sensing applications. The focus is one of coordinating when different hardware resources (e.g., WiFi, 3G) are turned on and off to save energy without hindering user experience. A lightweight annotation language and middleware service will be presented that can be used to build energy-efficient mobile sensing applications for Android.