Research
Overview
My research interests broadly lie in Human-centric Cyber-Physical Systems, Data Science, Generative AI, IoT, and Mobile and Ubiquitous Computing for smart cities. My work involves various techniques, such as large language models (LLMs), large sensor models (LSM), efficient machine learning, unsupervised machine learning, cross-domain data fusion, and large-scale system optimization & deployment. I have been investigating platforms across more than 300 cities on 2 continents with 2 billion E-commerce orders, 300 thousand vehicles, 500 thousand mobile devices, and 500 million residents involved.
Low-cost & Large-scale Mobile Sensing for Urban Upgrading
In recent years, rapid economic growth and human mobility have led to significant changes in urban environments, especially in developing countries. For example, the frequent establishment of new points of interest (POIs) and the construction of new buildings. Failing to capture these updates promptly brings negative impacts to smart cities, such as inaccurate map navigation and increased carbon emissions. By leveraging the unique capabilities of mobile devices (e.g., smartphones and smartwatches) carried by citizens, we conceptualize individuals as dynamic sensors for the city. I have designed an innovative framework for urban sensing, utilizing TB-level multi-modal data (e.g., GPS and WiFi). With this framework, I have identified the locations of shipping addresses and POIs across 366 cities. The findings have been published in TKDE’24, CIKM’23, CIKM’23, and SIGSPATIAL’22.
Generative AI for Smart Cities
Generative AI has transformed many research and application directions. In smart cities, Generative AI has the potential to address some long-standing challenges, such as optimizing resource allocation and enhancing urban planning. One of my works specifically focuses on developing generative AI models to generate urban infrastructure data. By harnessing the unique properties of spatio-temporal and sensory data, I have developed a model called DualSTGAN, which generates urban road maps from multi-modal sensor data. The results of this research have been published in Ubicomp’24. In another work, we designed an LLM framework for spatial language processing, and the results have been published in KDD’25. Recently, our work about generating spatio-temporal human mobility data with diffusion models has been accepted in SDM’25.
Foundation Model for Human Activity Sensing
Human activity sensing, or human behavior sensing, plays a crucial role in multiple applications, such as healthcare monitoring, fitness tracking, and smart home. The widespread adoption of mobile and wearable devices has opened up new opportunities for low-cost human activity sensing. However, two persistent challenges—domain shift and limited training data—have hindered the large-scale deployment of activity sensing models. My research focuses on developing foundational models for generalizable activity sensing by creating novel sensor augmentation techniques and sensor pretraining paradigms. The findings from this work have been published in Ubicomp’24.