1. Agentic AI System for No-Code Network Study in Healthcare (Ongoing)

    Network studies allow clinical researchers to run the same data analysis across multiple institutions or databases using a common data model, ensuring consistency and reproducibility. Traditionally, setting up such studies requires writing complex R code, which can be a barrier for clinical researchers with limited programming experience.

    Key Contributions of this project:

    • A user-friendly system that uses multiple AI agents to enable researchers to design network studies using natural language
    • Validation of final study specification for integrity and completeness

    Key Technologies: CrewAI framework, Python, Model Context Protocol (MCP), Strategus (OHDSI)

  2. Video Analytics System for Natural Language-Driven Surveillance

    This project designs a modular, AI-powered video analytics system capable of responding to natural language queries. It combines object detection (YOLO, GroundingDINO), tracking (DeepSORT), action recognition (SlowFast, TimeSformer), and fine-tuned VideoLLMs. A custom query agent translates user queries into subtasks and invokes models for real-time insights. The system targets industrial scenes (e.g., ports), addressing challenges like occlusion, poor lighting, and ambiguous motion patterns.

  3. SmrtFridge: IoT-based, User Interaction-driven Food Item & Quantity Sensing

    SmrtFridge is a smart fridge prototype that uses thermal and RGB sensing, combined with deep learning, to identify food items and estimate remaining quantity—without the need for RFID tags or user labeling. It leverages natural user interactions (e.g., door openings) to trigger video capture and processing. The system, triggered by natural interactions, achieves ~85% accuracy in classification and ~75% accuracy in quantity estimation. Its low-intrusion design makes it promising for real-world adoption in smart kitchens.

  4. CollabCam: Deep Learning based System for Energy-Efficient Pervasive Vision

    CollabCam is an edge-based vision system enabling collaborative inference in multi‑camera deployments. Each camera sends mixed-resolution frames, downsampled in shared FoVs, and collaboratively uses peer camera bounding boxes to boost detection. It cuts down network and energy usage by 25‑35%, while keeping accuracy losses under 2‑5% when compared to non‑collaborative approaches.