Home  /  Projects  /  Hobbies

Gyanig Kumar

I'm a Master's student in Computer Science at the University of Colorado Boulder, focusing on Human-Robot Interactions and Computer Vision research. Currently, I am advised by Prof. Bradley Hayes and Prof. Alessandro Roncone, working on in-context learning, VLM frameworks and multimodal interaction for human intent recognition.

I am developing a Vision-Language Framework that aligns human gaze (explicit cues) to enable adaptive and efficient robotic task planning in shared autonomy settings. In the past year, I worked on Active Preference Learning (APreL) for improving rewards using human-in-the-loop scenarios. I worked with various robotic platforms including 7-DoF manipulator robots (Sawyer) and Quadrupedal Robot (UnitreeGo1).

I started my research journey as a Research Assistant at the Indian Institute of Science (IISc) Bangalore under supervision of Prof. Pradipta Biswas. I developed gaze-tracking systems for automotive heads-up displays and applied Inverse Reinforcement Learning (IRL) to improve user intent prediction in robotic tasks such as pick-and-place and human-robot collaboration.

Over the years, I have developed proficiency in Gaze Estimation, Object Detection & Tracking, Collaborative Robotics, Self-Supervised Learning, XR-device development, and Reinforcement Learning. I've published papers at top-tier conferences including ACM THRI, ACM IUI and IEEE ICRA.

Email  /  CV  /  Scholar  /  Github  /  LinkedIn

I am applying for a full-time PhD position in the field of Human-Robot Interactions. Please reach out to me if our research interests align.

profile photo

Research Journey

A visual timeline of my academic and research experiences

Research

I'm interested in perception problems in robotics, especially in human-robot collaboration. Most of my research focuses on enhancing deep learning models, Modelling human intent recognition, developing multimodal frameworks. My contributions spans from automotive applications to mixed reality systems to robot planning. Some published papers are highlighted and the some are currently under review.

2025

Investigating Inverse Reinforcement Learning during Rapid Aiming Movement in Extended Reality and Human-Robot Interaction
ACM Transactions on Human-Robot Interaction (THRI), 2025
Mukund Mitra, Gyanig Kumar, P.P. Chakraborty, Pradipta Biswas
Investigation of inverse reinforcement learning techniques for rapid aiming movements in extended reality and human-robot interaction scenarios. This work explores how IRL can improve system understanding of human intent during rapid task execution in both virtual and collaborative robotic settings.
Comparing computer vision models for low resource dataset to develop a mixed reality based manual assembly assistant
Discover Robotics, 2025
Subin Raj, Bikram Karmakar, Gyanig Kumar, A. Mukhopadhyay, Rohit Chandrahas, Pradipta Biswas
Comparative study of computer vision models for generating realistic synthetic datasets to improve object detection in low-resource domains. This research evaluates different approaches for developing mixed reality-based manual assembly assistance systems using limited training data.

2024

Multimodal Target Prediction for Rapid Human-Robot Interaction
29th ACM Conference on Intelligent User Interfaces (ACM IUI), 2024
Mukund Mitra, A.A. Patil, G. Mothish, Gyanig Kumar, A. Mukhopadhyay, L.R.D. Murthy, P.P. Chakraborty, Pradipta Biswas
A multimodal approach for predicting human intent in robotic pick-and-place tasks using gaze tracking and deep learning. This work combines explicit gaze cues with implicit behavioral signals to improve prediction accuracy and enable more intuitive human-robot collaboration in rapid execution scenarios.
Enhanced Human-Robot Collaboration with Intent Prediction using Deep-IRL
IEEE International Conference on Robotics and Automation (ICRA), 2024
Mukund Mitra, Gyanig Kumar, P.P. Chakraborty, Pradipta Biswas
Leveraging Deep Inverse Reinforcement Learning for improved human-robot collaboration through intent prediction. This work demonstrates how learning from human demonstrations using Deep-IRL can enhance robot understanding of human preferences and enable more effective shared autonomy in collaborative manipulation tasks.
A Comparative Study on Image Translation GAN Models to Improve Object Detection in Low-Resource Domains
2024 International Conference on Vehicular Technology and Transportation Systems (ICVTTS), 2024
Yash Kumar Sahu, A. Mukhopadhyay, Gyanig Kumar, Ashok Kumar, Pradipta Biswas
Comprehensive study comparing GAN models for generating realistic synthetic datasets to improve object detection in low-resource domains. This research evaluates different image translation approaches for domain adaptation in scenarios where training data is limited or expensive to acquire.
Augmented reality and deep learning based system for assisting assembly process
Journal on Multimodal User Interfaces (JMUI), 2024
Subin Raj, L.R.D. Murthy, T.A. Shanmugam, Gyanig Kumar, A. Chakrabarti, Pradipta Biswas
Augmented reality-based assembly system using custom-trained object detection models with tailored eye gaze and hand tracking. This work demonstrates how combining AR visualization, deep learning for part recognition, and multimodal interaction can improve assembly process efficiency and reduce errors.

2022

Efficient Interaction with Automotive Heads-Up Displays using Appearance-based Gaze Tracking
Work-in-Progress Track of the 14th International Conference on Automotive User Interfaces and Interactive Vehicular Applications (AutomotiveUI), 2022
L.R.D. Murthy, Gyanig Kumar, Pradipta Biswas, M. Madan, S. Deshmukh
Appearance-based gaze tracking system for interactive automotive heads-up displays with gaze-based vs gesture-based on-road distraction detection. This work evaluates how gaze-based interaction can enable safer and more intuitive control of HUD elements while minimizing driver distraction during in-vehicle tasks.

Presentations

Gaze Estimation Presentation
End-to-end Frame to Gaze Estimation
Research Paper presentation Comprehensive Literature review and "End-to-end Frame-to-gaze" CVPR , 7000-level Recent Advancements in Computer Vision
Paper presentation.
POMDP Presentation
Partially Observed Markov Decision Processes for Robotics
Research Paper project for "Probability for Computer Science" class, led by Prof. Osita Onyejekwe | Dept. of Computer Science, University of Colorado Boulder
Presentation on POMDP frameworks and their applications in robotic decision-making.
DMU Paper
Reinforcement Learning for Mini Tetris
Research Paper project for "Decision Making Under uncertainty" class, led by Prof. Zachery Sunberg | Dept. of Aerospace, University of Colorado Boulder
Research paper on decision-making strategies and algorithms for handling uncertainty in autonomous systems.
UNESCO Report
UNESCO Report on AI and Education
UNESCO Publication
UNESCO report examining the role of artificial intelligence in education and learning systems. Represented my former Lab, demonstrating research projects.
BEL Industry Class Presentation
BEL Industry Class Presentation
Industry outreach / applied robotics & AI
Session covering practical adoption of AI-driven perception & planning pipelines in industrial settings; emphasized safety, latency constraints, and integration with legacy systems.
MRCV Class Presentation
MRCV Class – Multimodal Recent Advances in Computer Vision
Graduate seminar presentation
Explored multimodal fusion (vision-language-gaze), promptable VLM architectures, and evaluation protocols for embodied perception benchmarks.

Miscellanea

Projects

Ansh.AI: Conversational Diary Companion - 2nd Position in AI Track at HackCU11
Autonomous Driving Challenge with AWS DeepRacer - ROS 2 based autonomous vehicle
Lox: Interpreter for Python-like Language - End-to-end interpreter implementation
V-BIKES: Enhanced Safety Measures for Two-Wheeler Vehicles - IoT collision detection system

Teaching

Course Manager - CSCI 3302 Introduction to Robotics, University of Colorado Boulder (Spring 2025)
Teaching Assistant - CS3155 Principles of Programming Languages, University of Colorado Boulder (Summer 2025)
Part-Time CS Teacher - Polytechnic College Suriname (2022 - 2023)

Awards

Student Volunteer at ACC Denver Conference (2025)
2nd Position in AI Track at HackCU11, AWS Hackathon (2025)
Tata Millennium Scholarship - 800$ (2025)
Summer School Fellowship at CVIT, IIIT Hyderabad (2021)

Hobbies

Music Composition - Original compositions including "Urban Nest" (2024), exploring contemporary soundscapes
Photography - Landscape and nature photography, capturing moments from hiking trails and mountain vistas
Hiking & Outdoor Adventures - Exploring Colorado's trails including Anemone Trail and Rocky Mountain National Park
Travel Photography - Documenting visits to iconic locations like Mount Rushmore and various national parks