Back to projects

Final Year Project

Horizon UI

Horizon UI is a multimodal desktop control system that allows users to control computer actions using hand gestures and voice commands. It uses real-time hand tracking and speech recognition to create a touchless interaction layer for desktop environments.

3rd Place at Industrial Open House 2026

Horizon UI project thumbnail

What I built

Horizon UI is a multimodal desktop control system that allows users to control computer actions using hand gestures and voice commands. It uses real-time hand tracking and speech recognition to create a touchless interaction layer for desktop environments.

Why I built it

Desktop interaction is still heavily dependent on physical input devices, which limits hands-free workflows and accessibility-focused interaction patterns.

My role

Built the interaction pipeline, gesture mapping, voice command layer, desktop automation flow, and demo narrative for the final year project.

How it works

  • Camera feed is processed through OpenCV and MediaPipe for hand landmarks.
  • Gesture logic maps tracked landmarks into desktop commands.
  • Speech recognition layer listens for voice commands and triggers mapped actions.
  • PyAutoGUI executes desktop automation actions from validated inputs.

System architecture

  • Camera feed is processed through OpenCV and MediaPipe for hand landmarks.
  • Gesture logic maps tracked landmarks into desktop commands.
  • Speech recognition layer listens for voice commands and triggers mapped actions.
  • PyAutoGUI executes desktop automation actions from validated inputs.

Screenshots and demo

Horizon UI demo placeholder

Technical challenges

  • Maintaining responsive control while processing real-time video input.
  • Reducing false positives in gesture recognition during live demos.
  • Creating a command model that is flexible without being unpredictable.

Results

  • 3rd Place at Industrial Open House 2026
  • Multimodal gesture and voice control prototype

What I learned

  • Human-computer interaction projects need as much UX thinking as algorithmic accuracy.
  • Live demos improve when technical triggers are explainable and recoverable.

Explore the code

Explore the code, setup instructions, and technical documentation on GitHub.