Real-time eye tracking and voice-controlled web browsing with automated test script generation
EyeNav is a modular web interaction framework. It fuses real-time eye-tracking (with Tobii-Pro SDK) and on-device natural-language processing (using Vosk) within a Chrome extension and Python backend to deliver:
By orchestrating gaze-driven pointer control, voice-command parsing, and concurrent logging threads, EyeNav enables both interactive accessibility and behavior-driven development in web environments.
EyeNav implements the following core features:
Maps eye gaze to cursor movements using the Tobii Pro Nano and the
tobii-research
SDK.
Transcribes and interprets voice commands (click, input, scroll, navigate) with Vosk running locally.
Logs interactions in Gherkin syntax and replays them via Kraken & WebdriverIO.
Enable or disable any of the three subsystems (gaze, voice, test logger) independently.
Supports English and Spanish out of the box; additional languages can be downloaded and locale translation.
Clone the Repository
git clone https://github.com/TheSoftwareDesignLab/EyeNav.git
cd EyeNav
Backend Setup
Frtom the backend/
folder
python3 -m venv venv
source venv/bin/activate
pip install -r requirements.txt
Chrome Extension
chrome://extensions/
in Chrome (v114+)extension/
Start Backend
python backend/main.py
Load Web Page & Extension
Initiate Session
Generate Tests
tests/
directoryen
) and Spanish (es
) models is provided. To download more models, change the preferred language on Chrome.commands.json
, with the specified language code.href
id
className
xPath
Hands-free navigation for users with disabilities.
Generate and replay acceptance tests for regression.
Collect interaction data for consultants and researchers.
[TBD] Enable bots to navigate and test web UIs via gaze & speech.