Quantcast
Viewing latest article 14
Browse Latest Browse All 736

2025-01-15: Revolutionizing Eye Tracking and Data Collection: A Look at Project Aria Glasses

 Project Aria glasses, introduced by Meta (formerly Facebook), represent a significant leap in wearable devices. Announced as part of Meta's larger vision for the metaverse, these glasses are designed to capture rich contextual egocentric multi-modal data, providing a platform for developing augmented reality (AR) experiences grounded in real-world interactions. Aria glasses are built to collect data about the environment and user behavior while opening new frontiers for eye tracking, spatial navigation, and human-computer interaction research. NirdsLab recently became a research partner for the program researching contextualized advanced eye tracking. In this blog post, we will discuss the capabilities of Aria glasses and how we use them in eye-tracking research.

Image may be NSFW.
Clik here to view.
Figure 1: Components of the glasses (https://facebookresearch.github.io)

At the current stage of development at Meta, the Aria glasses take the form of a sensor array for future smart glasses, containing visual and non-visual sensors for multi-modal data collection. The sensory information captured by the glasses includes video, audio, locations, and motion from the wearer’s perspective. Table 1 summarizes the sensors available in Aria glasses.

Table 1: Different sensors and the description of their use.
SensorDescription
Mono scene cameraCaptures the world with a maximum peripheral vision with stereo overlap. Primarily used for the visual Simultaneous Localization and Mapping (SLAM) capabilities/functions.
RGB cameraCapture the point of view of the wearer in a higher resolution than the Mono Scene cameras.
Eye-Tracking cameraTwo monochromatic eye-facing cameras for eye-tracking.
Inertial measurement unit (IMU)Captures movements of the glasses through an accelerometer (linear) and a gyroscope (angular).
MagnetometerMeasures the ambient magnetic field.
Barometer and thermometerCaptures local air pressure and the thermometer.
MicrophoneMicrophone array comprising seven microphones, capable of capturing spatial audio at 48 kHz.
Global Navigation and Satellite System (GNSS) receiverSupports global location capturing with GPS and Galileo constellations.
Wi-Fi and Bluetooth transceiverTransceivers capable of regularly scanning and recording Received Signal Strength Indicator (RSSI) from Wi-Fi and Bluetooth beacons.

Aria glasses also offer the capability of customizing the data collection process by individually configuring each sensor, such as enabling/disabling or setting the sampling rate of sensors through a set of data collection profiles. Currently, the glasses provide 29 different recording profiles optimized for various applications. For instance, the recording profile for long-duration recordings disables most non-visual sensors and samples at lower rates to conserve battery usage, providing over 15 hours of recording time.

By combining the above sensory information, we can estimate the head, hand, gaze directions, and other movements of the wearer in a user study. This allows researchers to gather real-world data on how people perceive and interact with their physical and digital surroundings. With these capabilities, Aria glasses have applications in behavioral studies, accessibility research, and even personalized AR experience developments, offering insights into human interaction that were previously difficult to obtain.

Image may be NSFW.
Clik here to view.
Figure 2: Data collection view with project aria glasses. (source: https://www.projectaria.com/datasets/aea/)

Setting up Project Aria glasses

  1. Download and install the Desktop Companion App using the instructions available at https://facebookresearch.github.io/projectaria_tools/docs/ARK/ark_downloads#desktop.
  2. Log into the desktop companion app using the Project Aria credentials.
  3. Connect the Aria glasses to the desktop via USB.
  4. Once plugged in, the device will display as "Aria" and "active" under “My Device.”
  5. Select a profile to start recording (e.g., Profile0). See list of profiles: https://facebookresearch.github.io/projectaria_tools/docs/tech_spec/recording_profiles .

The Aria desktop dashboard provides the ability to record, process, transfer, and visualize data


Image may be NSFW.
Clik here to view.
Figure 3: Aria desktop application dashboard. (https://facebookresearch.github.io/projectaria_tools/docs/ARK/desktop_companion_app)

Recording with the Mobile Companion App

  1. Download and install the mobile companion app compatible with the device from https://facebookresearch.github.io/projectaria_tools/docs/ARK/ark_downloads.
  2. Plug the Project Aria glasses into the charger (this will automatically switch on the glasses) or switch on the glasses using the button.
  3. Log into the mobile companion app (first time).
  4. Select “Get Started” to begin the setup.
  5. The app will search for the nearby Project Aria glasses, and once discovered, the glasses will appear in the app.
  6. Select the listed glasses to pair with the companion app.
  7. Once pairing is completed, the app will ask to name the Aria glasses.
  8. Join a Wi-Fi network; the glasses must be plugged into a charger to complete the setup process.
  9. Once connected to Wi-Fi, the glasses will look for updates and update the glasses' OS.
  10. Once setup is complete, the mobile companion app dashboard will appear.
  11. The glasses are now set to record. Unplug the glasses and start recording using a preferred profile.

The mobile app provides limited functionalities for interacting with Aria glasses, including setting up recording profiles, managing recordings, and monitoring device status.

Image may be NSFW.
Clik here to view.
Figure 4: Aria mobile companion app dashboard. (https://facebookresearch.github.io/projectaria_tools/docs/ARK/mobile_companion_app)

Aria open datasets for new research opportunities

The Project Aria research team provides open datasets collected with Project Aria glasses for new research. Aria’s Aria Everyday Activities (AEA) dataset contains anonymized Aria sequences captured in various scenarios, such as cooking, playing games, or exercising. This dataset is a valuable resource for analyzing first-person perspectives of daily activities. 

The Aria Digital Twin (ADT)is an egocentric dataset recorded using Aria glasses, featuring comprehensive simulated ground truth data for devices, objects, and environments. This dataset establishes a new benchmark for egocentric machine perception research, advancing studies in areas such as 3D object detection and tracking, scene reconstruction and interpretation, sim-to-real learning, and human pose estimation. It also sparks the development of novel machine perception tasks tailored for AR applications.

TheAria Synthetic Environmentsdataset features interior layouts filled with 3D objects, simulated to match the sensor characteristics of Aria glasses. Unlike earlier 3D scene understanding datasets, which often lack sufficient scale for effective machine learning training, this dataset sets a new benchmark in size and scope. It opens up research opportunities in 3D scene reconstruction, object detection, and tracking.

Nymeria is the largest human motion dataset captured in natural settings, featuring diverse activities and locations. It is the first application to use multiple egocentric multimodal devices synchronized within a unified 3D metric framework. This dataset helps to advance research in egocentric human motion understanding, aiming to drive innovations in contextualized computing and future AR/VR technologies.

The HOT3D dataset addresses the challenge of understanding how humans use their hands to manipulate objects, a key area in computer vision research. This dataset opens new possibilities, such as transferring manual skills from experts to novices or robots, enabling AI assistants to interpret user actions.

The Digital Twin Catalog (DTC) dataset provides a highly detailed collection of sub-millimeter-level accurate 3D object models and source capture data for various reconstruction techniques. As reliance on 3D digital twins grows for e-commerce, entertainment, and advertising, DTC aims to reduce manual 3D model creation costs by enabling researchers to advance reconstruction techniques.

The Aria Everyday Objects (AEO) dataset provides a critical real-world validation tool for 3D object detection in egocentric data. It ensures that AI models trained on synthetic datasets like EFM3D and SceneScript can be generalized effectively to real-world scenarios beyond simulations.

Eye tracking with Project Aria glasses

Project Aria glasses can estimate gaze direction from data acquired from eye-tracking cameras. For this purpose, they have developed the Aria Machine Perception Service (MPS) to estimate gaze information during post-processing steps through the MPS Command Line Interface (MPS CLI) or the desktop application. It is important to note that a recording profile with eye-tracking cameras needs to be enabled to access the eye-tracking information. 

When eye tracking is combined with other spatial information, we can interpret the visual attention and environmental interactions of users in an environment. With Aria glasses, wireless wearable glasses eliminate device-level constraints such as movement restrictions due to wiring in multi-user eye-tracking with dedicated devices. In multi-user studies, project aria tools facilitate time-synchronized recordings with multiple project aria glasses. Further, in-built sensors allow us to capture additional events beyond visual attention, such as user movements and auditory events, making the glasses a robust platform for research in user behaviors and assistive technologies. 

Despite these benefits, Aria glasses do not provide gaze estimations in real time due to their reliance on cloud-based machine perception services. As a result, we cannot integrate it with our research applications, such as real-time multi-user eye-tracking dashboards and gaze-driven navigation applications.

Image may be NSFW.
Clik here to view.
Figure 5: Wearer’s point of view captured using the RGB camera during a user study in the NirdsLab. The green dot shows the gaze of the wearer.

After recording a session, the glasses provide a VRS file containing videos of the wearer's view (RGB camera), videos of two eyes (eye-tracking camera), and thumbnail images of the wearer’s scenery. The VRS file is used for post-processing to obtain gaze data, hand tracking data, and Simultaneous Localization and Mapping (SLAM) data from the MPS services.

Research with Project Aria glasses in the NirdsLab

Accessibility research

NirdsLab, in collaboration with the Accessible Computing Lab, is using Project Aria for accessibility research to enable the development of assistive technologies tailored for individuals with vision impairments. We collect eye tracking data, head movement data using integrated IMUs, and conversational data with built-in microphones to understand the visual behavior of people with visual impairments.  Our goal is to explore and analyze the data to identify unique eye-tracking patterns inherent to the visually impaired community, based on which we propose design implications for accessible egocentric vision systems, including improved calibration approaches.

Multi-user research

In multi-user research, We use Project Aria glasses to explore the collaborative and social interactions in shared environments. Their ability to synchronize data from multiple devices provides insights into group dynamics, joint attention, and collective decision-making. For example, we plan to study joint visual attention in a collaborative visual task on a screen. We will analyze eye-tracking data collected from the glasses and observe the collaborative activity using the participants' voice recordings. This research can potentially extend the designing of effective multi-user systems in AR/VR applications. These capabilities position Project Aria as a powerful tool for advancing research multi-user interactions, paving the way for innovative applications in diverse domains.

-Bhanuka (@mahanama94)  Kumushini (@KumushiniT)




Viewing latest article 14
Browse Latest Browse All 736

Trending Articles