Meta Ray-Ban Smart Glasses Recordings Reviewed by Human Contractors, Triggering Privacy Scrutiny
Investigations reported by Swedish outlets Svenska Dagbladet and Göteborgs-Posten found that recordings captured by Meta Ray-Ban smart glasses—including video and audio—are being reviewed by human contractors as part of AI training and quality assurance workflows. Workers employed by Sama, a Meta subcontractor in Nairobi, Kenya, described routinely handling highly sensitive content inadvertently recorded by users, including bathroom visits, undressing, sex/pornography, and private conversations, as well as incidental capture of bank cards and other identifying details; interviewees said they feared reprisals for raising concerns and described strict on-site controls intended to prevent leaks.
Following the reporting, the UK’s privacy regulator, the Information Commissioner’s Office (ICO), confirmed it is contacting Meta to ask questions about the devices and associated data-handling practices. While Meta’s terms reportedly disclose that some interactions may be reviewed by humans to improve the system, the reporting and worker accounts suggest the review pipeline can include intimate or identifying moments that wearers may not expect to be viewed by third parties, raising regulatory and reputational risk around consent, transparency, and safeguards for bystander and user privacy.
Timeline
Mar 5, 2026
UK ICO opens inquiries with Meta over Ray-Ban privacy concerns
The UK Information Commissioner's Office said it was contacting Meta to seek information on how the company complies with UK data protection obligations for its Ray-Ban Meta AI smart glasses. The regulator called the allegations concerning and highlighted requirements for transparency about what data is collected, how it is used, and who can access it.
Mar 5, 2026
Meta points to policies allowing human review of smart-glasses content
Following questions about the investigation, Meta said user content may be reviewed in certain circumstances to improve its AI products and pointed to its privacy policies and terms of use. The company said users can manage and delete recordings, while its policies describe cloud processing, human review, and sharing with third-party vendors and service providers.
Mar 5, 2026
Swedish media investigation reveals contractor review of Ray-Ban footage
A joint investigation by Svenska Dagbladet and Göteborgs-Posten reported that Meta subcontractor workers at Sama in Nairobi were reviewing audio and video captured by Ray-Ban Meta smart glasses to label data for AI systems. Workers said the material included highly sensitive content such as intimate moments, bathroom use, private conversations, and visible financial information, and that anonymization measures did not always work.
See the full picture in Mallory
Mallory subscribers get deeper analysis on every story, including:
Who’s affected and how
Deep-dive technical analysis
Actionable next steps for your team
IPs, domains, hashes, and more
Ask questions and take action on every story
Filter by topic, classification, timeframe
Get matching stories delivered automatically
Related Entities
Organizations
Sources
Related Stories

Privacy Risks of Smart Glasses in Healthcare Environments
Smart eyewear devices such as Meta Ray Ban glasses, equipped with microphones, cameras, and AI connectivity, present significant privacy and data security risks when used in hospital settings. These devices can inconspicuously record or livestream protected health information (PHI), including patient images and conversations, often without the knowledge or consent of those being recorded. The presence of a small LED indicator is insufficient as a safeguard, especially since third-party products exist to obscure the light, making unauthorized recording even harder to detect. Healthcare organizations face challenges as these are often unmanaged devices brought in by patients or staff, bypassing institutional controls and oversight. The direct connectivity of these glasses to social media platforms like Facebook and Instagram increases the risk of inadvertent or malicious disclosure of sensitive information, potentially violating HIPAA/HITECH regulations. The inconspicuous nature of smart glasses differentiates them from more obvious recording devices like smartphones, heightening the risk of unnoticed privacy breaches in clinical environments.
1 months ago
Expansion of AI-Enabled Camera Surveillance Raises Privacy and Biometric Identification Concerns
The New York Metropolitan Transportation Authority (MTA) is testing new subway gates that use **AI-powered cameras** to capture short recordings when riders are suspected of fare evasion and to generate a physical description that is transmitted to the MTA, prompting criticism from privacy advocates concerned about persistent monitoring in public transit. The MTA has also solicited vendor input for systems using computer vision and AI to detect “unusual or unsafe behaviors,” reflecting broader growth in surveillance deployments across New York City. In parallel, consumer **AI smart glasses** are re-emerging with built-in cameras and microphones, intensifying concerns that everyday wearables can enable covert recording and downstream biometric identification. Reporting highlighted that footage from *Ray-Ban Meta* smart glasses can be paired with external facial-recognition services to identify strangers, and noted policy issues such as cloud storage of wake-word voice recordings (potentially retained up to a year) and uncertainty about future features like on-device facial recognition; retailers in New York (e.g., Wegmans and others) are also expanding facial-recognition use, underscoring the convergence of AI, biometrics, and surveillance in both public and commercial spaces.
1 months ago
Nearby Glasses Android App Detects Smart Glasses via BLE Manufacturer IDs
An Android app called **Nearby Glasses** was released to alert users when certain camera-equipped smart glasses are nearby by scanning **Bluetooth Low Energy (BLE)** advertising traffic for **manufacturer company IDs** (Bluetooth SIG-assigned identifiers) rather than device names, MAC addresses, or service UUIDs, which can be inconsistent or randomized. The app, developed by **Yves Jeanrenaud** (Darmstadt University of Applied Sciences), runs as a foreground service and triggers notifications when detected devices meet a configurable RSSI threshold (reported default around `-75 dBm`, roughly 10–15 meters in open space). Reported monitored IDs include Meta/Luxottica and Snapchat-related identifiers, and users can add custom hex values to expand detection. Reporting highlighted that the approach can generate **false positives**, particularly from other Bluetooth devices made by the same vendors (e.g., other Meta hardware), and the project documentation warns against using detections to harass people. The app’s emergence is framed as a response to growing privacy concerns and reported misuse of smart glasses for **non-consensual recording** and demonstrations of **real-time facial recognition** using glasses paired with public data; coverage also notes legal and regulatory risk areas (e.g., state wiretapping laws) and ongoing debate over the effectiveness of recording indicator lights and the potential for those indicators to be disabled.
1 months ago