Meta has confirmed plans to use employee keystroke data to train its artificial intelligence models, igniting a fierce debate over workplace privacy and ethics. The social media giant aims to leverage internal productivity and behavioral metrics to improve AI systems, a move that underscores the growing value of proprietary data in the race for AI dominance.
Critics argue that monitoring keystrokes goes beyond reasonable oversight, potentially violating employee trust and privacy. Labor advocates warn that such data could be used to surveil worker performance rather than solely for AI development. Meta, however, maintains that the initiative is transparent and will comply with data protection regulations.
The controversy arrives as companies increasingly seek novel data sources to train more powerful AI. Privacy experts say this could set a precedent for how corporations collect and use employee-generated data. "This is a slippery slope," noted one analyst. "If keystrokes become fair game, what's next? Eye-tracking or voice recordings?"
Meta has not disclosed whether employees can opt out or how the data will be anonymized. The company faces growing calls for clearer policies and consent mechanisms.