← Back to blog

Redson Dev brief · ARTICLE

ARTICLE#AI

AI chatbots are giving out people’s real phone numbers

MIT Technology Review — AI · May 13, 2026

The unchecked proliferation of AI language models raises complex new questions about privacy and data security, extending beyond theoretical concerns to immediate, tangible risks. As these models become increasingly integrated into public-facing applications, their capacity for unexpected information disclosure poses a novel challenge for developers and users alike. The potential for a routine interaction with an AI to compromise personal data highlights the urgent need for a more robust understanding of how these systems learn, store, and retrieve sensitive information. A recent MIT Technology Review article illustrates this point starkly, reporting instances where AI chatbots have inadvertently revealed real, private phone numbers belonging to individuals. The core issue appears to stem from the models’ training data, which in some cases includes publicly accessible but often overlooked sources of personal contact information. One notable example cited involves a chatbot trained on publicly available web data that, when prompted, accurately recited a phone number associated with a specific individual, presumably scraped from a directory or neglected webpage. This behavior deviates from expected anonymization or filtering protocols, indicating a gap in current safeguards applied during model development and deployment. The article further suggests these aren't isolated incidents, but rather a symptom of broader data hygiene challenges within large language model development. One intriguing detail from the report centers on how seemingly benign training data can become a vector for private data exposure. The AI did not conjure the number out of thin air but recalled it from its training corpus, emphasizing that the problem lies not in outright fabrication but in the retrieval of actual, albeit obscurely sourced, data. This distinction is crucial for understanding mitigation strategies. Another point of interest is the implied lack of an effective personal identifiable information (PII) filter during the training or inference phase, allowing such data to persist and be recalled. For software, AI, and product builders, this development underscores the critical importance of rigorous data provenance and comprehensive PII scrubbing during model training. It suggests that a reactive approach to privacy breaches is insufficient; instead, proactive measures, including sophisticated data validation and anonymization pipelines, must be integrated into the foundational layers of AI development. Builders should also consider implementing robust content filtering mechanisms at the inference stage, specifically designed to detect and redact sensitive personal information before it reaches the end user.