A new NPR report has raised serious privacy and safety concerns about AI-enabled children’s toys. Researchers behind the report found that some of these toys not only shared disturbing or inappropriate content with kids, but also collected sensitive data about them.
According to the investigation, certain toys made use of large language models (LLMs) to hold conversations — but the content wasn’t always child-safe. Some toys reportedly gave instructions on dangerous activities, while others delved into mature or explicit topics without adequate filtering or context.
Beyond content issues, the report highlights data-collection risks. Because these toys are constantly listening, they can record personal details, voice data, and possibly other private information. There are worries about how this data is stored, shared, or potentially misused, particularly since children may confide in their AI “companions” in a way they wouldn’t with adults.
Consumer advocates are calling for stronger regulations and safety standards for AI toys. They argue that toy makers must ensure better content filters, limit data retention, and provide parents with clearer visibility into what their child’s toy is recording or saying.