ReportWire

Tag: CSAM

  • Maryland man working at DC school faces child exploitation charge – WTOP News

    [ad_1]

    Dandre Eric Davis, of Prince George’s County, who worked at a D.C. school, has been charged in a child sex abuse materials case, prosecutors said.

    A Prince George’s County, Maryland, man who worked at Luke C. Moore Opportunity Academy in Northeast D.C. has been charged with distribution of child sexual abuse material.

    Dandre Eric Davis, 31, of Suitland, exchanged messages beginning last month with someone using a dating app and later continued those communications on the encrypted messaging platform Telegram, according to federal prosecutors.

    Court documents indicate Davis discussed illegal interests and ultimately shared videos depicting child sexual abuse. Federal prosecutors said Davis described himself in those online conversations as a “kinky perv,” and that Davis said he was interested in children between 7 and 15 years old.

    The complaint was unsealed Monday by U.S. Attorney Jeanine Pirro, who was joined by FBI Assistant Director in Charge Darren Cox of the Washington Field Office.

    According to the U.S. Attorney’s Office for D.C., the case is being investigated by the FBI’s Child Exploitation and Human Trafficking Task Force as a part of Project Safe Childhood, a nationwide initiative to combat child sexual exploitation and abuse.

    Rodney Wormsley, principal at Luke C. Moore Opportunity Academy, told students and parents that there are currently no indicators that Luke C. Moore students were involved in the incident in an email statement on Tuesday.

    “DC Public Schools (DCPS) treats all allegations of employee misconduct with the utmost seriousness and is following all required protocols and cooperating with the law enforcement,” Wormsley said in the statement. “The staff member is out of the building on leave and will not report to school while this matter is under review.”

    WTOP has reached out to Davis for comment.

    Get breaking news and daily headlines delivered to your email inbox by signing up here.

    © 2026 WTOP. All Rights Reserved. This website is not intended for users located within the European Economic Area.

    [ad_2]

    Matt Small

    Source link

  • Amazon discovered a ‘high volume’ of CSAM in its AI training data but isn’t saying where it came from

    [ad_1]

    The National Center for Missing and Exploited Children said it received more than 1 million reports of AI-related child sexual abuse material (CSAM) in 2025. The “vast majority” of that content was reported by Amazon, which found the material in its training data, according to an investigation by Bloomberg. In addition, Amazon said only that it obtained the inappropriate content from external sources used to train its AI services and claimed it could not provide any further details about where the CSAM came from.

    “This is really an outlier,” Fallon McNulty, executive director of NCMEC’s CyberTipline, told Bloomberg. The CyberTipline is where many types of US-based companies are legally required to report suspected CSAM. “Having such a high volume come in throughout the year begs a lot of questions about where the data is coming from, and what safeguards have been put in place.” She added that aside from Amazon, the AI-related reports the organization received from other companies last year included actionable data that it could pass along to law enforcement for next steps. Since Amazon isn’t disclosing sources, McNulty said its reports have proved “inactionable.”

    “We take a deliberately cautious approach to scanning foundation model training data, including data from the public web, to identify and remove known [child sexual abuse material] and protect our customers,” an Amazon representative said in a statement to Bloomberg. The spokesperson also said that Amazon aimed to over-report its figures to NCMEC in order to avoid missing any cases. The company said that it removed the suspected CSAM content before feeding training data into its AI models.

    Safety questions for minors have emerged as a critical concern for the artificial intelligence industry in recent months. CSAM has skyrocketed in NCMEC’s records; compared with the more than 1 million AI-related reports the organization received last year, the 2024 total was 67,000 reports while 2023 only saw 4,700 reports.

    In addition to issues such as abusive content being used to train models, AI chatbots have also been implicated in several dangerous or tragic cases involving young users. OpenAI and Character.AI have both been sued after teenagers planned their suicides with those companies’ platforms. Meta is also being sued for alleged failures to protect teen users from sexually explicit conversations with chatbots.

    [ad_2]

    Anna Washenko

    Source link