The filing, posted on September 19, is closely redacted and Hive cofounder and CEO Kevin Guo advised MIT Know-how Evaluate that he couldn’t focus on the small print of the contract, however confirmed it entails use of the corporate’s AI detection algorithms for youngster sexual abuse materials (CSAM).
The submitting quotes data from the Nationwide Heart for Lacking and Exploited Kids that reported a 1,325% improve in incidents involving generative AI in 2024. “The sheer quantity of digital content material circulating on-line necessitates using automated instruments to course of and analyze information effectively,” the submitting reads.
The primary precedence of kid exploitation investigators is to search out and cease any abuse at present occurring, however the flood of AI-generated CSAM has made it tough for investigators to know whether or not photographs depict an actual sufferer at present in danger. A instrument that would efficiently flag actual victims could be a large assist once they attempt to prioritize instances.
Figuring out AI-generated photographs “ensures that investigative sources are centered on instances involving actual victims, maximizing this system’s influence and safeguarding weak people,” the submitting reads.
Hive AI gives AI instruments that create movies and pictures, in addition to a spread of content material moderation instruments that may flag violence, spam, and sexual materials and even determine celebrities. In December, MIT Know-how Evaluate reported that the corporate was promoting its deepfake-detection expertise to the US navy.