Cloudflare, a cloud company that provides security and content distribution services, provided CDN (Content Delivery Network) services to sites that host Child Sexual Abuse Material (CSAM). CDN refers to a group of geographically distributed servers that host site data to provide faster access to people visiting that website. L1ght, an Israel-based company working to provide a safe environment to children online, shared a report with TNW containing details of the websites featuring harmful content. We’re not disclosing the names of these sites for safety reasons. The firm claimed it reached out to Cloudflare in November, but is yet to receive a response. L1ght estimates the cloud company’s CDN hosts tens to hundreds of websites featuring exploitative child networks. Experts from the Tel-Aviv based firm actively look for email and social media accounts that are exchanging exploitative material. Then the company’s algorithm scrouges the internet to find sites that might be hosting such content. L1ght’s AI has been trained to catch text, images, or video that relates to child exploitation. [Read: Microsoft’s new message scanning tool can help identify sexual predators in chatrooms] In a statement to TNW, Cloudflare said CSAM is repugnant and illegal, and the company doesn’t tolerate it. It has also taken action on reported domains:  The company didn’t comment on how many websites it booted off its platform or when it took the action. It pointed us to a blog post published last December announcing the launch of a tool to catch CSAM. Around the same time, the New York Times published a report pointing out more than three popular CSAM websites used Cloudflare’s protection services to obscure their web addresses for years. Later, the company complied with norms and removed these sites. Cloudflare added it has provided 5,208 reports to the National Center for Missing or Exploited Children (NCMEC) and removed 5,428 domains from its service to date.  The fact that L1ght believes there are more sites on the CDN – and that Cloudflare removed thousands of domains from its service – indicates that this is a bigger problem than the company might imagine. Hany Farid, a professor of computer science at Berkeley, who helped Microsoft create PhotoDNA, a technology that helps in removing exploitative images of children, said it’s important for companies like Cloudflare to develop tools that detect and remove CSAM. Globally, major companies are joining the fight to remove CSAM from the web. In 2018, Facebook introduced a new machine learning-based technique to quarantine such content from its platform. Last November, the company said it removed over 11.6 million abusive posts between July and September 2019. Earlier this month, Microsoft introduced a new free tool to scan online chatrooms and detect predators. It took us almost 10 years to get the major social media companies to take this issue seriously. I hope that it doesn’t take another 10 years to get internet infrastructure companies to do the same.