According to the findings of an alarming report released on Wednesday, Instagram algorithms recommendation connected and even supported a “vast paedophile network” that advertised the sale of illegal “child-sex material” on the platform.
According to researchers at Stanford University and the University of Massachusetts Amherst who spoke to the Wall Street Journal, Instagram users could search by hashtags related to child sex abuse, including graphic terms like #pedowhore, #preteensex, #pedobait, and #mnsfw, which stands for “minors not safe for work.”
As per the researchers, the hashtags directed users to accounts that allegedly offered to sell pedophilic materials via “menus” of content that included videos of children harming themselves or engaging in bestiality. Some accounts allowed buyers to “commission specific acts” or arrange “meet ups,” the Journal said.
Sarah Adams, a Canadian social media influencer and activist who calls out online child exploitation, informed the Journal that Instagram algorithms recommendation impacted her.
Adams said one of her followers flagged a distressing Instagram account in February called “incest toddlers,” which had an array of “pro-incest memes.” The mother of two said she interacted with the page only long enough to report it to Instagram.
After the brief interaction, Adams said she learned from concerned followers that Instagram had begun recommending the “incest toddlers” account to users who visited her page.
Meta confirmed to the Journal that the “incest toddler” account violated its policies.
When reached for comment by The Post, a spokesperson for Instagram’s parent company, Meta, said it has since restricted the use of “thousands of additional search terms and hashtags on Instagram.”
In addition , the spokesperson said that Meta is “continuously exploring ways to actively defend against this behavior” and has “set up an internal task force to investigate these claims and immediately address them.”
“Child exploitation is a horrific crime,” a Meta spokesperson said in a statement. “We work aggressively to fight it on and off our platforms, and to support law enforcement in its efforts to arrest and prosecute the criminals behind it.” Meta pointed to its extensive enforcement efforts related to child exploitation.
The company said it disabled more than 490,000 accounts that violated its child safety policies in January and blocked more than 29,000 devices for policy violations between May 27 and June 2. Meta also took down 27 networks that spread abusive content on its platforms from 2020 to 2022.
“Large-scale communities promoting criminal sex abuse”
The Journal noted that researchers at both Stanford and UMass Amherst discovered “large-scale communities promoting criminal sex abuse” on Instagram.When the researchers set up test accounts to observe the network, they began receiving “suggested for you” recommendations to other accounts that purportedly promoted pedophilia or linked to outside websites.
“That a team of three academics with limited access could find such a huge network should set off alarms at Meta,” Alex Stamos, the head of the Stanford Internet Observatory and Meta’s former chief security officer, told the Journal.
Stamos called for Meta to “reinvest in human investigators.”
Meta said it already hires specialists from law enforcement and collaborates with child safety experts to ensure its methods for combating child exploitation are up to date.
In another bizarre development, the Journal noted that Instagram’s algorithm had previously sent users a pop-up notification warning that certain searches on the platform would yield results that “may contain images of child sexual abuse.”
The screen purportedly directed users to either “get resources” on the topic or “see results anyway.”
The report said Meta disabled the option allowing users to view the results anyway, but has declined to reveal why it was ever offered in the first place.The Journal’s findings came as Meta and other social media platforms face ongoing scrutiny over their efforts to police and prevent the spread of abusive content on their platforms.
In April, a group of law enforcement agencies that included the FBI and Interpol warned that Meta’s plans to expand end-to-end encryption on its platforms could effectively “blindfold” the company from detecting harmful content related to child sex abuse.