There are several clicks from innocent videos from childhood to pedophilia
Am Sure you can learn successful Blogging secrets through me, am also sure you can learn how to make money online with the help of my updates why not leave your email behind let me show you how.
Many people simply skip content, but there are those who will be looking for a child from the video, will be asked to talk to him, talk about it, parents have something to worry about when their kids post video on social networks.
YouTube will not disable your child video algorithm, though pedophiles on the web make it easy to access many sites. Obviously, the decisions of a large video channel are sharp and “dancing” on the blade that shares business ethics, writes the New York Times. Christina T. does not doubt when her 10-year-old daughter and her girlfriend set up a YouTube video showing how the two rebeccatsi in the pool in the backyard. “The video is innocent, it’s not a big deal,” says Christina.
A few days later, her daughter shared exciting news: the video has thousands of visits. Soon after, it was seen 400,000 times – stunning baby video in a swimsuit and her girlfriend. “I was watching the video again and I was scared by the number of visits,” says Christina, cited in the New York Times publication. And her concern is not without reason.
YouTube’s referral system – which controls most billions of video views on the platform – has started showing the video to users watching other videos with children of unhealthy interests. These people searched for erotic content with the participation of children who were partially dressed. Hair Recognition Algorithm YouTube offers every visitor videos from all over their archives, and seeing children in the pool in the backyard is just one of the many harmless home videos of joyous and harmless families, researchers say.
In many cases, however, the video channel algorithm directs similar videos to users who have previously viewed something else: sexual content. The result is that, according to experts, children are sexually sexual. “The YouTube Algorithm connects these records,” said Jonas Kaiser, one of the three researchers from the Internet and the International Harvard Internet Center, who encountered videos while exploring YouTube’s influence in Brazil. “It’s a scary thing.”
Christine Daughter’s video has been promoting YouTube systems for months after the company is warned that there is a problem in the case. In February, however, there is evidence that malicious entities used the comment section to target other pedophiles to the content in question. In February, YouTube solved the issue as very troublesome and deactivated comments from many children’s videos. But the referral system, which remains operational, continues to collect dozens of such videos and offers them huge audiences.
And it is not quite clear how he decided to promote them. Progress in Proposals Consumers, including those with unhealthy interests, should not look for child porn videos to watch them. Only the platform offers them. Consistency is simple: a user looking at erotic videos gets a video recommendation with the participation of women who are noticeably younger in each subsequent video, and then women who are defective in childrens clothing. In the end, some users can get video suggestions for girls aged 5 to 6 wearing swimsuits.
Moreover, there is an element of increased extremism in the structure of the proposal. For example, if someone looks at a bicycle video, bidding starts with regular bicycle footage and will eventually offer dangerous bicycle stops. By itself, any video can be completely innocent, it can be a small home movie recorded by a child.
But, grouped in the described way, they get a new meaning. “I’m really afraid,” Christina says. Attempts to Restrictions In response to numerous queries, YouTube decides to remove some videos with similar content, but many others remain, including those filed with fraudulent accounts. The recommendation system has also changed, though YouTube argues that this is probably a result of routine algorithm changes and not thoughtful policies.
Jennifer O Connor, YouTube’s Product Manager for Trust and Security, says the company is committed to stopping child exploitation on its platform and is constantly working to improve law enforcement. “Child protection is the first in our list,” she says. Business interests, However, there is no guarantee that the referral system will not restart children’s videos with pedophiles. The company says the recommendations are the biggest factor in generating traffic to the video channel.
They are a generator of 70% of traffic on the construction site. Moreover, removing certain content items from the content may harm the record creators. It is the whole ecosystem of “partners” who deliberately participate in the recording of a home video and announce them for making money. The company just says it will limit the recommendations for videos that, according to the company’s claims, bring children into danger. Kaiser and his team have shown progress – with the participation of Jason Cordova and Adrian Rauhflais – an experiment with YouTube recommendations.
The server opens their videos, then follows the best YouTube recommendations on what to watch. Implementing this experiment a thousand times allows scientists to do something like a map as the platform directs their users. Following recommendations for watching video with sexual content, scientists note that videos are becoming more bizarre or extreme, placing more emphasis on young people.
For example, women’s sex videos sometimes lead to videos of women in underwear or nursing where girls sometimes recall their age: 19, 18 and even 16. Some women accept donations from a “boyfriend” or invite them to watch their “personal” videos where they are naked. After a few clicks, some convert children to pose in childhood clothing. It’s easy to get videos with the right kids, like swimsuits.
Jodara Cordova, who studies the spread of online pornography, says she is aware of what is happening. Every single video may be pretty innocent – maybe it’s loaded by parents who want to share their family movies. But YouTube’s algorithm, part of training for users seeking erotic content, will soon “smell” videos and treat them as a destination for people of very different types of interests.
Normalization The mechanism is simple and inconspicuous. “It’s an incredibly powerful tool and people simply tone it,” says Stephen Blumenthal, a London psychologist who deals with people with deviant sexual interests and behavior. Progression, which comes from erotic content with adults in childhood records, clouds the boundary of the mistake.
Gradually, the limit of intolerance towards pedophilia is eliminated, says psychologist. “This is a process of normalization,” explains Marcus Rodgers, a purdue psychologist investigating child pornography. Many people who have accessed such content simply miss it. But there are those who will look for a kid from the video, want to talk to him, talk to him, says Rogers.
Such people encourage children to pose, photograph, create sexual images – and record and publish them. Parents have what to worry about. For now, Christina has banned her daughter from publishing video on a video platform. But is that enough?
There are several clicks from innocent videos from childhood to pedophilia