The first time I paid attention to all the chatter about TikTok was when I read the NYT article on how its recommendation engine works. The numbers to chase were time watched and retention - which are sensible things to do, but how it gets done is where the problem lies and gives cause for concern.
...watch time is key. The algorithm tries to get people addicted rather than giving them what they really want,” said Guillaume Chaslot, the founder of Algo Transparency, a group based in Paris that has studied YouTube’s recommendation system and takes a dark view of the effect of the product on children, in particular. Mr. Chaslot reviewed the TikTok document at my request.
“I think it’s a crazy idea to let TikTok’s algorithm steer the life of our kids,” he said. “Each video a kid watches, TikTok gains a piece of information on him. In a few hours, the algorithm can detect his musical tastes, his physical attraction, if he’s depressed, if he might be into drugs, and many other sensitive information. There’s a high risk that some of this information will be used against him. It could potentially be used to micro-target him or make him more addicted to the platform.”
Comments