Frequent users of YouTube have higher levels of loneliness, anxiety, and depression according to researchers from the Australian Institute for Suicide Research and Prevention (AISRAP).
Dr Luke Balcombe and Emeritus Professor Diego De Leo from Griffith University’s School of Applied Psychology and AISRAP sought to understand both the positive and negative impacts of the world’s most used streaming platform on mental health.
They found the most negatively affected individuals were those under 29 years of age, or who regularly watched content about other people’s lives.
Lead author Dr Luke Balcombe said the development of parasocial relationships between content creators and followers could be cause for concern, however some neutral or positive instances of creators developing closer relationships with their followers also occurred.
“These online ‘relationships’ can fill a gap for people who, for example, have social anxiety, however it can exacerbate their issues when they don't engage in face-to-face interactions, which are especially important in developmental years,” he said.
“We recommend individuals limit their time on YouTube and seek out other forms of social interaction to combat loneliness and promote positive mental health.”
Dr Balcombe said the amount of time spent on YouTube was often a concern for parents, who struggled to monitor their children’s use of the platform for educational or other purposes.
For the purpose of the study, over two hours per day of YouTube consumption was classed as high frequency use and over five hours a day as saturated use.
The study also determined more needed to be done to prevent suicide-related content being recommended to users based on algorithms for suggested viewing.
While ideally, people shouldn’t be able to search for these topics and be exposed to methods, the YouTube algorithm does push recommendations or suggestions based on previous searches, which can send users further down a disturbing ‘rabbit hole’.
Users can report this type of content, but sometimes it may not be reported, or it could be there for a few days or weeks and with the sheer volume of content passing through, it's almost impossible for YouTube’s algorithms to stop all of it.
If a piece of content is flagged as possibly containing suicide or self-harm topics, YouTube then provides a warning and asks the user if they want to play the video.
“With vulnerable children and adolescents who engage in high frequency use, there could be value in monitoring and intervention through artificial intelligence,” Dr Balcombe said.
“We’ve explored human–computer interaction issues and proposed a concept for an independent-of-YouTube algorithmic recommendation system which will steer users toward verified positive mental health content or promotions.
“YouTube is increasingly used for mental health purposes, mainly for information seeking or sharing and many digital mental health approaches are being tried with varying levels of merit, but with over 10,000 mental health apps currently available, it can be really overwhelming knowing which ones to use, or even which ones to recommend from a practitioner point of view.
“There is a gap for verified mental health or suicide tools based on a mix of AI-based machine learning, risk modelling and suitably qualified human decisions, but by getting mental health and suicide experts together to verify information from AI, digital mental health interventions could be a very promising solution to support increasing unmet mental health needs.”
0 Comments