Last Updated on 25th June 2024
Read the script below
Natalie: Hello, everyone, and welcome back to another Safeguarding Soundbites. I’m Natalie, and this week, I’m joined by Tyla. We’ve got quite a line up of stories today, don’t we?
Tyla: Absolutely, Natalie. Buckle up, everyone, because we’re diving headfirst into some shocking revelations and concerns in the tech world.
Natalie: Our first story sheds light on the darker side of online marketplaces. A recent study has uncovered that over a third of ads on Facebook Marketplace may be scams. The TSB retail banking chain’s fraud team went a step further and contacted 100 sellers on the platform. The shocking findings indicate that UK buyers are losing over £160,000 daily due to fraudulent listings. TSB found that 34% of the ads were scams, with sellers employing tactics commonly used by fraudsters. These include directing experts (who they thought were buyers) to fake websites, refusing in-person viewings, and demanding advanced fees.
Tyla: It’s disheartening to see such a widespread issue, Natalie. It’s important we all keep an eye out and report scams immediately. And highlight to the young people in your care the potential pitfalls of using online marketplaces like Facebook.
Natalie: Absolutely. Switching gears now and internal Meta documents about child safety have come to light. The documents reveal that Meta was not only aware of inappropriate and sexually explicit content being shared between adults and young people but also marketed its platforms to children. Disturbingly, Meta employees raised concerns internally about the exploitation of children on their private messaging platforms. The documents indicate that Meta intentionally recruited children to Messenger, limiting safety features in the process. This has raised significant concerns about Meta’s commitment to protecting children online.
Tyla: This is deeply troubling, Natalie. And the Internet Watch Foundation’s report adds another layer to those concerns. They found a 66% increase in self-generated child sexual abuse material (CSAM) featuring children under 10 on more than 100,000 webpages in the last year. With Meta’s move towards end-to-end encryption, there are fears that reporting such material could become even more challenging.
Natalie: Absolutely, Tyla. Let’s hope there’s a serious re-evaluation of priorities to ensure the safety of our youngest internet users.
Tyla: Now, on a different note, gamers, listen up. A recent study has shown that prolonged exposure to high-intensity sound in gaming may put players at risk of irreversible hearing loss and tinnitus. The study emphasizes the need for education and awareness to promote safe listening habits among gamers. Guidelines suggest permissible exposure limits, and the data indicates that, on average, sound levels often exceed these limits.
Natalie: It’s a wake-up call for the gaming community to prioritize their auditory health. Awareness and education can go a long way in preventing irreversible damage. And make sure the young people in your care are aware of the potential problems, especially as they may be using headphones.
Tyla: Definitely. Next up, a study by Fuse, the Centre for Translational Research in Public Health at Teesside University, and Newcastle University reveals a concerning link between young people’s consumption of energy drinks and mental health risks. Despite most supermarkets banning sales to those under 16, availability in local shops remains a concern. Teachers have expressed worries about the impact on behaviour and concentration in students who consume these drinks.
Natalie: This is a call to action for regulatory bodies to address the sale of energy drinks to young people. Their health and well-being should always be a top priority.
Tyla: 100%. And we often see advertising and marketing for energy drinks that seems to be geared towards young people.
Natalie: That’s true. Energy drinks…and vaping too!
Tyla: Absolutely. A quick ad break now and then we’ll be back with more news.
[Advert]
Welcome back to Safeguarding Soundbites. We have one more story that hits close to home, showing the far-reaching consequences of budget shortfalls. A UK council is on the brink of bankruptcy due to a staggering £60 million gap in special education funding.
Tyla: That’s right, Natalie. This issue is part of a larger problem. Most councils in England have been overspending on special needs and disabilities (SEND) since 2015. This overspending stems from the government’s extension of the age range for young people qualifying for SEND support, without providing councils the necessary funding to meet these new demands.
Natalie: It’s a dire situation. Since 2018, eight councils in England have issued section 114 notices, essentially declaring themselves bankrupt because they lack the resources to balance their budgets.
Tyla: The Local Government Association is calling on the government to take immediate action. They’re urging the government to write off all high-needs deficits and ensure that councils don’t have to cut other essential services to balance their budgets.
Natalie: Let’s hope for swift action to prevent further councils from facing financial crises and to ensure that essential services, including education, receive the support they deserve.
And now over to our Safeguarding Success story of the week! The British Standards Institution has published the first international guidelines on AI safety, with a focus on safeguards. It comes amidst the debate around AI and how to regulate safety on the fast-moving technology.
BSI says the new guidelines are designed to enable the safe, secure and responsible use of artificial intelligence across society.
Tyla: Well, that’s positive. We hear a lot about concerns around AI and how quickly it’s developing.
Natalie: We do and in fact research by BSI shows that around 61% of people in the UK and globally want international guidelines on the use of AI. It also shows that nearly two fifths of people around the world use AI in their jobs every day!
Tyla: Wow, that’s a lot! Listeners, if you want to learn more about AI, in particular chatbots, you can visit our website at ineqe.com or one of our suite of safeguarding apps, including our Safer School Apps.
Natalie: That’s all from us this week, join us next time for more safeguarding news and alerts.
Tyla: Thanks for listening and –
Both: Stay safe.
Join our Online Safeguarding Hub Newsletter Network
Members of our network receive weekly updates on the trends, risks and threats to children and young people online.
Pause, Think
and Plan
Guidance on how to talk to the children in your care about online risks.
Visit the Home Learning Hub!
The Home Learning Hub is our free library of resources to support parents and carers who are taking the time to help their children be safer online.