(3 weeks, 5 days ago)
Westminster HallWestminster Hall is an alternative Chamber for MPs to hold debates, named after the adjoining Westminster Hall.
Each debate is chaired by an MP from the Panel of Chairs, rather than the Speaker or Deputy Speaker. A Government Minister will give the final speech, and no votes may be called on the debate topic.
This information is provided by Parallel Parliament and does not comprise part of the offical record
I beg to move,
That this House has considered online safety for children and young people.
Just give me one second to get my notes in order, Mr Dowd.
The hon. Lady has called a debate on a really important issue. Could she set out why she thinks that now is a really important time to discuss this vital topic?
I will—and I thank the hon. Lady for her intervention.
It is a pleasure to serve under your chairmanship, Mr Dowd. It is my great honour to open this debate on online safety for our children. I welcome the Minister answering for the Department for Science, Innovation and Technology, and the shadow Minister, the hon. Member for Runnymede and Weybridge (Dr Spencer), answering for the official Opposition. I tabled this as my first debate in Westminster Hall, because I believe this issue is one of the most defining of our time. I promised parents and children in my constituency of Darlington that I would tackle it head-on, so here I am to fulfil that promise.
I would like to put on the record that I have long been inspired by the strength of the parents of Bereaved Families for Online Safety—a group of parents united by the unbearable loss of their children and by their steadfast commitment to get stronger online protections to prevent more children’s deaths. I say to Ellen, who is here with us this afternoon: thank you for your courage—you have experienced unimaginable pain, and I will do everything I can to prevent more parents from going through the same.
The consensus for action on this issue has been built, in no small part due to the incredible drive of parents to campaign for justice. It is felt in every corner of the country, and it is our job as a Government to step in and protect our children from online harm. In my constituency of Darlington, at door after door right across the town and regardless of background, income or voting intention, parents agreed with me that it is time to act to protect our children. I am taking this issue to the Government to fight for them.
I am standing up to amplify the voice of the girl who sends a picture of herself that she thought was private but arrives at school to find that it has been shared with all her peers; she is not only mortified but blamed, and the message cannot be unsent. I am standing up to amplify the voice of the boy who gets bombarded with violent, disturbing images that he does not want to see and never asked for, and who cannot sleep for thinking about them. I am standing up for the mother whose son comes home bruised and will not tell her what has happened, but who gets sent a video of him being beaten up and finds out that it was organised online. I am standing up for the father whose daughter refuses to eat anything because she has seen video after video after video criticising girls who look like her. I say to all those who have raised the alarm, to all the children who know something is wrong but do not know what to do, and to all those who have seen content that makes them feel bad about themselves, have been bullied online, have seen images they did not want to see or have been approached by strangers: we are standing up for you.
I congratulate my hon. Friend on securing this debate on online safety for children and young people. I have a keen personal interest, as a father of two young children. Earlier this year, Ofcom published 40 recommendations about how to improve children’s safety online, including through safer algorithms, and the Government rightly pointed to the role that technology companies can play in that. Does my hon. Friend agree that these companies must take their responsibilities much more seriously?
I absolutely agree that the companies must take those responsibilities seriously, because that will be the law. I am keen that we, as legislators, make sure that the law is as tight as it possibly can be to protect as many children as possible. We will never be able to eradicate everything online, and this is not about innovation. It is about making sure that we get this absolutely right for the next generation and for those using platforms now, so I thank my hon. Friend for his intervention.
The first meeting I called when I was elected the MP for Darlington was with the headteachers of every school and college in my town. I asked them to join together to create a town-wide forum to hear the voices of children and young people on what needs to change about online safety. The first online safety forum took place a couple of weeks ago, and the situation facing young people—year 10s, specifically—is much worse than I had anticipated.
The young people said that online bullying is rife. They said it is common for their peers to send and doctor images and videos of each other without consent, to spread rumours through apps, to track the locations of people in order to bully them through apps, to organise and film fights through apps, to be blackmailed on apps, to speak on games and apps to people they do not know, and to see disturbing or explicit images unprompted and without searching for them. They also said it is common to see content that makes them feel bad about themselves. This has to stop.
The last Government’s Online Safety Act 2023 comes into force in April 2025. The regulator, Ofcom, will publish the children’s access assessments guidance in January 2025. This will give online services that host user-generated content, search services and pornography services in the UK three months to assess whether their services are likely to be accessed by children. From April 2025, when the children’s codes of practice are to be published, those platforms and apps will have a further three months to complete a children’s risk assessment. From 31 July 2025, specific services will have to disclose their risk assessments to Ofcom. Once the codes are approved by Parliament, providers will have to take steps to protect users. There is to be a consultation on the codes in spring 2025, and I urge everybody interested in the topic—no matter their area of expertise or feelings on it—to feed into that consultation. The mechanism for change is in front of us, but my concern is that the children’s codes are not strong enough.
I congratulate my hon. Friend on securing this important debate. Could she comment on the use of artificial intelligence to create child sexual abuse materials? That is a key issue now. Many years ago, I trained with the National Society for the Prevention of Cruelty to Children as a child protection officer, and what I learned back then is that we have to get ahead of all the technologies in order to deal with the challenges effectively. Does she have any thoughts on that point? She may be coming to it in her own remarks.
I thank my hon. Friend for raising that great threat. My area of expertise on the issue is children’s and service users’ voices. There is definitely space for Ofcom and the Government to try to regulate the illegal manufacturing of images through AI. When I asked children in my constituency whether they had ever seen something that they knew was made by AI, they said yes—they had seen images of people that they knew were not real—but the notifications and warnings to tell them that it was AI were not as explicit as they could be. In other words, they could tell for themselves, but the notifications were not comprehensive enough for other children, who may not have noticed. This is a real danger.
There will always be content created online that we cannot police. We have to accept—as we do with any other piece of legislation—that there will be criminal actors, but I have called this debate because there are ways to protect children from harmful content, including by using the right age verification model. I am keen to focus my contribution on trying to protect children from content, in the round, that is harmful to them.
As I said before, the mechanism for change is in front of us, but my concern is that the children’s codes are not strong enough. The children in my town have told me—and I am sure everybody here knows it—that the current age verification requirements are easily passed through, and that content on some sites is deeply disturbing and sent to them without them asking for it. That means that the sites are hosting content that is deeply disturbing for children, and that the age verification is not fit for purpose. We need to talk either about stopping those sites from hosting that content, which is very difficult, or about changing the age verification process.
I want to talk about the scale of the problem that the hon. Lady touches on. The Children’s Commissioner for England reveals that 79% of children under 18 have encountered violent pornography before the age of 18, with the average age of first exposure being 13. Everything the hon. Lady is saying is very important, but this is not a niche problem; it is something that parents in Winchester have spoken to me about repeatedly in the four months since I was elected.
It is indeed prolific, for all our children—the whole generation. It is interesting that, among the different experts I have spoken to, there is consensus; the argument has been won that children are unsafe online and that is affecting them deeply, across the country. It is our job—it falls to legislators—to rectify the issue. I do not wish to defend online platforms, but they will do what the law tells them to do. They want to operate in this country. They want to make money. There is nothing wrong with that; they just have to adhere to the law. It is our job to make sure that the law is tight to protect our children. That is the crux of the issue.
My hon. Friend is powerfully illustrating the responsibility on all of us to step up to the needs of this moment. Parents in my constituency—at schools including William Ransom and Samuel Lucas—have been leading the way in taking further proactive action, signing up to a smartphone-free pledge to delay the age at which their young people have access to smartphones. Hundreds across the constituency have already signed up to the pledge. Does my hon. Friend agree that that underlines the strength of parental feeling on online safety and some of the wider associated issues, and that it highlights our responsibility to legislate—not just to celebrate the benefits of technology, but to do all we can to protect young people from the very real dangers it presents, too?
A smartphone-free pledge is a great idea, and I will take it to Darlington. Parents are further down the line than we are on this; children are further down the line than we are; campaign groups are further down the line than we are. We are lagging behind. We have taken action—the last Government passed the Online Safety Act. I think it is time for us to make sure that there is nothing missing from that Act. In my view, there are some areas where we could go further.
Children in Darlington have said to me that they are getting these unsolicited images—from the algorithms. These images are being fed to them. They are not from strangers, or bogeymen from another country, although that might happen. The most common complaint is that the algorithm is feeding them content that they did not ask for, and it is deeply disturbing, whether it is violent, explicit or harmful. Once they have seen it, they cannot unsee it.
That is why I am arguing to strengthen the codes. I am not sure that we should be retrofitting harmful apps with a code that may or may not work, and having to tweak a few bits of the algorithm to check whether it will actually protect our children. I think we can take stronger action than that.
Numerous mental health charities and a number of civil society experts have raised with me that there are powers within the Online Safety Act that must be used by the regulator. Indeed, the Secretary of State for DSIT made it very clear last week that he backed the Act and those powers. Does my hon. Friend agree that the regulator could and should act with more powers than it has?
I am loath to tell Ofcom that it does not have enough power. As I understand it, the powers are there, but we need to be explicit, and they need to be strengthened. How do we do that? The reason I outlined the timelines is that the time to act is now. We have to explicitly strengthen the children’s codes.
There are many ways to skin a cat, as they say, but one of the simpler ways to do this would be to outline the audience that the apps want to market to. Who is the base audience that the apps and platforms are trying to make money from? If that is explicitly outlined, the codes could be applied accordingly, and strengthened. If children are the target audience, we can question some of the things on those apps and whether the apps are safe for children to use in and of themselves.
With children able to access online content a lot more easily nowadays, many of my Slough constituents feel that it is critical that the content itself is appropriate and safe. Does my hon. Friend share my concerns about the rise of extreme misogynistic content and its impact on young people, especially considering that research has shown that it is actually amplified to teens?
I thank my hon. Friend for raising the really important—indeed, deeply concerning—issue of the rise of anti-women hate, with the perpetrators marketing themselves as successful men.
What we are seeing is that boys look at such videos and do not agree with everything that is said, but little nuggets make sense to them. For me, it is about the relentless bombardment: if someone sees one video like that, they might think, “Oh right,” and not look at it properly, but they are relentlessly targeted by the same messaging over and over again.
That is true not just for misogynistic hate speech, but for body image material. Girls and boys are seeing unrealistic expectations of body image, which are often completely fake and contain fake messaging, but which make them reflect on their own bodies in a negative way, when they may not have had those thoughts before.
I want to drive home that being 14 years old is tough. I am really old now compared with being 14, but I can truly say to anybody who is aged 14 watching this: “It gets better!” It is hard to be a 14-year-old: they are exploring their body and exploring new challenges. Their hormones are going wild and their peers are going through exactly the same thing. It is tough, and school is tough. It is natural for children and young people to question their identity, their role in the world, their sexuality, or whatever it is they might be exploring—that is normal—but I am concerned that that bombardment of unhealthy, unregulated and toxic messaging at a crucial time, when teenagers’ brains are developing, is frankly leading to a crisis.
I return to an earlier point about whether the parts of apps or platforms that children are using are actually safe for them to use. There are different parts of apps that we all use—we may not all be tech-savvy, but we do use them—but when we drill into them and take a minute to ask, “Is this safe for children?”, the answer for me is, “No.”
There are features such as the live location functionality, which comes up a lot on apps, such as when someone is using a maps app and it asks for their live location so they can see how to get from A to B. That is totally fine, but there are certain social media apps that children use that have their live location on permanently. They can toggle it to turn it off, but when I asked children in Darlington why they did not turn it off, they said there is a peer pressure to keep it on—it is seen as really uncool to turn it off. It is also about being able to see whether someone has read a message or not.
I then said to those children, “Okay, but those apps are safe because you only accept people you know,” and they said, “Oh no, I’ve got thousands and thousands of people on that app, and it takes me ages to remove each person, because I can’t remember if I know them, so I don’t do it.” They just leave their location on for thousands of people, many of whom may be void accounts, and they do not even know if they are active any more. The point is that we would not allow our children to go into a space where their location was shown to lots of strangers all the time. Those children who I spoke to also said that the live location feature on some of these apps is leading to in-person bullying and attacks. That is absolutely horrifying.
On that point, is the hon. Member aware that if someone toggles their location off on Snapchat, for example, it constantly—in fact, every time the app is opened—says, “You’re on ghost mode. Do you want to turn your location back on?” So every single time someone opens the app, it tries to convince them to turn their location back on.
I thank the hon. Member for raising that issue, because there are lots of different nudge notifications. We can understand why, because it is an unregulated space and the app is trying to get as much data as possible—if we are not paying for the service, we are the service. We all know that as adults, but the young people and children who we are talking about today do not know that their data is what makes them attractive to that app.
I thank my hon. Friend for allowing me to intervene again. In my previous role as head of public policy at the British Computer Society, the one thing that my colleagues and I talked about a lot was the lack of focus on education in the Online Safety Act. I commend the previous Government for passing that legislation, which was very brave. The Act has tried to do some wonderful things, but what is missing is that we have failed to empower a generation of young people to act safely online, to be able to take back the power and say, “No, I am not going to do that.” We have failed in that so far. How do we build that in for the future?
Order. I would like to bring to the attention of Members that we have had a huge number of interventions and we are 20 minutes into the debate. The Minister and Opposition spokesperson will get up at just after half past 3. It is a matter for the speaker whether she takes more interventions, but that does mean that the amount of time for those who have asked to speak will be significantly more restricted than I originally planned. That is just a housekeeping matter to be aware of. There is also an issue about the length of interventions: they are getting a bit long. On a matter of this importance, I do not want to restrict interventions and contributions, but I ask Members to please bear that in mind.
Okay, I will make progress. On the live location element, which I have discussed, I am not sure that there is any advantage in children using that, unless it is a specifically regulated live location app where the parents have given consent for their child.
I do not know whether chatting to strangers on games is suitable for children. Adding peers to a group and enjoying playing with them on games is fine, but there could be strangers from other countries, with no indication of their age. One child told me that he had found out, after about three weeks, that the person he had been playing with was a 50-year-old man on another continent. That man was probably mortified, as was the child, and they stopped playing together. Why are we leaving it up to them? That is such a high-risk strategy for those apps; we need to think about that.
It is down to Parliament to decide what is safe for our children, and to enforce it. Asking platforms to mark their own homework and police themselves will undoubtedly lead to more children seeing inappropriate, harmful content and sharing it with others. I would like the Government to strengthen the children’s codes, and consider changing the onus from reactive safety measures that make apps safe for children, when we suspect they are children, to proactively making apps or platforms safe for all children in the first place, and creating adult-only apps that require strong age verification, because adults can consent to giving their data.
A number of ways to protect children online are being debated, as I am sure we will hear this afternoon. I feel strongly that retrofitting apps once children have been exposed to harmful content or strangers, or have shared things they should not, is not the safest or most effective way to do this. A number of options around age verification are on the table, but I would like the Government to consider that being a child is tough and that children have a right to make mistakes. The issue is that those mistakes involve mass communications to peers and a permanent digital footprint, because someone has consented, aged 13, to give away their data.
We need to see whether any child can consent to give away their data, and therefore whether apps that identify their audience as children should be allowed to keep data at all. Should children be in chatrooms with strangers across the world? Should children be allowed to share their live location with strangers or people they have accepted as contacts? Should children be allowed to view unregulated livestreams or addictive-by-design content? Those questions have been raised not only by children themselves but by parents and national advocacy charities and leaders in this space. There is a consensus that we have to take action on this issue, so let us make the most of it.
Order. I remind Members that they should bob if they wish to be called in the debate.
Thank you, Mr Dowd. I am grateful to the Minister for her response.
We have had an insightful and cohesive debate, and I thank all Members for their time and expertise. It is clear to me—and, I am sure, to all of us—that innovation has outstripped legislation, leaving our children and young people shouting for help. Crime is organised and exacerbated on these platforms, and the police cannot stop it without our help. Twenty-four-hour access means that content and bullying have caused school refusals, and our educators cannot teach our children without our help.
Children and young people never share everything with their parents, but the sheer quantity of material, along with the functions of content providers, means that parents cannot protect their children without our help. Children’s mental health services are drowning after huge surges in the number of those needing support. Many issues are caused or exacerbated by online platforms, and our NHS cannot get our children well without our help. Today has demonstrated cross-party agreement for action, as well as agreement that this is one of the great issues of our time. We have our consensus, so now let us use it.
Question put and agreed to.
Resolved,
That this House has considered online safety for children and young people.