(2 weeks, 3 days ago)
Westminster HallWestminster Hall is an alternative Chamber for MPs to hold debates, named after the adjoining Westminster Hall.
Each debate is chaired by an MP from the Panel of Chairs, rather than the Speaker or Deputy Speaker. A Government Minister will give the final speech, and no votes may be called on the debate topic.
This information is provided by Parallel Parliament and does not comprise part of the offical record
It is every parent’s worst nightmare to lose a child—imagine losing them and not knowing why they are gone. Ellen Roome is Jools’ mum. She deserves answers but, unbelievably, she is not allowed access to the data that might provide them, which is so wrong.
This petition is for Jools’ law, which would allow parents to have access to their child’s online data in specific circumstances. Jools Sweeney was hugely loved and is greatly missed by his family and community. In actuality, Jools’ law would present a small amendment of no more than 100 words to the Online Safety Act 2023; the amendment and Jools’ law would appear in section 101 of chapter 4, which is titled “Information powers and information notices”.
The Act currently outlines the powers that a senior coroner has in relation to instructing Ofcom to issue a notice to online platforms to provide data in relation to the death of a child. Section 101 of the Act will be amended by clause 122 of the Data (Use and Access) Bill to strengthen the powers that Ofcom has to prevent the deletion of a child’s data when a notice has been given and issued to the regulated platforms that the child has died.
The progress made in this policy area is testament to the parents of children who are no longer with us and to their incredible strength and work. I thank those present and those watching for everything they have done to protect our children. We need to build on this work to allow parents access to data without the need for a second inquest.
A further amendment would allow for Ofcom to be notified as a routine course of action in the event of future tragedies of child suicide or unexplained deaths. That would alleviate the risk of vital answers to parents’ inevitable questions being deleted, and mitigate the reality of it being solely the responsibility of the parents to request the data in those painful early days of grief. The authorities should initiate a data notice in the event of a child’s death to protect those answers from being lost. The amendment, while small in word count, would be transformational to the rights and experiences of bereaved families. We in this place would be hard pressed to find a parent, or indeed anyone touched by the darkness of a child’s death, who would not support the measures.
Ellen Roome is Jools’ mum. She and I have bonded over our shared belief that there can be a future where our children are safe online, and that there must be a future where every child, in every corner of our great country, is protected from online harm. We are bonded by the fact that we will continue to shout loud until that becomes a reality. Ellen has asked me to read her statement about her work to get to this point—her story. That is a great privilege, and I will read her statement in its entirety without taking interventions, as a mark of respect for her incredible work in this area and for all those she speaks for who have experienced such intolerable pain. The work of the Bereaved Families for Online Safety group has already changed the law, and for that they should be immensely proud.
Before I read Ellen’s statement, I must pay tribute to her. We all hope that if we were put in Ellen’s position or faced with her reality, we would stand up and fight for change. The hard truth is that most people cannot, but Ellen Roome is not most people: she is exceptional. Her grit, tenacity and determination to turn her pain into purpose and progress, and to fight for answers for her family, for all those who knew and loved Jools and for those who have found themselves in the same terrible situation, is truly remarkable. More than that, it is Ellen’s warmth, openness and grace that I have been moved by.
These are her own words:
“It only takes one person to make a stand for morality and justice; in this case, that’s me. However, I'm supported by thousands and thousands of people across the world who think it is morally wrong that I am not entitled to see my child’s social media data, which might provide answers as to why my 14-year-old son chose to end his own life.
When I launched the petition, I asked that ‘Parents should have the right to full access to their child’s social media accounts either whilst they are still alive (to protect them) or if they die, as in my case’. It hadn’t crossed my mind that the parents might be the perpetrators. I now understand this could be the case and hope the Online Safety Bill and Ofcom can protect live children online. However, in my case and that of other parents, when the child has died, who are we protecting? The predators on these platforms? Social Media companies? Surely, I should have the right to look for answers to his cause of death. Jools’ young friends struggle to understand why he is no longer here. The ripple effect of his death is felt not only by us as his parents, but also Jools’ friends, teachers, and everyone in his life was so shocked as to why he ended his life – we deserve possible answers or at least to try for answers.
I am his parent, and he is a minor. As a child, he consented to terms and conditions that permitted social media companies to control his online data. I’m unaware of any other legal context in which a 13-year-old can authorise a legal document, such as terms and conditions.
I have always said that I do not know if it was social media that caused my son to end his own life; however, as a parent, I feel I should morally and humanely have the right to that data to give me possible answers as there was nothing offline which seemed to be an issue to Jools. He was not bullied; he was doing well at school and had many friends. There didn’t appear to be body issues, and whilst he didn’t like his floppy hair or chin, we are unaware of anything else that could be of concern. Yes, he had a cheeky side to him, as do a lot of teenagers, but he was a great kid who loved his parents, and his parents loved him VERY MUCH. I fight now for the right to possible answers as to why my son is no longer alive. I have always thought this to be an online challenge gone wrong.
Many MPs feel that the data bill will solve this issue. It won’t help me or other parents who are in the same awful boat as me. The data bill will allow a coroner the right to access this data in future deaths of children, BUT only if the coroner or the police request it. How do we stop future cases where neither the police nor the coroner asked to see this information? This is what happened in Jools’ case.
As a bereaved parent, I was barely breathing myself after the death of Jools, and I was in no fit state to ask or even think of asking the police and or coroner for this information. This could easily happen again with the new data bill. Also, retrospectively, we cannot obtain this information without applying to the High Court for a second inquest. My lawyer has quoted that it will cost me up to £86,000 to hopefully succeed in the high court, but that seems so wrong to have to find this level of legal fees, which is beyond the reach of almost all bereaved parents, to start looking into missing online activity and what was going on. Also, what a waste of legal professionals and staff involved with a new inquest. I’m just asking for data which I feel should be available to me as his parent. However, I’m not allowed to see it, which is wrong.
I hope this will be a good debate. But please remember that as a member of the Bereaved Families for Online Safety group, I represent many other families in the same awful situation and want to try for answers as to why their precious children are no longer alive.
If this had been your child, you would want answers too. I don’t want any other family to be in this hideous position, which will forever affect us all: our family, Jools’ friends, his teachers, everyone in Jools’ life, and their families, forever.”
Those are Ellen’s words. Ellen’s campaign for justice is rare. As a new MP, I may be forgiven for my perceived naivety, but to me Ellen’s campaign poses a binary choice for us—there is no grey area—so I ask that the Minister does everything in her power to help those seeking answers now, whose cases may not be supported through new measures. It is simply wrong that information that may offer clarity and peace to parents who face a new reality without their child is denied them. It is simply wrong that parents who are living in that unenviable reality now face the colossal emotional and financial burden of a second inquest to discover whether the information exists at all. Ellen Roome is Jools’ mum, a campaigner, a leader and a mother, and Ellen Roome is right.
I am very grateful to be able to speak in this debate, which was prompted by Ellen Roome’s petition, although I am extremely sorry that any of us needs to be here at all. I pay tribute to Ellen and all the other families in the Bereaved Families for Online Safety network for their tireless campaigning.
A week before Christmas, I sat in a Committee Room with Ellen and senior representatives from all the major tech firms, including Meta, TikTok, YouTube and Snap. One conversation that morning will stay with me for a very long time—a conversation that I can describe only as harrowing, shocking and deeply depressing. Sitting alongside two other heartbroken parents who have also lost their children because of online harms, Ellen confronted the representatives of TikTok and Instagram, pleading with them to release information that could give her some peace of mind following the death of her beloved son, Jools. There can be nothing worse for a parent than losing a child, but to lose a child and not understand how or why must compound that agony.
Ellen does not know why Jools died. Unlike many other children and young people, he was not being bullied online and did not seem to have any mental health issues. All Ellen wants is to find out what her son was looking at online before he died; it might shed some light on this tragedy that has clearly caused immeasurable grief. It was infuriating to listen to the tech firms’ pathetic excuses that morning about why they could not or would not release the data that Ellen is asking for.
There was—there is—no good reason not to release that data. Jools is no longer with us, so claiming data protection seems frankly pointless. TikTok said that it would be fined for releasing the data, but my question is: by whom? Who is going to press charges against a global tech company for supporting the request of a bereaved mother? Who in their right mind would think that a court case on that point would help anyone?
As we have heard from the hon. Member for Sunderland Central (Lewis Atkinson), some social companies have behaved differently in such cases. It is quite clear, however, why some will not agree to release that data: it is a pathetic attempt to avoid the potential bad publicity that will follow if it becomes clear that Jools’ short life ended after taking part in a social media challenge, which is one possibility. It is about protecting the reputation of those social media companies. It is about the accountants who fear the lawsuits. In short, I suggest it is about money. The absence of humanity, care and compassion in that room before Christmas was palpable and I applaud Ellen for having the courage to come back here today.
I can see no reason why tech companies cannot immediately release the data that these devastated parents are asking for. I fully support Ellen and all the other parents in their attempts to get Jools’ law on the statute books. In the meantime, I plead with Instagram and TikTok to not wait for a legal challenge, but just release the data: find your inner human and do something decent; imagine if it were your child.
Under UK law we have clear, legal processes for handling physical estates after death. It is high time that we establish clear protocols for the digital estates that are left behind, particularly the digital estates left by young people. The law must catch up with the world we are living in. Current provisions, such as Facebook’s legacy contact feature, are not sufficient, because they rely on a child making a decision while they are still alive, often without fully understanding the implications, as has been mentioned. It is also quite possible that, if children were asked whether their parents could have full access to all of their digital online life in the event of their death, they would say no. Without formal, legal access arrangements, parents are left with no way of viewing their children’s account.
I was reading up on that issue in preparation for this debate and I came across some amendments to the Data (Use and Access) Bill that would require those huge providers and tech companies to have a complaints procedure, where parents could appeal to their better nature for the release of the data, but if they were refused it, there would be a proper complaints procedure. Does the hon. Lady think that goes far enough?
No, I do not believe that goes far enough. There should be a legal right to access that data without having to go through any complaints process, particularly at a time when one is struggling with the worst bereavement imaginable.
The petition seeks to address that gap in law and ensure that, in the tragic event of a child’s death, parents have the right to access their child’s account to gain closure, to preserve memories and to ensure that harmful content is removed. I support the addition of Jools’ law into the Online Safety Act, and I urge the Government to do whatever they can to apply it retrospectively for those who have campaigned on this issue.
What Ellen’s family have been through is the absolute worst imaginable, but tens of thousands of families up and down the country are struggling with the impact of social media on their children and teenagers. Those children are addicted to their screens because of the wicked algorithms that lure them in; cowed by bullies who can intimidate them in their own bedrooms late at night; struggling with their body image because they do not look like the influencers they watch; depressed because their lives do not resemble the doctored, airbrushed Instagram image of perfection; and brainwashed by influencers who spew toxic messages through their pages.
The damaging impact of social media on our children is vast. Medical professionals from all disciplines tell us regularly of the harms children are experiencing from hour after hour spent glued to a screen. Their physical health is damaged, their mental health even more so, and even their ability to communicate and socialise with other humans is changing.
Some aspects of how Ofcom has said it will take these matters forward are to be welcomed, but I absolutely agree with the underlying sentiment of the hon. Lady’s comment. Currently, what has been set out does not go anywhere near far enough. As representatives of our communities and of the families who want to do everything possible to keep young people safe from online harm, it is our responsibility to ensure that we are holding Ofcom accountable for being far more ambitious about how it can most creatively and robustly deploy the powers that we are giving it to keep young people safe.
I thank my hon. Friend for his impressive and articulate outlining of the debate so far. Will he join my calls for Ofcom to strengthen the upcoming children’s code and, as the code is not yet published, to use this opportunity to include functionality, a stronger dynamic risk assessment—a live document that will be constantly updated—and the measures that my hon. Friend has laid out for the smaller and riskier platforms?
I concur wholeheartedly. My hon. Friend has been a tireless campaigner on this issue, both in our debate today and throughout the time I have known her—a very short time, but an impressive one none the less. As she rightly points out, the children’s code is a real opportunity to do right by the intentions of the legislation and by the collective ambition that we are discussing today. From my hon. Friend to the children’s commissioner, campaigners on the issue are pretty united about the opportunity that a more ambitious code could deliver for safeguarding young people.
For far too long, we have allowed young people to be exposed to a level of harm online that we would not tolerate in any other aspect of life. It is potentially understandable, but not excusable, that as legislators we are sometimes more comfortable imposing restrictions or acting in areas where we have more direct lived experience, as in the Children’s Wellbeing and Schools Bill or the Tobacco and Vapes Bill. Those are tangible things that we are comfortable and used to voting and making laws on, whereas online harm can sometimes feel a bit more nebulous and a bit tougher. However, that is no excuse not to act. The failure to act is written across the tragedies experienced by so many families across the country and so many campaigners in the room today. We must do better, and we have to make sure that this is the Parliament in which we do.
(2 months ago)
Westminster HallWestminster Hall is an alternative Chamber for MPs to hold debates, named after the adjoining Westminster Hall.
Each debate is chaired by an MP from the Panel of Chairs, rather than the Speaker or Deputy Speaker. A Government Minister will give the final speech, and no votes may be called on the debate topic.
This information is provided by Parallel Parliament and does not comprise part of the offical record
I beg to move,
That this House has considered online safety for children and young people.
Just give me one second to get my notes in order, Mr Dowd.
The hon. Lady has called a debate on a really important issue. Could she set out why she thinks that now is a really important time to discuss this vital topic?
I will—and I thank the hon. Lady for her intervention.
It is a pleasure to serve under your chairmanship, Mr Dowd. It is my great honour to open this debate on online safety for our children. I welcome the Minister answering for the Department for Science, Innovation and Technology, and the shadow Minister, the hon. Member for Runnymede and Weybridge (Dr Spencer), answering for the official Opposition. I tabled this as my first debate in Westminster Hall, because I believe this issue is one of the most defining of our time. I promised parents and children in my constituency of Darlington that I would tackle it head-on, so here I am to fulfil that promise.
I would like to put on the record that I have long been inspired by the strength of the parents of Bereaved Families for Online Safety—a group of parents united by the unbearable loss of their children and by their steadfast commitment to get stronger online protections to prevent more children’s deaths. I say to Ellen, who is here with us this afternoon: thank you for your courage—you have experienced unimaginable pain, and I will do everything I can to prevent more parents from going through the same.
The consensus for action on this issue has been built, in no small part due to the incredible drive of parents to campaign for justice. It is felt in every corner of the country, and it is our job as a Government to step in and protect our children from online harm. In my constituency of Darlington, at door after door right across the town and regardless of background, income or voting intention, parents agreed with me that it is time to act to protect our children. I am taking this issue to the Government to fight for them.
I am standing up to amplify the voice of the girl who sends a picture of herself that she thought was private but arrives at school to find that it has been shared with all her peers; she is not only mortified but blamed, and the message cannot be unsent. I am standing up to amplify the voice of the boy who gets bombarded with violent, disturbing images that he does not want to see and never asked for, and who cannot sleep for thinking about them. I am standing up for the mother whose son comes home bruised and will not tell her what has happened, but who gets sent a video of him being beaten up and finds out that it was organised online. I am standing up for the father whose daughter refuses to eat anything because she has seen video after video after video criticising girls who look like her. I say to all those who have raised the alarm, to all the children who know something is wrong but do not know what to do, and to all those who have seen content that makes them feel bad about themselves, have been bullied online, have seen images they did not want to see or have been approached by strangers: we are standing up for you.
I congratulate my hon. Friend on securing this debate on online safety for children and young people. I have a keen personal interest, as a father of two young children. Earlier this year, Ofcom published 40 recommendations about how to improve children’s safety online, including through safer algorithms, and the Government rightly pointed to the role that technology companies can play in that. Does my hon. Friend agree that these companies must take their responsibilities much more seriously?
I absolutely agree that the companies must take those responsibilities seriously, because that will be the law. I am keen that we, as legislators, make sure that the law is as tight as it possibly can be to protect as many children as possible. We will never be able to eradicate everything online, and this is not about innovation. It is about making sure that we get this absolutely right for the next generation and for those using platforms now, so I thank my hon. Friend for his intervention.
The first meeting I called when I was elected the MP for Darlington was with the headteachers of every school and college in my town. I asked them to join together to create a town-wide forum to hear the voices of children and young people on what needs to change about online safety. The first online safety forum took place a couple of weeks ago, and the situation facing young people—year 10s, specifically—is much worse than I had anticipated.
The young people said that online bullying is rife. They said it is common for their peers to send and doctor images and videos of each other without consent, to spread rumours through apps, to track the locations of people in order to bully them through apps, to organise and film fights through apps, to be blackmailed on apps, to speak on games and apps to people they do not know, and to see disturbing or explicit images unprompted and without searching for them. They also said it is common to see content that makes them feel bad about themselves. This has to stop.
The last Government’s Online Safety Act 2023 comes into force in April 2025. The regulator, Ofcom, will publish the children’s access assessments guidance in January 2025. This will give online services that host user-generated content, search services and pornography services in the UK three months to assess whether their services are likely to be accessed by children. From April 2025, when the children’s codes of practice are to be published, those platforms and apps will have a further three months to complete a children’s risk assessment. From 31 July 2025, specific services will have to disclose their risk assessments to Ofcom. Once the codes are approved by Parliament, providers will have to take steps to protect users. There is to be a consultation on the codes in spring 2025, and I urge everybody interested in the topic—no matter their area of expertise or feelings on it—to feed into that consultation. The mechanism for change is in front of us, but my concern is that the children’s codes are not strong enough.
I congratulate my hon. Friend on securing this important debate. Could she comment on the use of artificial intelligence to create child sexual abuse materials? That is a key issue now. Many years ago, I trained with the National Society for the Prevention of Cruelty to Children as a child protection officer, and what I learned back then is that we have to get ahead of all the technologies in order to deal with the challenges effectively. Does she have any thoughts on that point? She may be coming to it in her own remarks.
I thank my hon. Friend for raising that great threat. My area of expertise on the issue is children’s and service users’ voices. There is definitely space for Ofcom and the Government to try to regulate the illegal manufacturing of images through AI. When I asked children in my constituency whether they had ever seen something that they knew was made by AI, they said yes—they had seen images of people that they knew were not real—but the notifications and warnings to tell them that it was AI were not as explicit as they could be. In other words, they could tell for themselves, but the notifications were not comprehensive enough for other children, who may not have noticed. This is a real danger.
There will always be content created online that we cannot police. We have to accept—as we do with any other piece of legislation—that there will be criminal actors, but I have called this debate because there are ways to protect children from harmful content, including by using the right age verification model. I am keen to focus my contribution on trying to protect children from content, in the round, that is harmful to them.
As I said before, the mechanism for change is in front of us, but my concern is that the children’s codes are not strong enough. The children in my town have told me—and I am sure everybody here knows it—that the current age verification requirements are easily passed through, and that content on some sites is deeply disturbing and sent to them without them asking for it. That means that the sites are hosting content that is deeply disturbing for children, and that the age verification is not fit for purpose. We need to talk either about stopping those sites from hosting that content, which is very difficult, or about changing the age verification process.
I want to talk about the scale of the problem that the hon. Lady touches on. The Children’s Commissioner for England reveals that 79% of children under 18 have encountered violent pornography before the age of 18, with the average age of first exposure being 13. Everything the hon. Lady is saying is very important, but this is not a niche problem; it is something that parents in Winchester have spoken to me about repeatedly in the four months since I was elected.
It is indeed prolific, for all our children—the whole generation. It is interesting that, among the different experts I have spoken to, there is consensus; the argument has been won that children are unsafe online and that is affecting them deeply, across the country. It is our job—it falls to legislators—to rectify the issue. I do not wish to defend online platforms, but they will do what the law tells them to do. They want to operate in this country. They want to make money. There is nothing wrong with that; they just have to adhere to the law. It is our job to make sure that the law is tight to protect our children. That is the crux of the issue.
My hon. Friend is powerfully illustrating the responsibility on all of us to step up to the needs of this moment. Parents in my constituency—at schools including William Ransom and Samuel Lucas—have been leading the way in taking further proactive action, signing up to a smartphone-free pledge to delay the age at which their young people have access to smartphones. Hundreds across the constituency have already signed up to the pledge. Does my hon. Friend agree that that underlines the strength of parental feeling on online safety and some of the wider associated issues, and that it highlights our responsibility to legislate—not just to celebrate the benefits of technology, but to do all we can to protect young people from the very real dangers it presents, too?
A smartphone-free pledge is a great idea, and I will take it to Darlington. Parents are further down the line than we are on this; children are further down the line than we are; campaign groups are further down the line than we are. We are lagging behind. We have taken action—the last Government passed the Online Safety Act. I think it is time for us to make sure that there is nothing missing from that Act. In my view, there are some areas where we could go further.
Children in Darlington have said to me that they are getting these unsolicited images—from the algorithms. These images are being fed to them. They are not from strangers, or bogeymen from another country, although that might happen. The most common complaint is that the algorithm is feeding them content that they did not ask for, and it is deeply disturbing, whether it is violent, explicit or harmful. Once they have seen it, they cannot unsee it.
That is why I am arguing to strengthen the codes. I am not sure that we should be retrofitting harmful apps with a code that may or may not work, and having to tweak a few bits of the algorithm to check whether it will actually protect our children. I think we can take stronger action than that.
Numerous mental health charities and a number of civil society experts have raised with me that there are powers within the Online Safety Act that must be used by the regulator. Indeed, the Secretary of State for DSIT made it very clear last week that he backed the Act and those powers. Does my hon. Friend agree that the regulator could and should act with more powers than it has?
I am loath to tell Ofcom that it does not have enough power. As I understand it, the powers are there, but we need to be explicit, and they need to be strengthened. How do we do that? The reason I outlined the timelines is that the time to act is now. We have to explicitly strengthen the children’s codes.
There are many ways to skin a cat, as they say, but one of the simpler ways to do this would be to outline the audience that the apps want to market to. Who is the base audience that the apps and platforms are trying to make money from? If that is explicitly outlined, the codes could be applied accordingly, and strengthened. If children are the target audience, we can question some of the things on those apps and whether the apps are safe for children to use in and of themselves.
With children able to access online content a lot more easily nowadays, many of my Slough constituents feel that it is critical that the content itself is appropriate and safe. Does my hon. Friend share my concerns about the rise of extreme misogynistic content and its impact on young people, especially considering that research has shown that it is actually amplified to teens?
I thank my hon. Friend for raising the really important—indeed, deeply concerning—issue of the rise of anti-women hate, with the perpetrators marketing themselves as successful men.
What we are seeing is that boys look at such videos and do not agree with everything that is said, but little nuggets make sense to them. For me, it is about the relentless bombardment: if someone sees one video like that, they might think, “Oh right,” and not look at it properly, but they are relentlessly targeted by the same messaging over and over again.
That is true not just for misogynistic hate speech, but for body image material. Girls and boys are seeing unrealistic expectations of body image, which are often completely fake and contain fake messaging, but which make them reflect on their own bodies in a negative way, when they may not have had those thoughts before.
I want to drive home that being 14 years old is tough. I am really old now compared with being 14, but I can truly say to anybody who is aged 14 watching this: “It gets better!” It is hard to be a 14-year-old: they are exploring their body and exploring new challenges. Their hormones are going wild and their peers are going through exactly the same thing. It is tough, and school is tough. It is natural for children and young people to question their identity, their role in the world, their sexuality, or whatever it is they might be exploring—that is normal—but I am concerned that that bombardment of unhealthy, unregulated and toxic messaging at a crucial time, when teenagers’ brains are developing, is frankly leading to a crisis.
I return to an earlier point about whether the parts of apps or platforms that children are using are actually safe for them to use. There are different parts of apps that we all use—we may not all be tech-savvy, but we do use them—but when we drill into them and take a minute to ask, “Is this safe for children?”, the answer for me is, “No.”
There are features such as the live location functionality, which comes up a lot on apps, such as when someone is using a maps app and it asks for their live location so they can see how to get from A to B. That is totally fine, but there are certain social media apps that children use that have their live location on permanently. They can toggle it to turn it off, but when I asked children in Darlington why they did not turn it off, they said there is a peer pressure to keep it on—it is seen as really uncool to turn it off. It is also about being able to see whether someone has read a message or not.
I then said to those children, “Okay, but those apps are safe because you only accept people you know,” and they said, “Oh no, I’ve got thousands and thousands of people on that app, and it takes me ages to remove each person, because I can’t remember if I know them, so I don’t do it.” They just leave their location on for thousands of people, many of whom may be void accounts, and they do not even know if they are active any more. The point is that we would not allow our children to go into a space where their location was shown to lots of strangers all the time. Those children who I spoke to also said that the live location feature on some of these apps is leading to in-person bullying and attacks. That is absolutely horrifying.
On that point, is the hon. Member aware that if someone toggles their location off on Snapchat, for example, it constantly—in fact, every time the app is opened—says, “You’re on ghost mode. Do you want to turn your location back on?” So every single time someone opens the app, it tries to convince them to turn their location back on.
I thank the hon. Member for raising that issue, because there are lots of different nudge notifications. We can understand why, because it is an unregulated space and the app is trying to get as much data as possible—if we are not paying for the service, we are the service. We all know that as adults, but the young people and children who we are talking about today do not know that their data is what makes them attractive to that app.
I thank my hon. Friend for allowing me to intervene again. In my previous role as head of public policy at the British Computer Society, the one thing that my colleagues and I talked about a lot was the lack of focus on education in the Online Safety Act. I commend the previous Government for passing that legislation, which was very brave. The Act has tried to do some wonderful things, but what is missing is that we have failed to empower a generation of young people to act safely online, to be able to take back the power and say, “No, I am not going to do that.” We have failed in that so far. How do we build that in for the future?
Order. I would like to bring to the attention of Members that we have had a huge number of interventions and we are 20 minutes into the debate. The Minister and Opposition spokesperson will get up at just after half past 3. It is a matter for the speaker whether she takes more interventions, but that does mean that the amount of time for those who have asked to speak will be significantly more restricted than I originally planned. That is just a housekeeping matter to be aware of. There is also an issue about the length of interventions: they are getting a bit long. On a matter of this importance, I do not want to restrict interventions and contributions, but I ask Members to please bear that in mind.
Okay, I will make progress. On the live location element, which I have discussed, I am not sure that there is any advantage in children using that, unless it is a specifically regulated live location app where the parents have given consent for their child.
I do not know whether chatting to strangers on games is suitable for children. Adding peers to a group and enjoying playing with them on games is fine, but there could be strangers from other countries, with no indication of their age. One child told me that he had found out, after about three weeks, that the person he had been playing with was a 50-year-old man on another continent. That man was probably mortified, as was the child, and they stopped playing together. Why are we leaving it up to them? That is such a high-risk strategy for those apps; we need to think about that.
It is down to Parliament to decide what is safe for our children, and to enforce it. Asking platforms to mark their own homework and police themselves will undoubtedly lead to more children seeing inappropriate, harmful content and sharing it with others. I would like the Government to strengthen the children’s codes, and consider changing the onus from reactive safety measures that make apps safe for children, when we suspect they are children, to proactively making apps or platforms safe for all children in the first place, and creating adult-only apps that require strong age verification, because adults can consent to giving their data.
A number of ways to protect children online are being debated, as I am sure we will hear this afternoon. I feel strongly that retrofitting apps once children have been exposed to harmful content or strangers, or have shared things they should not, is not the safest or most effective way to do this. A number of options around age verification are on the table, but I would like the Government to consider that being a child is tough and that children have a right to make mistakes. The issue is that those mistakes involve mass communications to peers and a permanent digital footprint, because someone has consented, aged 13, to give away their data.
We need to see whether any child can consent to give away their data, and therefore whether apps that identify their audience as children should be allowed to keep data at all. Should children be in chatrooms with strangers across the world? Should children be allowed to share their live location with strangers or people they have accepted as contacts? Should children be allowed to view unregulated livestreams or addictive-by-design content? Those questions have been raised not only by children themselves but by parents and national advocacy charities and leaders in this space. There is a consensus that we have to take action on this issue, so let us make the most of it.
Order. I remind Members that they should bob if they wish to be called in the debate.
Thank you, Mr Dowd. I am grateful to the Minister for her response.
We have had an insightful and cohesive debate, and I thank all Members for their time and expertise. It is clear to me—and, I am sure, to all of us—that innovation has outstripped legislation, leaving our children and young people shouting for help. Crime is organised and exacerbated on these platforms, and the police cannot stop it without our help. Twenty-four-hour access means that content and bullying have caused school refusals, and our educators cannot teach our children without our help.
Children and young people never share everything with their parents, but the sheer quantity of material, along with the functions of content providers, means that parents cannot protect their children without our help. Children’s mental health services are drowning after huge surges in the number of those needing support. Many issues are caused or exacerbated by online platforms, and our NHS cannot get our children well without our help. Today has demonstrated cross-party agreement for action, as well as agreement that this is one of the great issues of our time. We have our consensus, so now let us use it.
Question put and agreed to.
Resolved,
That this House has considered online safety for children and young people.