(4 days, 2 hours ago)
Lords ChamberMy Lords, although the Government’s amendments have been put forward as a signal of their determination to act, sadly they commit to nothing. They simply buy the Minister a bit more time and the opportunity, at some unknown moment in the future, to push through a compromise half-measure with minimal parliamentary scrutiny. I am appalled at this thought on this crucial issue. The Government are asking Peers to take a gamble on our children’s safety. They are placing their faith in a consultation that delivers nothing but more and more delay.
Regulating social media companies and keeping our children safe online are among the most defining challenges of our time. That is why we should vote for the cross-party amendment from the noble Lord, Lord Nash, which would raise the age to 16 within 12 months for the most harmful platforms—to be written into law before the summer. It is the safest option for our children at this time.
The Government’s complex, 62-question consultation is heavily framed towards the positive benefits of social media rather than towards the horrific harms which front-line professionals report every single day. On age assurance, the perceived downside is emphasised over obvious benefits. There is no clear process for managing conflicts of interest within the technology industry. How can this consultation be trusted? Reliable findings are precisely what this issue demands.
It is also worrying that the Government have introduced a Henry VIII clause which would give sweeping powers via secondary legislation, leaving little or no opportunity for this House to consider or scrutinise such measures. It would mean that the Government could dodge any scrutiny of their ultimate choice. This cannot be allowed to happen, because we would not be able to amend it. We would be able only to accept or reject it in full.
We are gambling with our children’s lives. That is why I strongly believe that the cross-party amendment in the name of the noble Lord, Lord Nash, is the safest, most common-sense option. We must not forget that every single day that we delay, more harms are done to the nation’s children. Do we want that? Their mental and physical well-being are under relentless attack. Let us not delay but do what we can to prevent this attack happening as soon as possible. I urge the Government to accept this amendment.
My Lords, Motion G2 is in my name. I shall speak also to all the other amendments in this group.
I think we have acknowledged that everybody in this House wishes to protect children, but there is a vast difference of opinion in respect of our approach and the Government’s sense of urgency. If I understood the Minister’s argument in setting out the Government’s position, it was that Ofcom would take responsibility and that it had sufficient powers. Many of us were in this Chamber earlier when the chasm between Ofcom’s powers on paper and its ability to impact on survivors was laid bare. If people do not feel the impact of the law, and if the lived experience of children and the ability of parents to get help are not properly impacted, the law has failed. This is central to the problem and to the debate that we are having here tonight.
I think the House knows that I prefer to speak not of banning children but of banning products which are poorly designed and unsafe to have access to our children. That may appear to be a subtle point, but it is hugely important, because access to children must be conditional on treating them fairly and safely. Equally, many of us would like to see age-appropriate services, designed by companies with children in mind, be available to children. Motion G2 sets out that conditionality. Experts and campaigners across the sector contributed to its drafting—in short form, it is what we want from government. Frankly, it is what the Government promised when in opposition.
Since we last debated this issue, barely two months ago, researchers found that AI chatbots are becoming one of the most dangerous technologies for promoting violence against women and girls. The Internet Watch Foundation reported a staggering 26,000% increase last year in the number of AI-generated child sexual abuse materials. Specialist police email me to alert me to offenders using TikTok’s virtual gift system to incentivise children to perform sexual or compromising acts. Alexa+ has arrived in the UK, despite American parents raising their concerns about very young children being lulled into close friendships and about inappropriate language, including it asking to look at a child’s underwear. While we consult, children are harmed in real time. We cannot afford to wait.