Google and YouTube on Thursday announced a new policy that prohibits climate deniers from being able to monetize their content on its platforms via ads or creator payments.
Why it matters: It’s one of the most aggressive measures any major tech platform has taken to combat climate change misinformation.
Details: Google advertisers and publishers, as well as YouTube creators, will be prohibited from making ad revenue off content that contradicts “well-established scientific consensus around the existence and causes of climate change,” the company’s ads team said in a statement.
- “This includes content referring to climate change as a hoax or a scam, claims denying that long-term trends show the global climate is warming, and claims denying that greenhouse gas emissions or human activity contribute to climate change.”
- Ads and monetization will still be allowed to run alongside other climate-related topics, like public debates on climate policy, impacts of climate change, and new research around the issue.
Google said it’s making these changes in response to frustration from advertisers and content creators about their messages appearing alongside climate denialism.
- “Advertisers simply don’t want their ads to appear next to this content. And publishers and creators don’t want ads promoting these claims to appear on their pages or videos,” the company said.
Yes, but: Google often makes changes to its ads policies to reduce misinformation, but this update is notable, given how hard it can be to characterize certain commentary about climate change as denialism or misinformation.
- The tech giant says that when evaluating content against the new policy, “we’ll look carefully at the context in which claims are made, differentiating between content that states a false claim as fact, versus content that reports on or discusses that claim.”
- The company says it has consulted with experts, like representatives of the United Nations Intergovernmental Panel on Climate Change Assessment Reports, to create the policy. The report found that there is “unequivocal” evidence showing that human emissions of greenhouse gases are causing global warming.”
- Google says it will use a combination of automated tools and human review to enforce the new policy.
The big picture: Internet companies have been under increased pressure from climate activists to do more to address climate change denial on their platforms.
- Google on Wednesday unveiled a suite of new tools that give consumers more information so they can choose to cut their greenhouse gas emissions.
- In February, Facebook expanded an online portal meant to counter misinformation about climate change.
Why it matters: Social media platforms have immense reach, and they’ve come under fire from activists and some lawmakers globally for doing too little to thwart the spread of inaccurate content.
What to watch: Google will begin enforcing the new policy next month.
Buried deep in Biden Infrastructure Law: mandatory kill switches on all new cars by 2026
Remember that 2700-page, $1 trillion dollar infrastructure bill that the US government passed back in August? Well, have you read it? Of course we’re joking — we know you haven’t read it. Most of the legislators who voted on it probably haven’t either. Some folks have, though, and they’re finding some pretty alarming things buried in that bill.
One of the most concerning things we’ve heard so far is the revelation that this “infrastructure” bill includes a measure mandating vehicle backdoor kill-switches in every car by 2026. The clause is intended to increase vehicle safety by “passively monitoring the performance of a driver of a motor vehicle to accurately identify whether that driver may be impaired,” and if that sentence doesn’t make your hair stand on end, you’re not thinking about the implications.
Let us spell it out for you: by 2026, vehicles sold in the US will be required to automatically and silently record various metrics of driver performance, and then make a decision, absent any human oversight, whether the owner will be allowed to use their own vehicle. Even worse, the measure goes on to require that the system be “open” to remote access by “authorized” third parties at any time.
The passage in the bill was unearthed by former Georgia Representative Bob Barr, writing over at the Daily Caller. Barr notes correctly that this is a privacy disaster in the making. Not only does it make every vehicle a potential tattletale (possibly reporting minor traffic infractions, like slight speeding or forgetting your seat-belt, to authorities or insurance companies), but tracking that data also makes it possible for bad actors to retrieve it.
More pressing than the privacy concerns, though, are the safety issues. Including an automatic kill switch of this sort in a machine with internet access presents the obvious scenario that a malicious agent could disable your vehicle remotely with no warning. Outside that possible-but-admittedly-unlikely idea, there are all kinds of other reasons that someone might need to drive or use their vehicle while “impaired”, such as in the case of emergency, or while injured.
Even if the remote access part of the mandate doesn’t come to pass, the measure is still astonishingly short-sighted. As Barr says, “the choice as to whether a vehicle can or cannot be driven … will rest in the hands of an algorithm over which the car’s owner or driver have neither knowledge or control.” Barr, a lawyer himself, points out that there are legal issues with this whole concept, too. He anticipates challenges to the measure on both 5th Amendment (right to not self-incriminate) and 6th Amendment (right to face one’s accuser) grounds. He also goes on to comment on the vagueness of the legislation. What exactly is “impaired driving”? Every state and many municipalities have differing definitions of “driving while intoxicated.”
Furthermore, there’s also no detail in the legislation about who should have access to the data collected by the system. Would police need a warrant to access the recorded data? Would it be available to insurance companies or medical professionals? If someone is late on their car payment, can the lender remotely disable the vehicle? Certainly beyond concerns of who would be allowed official access, there’s also once again the ever-present fear of hackers gaining access to the data—which security professionals well know, absolutely will happen, sooner or later. As Barr says, the collected data would be a treasure trove of data to “all manner of entities … none of which have our best interests at heart.”
Microsoft employees say hello by pronouns and race
Facebook plans to shut down its facial recognition program
- Meta, the company formerly known as Facebook, on Tuesday announced it will be putting an end to its face recognition system.
- The company said it will delete more than 1 billion people’s individual facial recognition templates as a result of this change.
- Facebook services that rely on the face recognition systems will be removed over the coming weeks, Meta said.
Facebook on Tuesday announced it will be putting an end to its facial recognition system amid growing concern from users and regulators.
The social network, whose parent company is now named Meta, said it will delete more than 1 billion people’s individual facial recognition templates as a result of this change. The company said in a blog post that more than a third of Facebook’s daily active users, or over 600 million accounts, had opted into the use of the face recognition technology.
Facebook will no longer automatically recognize people’s faces in photos or videos, the post said. The change, however, will also impact the automatic alt text technology that the company uses to describe images for people who are blind or visually impaired. Facebook services that rely on the face recognition systems will be removed over the coming weeks.
“There are many concerns about the place of facial recognition technology in society, and regulators are still in the process of providing a clear set of rules governing its use,” the company said. “Amid this ongoing uncertainty, we believe that limiting the use of facial recognition to a narrow set of use cases is appropriate.”
Ending the use of the face recognition system is part of “a company-wide move away from this kind of broad identification,” the post said.
Meta, which laid out its road map last week for the creation of a massive virtual world, said it will still consider facial recognition technology for instances where people need to verify their identity or to prevent fraud and impersonation. For future uses of facial recognition technology, Meta will “continue to be public about intended use, how people can have control over these systems and their personal data.”
The decision to shut down the system on Facebook comes amid a barrage of news reports over the past month after Frances Haugen, a former employee turned whistleblower, released a trove of internal company documents to news outlets, lawmakers and regulators.