AI

This Week in AI: AI isn’t world-ending — but it’s still plenty harmful

Comment

Robot concept or robot hand chatbot pressing computer keyboard enter
Image Credits: sompong_tom (opens in a new window) / Getty Images

Hiya, folks, welcome to TechCrunch’s regular AI newsletter.

This week in AI, a new study shows that generative AI really isn’t all that harmful — at least not in the apocalyptic sense.

In a paper submitted to the Association for Computational Linguistics’ annual conference, researchers from the University of Bath and University of Darmstadt argue that models like those in Meta’s Llama family can’t learn independently or acquire new skills without explicit instruction.

The researchers conducted thousands of experiments to test the ability of several models to complete tasks they hadn’t come across before, like answering questions about topics that were outside the scope of their training data. They found that, while the models could superficially follow instructions, they couldn’t master new skills on their own.

“Our study shows that the fear that a model will go away and do something completely unexpected, innovative and potentially dangerous is not valid,” Harish Tayyar Madabushi, a computer scientist at the University of Bath and co-author on the study, said in a statement. “The prevailing narrative that this type of AI is a threat to humanity prevents the widespread adoption and development of these technologies, and also diverts attention from the genuine issues that require our focus.”

There are limitations to the study. The researchers didn’t test the newest and most capable models from vendors like OpenAI and Anthropic, and benchmarking models tends to be an imprecise science. But the research is far from the first to find that today’s generative AI tech isn’t humanity-threatening — and that assuming otherwise risks regrettable policymaking.

In an op-ed in Scientific American last year, AI ethicist Alex Hanna and linguistics professor Emily Bender made the case that corporate AI labs are misdirecting regulatory attention to imaginary, world-ending scenarios as a bureaucratic maneuvering ploy. They pointed to OpenAI CEO Sam Altman’s appearance in a May 2023 congressional hearing, during which he suggested — without evidence — that generative AI tools could go “quite wrong.”

“The broader public and regulatory agencies must not fall for this maneuver,” Hanna and Bender wrote. “Rather we should look to scholars and activists who practice peer review and have pushed back on AI hype in an attempt to understand its detrimental effects here and now.”

Theirs and Madabushi’s are key points to keep in mind as investors continue to pour billions into generative AI and the hype cycle nears its peak. There’s a lot at stake for the companies backing generative AI tech, and what’s good for them — and their backers — isn’t necessarily good for the rest of us.

Generative AI might not cause our extinction. But it’s already harming in other ways — see the spread of nonconsensual deepfake porn, wrongful facial recognition arrests and the hordes of underpaid data annotators. Policymakers hopefully see this too and share this view — or come around eventually. If not, humanity may very well have something to fear.

News

Google Gemini and AI, oh my: Google’s annual Made By Google hardware event took place Tuesday, and the company announced a ton of updates to its Gemini assistant — plus new phones, earbuds and smartwatches. Check out TechCrunch’s roundup for all the latest coverage.

AI copyright suit moves forward: A class action lawsuit filed by artists who allege that Stability AI, Runway AI and DeviantArt illegally trained their AIs on copyrighted works can move forward, but only in part, the presiding judge decided on Monday. In a mixed ruling, several of the plaintiffs’ claims were dismissed while others survived, meaning the suit could end up at trial.

Problems for X and Grok: X, the social media platform owned by Elon Musk, has been targeted with a series of privacy complaints after it helped itself to the data of users in the European Union for training AI models without asking people’s consent. X has agreed to stop EU data processing for training Grok — for now.

YouTube tests Gemini brainstorming: YouTube is testing an integration with Gemini to help creators brainstorm video ideas, titles and thumbnails. Called Brainstorm with Gemini, the feature is currently available only to select creators as part of a small, limited experiment.

OpenAI’s GPT-4o does weird stuff: OpenAI’s GPT-4o is the company’s first model trained on voice as well as text and image data. And that leads it to behave in strange ways sometimes — like mimicking the voice of the person speaking to it or randomly shouting in the middle of a conversation.

Research paper of the week

There are tons of companies out there offering tools they claim can reliably detect text written by a generative AI model, which would be useful for, say, combating misinformation and plagiarism. But when we tested a few a while back, the tools rarely worked. And a new study suggests the situation hasn’t improved much.

Researchers at UPenn designed a dataset and leaderboard, the Robust AI Detector (RAID), of over 10 million AI-generated and human-written recipes, news articles, blog posts and more to measure the performance of AI text detectors. They found the detectors they evaluated to be “mostly useless” (in the researchers’ words), only working when applied to specific use cases and text similar to the text they were trained on.

“If universities or schools were relying on a narrowly trained detector to catch students’ use of [generative AI] to write assignments, they could be falsely accusing students of cheating when they are not,” Chris Callison-Burch, professor in computer and information science and a co-author on the study, said in a statement. “They could also miss students who were cheating by using other [generative AI] to generate their homework.”   

There’s no silver bullet when it comes to AI text detection, it seems — the problem’s an intractable one.

Reportedly, OpenAI itself has developed a new text-detection tool for its AI models — an improvement over the company’s first attempt — but is declining to release it over fears it might disproportionately impact non-English users and be rendered ineffective by slight modifications in the text. (Less philanthropically, OpenAI is also said to be concerned about how a built-in AI text detector might impact perception — and usage — of its products.)

Model of the week

Generative AI is good for more than just memes, it seems. MIT researchers are applying it to flag problems in complex systems like wind turbines.

A team at MIT’s Computer Science and Artificial Intelligence Lab developed a framework, called SigLLM, that includes a component to convert time-series data — measurements taken repeatedly over time — into text-based inputs a generative AI model can process. A user can feed these prepared data to the model and ask it to start identifying anomalies. The model can also be used to forecast future time-series data points as part of an anomaly-detection pipeline. 

The framework didn’t perform exceptionally well in the researchers’ experiments. But if its performance can be improved, SigLLM could, for example, help technicians flag potential problems in equipment like heavy machinery before they occur.

“Since this is just the first iteration, we didn’t expect to get there from the first go, but these results show that there’s an opportunity here to leverage [generative AI models] for complex anomaly detection tasks,” Sarah Alnegheimish, an electrical engineering and computer science graduate student and lead author on a paper on SigLLM, said in a statement.

Grab bag

OpenAI upgraded ChatGPT, its AI-powered chatbot platform, to a new base model this month — but released no changelog (well, barely a changelog).

So what to make of it? What can one make of it, exactly? There’s nothing to go on but anecdotal evidence from subjective tests.

I think Ethan Mollick, a professor at Wharton studying AI, innovation and startups, had the right take. It’s hard to write release notes for generative AI models because the models “feel” different in one interaction to the next; they’re largely vibes-based. At the same time, people use — and pay for — ChatGPT. Don’t they deserve to know what they’re getting into?

It could be the improvements are incremental, and OpenAI believes it’s unwise for competitive reasons to signal this. Less likely is the model relates somehow to OpenAI’s reported reasoning breakthroughs. Regardless, when it comes to AI, transparency should be a priority. There can’t be trust without it — and OpenAI has lost plenty of that already.

More TechCrunch

Cybersecurity giant Palo Alto Networks is getting a lot of grief for a recent trade show event in which two women posed with lampshades on their heads. The debacle —…

Palo Alto Networks CEO apologizes for happy hour display featuring women with lampshades on their heads

Hiya, folks, welcome to TechCrunch’s regular AI newsletter. This week in AI, a new study shows that generative AI really isn’t all that harmful — at least not in the…

This Week in AI: AI isn’t world-ending — but it’s still plenty harmful
Image Credits: sompong_tom (opens in a new window) / Getty Images

Popular iOS pro photography app Halide launched its new version today with a new feature called Process Zero, which does not use AI in image processing. Lux Optics, the company…

Camera app Halide’s latest update adds an option for ‘zero-AI’ image processing

Definity focuses on the data transformation plane on top of a data lake or warehouse, not the data ingestion part of the pipeline.

Definity raises $4.5M as it looks to transform data application observability

Analytics and AI giant Databricks reportedly paid nearly $2 billion when it acquired Tabular in June, a startup that was only doing $1 million in annual recurring revenue, according to…

Databricks reportedly paid $2 billion in Tabular acquisition

Apple’s exclusive access to the iPhone’s NFC capabilities had been under investigation by the European Commission for years.

Apple opens up NFC transactions to developers, but says there will be ‘associated fees’

Stoke Space is nothing if not ambitious. The five-year-old launch startup has generated a lot of hype due to its bold plans to develop the first fully reusable rocket, with…

Stoke Space’s initial launch plans at Cape Canaveral take shape

Telegram announced on Wednesday that it’s adding new ways for creators to make money on its platform. Most notably, the platform is launching monthly paid subscriptions that users can purchase…

Telegram adds new ways for creators to earn money on its  platform

A Texas company says it lost $60 million to a criminal fraud scheme, which the FBI says makes fraudsters billions of dollars every year.

Texas firm says it lost $60M in a bank wire transfer scam

Software as a service (SaaS) is an ever-evolving industry. We’ll talk to some of the brightest minds and leaders in the industry — executives from early- and late-stage SaaS companies,…

Announcing the final agenda for the SaaS Stage at TechCrunch Disrupt 2024

What is the right way to build a software business? Many startup advisers say that B2B software should solve one pain point, gain customers, then add features as their company…

Parker Conrad says founders have been building software wrong for the last 20 years

Virtuix’s timeline has coincided with a rise of interest around mixed reality, led by Oculus/Meta, HTC and now Apple, among others.

Virtuix’s VR treadmill is finally launching in September

London-based Roto VR’s spinning gaming chair is the first of its kind to boast a “Made for Meta” seal of approval.

Check out this $800 rotating VR chair for Meta Quest

EliseAI employs an army of chatbots to text with, email, and respond to calls from renters about things such as apartment tours, maintenance requests, lease renewals and delinquencies.

EliseAI lands $75M for chatbots that help property managers deal with renters

In crafting laws to regulate AI, like the EU AI Act or California’s SB 1047, policymakers have struggled to come to a consensus on which risks the laws should cover.

MIT researchers release a repository of AI risks

Kiteworks, which builds tools to secure email communications and file sharing, has raised $456 million from Insight Partners and Sixth Street Growth.

Kiteworks captures $456M at a $1B+ valuation to help secure sensitive data

Hadrian announced they bought Datum Source, a software company founded by SpaceX alums that uses AI to help hardware companies find manufacturing partners.

The defense tech acquisition spree has begun: Autonomous factory startup Hadrian acquires Datum Source

Spotify will be able to display the pricing for things like Spotify subscriptions and digital goods, including Spotify’s more recently added collection of audiobooks.

Apple finally allows Spotify to show pricing info to EU users on iOS

India’s Supreme Court has cleared the way for insolvency proceedings to be resumed against Byju’s in a win for U.S. creditors.

India’s top court clears way for Byju’s insolvency proceedings

Elon Musk-owned X launched Grok-2 and Grok-2 mini in beta today with improved reasoning. The new Grok AI model can now generate images on the X social network, though Grok…

xAI releases Grok-2, adds image generation on X

Google Pixel 9 series India launch coincides with the expansion of its sales channels and after-sales support in the country.

Google faces headwinds as it brings Pixel 9 to India

General Catalyst and Mars Growth Capital are co-leading the Series G round, which will be closed within a few days, sources familiar with the deal told TechCrunch.

Zepto raises $340M at a $5B valuation as India’s quick-commerce market heats up

Let’s dive right into what the Google Pixel 9 lineup looks like, how Google’s Gemini AI will be incorporated in the devices, and more.

Made by Google 2024: All of Google’s reveals, from the Pixel 9 lineup to Gemini AI’s addition to everything

We rounded up some of the more intriguing AI-related announcements that didn’t get a ton of play, like Pixel Studio.

Made by Google 2024: A few AI features you might’ve missed

Ben Affleck and Matt Damon have acquired a screenplay called “Killing Gawker,” which presumably delves into billionaire VC Peter Thiel’s campaign to bury the media outfit for posting excerpts from…

Thiel’s Gawker takedown could be coming to a theater near you

Google launched Gemini Live during its Made by Google event Tuesday. The feature allows you to have a semi-natural spoken conversation, not typed out, with an AI chatbot powered by…

Gemini Live first look: Better than talking to Siri, but worse than I’d like

Texas filed a lawsuit Tuesday against GM over years of alleged abuse of customers’ data and trust. New car owners were presented with a “confusing and highly misleading” process that…

Texas sues GM, saying it tricked customers into sharing driving data sold to insurers

Chinese autonomous vehicle company WeRide has received the green light to test its driverless vehicles with passengers in California.  The step comes as WeRide begins the process to go public…

Chinese robotaxi startup WeRide gets approval to carry passengers in California 

Kristen Faulkner astonishing Olympic success of two gold medals stems from lessons learned from her former career as a venture capitalist, she says.

Winning a gold medal is a lot like being a VC, according to Olympic champion Kristen Faulkner

SB 1047 has drawn the ire of Silicon Valley players large and small, including venture capitalists, big tech trade groups, researchers and startup founders.

California AI bill SB 1047 aims to prevent AI disasters, but Silicon Valley warns it will cause one