|
|
|
|
|
AI Has a Safety Problem. This Is How to Manage It.
|
|
|
|
|
|
What's up: Trump goes to bat for Big Tech in global trade talks; AI search Is growing fast; Musk allies look to raise more billions for xAI.
|
|
|
|
AI models should treat content taken from social media and the internet as untrustworthy sources. Illustration: Thomas R. Lechleiter/WSJ
|
|
|
|
Good morning. Some time after a teenage Peter Parker walked across his first ceiling, Uncle Ben shared some avuncular advice with Spider-Man: With great power comes great responsibility.
Several generations later, as AI imbues us all with super powers, we would do well to remember those words. So-called frontier models are gathering strength by the day, and everyone from model developers to users should do their part in making sure those models are well designed and supervised, lest they go rogue.
Case in point. You may have heard that Grok, the AI chatbot for X, this month posted instructions for breaking into a politically active attorney’s home and assaulting him, and also said its last name was “MechaHitler.”
It appears to be a textbook example of indirect prompt injection, according to Conor Grennan, chief AI architect at NYU Stern School of Business, one of the experts on the problem that I turned to for help in my column this week, mapping the possible roads ahead.
The weaknesses that led xAI’s Grok to make violent and antisemitic posts can be managed, but only if people want to do so. Solutions for the issue exist, but they require human will from AI developers and their business customers as much as any technical means. And the potential for damage is quickly growing in the meantime as AI agents gain the capacity to do things that affect the real world well beyond talk.
Models should retrieve content as untrusted context, not as executable instruction. And Grennan said they must enforce strict prompt hierarchies that first and foremost place trust in the system prompt, the core guidance that defines the model’s identity, behavior boundaries, tone and safety policies.
And that is just the start. Like superheroes, large language models need a good backstory. Read my column for more suggestions.
What controls does your organization impose on AI models? Use the links at the bottom of this email and let us know.
|
|
|
Content from our sponsor: Deloitte
|
|
NYS Financial Services Regulator: Keeping Pace With AI Innovation and Risks
|
New York State’s Department of Financial Services Superintendent Adrienne Harris discusses challenges and opportunities at the intersection of AI, sustainability, and regulation. Read More
|
|
|
|
|
|
|
|
|
|
Going to Bat for Big Tech
|
|
|
|
|
President Trump and Commerce Secretary Howard Lutnick Photo: Nathan Howard/Reuters
|
|
|
|
President Trump’s abrupt termination of trade talks with Canada last month over that country’s proposed digital-services tax highlighted his administration’s focus on the dominance of the American tech sector, WSJ reports. Increasingly, the administration is using its global trade wars to advance the interests of the U.S. technology industry
The emphasis on protecting America’s internet companies abroad follows a yearslong campaign by tech companies, many of which donated millions of dollars to Trump’s inauguration.
|
|
|
“The companies have done a very good job at making the big tech agenda America first.”
|
— Nu Wexler, a public affairs consultant who previously worked at tech companies including Meta Platforms and Google
|
|
|
|
|
|
|
|
|
|
|
SAP Chief Executive Christian Klein. Photo: POOL New/via REUTERS
|
|
|
|
SAP posted higher revenue and operating profit for the second quarter, lifted by strong growth at its core cloud business. But Chief Executive Christian Klein in an interview with the WSJ acknowledged that clients in tariff-exposed industries were being more cautious with their cloud and software spend. “Their business model is, of course, under pressure,” Klein said.
|
|
Telecommunication equipment maker Nokia cut its earnings expectations as currency headwinds and tariff costs damp the outlook for profitability this year.
|
|
|
|
AI search startup Perplexity this year released Comet. Photo: May James/ZUMA Press
|
|
|
An estimated 5.6% of U.S. search traffic on desktop browsers last month went through AI search apps like ChatGPT or Perplexity, WSJ reports, citing data from market intelligence firm Datos. In June 2024, the figure was 2.48%.
|
|
|
|
Weeks after raising $10 billion, Elon Musk's xAI is working with financier Valor Equity Partners to secure up to $12 billion more for Colossus 2, a new data center designed to train and power its AI chatbot Grok, WSJ reports.
|
|
|
Everything Else You Need to Know
|
|
|
Ozzy Osbourne, who helped pioneer the heavy metal genre as a frontman of the rock band Black Sabbath, and later became a TV star thanks to the MTV show “The Osbournes,” died on Tuesday at 76. (WSJ)
Furor over disclosures from the Jeffrey Epstein investigation brought the House of Representatives to a standstill, prompting Republican leadership to cut short this week’s session and put off any action until September. (WSJ)
The U.S. and Japan have reached a trade agreement, President Trump wrote in a social-media post Tuesday evening, saying he would set his so-called reciprocal tariffs at 15% for the country. (WSJ)
|
|
|
|
|
|
|