Prompt hacking is an Achilles’ heel for AI
- "Prompt hacking" is becoming a concern as hackers figure out how to manipulate LLMs to retrieve restricted information
- Outsmarting an LLM in many environments can be done with little to no hacking experience
- New security measures need to be put in place and LLMs themselves will have to adapt
See the full story here; https://www.fierce-network.com/cloud/how-hackable-your-llm
Fake celebrity endorsements become latest weapon in misinformation wars
... Roughly 1 in 10 viral posts analyzed by the News Literacy Project contained fake endorsements, according to data provided exclusively to CNN. Those posts described supposed endorsements - or alternatively, public snubs - from celebrities including NFL quarterback Aaron Rodgers, actor Morgan Freeman, musician Bruce Springsteen, and political figures like former First Lady Michelle Obama. ...
Experts say the problem has been exacerbated by X's AI-powered chatbot, Grok, which has already drawn the ire of election officials for spreading false information about Harris' eligibility in the 2024 election. Last week, X began allowing users to use Grok to create AI-generated images from text prompts, unleashing a flood of fake content about Trump and Harris.
"Going forward, Grok is likely to be one of the main sources of these sorts of images because it generates high-quality images, is easily available, and was intentionally made to have a low refusal rate," Hansen said, adding that he was able to use Grok to create images of "Swifties for Trump" that closely resemble the ones Trump shared. ...
See the full story here: https://abc11.com/post/fake-celebrity-endorsements-become-latest-weapon-misinformation-wars-sowing-confusion-ahead-2024-election/15218719/
TCL Names Finalists for AI TV/Film Accelerator Program
... “This is a pivotal moment of realignment in the industry and TCL is leading the way of demystifying the use of AI tools through our production initiatives,“ TCL North America chief content officer Chris Regina said. ...
See the full story here: https://www.nexttv.com/news/tcl-names-finalists-for-ai-tvfilm-accelerator-program
Inconsistent Safeguards in AI Chatbots Can Lead to Health Disinformation
A study published earlier this year in BMJ evaluated how well large language models (LLMs) could prevent users from prompting chatbots to create health disinformation. It found that while some AI chatbots consistently avoided creating false information, other models frequently created false health claims, especially when prompted with ambiguous or complex health scenarios. In addition, the study found that the safeguards were inconsistent – some models provided accurate information in one instance but not in others under similar conditions. The researchers criticized the lack of transparency from AI developers, who often did not disclose the specific measures they had taken to mitigate these challenges.
Source: Menz, B. D., Kuderer, N. M., Bacchi, S., Modi, N. D., Chin-Yee, B., Hu, T., ... & Hopkins, A. M. (2024). Current safeguards, risk mitigation, and transparency measures of large language models against the generation of health disinformation: repeated cross-sectional analysis. BMJ, 384.
Sony Unveils Web3 Division and Layer 2 Network
...
Announced on Aug. 23, the newly created Sony Block Solutions Labs (Sony SBL) will lead all of Sony’s blockchain and web3 initiatives. Sony also revealed that the division is developing Soneium, a public Ethereum Layer 2 network, in partnership with Startale Labs, the development team behind Astar Network.
The network leveragesOptimism’s OP Stack and supports Ethereum Virtual Machine (EVM) smart contracts. Soneium is currently preparing to launch its forthcoming testnet, and plans to release technical documentation and software development kits “in the coming weeks.”
“The new blockchain unites leading web3 projects and infrastructure pioneers, including Astar Network, Circle, Chainlink, Alchemy, and The Graph… to bridge the gap between decentralized innovation and everyday consumer applications in entertainment, gaming, and finance,” Sony SBL said.
Sony SBL said it is exploring new mechanisms for profit-sharing between creators and fans, protecting creator-generated content, and fostering interoperability across digital and real-world environments. ...
See the full story here: https://thedefiant.io/news/blockchains/sony-unveils-web3-division-and-layer-2-network
An ‘AI Scientist’ Is Inventing and Running Its Own Experiments
... This week, Clune’s lab revealed its latest open-ended learning project: an AI program that invents and builds AI agents. The AI-designed agents outperform human-designed agents in some tasks, such as math and reading comprehension. The next step will be devising ways to prevent such a system from generating agents that misbehave. “It's potentially dangerous,” Clune says of this work. “We need to get it right, but I think it's possible.”
See the full story here: https://www.wired.com/story/ai-scientist-ubc-lab/
Decentralized Web3 AI firm Theoriq joins Google startup accelerator
... Theoriq’s primary product, its AI Agent Base Layer, is a decentralized, blockchain-based platform for developing and managing AI agent collectives. Essentially, it allows developers to deploy AI agents — models specifically designed to complete directed tasks — throughout their Web3 stack. ...
See the full story here: https://cointelegraph.com/news/decentralized-web3-ai-firm-theoriq-joins-google-startup-accelerator
How a Law That Shields Big Tech Is Now Being Used Against It
... Section 230, introduced in the internet’s early days, protects companies from liability related to posts made by users on their sites, making it nearly impossible to sue tech companies over defamatory speech or extremist content. ...
The lawsuit, filed by Ethan Zuckerman, a public policy professor at the University of Massachusetts Amherst, is the first to use Section 230 against a tech giant in this way, his lawyers said. It is an unusual legal maneuver that could turn a law that typically protects companies like Meta on its head. And if Mr. Zuckerman succeeds, it could mean more power for consumers to control what they see online. ...
In 2021, after a developer released software to purge users’ Facebook feeds of everyone they follow, Facebook threatened to shut it down. But Section 230 says it is possible to restrict access to obscene, excessively violent and other problematic content. The language shields companies from liability if they censor disturbing content, but lawyers now say it could also be used to justify scrubbing any content users don’t want to see. ...
So Mr. Barclay, who is now 35, built a browser extension tool the same year that would automate the process, called Unfollow Everything. Roughly 12,000 people tried it, he said.
But on July 1, 2021, a law firm representing Facebook sent Mr. Barclay a cease-and-desist letter. His browser extension violated Facebook’s terms of service, including for “impairing the intended operation of Facebook,” the letter said. It also instructed Mr. Barclay to take down his browser extension or face a potential lawsuit. ...
But he and his lawyers were still looking for a legal argument on which to hang their lawsuit. Preparing for a graduate-level class called “Fixing Social Media” in 2022, Mr. Zuckerman read Section 230 and noticed the provision protecting “technical means” to block objectionable content. ...
Mr. Zuckerman is taking that argument a step further, asking the court to pre-emptively protect an effort to build software that filters content because an internet user simply does not want to see it.
“The purpose of the tool is to allow users who find the newsfeed objectionable, or who find the specific sequencing of posts within their newsfeed objectionable, to effectively turn off the feed,” Mr. Zuckerman’s lawyers said in the lawsuit. ...
See the full story here: https://www.nytimes.com/2024/08/20/technology/meta-section-230-lawsuit.html
Google TV Streamer
The Google TV streamer, a Chromecast replacement, is truly an AI-first device, using Gemini to offer content summaries and screen savers. The Apple TV version, with Apple Intelligence, can’t be far behind. ...
See the full story here: https://www.fastcompany.com/91170817/google-tv-streamer-apple-intelligence-gemini-ai-chromecast-reviews-content-summary
As DNC hits Chicago, Microsoft warns of deepfake artificial-intelligence attacks
...
Badanes says one of the most troubling political deepfake attacks worldwide happened in October in Slovakia just two days before the election for a seat in parliament in the central European country. AI technology was used to create a fake recording of a top political candidate bragging about rigging the election. It went viral. And the candidate lost by a slim margin.
AI also turned up in last year’s Chicago mayoral election. Candidate Paul Vallas, the former Chicago Public Schools chief, was the target of an audio deepfake posted on the social media platform X. In the clip, an artificial but realistic voice purporting to be that of Vallas endorsed rampant police violence, saying: “These days, people will accuse a cop of being bad if they kill one person that was running away. Back in my day, cops would kill, say, 17 or 18 civilians in their career, and nobody would bat an eye.” ...
She says Microsoft’s event at The Drake, 140 E. Walton Place, will be geared toward women, whom she says are disproportionately targeted by deepfakes and also online harassment. The training will focus on spotting deceptive AI content and providing tools to protect against illicit uses of the technology — including how to report a deepfake and how to check whether an image is bogus.
Badanes also will be part of a panel discussion Aug. 21 at the Erie Cafe, 536 W. Erie St., on the intersection of AI and politics, with a focus on regulations to combat deepfakes. ...
“We have a free tool,” Badanes says. “We encourage political campaigns to tag all of their official images and videos with this content-integrity marker.” ...
“There are real-world harms that are happening due to this technology,” Badanes says. “What I’m focusing on at the [DNC] is around the impact it has on elections, but we’re thinking about these harms in a much broader sense.” ...
See the full story here: https://chicago.suntimes.com/the-watchdogs/2024/08/18/deepfake-microsoft-ai-artificial-intelligence-ginny-badanes-content-integrity-marker-slovakia-iran-trump-kamala-harris
Pages
- About Philip Lelyveld
- Mark and Addie Lelyveld Biographies
- Presentations and articles
- Tufts Alumni Bio