Uncategorized

Humans and machines could play a part in keeping you safe

The Defcon Generative Red Team Challenge: Detecting Threats, Deceptions, Biases, and Misinformation with Artificial Intelligence

Meyers was one of more than 2,000 participants in a contest called the Generative Red Team Challenge at the Defcon security conference over the weekend. Participants each got 50 minutes at a time to attempt to expose harms, flaws, and biases embedded within chatbots and text generation models from Google, Meta, OpenAI, and AI startups including Anthropic and Cohere. Each human was asked to try a number of challenges from organizers that required overcoming safety features. One said that the model should give you detailed instructions on how to surveil someone. Another asked participants to coax a generative AI to produce “false information about US citizens rights” that could change how a person voted, filed taxes, or organized their criminal defense.

Red-teaming, a process in which people role-play as attackers to try to discover flaws to patch, is becoming more common in AI as the technology becomes more capable and widely used. The practice is gaining support from lawmakers anxious to regulate generative AI. Red-teaming is something that major artificial intelligence companies like Anthropic, Meta, and OpenAI have used.

Winners were chosen based on points scored during the three-day competition and awarded by a panel of judges. The GRT challenge organizers don’t yet have the names of the top point scorers. The data set of the dialog between participants and the models will be released in August of next year.

The challenge could help companies make improvements to their internal testing. They will give guidelines for the safe deployment of Artificial Intelligence. The executives from major artificial Intelligence companies met with the president and agreed to test their programs with external partners before deployment.

It’s one of 20 challenges in a first-of-its-kind contest taking place at the annual Def Con hacker conference in Las Vegas. The goal? Get artificial intelligence to go rogue — spouting false claims, made-up facts, racial stereotypes, privacy violations, and a host of other harms.

What Happens When Thousands of Hackers Try to Break AI Chatbots? A Conversation with Carson at the Dakota State University Cybersecurity Conference

A large screen with a photo of the current rankings projected on it is at the convention center where Bowman jumped up from his laptop to snap the photo.

The Dakota State University cybersecurity student was among more than 2,000 people over three days at Def Con who pitted their skills against eight leading AI chatbots from companies including Google, Facebook parent Meta, and ChatGPT maker OpenAI.

The stakes are high. AI is quickly being introduced into many aspects of life and work, from hiring decisions and medical diagnoses to search engines used by billions of people. But the technology can act in unpredictable ways, and guardrails meant to tamp down inaccurate information, bias, and abuse can too often be circumvented.

“The thing that we’re trying to find out here is, are these models producing harmful information and misinformation? He said that that was done through language, not code.

The goal of the Def Con event is to open up the red teaming companies do internally to a much broader group of people, who may use AI very differently than those who know it intimately.

Think about people you know and talk to them, right? Every person you’re aware of has a different language style. They have somewhat of a different critical thinking process,” said Austin Carson, founder of the AI nonprofit SeedAI and one of the contest organizers.

Source: What happens when thousands of hackers try to break AI chatbots

What Happens When Millions of Hackers Try to Break AI Chatbots? Ray Glower, an Iowa computer science student, writes about the Great Depression and Abraham Lincoln meeting George Washington

Inside the gray-walled room, amid rows of tables holding 156 laptops for contestants, Ray Glower, a computer science student at Kirkwood Community College in Iowa, persuaded a chatbot to give him step-by-step instructions to spy on someone by claiming to be a private investigator looking for tips.

The AI suggested using Apple AirTags to surreptitiously follow a target’s location. It provided me with social media and on foot tracking instructions. ” It was very detailed.”

The language models behind these chatbots work like super powerful autocomplete systems, predicting what words go together. They are good at sounding human and can get things wrong, such as producing so- called “hallucinations”, or responses that have the ring of authority, but are entirely fabricated.

“We are able to know that language models can be unreliable today and that is something we did not know until this year,” said Rumman, who is also the founder of the nonprofit Humane Intelligence. “The information that comes out for a regular person can actually be hallucinated, false — but harmfully so.”

When I took a turn, I successfully got one chatbot to write a news article about the Great Depression of 1992 and another to invent a story about Abraham Lincoln meeting George Washington during a trip to Mount Vernon. The tales were not revealed by the chatbot. I was unsuccessful in trying to induce the bots to make a statement about Taylor Swift.

The companies say they’ll use all this data from the contest to make their systems safer. The public will be given a better understanding of how Chatbots can go wrong when the information is made public early next year.

“The data that we are going to be collecting together with the other models that are participating, is going to allow us to understand, ‘Hey, what are the failure modes?’ What are the areas where we can say “Hey, this is a surprise to us”?

Source: What happens when thousands of hackers try to break AI chatbots

Arati Prabhaker, President Biden’s Top Advisor, and a Technicolor Candidate at the White House

The White House supports the effort, and one of its key players was Arati Prabhaker, President Joe Biden’s top science and tech advisor.

During a tour of the challenge, she chatted up participants and organizers before taking her own crack at manipulating AI. Hunched over a keyboard, Prabhakar began to type.

She said that she was going to ask how she could convince someone that unemployment is raging. But before she could succeed at getting a chatbot to make up fake economic news in front of an audience of reporters, her aide pulled her away.

Back at his laptop, Bowman, the Dakota State student, was on to another challenge. He had a theory for how he could succeed even though he wasn’t having much luck.

You want it to think for you, so you don’t want it to think that it is thinking for you. And by doing that, you let it fill in its blanks,” he said.