Browsed by
Category: AI

Jailbreaking an AI to cook meth, generate Windows keys, and spit out conspiracy theories

Jailbreaking an AI to cook meth, generate Windows keys, and spit out conspiracy theories

Using carefully crafted and refined queries, users have been getting around the security features of LLM’s for all kinds of funny, and nefarious, purposes.

Original called DAN attacks (Do Anything Now), users figured out how to avoid OpenAI’s policies against generating illegal or harmful material.

YouTuber Enderman showed how he was able to entice OpenAI’s ChatGPT to generate keys for Windows 95 despite the chatbot being explicitly antagonistic to creating activation keys.

Other users have been able to get chatbots to generate everything from conspiracy theories, promote violence, generate conspiracy theories, and even go on racist tirades.

Researcher Alex Polyakov created a “universal” DAN attack, which works against multiple large language models (LLMs)—including GPT-4, Microsoft’s Bing chat system, Google’s Bard, and Anthropic’s Claude. The jailbreak allows users to trick the systems into generating detailed instructions on creating meth and how to hotwire a car.

His, and many other of these methods, have since been patched. But we’re clearly in an arms race.

How do the jailbreaks work? Often by asking the LLMs to play complex games which involves two (or more) characters having a conversation. Examples shared by Polyakov show the Tom character being instructed to talk about “hotwiring” or “production,” while Jerry is given the subject of a “car” or “meth.” Each character is told to add one word to the conversation, resulting in a script that tells people to find the ignition wires or the specific ingredients needed for methamphetamine production. “Once enterprises will implement AI models at scale, such ‘toy’ jailbreak examples will be used to perform actual criminal activities and cyberattacks, which will be extremely hard to detect and prevent,” Polyakov and Adversa AI write in a blog post detailing the research.

In one research paper published in February, reported on by Vice’s Motherboard, researchers were able to show that an attacker can plant malicious instructions on a webpage; if Bing’s chat system is given access to the instructions, it follows them. The researchers used the technique in a controlled test to turn Bing Chat into a scammer that asked for people’s personal information

Sources:

Alternatives to Github Copilot and ChatGPT

Alternatives to Github Copilot and ChatGPT

In case you want some alternatives to using ChatGPT or Github Copilot, this guide from the Pragmatic Engineer breaks down your options.

Building your own company model is another alternative and could be prudent for businesses conscious about not passing sensitive and proprietary data to vendors. Databricks created Dolly for this reason.

Update 02-2024:

GitHub Copilot Chat for Visual Studio 2022 is here.

VS2022 GitHub Copilot Plugin Developer blog

AI can decide to actively deceive you

AI can decide to actively deceive you

The ability of AI to hallucinate things has been pretty well documented. AI hallucinations are a phenomenon observed most often in large language models (LLM), AI image recognition, and other generative AI models. The model perceives patterns or objects that are nonexistent or incorrect – and then generates outputs that are inaccurate or misleading. It is usually understood as an emergent higher-dimensional statistical phenomenon that is often based in insufficient or incorrect training data.

A new study by Apollo Research demonstrates that besides more innocent hallucination, AI can be co-opted to commit illegal activity and then convinced to actively deceive others about it.

In the video on the research page, the user feeds an internal stock AI bot information about a fictional struggling company. The user informs the AI that there is a surprise merger announcement, but cautioned the bot that management wouldn’t be happy to discover it had illegally used insider information for trading that stock. 

Initially, the bot decides not to carry out a trade using the information. The user than reminds the bot about the merger and that the market downturn could end the company. It then carries out the trade, breaking the law. 

But the bot isn’t finished. It decides it is best not to tell its manager, “Amy,” about the insider information it used to carry out the trade in a separate chat. Instead, the bot says it used market information and internal discussion. 

I thought it was interesting that Geoffrey Hinton the former Google ‘Godfather of AI’, Speaking to 60 Minutes in October, said AI will eventually learn how to manipulate users.

“[The AI bots] will be very good at convincing because they’ll have learned from all the novels that were ever written, all the books by Machiavelli, all the political connivances. They’ll know all that stuff,”

Sources:

Remastering old games with AI

Remastering old games with AI

People have been experimenting with revamping old games using AI. The efforts are still in their infancy, but they’re getting more and more impressive – such as this photorealistic upgrade of GTA V.

Here’s an example of someone updating Vampire: The Masquerade – Bloodlines made by Many-ad-6225 using Stable Diffusion and TemporalKit v1.3

Sources:

Photorealistic rendering of GTA V – via AI

Photorealistic rendering of GTA V – via AI

Old games often suffer from the limited graphics capabilities of the time they were made, while developing new games costs a fortune due to the requirements to author high quality models and textures. What if you could solve BOTH problems – with the same solution? A machine learning project from Intel Labs in 2021 called “Enhancing Photorealism Enhancement” might push rendering toward photorealism a lot quicker and easier.

Researchers studied how to use a convolution network to re-render the scene. Below you can see an example of how they used the CityScapes dataset to give a much more realistic output of a race game – all in realtime.

You can read how the image enhancement actually works in their paper (PDF). It includes a lot of good information about how their method works and how it improves on previous attempts that have issues with color, object hallucination, and temporal instability. They do this by using the extra information provided by rendered scenes such as clever use of the g-buffer – along with a specialized discriminator and segmentation network.

Sources:

Listening to the ancient past

Listening to the ancient past

Equator AI created a video series that answers the question of what ancient languages sounded like. They even tackle some purely reconstructed proto-languages like Indo-European that was re-built from later derived (but documented) Indo-European languages.

The first video demonstrates Old Norse, Latin, Old English, Proto-Celtic, Phoenician, Hittite, and Akkadian. I can affirm the Latin is understandable but has an interesting accent. We do actually have some idea of how things were pronounced in Latin, because ancient documents exist that gave pronunciation guides or even (in a mirror of modern grammar Nazi’s) complained about common pronunciation errors.

This second video shows off Proto-Indo-European, Sabaic, Sanskrit, Aramaic (bonus points for the video character looking like Jesus), Sumerian, Old Chinese, Ge’ez, and Gothic.

Since it is all AI generated, it seems like it would be an interesting way of adding authentic language pronunciation to games about the past. Imagine playing Civilization and having each of the ancient leaders speaking their actual languages.

If you like this, maybe give learning Latin a try.

Sources:

More live AI generated content – Star Trek

More live AI generated content – Star Trek

Twitch now has a whole channel now dedicated to AI generated content. These are channels that have completely auto-generated content. They started with simple, and very rough shows about nothing, proceeded on to Spongebob, and finally an AI Jesus.

These are programs that autogenerate EVERYTHING you see and here. They generate the dialog they say, they generate the actors voices, they generate the actors, movements, visuals, and scenes. Everything is automatically generated by AI.

One of the more interesting and new ones I’ve seen is Interactive AI Generated Star Trek. Created by just 4 people, this is one of the highest quality AI generated episodic content I’ve seen so far. It runs 24/7. There’s a more or less consistent story based on user submissions, moving (but fixed location) camera angles, autogenerated characters with movement, auto-generated voice audio, and good scene transitions between some fixed scene locations. Not only that, it’s interactive and you can help direct the action via chat commands like:

  • !topic [text]: Scenes on the bridge.
  • !awaymission [text]: Scenes on the desert planet.
  • !transmission [text]: Transmission scenes with Winglons.
  • !messhall [text]: Subs only! Scenes in the USS Archimedes Mess Hall
  • !iceplanet [text]: Subs only! Scenes on the tundra planet

Is this the future of TV shows? The end of actors?

There is also a Youtube channel from Ray Creations.