How Does AI Generated Art Work

What Is AI Art?
AI art is any type of digital art created by entering prompts into an AI art generator, which employs large language models trained on millions of data points. These models reference this data to process a request and produce the desired result. AI art generators can then take written or visual prompts and craft new text, images or audio clips for marketing materials, music videos and professional websites, among other mediums.

How Does AI Image Generating Work?
While the visual distinction between a sandwich and a pineapple may be clear to you, it’s a trickier task for a computer. Researchers must feed machines millions upon millions of images. They annotate the data sets, so the machine has a text reference, and the device is tweaked and calibrated until it can recognize the photos. Thus, it begins to parse pineapples from sandwiches and can start to make its own.

These new tools are mind-boggling. And public. On September 28, 2022, OpenAI made a staggering announcement: DALLE-2 would be open and accessible to all, paving the way for a range of AI art generators to become available for public use.

The year 2022 alone saw the rise of DALL-E 2, Midjourney, Craiyon and the often controversial Stable Diffusion. If you’re looking for a bit of experimentation, these programs have free trials. Once you run out, you’ll have to subscribe to Midjourney or buy into DALL-E’s token system.

How to Use Midjourney
Midjourney’s image generator runs through the ever-popular Discord. Once you join the official server, you’ll have to navigate to Midjourney’s newbie server.

From here, the process is as simple as typing, with imagination to boot. Type “/imagine” in the chat bar and enter your prompt — this cue tells the server you’re looking for an image. For our purposes, let’s try: “Homer Simpson in an arcade.”

As soon as you hit enter, Midjourney begins to generate. Unlike other AI generators, you watch Midjourney develop in real time, likely around 30 seconds.

Midjourney Homer Simpson original
Image: Jacob Vaus / Midjourney
On the first go, the results can be confusing. Midjourney gives you a grid of images with several commands beneath it. U means “upscale,” and V means “Variation.” The number corresponds clockwise. The top left is U1, and the bottom right is U4.

Motivational Quotes in Urdu

Islamic Quotes

Beautiful Quotes for Your Loved Ones

Sad Quotes in English

If you like an image and want to see it in HD, you can choose the corresponding U to add an increased definition. If you like the idea of an image, but want to see what else it can do, click the corresponding V. Midjourney will create four different images with the same baseline.

Clicking V4 presents me with four new options:

Midjourney Homer Simpson variations
Image: Jacob Vaus / Midjourney
Midjourney’s generations are not as lifelike as DALL-E’s. While you can tailor the input with specific words, for instance “photorealistic,” “35mm” or “HDR,” I’ve found Midjourney much better suited for art.

If I like the top left, I select “U1” to upscale.

Midjourney Homer Simpson with upscaling
Image: Jacob Vaus / Midjourney
Again, I can make variations. However, if I like the image, I can “Upscale to Max” to make it as HD as possible. This is the final, high-quality result:

Midjourney Homer Simpson final result
Image: Jacob Vaus / Midjourney
Don’t be discouraged by Midjourney’s less-than-photorealistic results. While this program may not be as poised for visual trickery, it generates lovely, intriguing artwork.

RELATED READING
Stop Freaking Out About Generative AI

How to Use DALL-E
DALL-E is the program you boot up with your friends on Friday night to show them what AI can do. The results are often less cerebral and artistic and more outright fun.

The process itself is quite simple. You’ll arrive at the image-generation bar when you create an account for Open AI. From there, all you have to do is type. If you can’t think of anything, click ‘Surprise Me.’

Let’s try our old prompt: “Homer Simpson in an arcade.” Rather than watching the images form, you’ll track a loading bar, with some reference images below. Within 30 seconds, you’ll have your output.

DALL-E Homer Simpson in an arcade initial try
Image: Jacob Vaus /DALL-E
The results are typically shocking. You may wonder: What else can it do? By inputting the same prompt, you can see the limitless possibilities.

DALL-E Homer Simpson in an arcade second try
Image: Jacob Vaus /DALL-E
The words will often be jumbled, the figures demented, but it’s a small price to pay for images of this quality.

DALL-E Homer Simpson in an arcade, third attempt
Image: Jacob Vaus / DALL-E
It’s important to note that DALL-E does much better with characters than humans. The program recognizes pop culture icons better and seemingly struggles to create lifelike figures with clear faces. For instance, the prompt: “A fifty-year-old plays double dutch.”

DALL-E image of woman playing double dutch jump rope
Image: Jacob Vaus /DALL-E
Disturbing, I know.

How to Use Stable Diffusion
Stable Diffusion is another simple process — with a twist. Your prompt is what you want to see in the image, while your negative prompt is what you don’t want to see. If I want a seedy night image, I’ll try “Daylight” as a negative prompt. The loading screen creates output in a similar time frame.


Comments

Leave a comment

Design a site like this with WordPress.com
Get started