In the second part of our 4-chapters exploration of V6, we will look at the very core of Midjourney—its understanding of language—i.e., prompts.

First, we will experiment with simple and complex Text Prompts, and then dive into Image Prompting, and see how the new model recognizes and reimagines (and blends!) existing pictures, and how to control the outcome with the Image Weight parameter.

Prompting In V6

The official announcement says that prompting in V6 differs significantly from how we learned to communicate with previous Midjourney models. Up to a point where you will ”need 'relearn' how to prompt.”

V5.2
V6

Word Order + Optimal Prompts

To start with prompting, let’s experiment with the word order to see its influence on the outcome. For this test, I used four variants of the same prompt with different word/token order. First, I applied the same --seed to each generation to get the results as similar as possible:

And even though there are slight differences between the four images, the overall result didn’t change that much from variant to variant. However, this happened when I removed the --seed and let each generation be unique:

But the true importance of the word order (and the benefit of using Optimal Prompts) becomes obvious with less powerful Midjourney styles than that of Aaron Horkey.

Here are just a couple of examples of how the right wording can change the outcome drastically:

Prompt Complexity

When exploring Midjourney's capability to understand your prompt, an important question to answer is: how complex can we make our prompt before MJ ”breaks” (i.e., stops reacting to new additions to the prompt or “loses” its previous parts).

Let's start with a simple prompt, and gradually add new tokens to it, while keeping the --seed paramter fixed for consistency.

Midjourney V6 made it to an impressive seven steps before parts of the prompt started “disappearing“. While V5.2 started losing tokens one step earlier.

Obviously, results can vary depending on different situations and prompts. However, in most of my experiments, the predecessor model tended to reach its limit at around 6–7 steps before ”breaking“, whereas the new model consistently managed to handle a couple of steps more before encountering similar issues.

Abstract Concepts

For the concluding test of this chapter, let’s see how both models—and their Raw Modes—interpret more abstract (and even nonsensical!) prompts.

Despite the dramatic official announcement from the Midjourney team, the prompting strategies I developed with earlier models proved highly effective in the new version. Nonetheless, V6’s advancements—in terms of quality, variability, and the AI's improved understanding of prompts—are unmistakable.

Image Prompting

Image Prompts, allowing the use of existing images as a basis for Midjourney to build upon, have been a part of the platform since V3. With each new model, this functionality has grown more powerful. So, what changes does V6 bring to Image Prompting?

V5.2
V6

Here to present to you the capabilities of V6—Francis D., my friend, muse, and the ideal subject for Midjourney’s Image Prompts!

Francis D. (2011)
Original photo

A major issue I've found with Image Prompts in earlier Midjourney versions was their limited ability to transform photorealistic images into something else. For instance, using a photographic portrait as a source image to create a non-photographic result often didn't fully work. The outcome would still lean towards photorealism.

Let's see if V6 brings any improvement in this area!

Another notable aspect of Image Prompts in V6 is the model's boldness in completely transforming the source image to meet the prompt's requirements. How about turning Francis D. into a lion? Or an owl? Or Stan Winston’s monster?

The reason why Francis is such a great model for Image Prompts is the combination of his standout facial features and the dramatic lighting that amplifies them. That’s the best recipe to help Midjourney “recognize,” or “read” a face to then reproduce it. But how about less obvious photographs? Or not photographs at all?

For this experiment, I began with a text part of the prompt that complemented the contents of the image. Then, I proceeded with prompts designed to transform the source picture into something entirely different:

Throughout my experiments, an intriguing observation emerged regarding the text component of single-picture Image Prompts. When the prompt includes the image of a well-known individual, adding this person's name tends to “weaken” the impact of the other elements in the prompt. It appears that Midjourney's preconceived notion of the individual forces the AI to favor more traditional, real-life representations of that person.

Image Weight

Talking about Image Prompts, it's impossible not to mention the Image Weight, or --iw parameter. This setting in V6, ranging from 0.1 to 3 (to 2 in V5.2), dictates how closely Midjourney adheres to your source image in the generation process.

For this test, I used some photographs from my own archive:

Using these portraits as source pictures, let's explore how changing the --iw value affects Image Prompts in V6 in comparison to V5.2:

Considering that V5.2's maximum Image Weight is limited to 2, let's explore what the increased maximum for this parameter in V6 can achieve with the same images and prompts (just V6 samples for this one).

Blending Pictures

Midjourney’s /blend command is a fascinating capability of merging multiple existing images to create a new, unique result.

Unlike single-picture Image Prompts, blends don't need any text input. This means you can just put two or more images into Midjourney’s blender and let it do its magic—based entirely on the visual elements of the pictures.

To explore the nuances of the “blending behavior” of Midjourney V6, I again turned to my photographic archive.

Okay, it works truly well with portraits, but how about more abstract things, more difficult blends—using visually polar images with conflicting aesthetics?

Both, Image Prompts in general and Blending specifically open a whole new dimension of creative possibilities in Midjourney. With V6’s sensitivity and ability to recognize a source image’s main features to later reproduce them in an unusual mix, this universe of possibilities has just become much broader.

In the next chapter, we will learn to control our prompts and increase the predictability of Midjourney's generations with parameters. Continue reading →

/discuss
this Guide

If you like our Guides, you can help us maintain and expand Midlibrary and produce more regular educational content of higher quality. And keep it free for all!

Support Midlibrary on Patreon! →

All samples are produced by Midlibrary team using Midjourney AI (if not stated otherwise). Naturally, they are not representative of real artists' works/real-world prototypes.

Support Midlibrary on
Patreon →
Privacy policy for Comments section

Ver. 2.9.1

All content in the Midlibrary catalog is generated by the Midlibrary team using Midjourney AI. We do not feature real artists' images, artworks, or any copyrighted material in our catalog. The samples provided by Midlibrary are intended for educational and illustrative purposes only and are not representative of real artists' works or real-world prototypes. Midlibrary is a non-profit initiative, not affiliated with real artists or authors, aiming to educate and inspire through the demonstration of the technology's potential in creative explorations.
I understand, don't show this again
Encountered a bug?

We do our best to keep this website running as smoothly as possible. However, stuff happens, and we thank you for letting us know!

Thank you!
Midlibrary Groundskeeper has been notified.
✕ Close
Something went wrong while submitting the form. Please, check if you filled all fields.
We're here to help! If you're unable to resolve the issue, please, contact us.
Subscribe to Midlibrary Newsletter

We regularly publish new Midjourney Guides, compile new Style Tops, update the website, and have fun! Want to be the first to get Midlibrary news? Subscribe to our newsletter and never miss a thing!

Thank you for subscribing!
Please, expect emails from [email protected]. If you're not receiving our newsletter for a long enough time, please, check your Spam folder.
✕ Close
Something went wrong... Please, check if you filled all required fields.
If you're unable to resolve the issue, please, contact us.

Personal Libraries are available to our Patreon Community

Midlibrary Team

Learn more about the benefits of supporting us by becoming Midlibrary Patron—and start your Personal Library ↗︎

You have just become a Patron, and cannot log in?

Please, allow our team some time (usually not more than 24 hours) to set up your Personal Library.

You may be using different emails for your Patreon and Discord accounts. If that is the case, please, send your Discord email to [email protected].

If the issue perists, or you didn't get a response to your email, please, inform us via Bug Report form

Close

We are currently updating the Personal Libraires' infrastructure

Midlibrary Style Contributor

In the nearest future, it will allow you to access your Collections much quicker, add covers to them, tag the styles you save to quickly find them, and—most importantly—save your --sref (numerical) styles!

However, at the moment, logging in to your Library is unavailable. We apologize for the inconvenience. If you are a Midlibrary Patron, please, check this Patreon post ↗︎ for Personal Libraries status updates.

To start creating Collections and save favorite styles:

Log in with Discord →
Personal Style Collections

Learn more about Personal Style Libraries, saving favorite styles, and organizing them into Collections.

Learn more about supporting Midlibrary and the  benefits of joining our Patreon community →

Close