The year is drawing to a close, so let's take a retrospective glance at some of the stories we looked at months ago to update their content and do some necessary follow-ups.

In April, for example, we looked at the developments that Artificial Intelligence (AI) text-to-image application Midjourney went through in the previous three months to see how easy it was to design a fashion collection employing this application.

Midjourney seemed easy to use, but not perfect: as you may remember from that previous post, a prompt to generate a fashion collection inspired by Renaissance costumes easily confused the system that, rather than using the reference as prints or surface details for the designs, started mixing fashion models with paintings, at times creating blends of humans and canvases with uncanny results.

If you remember that post, hands, fingers and other details such as eyes and lips were also imperfect and the best solution at the time seemed to try and come up with models with their hands in their pockets to avoid seeing undesired distortions or fingers in excess.

The system also exhibited discrimination, primarily generating models with white features. When prompted to create Black models, it often resorted to pasting the head of a Black model onto a body, resulting in a loss of the fluidity observed in the generation of white models.

Generating crowds was also a nightmare, so prompting the system to come up with the front row at a fashion show ended up producing guests with multiple distortions, with several people sprouting additional limbs or combining in one head the features of two different people.

In mid-March, after Midjourney v5 was released, things radically improved: everything was better defined, there seemed to be more diversity when it came to models and some images produced may have almost passed for real ones.

Yet, even the best attempts at producing a coherent fashion collection were plagued by tricks that Artificial Intelligence played on users.

When a user employs an AI application, the latter starts responding to the prompts and learns what the user likes or dislikes, eventually understanding which is the user's favored image in response to a specific prompt.

This means that, from then on, it will also start suggesting other users that specific aesthetic in response to prompts that may contain the same words.

In a previous post, for example, we have seen how in Midjourney the word "monster" in the prompt - combined with terms such as "kawaii", "stuffed animals", "plush toys", "knitwear" and "fun moods" - generates colorful jumpers with conglomerations of undefined colorful monsters.

We also learnt the perils of mode collapse: in the context of generative models, this definition refers to a situation where a model fails to capture the diversity present in the training data and instead produces limited, repetitive, or biased outputs. This phenomenon is often associated with generative adversarial networks (GANs), a popular class of models used for generating realistic synthetic data.

In a GAN, there are two main components: a generator and a discriminator. The generator creates synthetic data samples, and the discriminator evaluates whether a given sample is real (from the training data) or fake (generated by the generator).

The generator aims to improve its output to fool the discriminator, while the discriminator tries to become better at distinguishing real from fake samples.

Mode collapse occurs when the generator manages to produce samples that fool the discriminator, but these samples are limited in diversity. Instead of capturing the entire range of variability in the training data, the generator focuses on a subset of patterns or modes. As a result, the generator might produce repetitive or unrealistic outputs that lack the richness and diversity of the true data distribution.

To address mode collapse, researchers and practitioners employ various techniques, such as adjusting the model architecture, optimizing hyperparameters, using different loss functions, or incorporating regularization methods.

Additionally, there have been advancements in GAN variants and training strategies designed to mitigate mode collapse and improve the overall stability and performance of generative models.

As months passed, more featured were introduced in Midjourney – such as the Vary Strong, Vary Subtle, Vary Region and Zoom Out options, that allow to introduce more variations in a generated image. More complex prompts and using other functions such as /describe and /blend, can help adding more variety.

Is it easier or more satisfying to create a fashion collection now using Midjourney? Well, I attempted to do a new one in October for Atelier Blancmange, so let's have a look at what went on behind the scenes (please note, the draft images for the collection go from Image 1 in this post to Image 18; the following image opens instead the real collection).

The collection - entitled "Be the Crane" - was showcased during the second edition of Artificial Intelligence Fashion Week (AIFW).

The starting point for the collection were Taipei's claw machine arcades, combined with SpongeBob SquarePants's episode "Skill Crane" (Season 4).
In this episode, Squidward attempts to win a prize from the claw crane that Mr. Krabs installs in the Krusty Krab.

He eventually manages to do so after SpongeBob tells him that the secret to win a prize is to "be the crane", even though there will be a rather bizarre twist in Squidward's fate.

I had in mind a playful and imaginative collection inspired by this whimsical world that moved from the vibrant colors of soft toys and of plastic figurines from gashapon vending machines as well, hoping to come up with designs that encapsulated the carefree mood of childhood fun.

So the main moods of the collection had to be joyful, possibly overloaded with sensorial stimuli with textile and surface embellishments and some soft elements that pointed at cute plush toys, but also at other items, including hard plastic figurines or food items such as snacks and candies, also distributed by such machines.

The first experiments were not successful: I dreamt of a proper runway on which models walked among claw machines, but the system was confused about what a claw machine is and mainly created glass cabinets filled with toys.

Only several trials later I managed to create a runway that featured some claw machines on one side and the audience on the other, but no models.

For what regarded the clothes, in most cases Midjourney created editorial images of models with a pile of plush toys in the background or wearing designs made with plush toys or integrating them.

See the first few images in this post and you will get an idea: as you can see they aren't that new at all fashion-wise, but resemble more colourful and also more infantile re-editions of Jean Charles de Castelbajac's (View this photo) and Moschino's (View this photo) late '80s teddy bear coats (the idea was later borrowed again by Jeremy Scott for Moschino's S/S 2014 collection).

This wasn't the effect I was trying to achieve, so I discarded most images and attempted to come up with knitted designs.
Yet also knitwear wasn't a great idea as the system produced a series of jumpers that looked like conglomerations of cute monsters.

Eye-catching and colorful, most of these jumpers were too similar to kids' garments designed by Midjourney and currently advertised and sold on sites trying to scam consumers who order the AI designed garment, but then receive something completely different.

Further attempts at insisting along these lines produced more childish designs, kind of cute, but verging more towards the Japanese "decora" aesthetic and style, in bright and bold shades, yet not necessarily what I was looking for.
Besides, most of these images didn't show models on runways, but they looked like shoots done in a studio.

After working along these lines for a few days in October I completely abandoned the idea and started thinking about how to use my main inspiration in other ways.

Eventually I decided to shift perspective and place the models inside the claw machine, as if they were surrounded by cute toys.
This seemed to open up another world with more options and finally Midjourney started generating images that seemed to fit my vision.

Midjourney overloaded the set of the runway with toys and plush animals, and started flattening the toys integrated in the garments, presenting them as quilted or padded elements (and in some cases as prints), features that in a real collection may provide a sense of comfort and care, further enhancing the wearer's experience.

Models at times still displayed a look aligned with the "decora" style, but, rather than in super bright Crayola-like shades, the system satisfied my request for a softer palette, verging more towards dusty pink and lilac, faded yellow, pale blue and pastel mint.

For the final section of the runway I abstracted the colorful cloud motif that decorated the set, magnified it and came up with a sort of squishy or melting motif.

Oversized, whimsical abstract shapes, puffed sleeves, balloon and cloud-like elements are employed in this section to reflect the playful aesthetic of the main theme.

The same motifs were replicated in a few menswear items (mainly jackets and tops), while a bomber kimono jacket combined the padded toy theme with the abstract cloud motif.

Shoes and bags were inspired by gashapon machines and decorated with plastic spheres that pointed at the capsules dispensed by these machines, or embellished with small figurines and toys.

The gashapon bag was a bit tricky as the system has learnt to create handbags filled with plastic balls (View this photo) and tends to recreate this aesthetics when in the prompt you use words such as "handbag", "plastic spheres" or "capsules" (View this photo - not rare to spot this version on Instagram accounts by other AI artists).

Yet that wasn't what I had in mind; I was indeed trying to achieve another configuration that vaguely evoked a pétanque ball carrier and in the end I managed to.

For what regards backstage images, front rows and street styles (complementary images requested by the rules of AIFW), I didn't have any of the problems I had in February/March when generating images with Midjourney.
Midjourney seemed to respond well to my vision and there were only minor imperfections (almost undetectable in some cases) when it came to hands, eyes or lips.

The level of details that the system managed to produce when prompted to create whimsical makeup and hair styles was definitely impressive.

Street style images also proved fun to make and easily combined the aesthetic of a street in a Taipei alley, with some futuristic and whimsical twists.

I found problematic instead coming up with Black female models for this collection: the system lacked indeed consistency when it came to creating Black models that fit with this pastel fantasy world and kept on generating either young Black women in "decora"-inspired designs or wearing tracksuits decorated with bright elements that didn't fit with the rest of the collection.

So, after a year of experimenting with Midjourney, what do I think, is it worth designing a fashion collection with the support of AI? Well, it is fun to do it, but it can still be frustrating and there are still several issues to consider.

First and foremost, if you're not feeding into the system your own images, the system may be stealing and collaging other people's work (yes, I know what you're thinking, there are designers in the real fashion industry copying and pasting things from other designers' works and nobody seems to care…point taken, but in this case we are considering the highs and lows of AI).

This means that, when you create a fashion collection with AI, you must keep your feet firmly on the ground and not think you have created something previously unseen and 100% new as, in the same instant you may be feeling triumphant, the system may have generated a very similar image for another user in another part of the world (consider how the designs I have generated with quilted and padded clouds look like the more sophisticated versions of some kids' clothes generated by Midjourney and sold on scam sites...).

If you want to create your fashion collection with AI, remember that a vast fashion history knowledge may help you spotting copies or derivative designs (think about the early designs Midjourney generated with my first prompts that seemed to borrow from JCDC and Moschino, as stated above).
So, rather than thinking you're a great fashion designer, marvel at the polished and sophisticated image you have obtained and wonder how you can transform, enhance or reinvent it (if you're using Midjourney play around with Vary Strong, Vary Subtle, Vary Region, Zoom Out and so on, as highlighted above).

Last but not least - and this is probably the most important point to take into consideration if you want to take part in an AI fashion competition - be careful about the monster you're feeding.
The more you train it with your vision, the more the system will learn to respond to those specific words in your prompt with that vision, which means that, at some point, it will learn to respond in more or less the same way to another user.

If you want something more exclusive and unique, start working on your idea, leave it for a while, and then return to it again in a more intensive way, but for a shorter period of time, maybe closer to the deadline of your competition.
In this way the system may not be able to immediately borrow from the latest knowledge it has gathered from you to generate similar images for another user.

That said, while creating something, you may already be borrowing from the results of another user's training session, so once you obtain an image, do a research on the Internet to see if that specific picture you have generated looks like someone else's and it's already popular on social media.

From this list of suggestions and advice you can easily realize that doing a fashion collection with Midjourney is not that fast, after all.
Sure, AI text-to-image applications allow us to quickly visualize something and, while they do allow us to save money on sets, props and teams, it takes several passages and stages to refine your vision and come up with something that in some ways mirrors what you had in mind.

Besides, consistency and coherence are vital if you are working on a collection and you must try and create different designs that should look as taken from the same show. So, try to focus on how to generate a variety of creations that may fit well in the same imaginary show.
Once you're done, expand your vision, adding AI videos, maybe an AI ad campaign, or imagine your brand ambassador, influencer or spokeperson as an AI twin, like Atelier Blancmange's Harumi K, one of the models from the "Be the Crane" collection.
Trained to talk about the collection and its inspirations, Harumi K is on Twise and you can interact with her vocally or type your questions and thoughts in a chatbox (mind you, at times, she still hallucinates...perils of AI).

So, will AI fashion be a reality at some point? Well, Maison Meta's AIFW seems to be a reality at the moment and the winners get the chance to enter the Maison Meta x Revolve AI Fashion Incubator and get their collections produced. Yet, even though AI proceeds at a fast pace, it will probably take a while before a prominent fashion house or brand designs and manufactures an entire collection using Artificial Intelligence.
Till then, just have fun with these systems or, as Alan Warburton states in his video essay "The Wizard of AI", "show me some AI candy!"
