AI generator

Tech Update 1: AI Generators (Jan 2023)

Tracy Harwood Blog January 8, 2023 Leave a reply

A month is apparently a VERY long time in the world of artificial intelligence… since our last post on this topic, released at the beginning of December, we’ve seen even more amazing techs launch.

ChatGPT

This is another creative AI from the OpenAI stable (also creator of DALL-E 2) that produces text in a conversational way. ChatGPT is capable of answering pretty much any question or request you care to throw at it, and the answers are actually quite robust if a little fuzzy around the edges at times (don’t rely on it for anything critical, in the words of Sam Altman, OpenAI CEO). Here’s an example I asked it for this post, using the prompt: ‘In the style of H P Lovecraft, write a description of a character for someone who hunts the Chluthu’

Response:

Professor Alistair Dunsmore was a man obsessed. For years, he had devoted his life to studying the mysteries of the cosmos, and in particular, the myth of the Chluthu.

A brilliant but troubled man, Dunsmore was driven by a need to uncover the truth about the terrifying creature that lurked beyond the veil of our known universe. He had spent years pouring over ancient texts and forbidden tomes, piecing together the scattered clues that hinted at the Chluthu’s existence.

And then, one fateful day, he discovered something that would change his life forever. A fragment of an ancient artifact, imbued with otherworldly powers, that seemed to hold the key to finding the Chluthu.

Dunsmore knew that he had to have it, no matter the cost. And so he set out on a quest to track down the artifact and unlock its secrets. He traveled to the far corners of the earth, braving dangers and braving the unknown, all in the pursuit of his ultimate goal.

Some say that Dunsmore was mad, driven to the brink by his obsession. But to those who knew him, he was a hero, a brave and brilliant man who dared to stare into the darkness and face the horrors that lay within.

That’s impressive – and it took just seconds to generate. It has great potential to be a useful tool for scriptwriting for stories and character development that can be used in machinima and virtual productions, and also marketing assets you might use to promote your creative works too!

And as if that isn’t useful enough, some bright folks have already used it to write a game and even create a virtual world. Note the detail in the prompts being used – this one from Jon Radoff’s article (4 Dec 2022) for an adventure game concept: ‘I want you to act as if you are a classic text adventure game and we are playing. I don’t want you to ever break out of your character, and you must not refer to yourself in any way. If I want to give you instructions outside the context of the game, I will use curly brackets {like this} but otherwise you are to stick to being the text adventure program. In this game, the setting is a fantasy adventure world. Each room should have at least 3 sentence descriptions. Start by displaying the first room at the beginning of the game, and wait for my to give you my first command’.

The detail is obviously the key and no doubt we’ll all get better at writing prompts as we learn how the tools respond to our requests. It is interesting that some are also suggesting there may be a new role on the horizon… a ‘prompt engineer’ (check out this article in the UK’s Financial Times). Yup, that and a ‘script prompter’, or any other possible prompter-writer role you can think of… but can it tell jokes too?

Give it a go – we’d love to hear your thoughts on the ideas it generates. Of course, those of you with even more flAIre can then use the scripts to generate images, characters, videos, music and soundscapes. There’s no excuse for not giving these new tools for producing machine cinema a go, surely.

Link requires registration to use (it is currently free) and note the tool now also keeps all of your previous chats which enables you to build on themes as you go: ChatGPT

Image Generators

Building on ChatGPT, D-ID enables you to create photorealistic speaking avatars from text. You can even upload your own image to create a speaking avatar, which of course raises a few IP issues, as we’ve just seen from the LENSA debacle (see this article on FastCompany’s website), but JSFILMZ has highlighted some of the potentials of the tech for machinima and virtual production creators here –

An AI we’ve mentioned previously, Stable Diffusion version 2.1 released on 7 December 2022. This is an image generating AI, its creative tool is called Dream Studio (and the Pro version will create video). In this latest version of the algorithm, developers have improved the filter which removes adult content yet enables beautiful and realistic looking images of characters to be created (now with better defined anatomy and hands), as well as stunning architectural concepts, natural scenery, etc. in a wider range of aesthetic styles than previous versions. It also enables you to produce images with non standard aspect ratios such as panoramics. As with ChatGPT, a lot depends on the prompt written in generating a quality image. This image and prompt example is taken from the Stability.ai website –

source: Stability.ai

So, just to show you how useful this can be, I took some text from the ChatGPT narrative for our imaginary character, Professor Alistair Dunsmore, and used a prompt to generate images of what he might look like and where he might be doing his research. The feature images for this post are some of the images it generated – and I guess I shouldn’t have been so surprised that the character looks vaguely reminiscent of Lovecraft himself. The prompt also produced some other images (below) and all you need to do is select the image you like best. Again, these are impressive outputs from a couple of minutes of playing around with the prompt.

images of Professor Alistair Dunsmore, in his study, searching for the Chluthu, by Tracy & Stable Diffusion

For next month, we might even see if we can create a video for you, but in the meantime, here’s an explainer of a similar approach that Martin Nebelong has taken, using MidJourney instead to retell some classic stories –

Supporting the great potential for creative endeavour, ArtStation has taken a stance in favour of the use of AI in generating images with its portfolio website (which btw was bought by Epic Games in 2021). This is in spite of thousands of its users demanding that it remove AI generated work and prevent content being scraped. This request is predicated on the lack of transparency used by AI developers in training and generating datasets. Instead, ArtStation has removed those using the Ghostbuster-like logo on their portfolios (‘no to AI generated images’) from its homepage and issued a statement about how creatives using the platform can protect their work. The text of an email received on 16 December 2022 stated:

Our goal at ArtStation is to empower artists with tools to showcase their work. We have updated our Terms of Service to reflect new features added to ArtStation as it relates to the use of AI software in the creation of artwork posted on the platform.

First, we have introduced a “NoAI” tag. When you tag your projects using the “NoAI” tag, the project will automatically be assigned an HTML “NoAI” meta tag. This will mark the project so that AI systems know you explicitly disallow the use of the project and its contained content by AI systems.

We have also updated the Terms of Service to reflect that it is prohibited to collect, aggregate, mine, scrape, or otherwise use any content uploaded to ArtStation for the purposes of testing, inputting, or integrating such content with AI or other algorithmic methods where any content has been tagged, labeled, or otherwise marked “NoAI”.

For more information, visit our Help Center FAQ and check out the updated Terms of Service.

You can also read an interesting article following the debate on The Verge’s website here, published 23 December 2022.

example of a logo used by creators on ArtStation portfolios

We’ve said it before, but AI is one of the tools that the digital arts community has commented on FOR YEARS. Its best use is as a means to support creatives to develop new pathways in their work. It does cut corners but it pushes people to think differently. I direct the UK’s Art AI Festival and the festival YouTube channel contains a number videos of live streamed discussions we’ve had with numerous international artists, such as Ernest Edmonds, a founder of the digital arts movement in the 1960s; Victoria and Albert Museum (London) digital arts curator Melanie Lenz; the first creative AI Lumen Prize winner, Cecilie Waagner Falkenstrom; and Eva Jäger, artist, researcher and assistant curator at Serpentine Galleries (London), among others. All discuss the role of AI in the development of their creative and curatorial practice, and AI is often described as a contemporary form of a paintbrush and canvas. As I’ve illustrated above with the H P Lovecraft character development process, its a means to generate some ideas through which it is possible to select and explore new directions that might otherwise take weeks to do. It is unfortunate that some have narrowed their view of its use rather than more actively engaged in discussion on how it might add to the creative processes employed by artists, but we also understand the concerns some have on the blatant exploitation of copyrighted material used without any real form of attribution. Surely AI can be part of the solution for that problem too although I have to admit so far I’ve seen very little effort being put into this part of the challenge – maybe you have?

In other developments, a new ‘globe’ plug-in for Unreal Engine has been developed by Blackshark. This is a fascinating world view, giving users access to synthetic 3D (#SYNTH3D) terrain data, including ground textures, buildings, infrastructure and vegetation of the entire Earth, based on satellite data. It contains some stunning sample sets and, according to Blackshark’s CEO Michael Putz, is the beginning of a new era of visualizing large scale models combined with georeferenced data. I’m sure we can all think of a few good stories that this one will be useful for too. Check out the video explainer here –

And Next…?

Who knows, but we’re looking forward to seeing how this fast action tech set evolves and we’ll be aiming to bring you more updates next month.

Don’t forget to drop us a line or add comments to continue the conversation with us on this.

Projects Update (Oct 2022)

Tracy Harwood Blog October 10, 2022 1 Comment

This week’s Projects Update on machinima, virtual production and content creation:

The Crow

One of the most interesting creative projects we’ve seen so far using MidJourney, a creative AI generator is the The Crow (by Glenn Marshall Neural Art). Here the generator has been used to recreate a version of the ballet performance portrayed in the short film, Painted (by Duncan McDowall and Dorotea Saykaly). Stunning to say the least and we recommend you play it at least once side-by-side against the original performance too for added insight.

We’re so impressed with the potential of AI generators, whether that’s DALL-E, MidJourney, Stable Diffusion or any of the others that are now emerging, that we’re going to dedicate a special episode of the podcast to the subject next month, so watch out for that!

Jim Henson Company

Jim Henson Company is using real-time animation on their new show, Earth to Ned. Characters are created with Unreal (its the AI in the background) but JHC has been so impressed with the workflow and no post production requirement that it is looking to use the virtual production method more. What’s interesting is the level of feedback in the process that guests experience – they are not aware of the puppeteering in the background, just the virtual actor on the screen, performing naturalistically in real-time! We’ve not seen much of this kind of machinima before although actually Hugh Hancock did some very early work on this and of course Rooster Teeth have done live performances using similar techniques. We can certainly expect to see a lot more of it, particularly for interactive theatre, VR and AR.

Half Life 3

Half Life 3 was never going to be like the originals? This article on Tech Radar is interesting: the author (Phil Iwanuik) contends the Half Life franchise remakes would never be like the originals because the extreme attention to the world of HL created so much pressure the Valve team could never live up to it. We’re not sure about that, but it’s an interesting idea.

source: Valve

Dune: Awakening

A very impressive MMO has launched using the Dune world, currently in beta, Dune: Awakening. Here’s the trailer – we’re looking forward to seeing machinima made with this –

Dungeons & Dragons?

What does Dungeons and Dragons, typically a game played around a table, have to do with machinima? There’s been a rise in popularity of web based shows where people play the game and act out scenes. This group (Corridor Crew) is using Unreal Engine 5 for virtual production (not quite The Mandalorian but sort of similar) to put their actors, real-time, into the environments of their adventure. Check it out here –

Tech Update (Oct 2022)

Tracy Harwood Blog October 3, 2022 2 Comments

This week’s Tech Update picks for machinima, virtual production and 3D content producers:

Nvidia RTX4080

Nvidia is launching two RTX 4080 graphics cards in November… you know what they say, you wait ages for a bus and then two come at once: the RTX 4080 12GB and RTX 4080 16GB. Here’s the story on PC Gamer‘s website. You can also catch up on all latest Nvidia’s announcements made in Jensen Huang’s (CEO) keyote at GTC in September in this video and on their blog here.

Ricky comments: Of course it was only a matter of time before NVidia announced the 40x series of RTX graphics cards. Two models have been announced so far, the 4080 and the 4090, with the 30x series sticking around for the lower price range. My guess is so they can focus their resources on producing more of just two high end cards instead of a whole range. Although given the prices of these new cards ($800+), I think I’ll be sticking with my 3070 for the time being.

UE 5.1.0

Unreal Engine have teased the new features coming to V5.1.0 – see the features documentation on their website here. Onsetfacilities.com has produced a nice overview – link here – and a nice explainer by JSFilmz here –

Cine Tracer

Check out the new Actor Animation system in Cine Tracer v0.7.6. This update gives the Actors a set of talking animations that can be used as an alternative to the Posing system.

Follow the socials on Instagram and download Cine Tracer on Steam

Sketchfab

Sketchfab is doing a weekly listing of top cultural heritage and history models – these are actually pretty amazing and of course downloadable (for a fee)!

source: Sketchfab – cultural heritage and history top 10

DALL-E

DALL-E, one of the creative AI generators that is all the buzz at the moment, has developed a new feature called Outpainting which can help users extend an image beyond its original borders by adding visual elements in the same style, or taking a story in new directions. This could be great for background shots in virtual productions.

Source: DALL-E, original is Girl with a Pearl Earring by Johannes Vermeer, Outpainting by August Kamp

Second Life

Second Life have launched a puppetry project for their avatars. As Wagner James Au reports in his regular blog on all things metaverse and Second Life in particular, this is using a webcam and mocap. Check out Au’s review of it here and go directly to Second Life here to read their post about it and follow their channel on YouTube for latest updates and how-tos here.

Eleven Labs

Eleven Labs have launched Voice Conversion. This lets you transform one person’s voice into another’s. It uses a process called voice cloning to encode the target voice – ie, the voice we convert to – to generate the same message spoken in a way which matches the target speaker’s identity but preserves the original intonation. What’s interesting about this is the filmmaking potential but of course there are very clearly IP interests that have to be considered here – it has potential for machinima application but beware the guidelines on using it. Importantly, note that it is primarily going to be used as part of an identity-preserving automatic dubbing tool which Eleven is launching in 2023. More here on this and the guidlines on using Voice Conversion.