Matrix

Tech Update: AI Generators (Feb 2023)

Tracy Harwood Blog February 5, 2023 1 Comment

Now the AI genie is in full flight, we’ve been anticipating the exponential growth in interest in creative applications – and also in the ethical and moral questions being asked. This month, we have not been disappointed! We start our review with some of the tools we’ve seen emerge and finish with a review of the legal situation that’s been taking shape over the last month, since our January update on the topic.

It takes Text-to-???

It seems most of the online world has bought into the hype around ChatGPT, and who can blame folks for wanting in on the action – it reached a million users faster than any platform has previously in the history of the internet ie., 5 days. Whilst it appears that Google and others have for once been caught sleeping on the job, Microsoft has stolen a march and helped OpenAI monetize its premium chat service for a mere $20/month (just a week after the extended partnership was announced and only if you are US-based) from which each partner will no doubt benefit massively. In the meantime, there has been a huge number of Chrome browser extensions launched based on ChatGPT for everything from search using voice commands, article summaries, writing Excel formulae, email assistance, LinkedIn comments management, to SEO optimization and a heap of other useful-ish applications. Go to the Chrome web store and search for the ones that will help with your creative pipeline – I’m sure someone somewhere will have thought of it before you.

I found a few uses for the YouTube summary assistants of which there are a couple of options, this being one (by Glasp) –

After adding the extension, it took a couple of seconds in total to transcribe the video, copy the text into my ChatGPT account and summarize an hour long interview I did with John Gaeta last year. This is the summary of that interview, which is a pretty good overview of what was discussed albeit the first part is almost verbatim from the intro –

The video is an interview with John Gaeta, who is known for creating the famous bullet time shot in the “Matrix” films. He won the Best Visual Effects Oscar for his work on the Matrix and co-founded Lucasfilm’s immersive entertainment division called ILMxLAB, where he acts as the Executive Creative Director. In the interview, he talks about his experience in creating a demo for the Sony PlayStation super computer, which was shown at Siggraph in 2000. He also mentions his interest in building big and complicated projects while also making entertainment products. Gaeta explains how the bullet time shot was a result of a philosophy they had during the Matrix trilogy of creating methods that might be used if one was making virtual reality. He also touches on how the rise of the internet and gaming helped audiences comprehend the shot better and how it carried on the underlying premise of the Matrix itself. (ChatGPT)

and here’s the full interview –

If, like me, you’re after nuggets and detail (my day job is as a researcher), then this won’t really help you but if you just want to get a sense of what’s being discussed, and you’re reviewing lots of material from various channels, or want a quick summary for promo, then its really a great way to generate an overview.

Creatively, though, we’re far more interested in the potential for Text-to-Otherstuff, such as 3D assets, video and 3D environments. Towards that end, although targeting game asset dev, Scenario.gg is a proposition (launched in 2019) that closed a round of significant seedcorn investment in January. With its creators’ backgrounds in gaming, AI and 3D technology, Scenario’s generative AI creates game assets using both image and text promps, albeit currently outputs are 2D images (see below). Its aim is to support creation of high fidelity assets, 3D models, sounds & music, animations, environments and more, based on users’ uploads of their own content (image and text description). The ownership model on generated content is interesting, which pushes the IP issue back to users since only images you have the right to use can be uploaded.

Scenario believes its product will cut creative production time for game artists (those who choose to work with AI). It surpassed .5M created images on 21 January, so is clearly gaining momentum. This is an interesting development, given comments by Aura Triolo (an Independent Games Festival 2023 judge and animation lead at Ivy Road) in an article covering AI devs for metaverse and games here (by Wagner James Au on his New World Notes blog, published in mid December). Triolo makes the point that time savings probably won’t be worth the effort given how much additional work is required to refine 3D models that AI generates, particularly in AAAs (as in the use of AI for procedural generation). That may well be true in a context where automation tools have been used for some time but this type of toolset will surely benefit thousands of indies, and not just in gaming but also machinima and virtual production. Time will tell.

source: Scenario in Techcrunch

Meta AI has published a paper that discusses taking their text-to-video (MAV) generator one step further to 4D (NeRFs or neural radiance fields), referring to it as MAV3D (Make-a-Video-3D). It optimizes scene appearance, density and motion consistency from text-to-video, generates a view from any camera location and angle and can be composited into any 3D environment. MAV3D does not require any 3D or 4D data. It is not yet available as a tool to use but here’s the paper to read. We look forward to hearing more about this in due course.

Text-to-fashion? Well, maybe not just yet, however ReadyPlayerMe, which is a cross platform avatar creator, has a new feature on its recently launched ‘labs’ web platform. Currently in beta and free, it allows you to customise avatar outfits using DALL-E’s generative AI art platform for text prompts. After faffing for a few minutes, I created this (but the hair is still a mess!) –


Text-to-music is an interesting area. There are no doubt going to be lots of training issues emerge with this type of AI, however, what fascinates me with Google‘s MusicLM is the ability to generate music from rich captions, using a ‘story mode’ (with a sequencer) and even from descriptions of paintings, places and epochs. I don’t think I’ve ever heard anything quite like the piece it generated for Munch’s The Scream, using a description by Iain Zaczek – not exactly melodic but certainly evocative of the artwork. It will also let you hum something and then apply a specific instrument to hear it played back, apparently. There is currently no API through which you can test your own ideas, however, but go to the github page here and check out the samples reported in the paper. Google it seems is only making the dataset of MusicCaps comprising 5.5k music-text pairs available, which includes rich text descriptions provided by human experts, and has obviously decided to let someone else create the API and take the rap with it. It will no doubt be one of many in due course, but there are some great ideas presented in the paper worth checking out.

Curation and Discovery

Curating content is one of the perenial problems of the internet – and its a problem that is getting more challenging because even with so much effort being put into the creator toolsets, no one is really paying much attention in the creator context of how work can be discovered (unless of course there is advertising embedded in it, which is a whole different agenda). One can only hope that when advanced AIs are embedded within search engines, new opportunities for content discovery will emerge – sadly, however, I suspect this will result in an even deeper quagmire, leaving it to the key platforms to find a way through. Related to which, Artstation has now improved its AI search and browsing filters – it can hide artwork generated with AI in search and marketplaces and thereby ‘make it easier to discover and connect with creators most relevant to you’, but strangely it doesn’t promote only work created with AI in search.

On the matter of curation, a website for AI generators has launched, called All Things AI – AI developers can submit their tool to the site for its potential inclusion. The site has been developed by Rick Waalders, and whilst there are numerous AI tools and services on the site, there’s not much information yet about its creator or indeed reviews of the AIs themselves. If the site takes off, it might just be the place to find the apps you want – time will tell. Until then, blogs such as Pinar Seyhan-Demirdag‘s Medium post, dated 11 January, are great sources for curated content. In this post, Pinar lists more than a dozen 3D asset and scene generation models – a very useful summary, thanks. Now, what we really need is a Fandom wiki for AIs…!

The Legals are Circling

What a busy month it has been in lawyerland.

On 17 January, in San Francisco US, a class action suit was filed on behalf of three artists. It claims that Stability AI’s Stable Diffusion and DreamStudio, MidJourney and DeviantArt have colluded in the use of an AI that has been trained on scraped content that infringes the rights of copyright holders (the AI being created by a company called LAION which has connections to Stability AI) and that the results of its application by users has a detrimental impact on the artists making profit from their own work as a consequence. One of the legal team has written a detailed blog about the action here, and here is the link to the action, should you want a quick scan through its 46 pages. The following day, 18 January, Getty Images stated that it has commenced proceedings against Stability AI in the High Court of Justice in London –

… Stability AI infringed intellectual property rights including copyright in content owned or represented by Getty Images. It is Getty Images’ position that Stability AI unlawfully copied and processed millions of images protected by copyright and the associated metadata owned or represented by Getty Images absent a license to benefit Stability AI’s commercial interests and to the detriment of the content creators.

Getty Images believes artificial intelligence has the potential to stimulate creative endeavors. Accordingly, Getty Images provided licenses to leading technology innovators for purposes related to training artificial intelligence systems in a manner that respects personal and intellectual property rights. Stability AI did not seek any such license from Getty Images and instead, we believe, chose to ignore viable licensing options and long‑standing legal protections in pursuit of their stand‑alone commercial interests.

No more specifics are available on the Getty case. The latter comments are particularly interesting, however, given its stance with creative contributors whom it has banned from uploading AI generated content, something we highlighted in our December blog post.

On the US class action, notwithstanding the technicalities of its description of how the AI works (which some have already questioned as being incorrect), the action is primarily about two aspects of copyright infringement – one related to a company which is ‘licensing images’ for use in training the AIs (that a couple of the image generating companies are using); the other is the specific use of an artist’s name to generate an image ‘in the style of …’ which suggests that person’s specific work, tagged presumably with their name, has been used without their permission to train the AI. Those using the images ‘in the style of x’ are referred to as ‘imposters’ whom it is being argued are contributing to the fake economy (which different governments are currently trying to control). The suit is not against the imposters but those who allow imposters to profit.

The action holds that the companies ‘scraping’ the images (which is a metaphor for how images are actually used in the diffusion process) could provide a means to seek permission from those artists used but it has not done so because it is ultimately expensive and takes time to do. The action is for compensation of damages for lost revenue and damage to brand identity of artists. The premise for this is that the companies being sued are generating huge amounts of money that is not finding its way to those contributors whose work is used in the training processes. The money flows are therefore the areas where the ‘fair use doctrine’ is being brought to bear.

The substantive legal issue, however, seems to centre on transformative works (from derivative works). Corridor Crew has produced a nice summary in a video with California State attorney Jake Watson explaining how AI probably DOES transform the work sufficiently for its use of copyrighted images to constitute FAIR USE. So it still comes down to what fairness means in terms of money flows, ultimately. Here’s the video –

And for a line by line commentary on another perspective of the class action blog post by one of the artists’ legal team, a response has been created by a group of ‘tech enthusiasts uninvolved in the case, and not lawyers, for the purpose of fighting misinformation’. I hesitated to include this link, especially since the author/s are anonymous and the contact link is to an ancient Simpsons video sniping at profiteering lawyers, but it makes some interesting points and is well referenced.

It generally sounds like there is a fix to this problem, and its one we’ve highlighted in previous posts on this topic, where AI generator platforms pay artists to use their work… oh, wait, isn’t Shutterstock already doing that, working with OpenAI’s DALL-E? Yes, and here is its generator – it pays the artists (I couldn’t find how much) and users pay for the images downloaded ($19/month for 10 images or 1 video). The wicked problem here, Techcrunch argues, is will folks be willing to pay for AI generated artwork, suggesting yes they will if the generator service has the best selection, pricing, discovery, and overall experience for the user and the artist. And, DreamStudio Pro is already a paid for service that folks are using.

Or, you can opt out of your art being included in Stable Diffusion’s AI, using the third party HaveIBeenTrained web service by an artist-based startup called Spawning. It looks as though there have been a few problems using this service according to the comments, such as being able to tag anyone’s imagery, but even more surprising was that not many seem to have viewed the support video in comparison to all the hoo-haa in the media on this topic. Check out this video –

Alternatively, you can just head over to LAION’s website and opt out there (scroll to the bottom of the page), following the long-winded GDPR processes under the EU reparation system.

In the meantime, Artstation has updated its T&Cs to make it clear that scraping, reselling or redistributing content is not permited and furthermore, it has committed to not licensing content to AI generator platforms for training purposes. Epic’s stance is always interesting to note, but its business model is not tied to just this one type of offer as the other platforms are.

And finally on the legals this month, we were intrigued to note that the US Copyright Office appeared to have cancelled the registration of the first AI generated graphic novel, called Zarya of the Dawn (our feature image for this article, by Kristina Kashtanova), claiming it had ‘made an error’ in the registration process… turns out that they are still ‘working on a response’, stating their portal is in beta. This was not before the artist had made an extensive response to the apparent cancellation through her lawyer, Van Lindberg. Its worth taking a few minutes to read the claims for originality, using MidJourney to support her creative processes, here. In sum, the response states –

the use of that tool does not diminish the human mind that conceived, created, selected, refined, cropped, positioned, framed, and arranged all the different elements of the Work into a story that reflects Kashtanova’s personal experience and artistic vision. As such, the Work is the result of human authorship

So this is yet another situation where the outcome is awaited. More next month for sure!

Completely Machinima S2 Ep 40 News (July 2022)

Tracy Harwood Podcast Episodes July 7, 2022 Leave a reply

Despite being encouraged to create an episode of just 15 minutes duration by one of our followers, the team have this month extended their coverage – hear Tracy, Damien and Phil discuss vtubing, Ricky’s Duke Henry the Red character in the game Evil Dead, the FTC’s proposed updates to social media guidelines, Unreal’s review of the Matrix Awakens Experience, John Gaeta’s latest exploits, metaquette, Reallusion’s iClone 8 and CC4 and a number of other exciting developments relevant to the world of real-time filmmaking and machinima.  Thankfully, you can use the timestamps to jump to the bits your most interested in!



YouTube Version of This Episode

Time stamps, links and show notes

1:34 Feedback from our followers: 3DChick, Al Scotch, Spentaneous, Mike Clements, Circu Virtu, Notagamer3d

8:14 Vtubing and Face Rig app (Steam), VTuber Studio – real time puppeteering using faceware

17:41 Interactive video on Vimeo, branching narrative storylines

20:23 Ukranian films of the recent war, showcased at Milan Machinima Film Festival website

24:17 Evil Dead and the character Duke Henry the Red played by Ricky Grove and IP generally, Ricky Grove on IMDb

Ricky vs Ricky

39:29 FTC updating ‘disclosures 101 for social media infuencers’ guide’ discussion, the relationship between brands, platforms and influencers and see also [Company] Rulez! (Phil Rice aka zsOverman & Evan Ryan aka Krad Productions).  Here is PC Gamers’ comments and proposals to update the guidelines

54:55 Competition updates: Nvidia Omniverse Machinima promo video ‘Top Goose’ | A NVIDIA Omniverse Machinima Short #MadeInMachinima, released 9 June 2022, and Unreal competition.  Here is Ben Tuttle’s The Amazing Comet (Unreal Engine/iClone) (4413 Media), released 9 June 2022.  And here’s a link to William Faucher’s YouTube channel.

screencap Ben Tuttle’s The Amazing Comet

58:03 Making of Unreal’s Matrix Awakens Experience, Behind the Scenes on The Matrix Awakens: An Unreal Engine 5 Experience, released 6 June 2022, and Tracy’s interview with John Gaeta, VFX, The Matrix films, released 17 February 2022.  Living Cities is a new metaverse mirror world, website link and Visual Effects Giant John Gaeta joins Inworld AI as Chief Creative Officer, 1 June 2022 and Inworld AI teaser video, released 28 April 2022

screencap Behind the Scene of Matrix Awakens Experience

1:00:16 Pooky Amsterdam’s blog on metaverse etiquette, called Metaquette website link

1:00:39 Reallusion’s iClone 8 character animation processes including Character Creator 4 Launch by Reallusion (released 26 May 2022) and iClone 8 Demo Video by Reallusion (released 26 May 2022)

1:13:08 Mesh to MetaHuman in Unreal Engine by Unreal Engine, released 9 June 2022

1:13:33 Meta’s new model allows creating photoreal avatars with an iPhone, 80.lv, 14 June 2022 and full research publication

1:14:02 Mocap with the MoCats: Livestreaming with Multiple Actors (Faceware & Movella/Xsens) by Faceware, released 7 June 2022

1:14:29 Love, Death & Robots, Jibaro character creation Love, Death + Robots | Inside the Animation: Jibaro, Netflix, released 9 June 2022 and Art Dump: stunning projects made for Love, Death + Robots, 80.LV website link

1:16:48 Jonathon Nimmons WriteSeen.com, launched June 2022, website connecting creative writers with industry professionals (upload written content, attach a video pitch, audio clips, video clips and a link to a prototype if required)

1:19:08 A word of thanks to our sponsors

Completely Machinima History with Ben Grussi: Halo Machinima

Tracy Harwood Podcast Episodes June 23, 2022 Leave a reply

In this episode, Ben reviews notable releases from the first days of Halo, including Steve Jobs reveal, the promotion of Nvidia’s GeForce2 GTS, Randall Glass’ classic Warthog Jump, the legendary Red Vs Blue release and others. See show notes and links on the CM website.



YouTube Version of this Episode

Show Notes and Links

0:47 May 1999, Steve Jobs (Apple) and Macintosh presentation reveal for Halo

Steve Jobs reveals Halo game, 1999

1:57 2000, Halo promo for NVIDIA Geforce2 GTS 

3:16 15 November 2001 Halo for Xbox released

3:40 First machinima for Halo, released 6 June 2002, Warthog Jump by Randall Glass 

5:04 Red vs Blue premieres, 1 April 2003, by Rooster Teeth Productions

7:04 Halo released for PC on 30 September 2003

7:31 Fire Team Charlie starts up in 2003

7:56 Halo PC Custom Edition released on 6 May 2004 and a recently made Halo PC Custom Edition Fan Documentary by Subpixel, released 17 April 2020

8:57 On 6 June 2004, the seed of Sponsors Vs Freeloaders is planted –

12:09 HaloTV.net was launched on 10 July 2004 – this channel (pre-YouTube) offered 24/7 Halo machinima primarily of clan matches with commentary plus had a dedicated Red Vs Blue channel

13:10 The Matrix (Halo Remix) by Halo1007 released (Halo Footage – Matrix Voice-Overs, released on 5 September 2004

Matrix (Remix)

14:01 The Heart is a Lonely Hunter by Dave Anderson, released 13 October 2004

14:55 Halo 2 released 9 November 2004 and Burnie Burns reports from Camp Bungie on the secret machinima test they doing in the game

16:26 The Codex series ran for 21 Episodes with an average run time 3-5 minutes and it’s prequel The Heretic, by Edgeworks Entertainment, released in 2005

17:39 This Spartan Life by Damien Lacedaemon aka Chris Burke premiered in May 2005

This Spartan Life

19:07 iGod… Holy Halo series by Fuzion Church, Crossroads Baptist Church and Chi-Ro Ministries, a Halo-based machinima to promote a Christian weekend retreat and to promote Christian principles in general – the episodes are covered on Halo.Bungie’s fan website (see Purple and Pink Mirrored), released 12 April 2005

20:55 First green screening machinima by AmandaJ3162, merging Sims 2 and Halo in You’re Fired, released 10 July 2005

You’re Fired

24:41 Company Rulz by Z-Studios and Krad Productions, a Microsoft game usage rules explainer, released in 2007 

Completely Machinima S2 Ep 36 News (May 2022)

Tracy Harwood Podcast Episodes May 4, 2022 2 Comments

In this episode, Phil leads the month’s news for machinima creators.  We kick off with the Nvidia’s Machinima Contest, the Omniverse Create and Machinima apps, GTC conference recordings, the addition of Reallusion’s ActorCore library to the Omniverse platform and ask ‘how well do laptops cope with RTX cards’ (what do you think)?  Damien reviews Concert Creator AI for piano with a rendition of Battlestar Galactica’s Kara Remembers piece, Tracy talks about another of her projects demonstrating creative applications of AI (the Art AI Festival) and Ricky highlight’s Phil Tippett’s completed Mad God film project.  Other film contests reviewed are the Real Time Shorts Challenge and the Milan Machinima Film Festival and the news of an upcoming Shakespearean performance in GTA Online by Sam Crane.  Tracy reviews Draxtor Dupres’ major pandemic machinima film project comparing virtual life in Second Life and Animal Crossing.  We discuss the Unreal 5 City Sample, Unity’s Enemies (hiding under the UE bushel) and Ricky’s Paperback Show.  Phil also teases us with rumours about GTA6.      



YouTube Version of this Episode

Show Notes & Links

2.35 Nvidia Omniverse Machinima contest, assets and prizes

4.18 Nvidia’s GTC conference recordings – worth the watch!

6.23 How are RTX cards coping with heat in laptops?

6.34 Nvidia Omniverse Create and Machinima apps

10.17 Reallusion’s ActorCore library now available for Nvidia Omniverse

12.51 Concert Creator AI creator for music and Damien’s animation of Kara Remembers (Battlestar Galactica) test

screencap: Damien’s test of Concern Creator AI

16.16 Art AI Festival in Leicester, creative application of AI including a YouTube channel of presentations

19.49 Mad God by Phil Tippett, film now completed and available on Shudder TV (subscription-based horror streaming channel)

screencap: Mad God (teaser)

22.43 Real Time Shorts Challenge competition (John MacInnes) overview

25.29 Milan Machinima Film Festival overview and Critics’ Choice Award, We Are Such Stuff As Dreams Are Made On (Shakespear’s Tempest performance) by Sam Crane (GTA Online) and current project to perform Hamlet in GTA Online

28.57 Virtual Cultures in Pandemic Times by Draxtor Depres, comparison of life in the metaverses, Second Life and Animal Crossing, during the pandemic

34.47 Matrix Awakens Experience City Sample available for PC

screencap: City Sample (Unreal Engine)

39.52 Unreal 5 update is great for machinima creators (Kim Libreri’s keynote and The Matrix Awakens: creating a world tech talk)

42.10 GTA 6 rumours!

43.20 Unity realtime cinematic, Enemies and Unity for machinima

screencap: Enemies (Unity) teaser

46.23 Don’t forget to read!  Phil’s recommendation: Stanislaw Lem’s Solaris, acquired from Ricky’s Paperback Show and book store

Completely Machinima Interview: John Gaeta, The Matrix

Tracy Harwood Podcast Episodes February 17, 2022 2 Comments

In this episode, Tracy talks to John Gaeta about his interests in machinima and real time filmmaking, The Matrix Awakens Experience, the influence of the bullet time shot, building the metaverse, future of storytelling in immersive environments, the potential of NFTs and his advice for indie creators.



YouTube Version of this Episode

Show Notes and Links

John Gaeta on IMDb

screencap: Matrix Resurrections (l-r Donald Mustard, Kim Libreri, John Gaeta, Keanu Reeves)

Pioneers in Machinima, Prologue by Kim Libreri, CTO Epic Games

SIGGRAPH 2000

The Matrix Awakens Experience video (UE5)

bullet time

UE5 Nanite

News coverage of Disney’s AR patent application

3D Game-Based Filmmaking: Art of Machinima (book published 2004) by Paul Marino

Time Stamps and Themes

3.32 Backstory to Matrix Awakens Experience

10.01 Game vs Experience

12.00 World building The Matrix

18.11 Why PS5 and Xbox X/S?

21.36 Bullet time – again!

26.04 Awakens as a time capsule – remaking the virtual Neo

29.03 Multi-modal storytelling is the future

33.06 Storytelling and AI

37.00 Real time vs cinema – a cross-over moment

39.39 Scalable metaverses, AR and AI – reflecting on Disney’s recent announcements

49.27 The space between bones and spaceship in 2001: A Space Odyssey – vis Paul Marino’s 2004 Art of Machinima book

52.10 The role of Nvidia’s Omniverse platform

54.46 NFTs and open markets

1.03.00 Advice to machinima and real-time creators: passion and knowledge!

1.08.00: Red or Blue?

Credits

Producer/Editor: Ricky Grove

Music credits: frankum’s Nebua Techno House. freesound.org Creative Commons