24 minutes 12 seconds
Speaker 1
00:00:00 - 00:00:40
In this video, I want to attempt to create a short film entirely using AI. I've seen a lot of really cool videos circulating on Twitter and I kind of want to figure out how they made them. So in this video, we're going to figure it out. Now, the point of this video is to a show you how you can make a short film using AI tools, but also be to show you how to kind of tie a whole bunch of different tools together because I think that's really where all of these new AI tools get really powerful is when you take this tool and you piece it together with this tool and you sprinkle on a little bit of this tool over here and you combine them all to get something really, really cool. And that's what I want to demonstrate in this video is just pulling all these pieces together and coming out with something completely unique.
Speaker 1
00:00:40 - 00:00:53
So I'm going to use AI to generate the script. I'm going to use AI to generate the video itself. I'm going to use AI to generate the voiceover. I'm going to use AI to generate the background music. The entire video from start to finish will be AI.
Speaker 1
00:00:53 - 00:01:13
The only thing that I'm not really going to use AI for is I'm going to pull all the elements together inside of DaVinci Resolve, inside of my video editor. So let's just jump straight into it. So the first thing I need is a script to a movie. Now this 1, I don't want any dialogue in it. I want it to be a sort of documentary style with a voiceover throughout the whole thing.
Speaker 1
00:01:13 - 00:01:32
And I also need to generate a shot list to go with that voiceover. So to get the script, the voiceover and the shot list, we're going to use GPT-4. So here's the initial prompt I'm going to use. I'd like to make a short film that's between 60 and 90 seconds long. The concept of the film is that it will be a documentary about an AI robot that saves civilization.
Speaker 1
00:01:32 - 00:01:48
The film will have no dialogue. The entire film will have a voiceover narrator that tells the story. Please generate a script for the narrator to read that shares the story. So I'm basically telling GPT-4 I don't want any dialogue, just tell a story and it will be in the form of a documentary. So this is gonna be step 1.
Speaker 1
00:01:48 - 00:02:09
Step 2 after this is that I'm going to have it generate a shot list for me. Let's go ahead and have it generate our narration script first. All right, so it generated the entire narration and it even gave us a shot list in the initial prompt. I didn't even have to ask it for a shot list. I'm going to have it generate a more detailed shot list now, but it actually already kind of envisioned what the whole film is going to look like right here.
Speaker 1
00:02:09 - 00:02:33
So my next prompt for each of the shots in the above film, please give a much more detailed description of the shot itself. Include as much detail as possible around the setting and any objects or people in the shot. So now each of these shots I'm going to get a little more detailed so that I can use them for my prompting later. Alright, so now we have a much more detailed shot list that I can use for prompting. I'm going to go ahead and create a new folder on my desktop here, call it AI short film.
Speaker 1
00:02:33 - 00:03:14
That way, any clips that I make, I'm going to save right to that folder there. So to generate these clips, I'm going to use runways, gen 2, and using gen 2, I'm going to have the AI try to generate each of these shots. So in Gen 2, in order to generate a shot, we simply type at Gen dash 2, and then we just literally paste our prompt in. So if I jump back over to our OpenAI here, here's our opening shot, a wide angle view of a dystopian cityscape in ruins with crumbling buildings, shattered glass, and abandoned vehicles strewn across the streets. Fires are still smoldering in the aftermath of the apocalypse, casting an eerie glow on the thick layers of smoke that hovers above the desolation.
Speaker 1
00:03:14 - 00:03:34
So I'm gonna copy this whole thing and let's see what Gen 2 gives us with that prompt. Paste that whole prompt in here. Alright so it took about 3 minutes and we have our first generated video. Let's go ahead and take a peek. Now I would say this is a pretty spot-on representation of that prompt here.
Speaker 1
00:03:34 - 00:03:59
Now 1 of the issues it's only 4 seconds. Based on our shot list, it needs to be 5 seconds, but that shouldn't be an issue. When I pull this 1 into DaVinci Resolve to edit it, I can actually slow down the pace slightly and stretch it out to 5 seconds. So now I'm going to go ahead and do the same thing with our next shot, except this time, our next shot needs to be about 9 seconds long. So what I'll probably do this time is actually generate 2 or 3 of the exact same shot and butt them up against each other.
Speaker 1
00:03:59 - 00:04:19
So let's see what happens. Let's go ahead and prompt it once here. A dimly lit underground laboratory filled with computers, wires, and advanced equipment. A team of scientists and engineers wearing tattered lab coats and protective gear work on an AI robot with a humanoid form. They are soldering circuits, adjusting mechanical joints, and inputting data into a computer terminal.
Speaker 1
00:04:19 - 00:04:35
Let's go ahead and copy this. Let's just go ahead and start by trying to generate the entire prompt here. I might need to break this prompt up, but let's see what we get by just plugging in this prompt in its entirety. First, This 1 only took about 2 minutes this time and let's see what we got. So we've got a bunch of people working in a laboratory.
Speaker 1
00:04:35 - 00:05:02
I'm not really seeing the humanoid robot. So maybe I'll add another prompt here and focus in more on the robot with the humanoid form here. I'm just going to take the second half of this prompt, a team of scientists and engineers wearing tattered lab coats work on an AI robot with a humanoid form, soldering circuits, adjusting mechanical joints. And let's see if we can get another scene with more of them working on the robot here. So we'll go at Gen 2, paste that in.
Speaker 1
00:05:02 - 00:05:18
All right, so here's what we got from that prompt. They're working on something. It doesn't really look like a humanoid robot. But once again, I'm going to try to steer it a little bit closer. Luckily, I can use all of these clips strung together to fill out the time that I need for this specific clip.
Speaker 1
00:05:18 - 00:05:42
So this time I'm simplifying the prompt a little bit, a team of scientists and engineers and white lab coats leaning over a table with a life-size humanoid robot on it. Let's see if I can get closer to what the sort of vision of this shot is. Okay. So I really haven't been able to get Gen 2 to generate the image that I'm looking for of a robot sort of on an operating table. So I'm going to jump over to Mid Journey and generate a still image of that.
Speaker 1
00:05:42 - 00:06:01
And then we'll kind of do a panning shot in the video of that still image to represent that piece of the script. So let's go ahead and do imagine. We'll grab our same prompt here and let's just plug it into mid journey. This time, I'm going to go ahead and give it an aspect ratio of 16, 9. So it fills our screen and let's see what we get from this.
Speaker 1
00:06:01 - 00:06:18
Okay, so I think I finally got what I'm looking for up here in image 1. We have a bunch of scientists kind of working over a robot here. So I'm gonna go ahead and upscale that 1. And now what I'm gonna do is I'm going to follow this exact same process for every single scene from our chat GPT script here. So far, I've got our opening shot and our second shot.
Speaker 1
00:06:18 - 00:06:50
I've got 123456 more shots to generate. I'm going to try to generate them with Runway Gen 2 first. If I can't get Runway Gen 2 to create an image that I really like, I'm going to jump over to mid journey, give it the same prompt and try to get mid journey to generate something that looks like what I want. So instead of boring you going back and forth and doing every single 1, I'm going to go and generate all of the shots real quick and then we'll jump back and complete the process. Okay, so I've gone ahead and generated all of the shots in the shot list here.
Speaker 1
00:06:50 - 00:07:15
You can see I've got a folder with a ton of different files in it. We have images here of different scenes from the script here and then I've got a whole bunch of different videos that I generated for different parts of the video. Now, some of these images here, because they're static, I want to give them a little bit of an animation. And there's also 1 shot that I didn't generate that. I think a different tool is going to generate better though.
Speaker 1
00:07:15 - 00:07:41
I'm going to walk through those right now. In case you're wondering, Gen 2 is currently in a closed beta. They are slowly letting more people in and I believe it'll be rolling out publicly pretty soon. If you don't have access to Gen 2, Check out this video that I made called actual AI text to video is finally here. That teaches you how to use model scope, which is a free version of what you essentially can do with gen 2.
Speaker 1
00:07:41 - 00:08:25
The only downside is it does add little shutter stock watermarks inside of the video, but you can use that and it does something very similar to what Gen 2 does completely free. Now 1 of the shots that I wasn't really able to get a good version of using Gen 2 was this 1 here a time lapse of the city slowly coming back to life, showing a succession of events such as buildings being reconstructed, plants growing through the rubble, and people returning to populate the urban landscape. So, I'm going to jump into Mid Journey and see if I can generate a starting image. I'm going to use this prompt from our shot list here of a dystopian cityscape and ruins. I'm going to paste this into Midjourney and use this as a starting shot for my time lapse.
Speaker 1
00:08:25 - 00:08:45
Make it an aspect ratio of 16, 9 here. All right, so we have some images here of a destroyed city. I think this top left 1 looks the best, so I'm going to go ahead and upscale that 1. And then I'm going to jump into this Genmo tool, which you can find at alpha.genmo.ai. And I'm going to upload the image we just created and use this as my starting image here.
Speaker 1
00:08:45 - 00:09:08
And we have some settings that we can do. So it's asking for my edit and what I'm going to do is I'm going to jump back to my prompt here and you can see we've got a time lapse of the city slowly coming back to life. Let's go ahead and try the prompt. Restore the city to a clean, vibrant, bustling town with happy people and colorful plant life. I'll go ahead and leave the exclusions there.
Speaker 1
00:09:08 - 00:09:28
So length, it says 9 seconds. So let's see the time lapse asks for specifically 9 seconds. So we'll go ahead and leave it at 9 seconds and let's go ahead and make the video and see what we get out of it. Look at that it actually kind of nailed it on the first try. You can see it really quickly change from this dystopian city to this colorful beautiful city with happy people walking around.
Speaker 1
00:09:28 - 00:09:59
I'm going to go ahead and download the video here and we've got our time-lapse for that scene. Inside my folder with all my video assets and image assets, some of these are still shot and I want to make some of these still shots a little more dynamic, a little more animated, make them feel a little more video-like. And there's a couple of ways I can do that. I can use a tool like Leia Pix which sort of adds more depth to the image and animates it so that it kind of looks more 3D. And then I can use a tool like Kyber which adds some interesting effects to it.
Speaker 1
00:09:59 - 00:10:18
So I'm going to go through some of these images and run them through LeiaPix, run them through Kyber, and have a bunch of variations and options that I can then use in my final video. So let's start with Leia. I'll go ahead and upload an image here. Let's start with this 1 of the robot looking at the screen. You can see how it kind of adds that 3D effect to it.
Speaker 1
00:10:18 - 00:10:43
I can change the animation, slow it down a little bit by changing it to 6 seconds here, speeding up by changing it to 1 second. I think I want somewhere in the middle here. Let's go to, yeah, let's go ahead and save it as about 4 and a half seconds. Animation style, You can go horizontal where it sort of goes back and forth a wide circle, a normal circle, a tall circle, a vertical or perspective, which is a newer 1 that I actually haven't seen yet. Let's go ahead and leave it on perspective.
Speaker 1
00:10:43 - 00:10:54
That looks pretty cool. But if it is on perspective, I think I want to slow it down even more. We can change the amount of motion to more or less. So it's a lot more subtle. Put it right in the middle at regular or put it at more.
Speaker 1
00:10:54 - 00:11:07
I think I like regular. Regular seems to be a nice blend there. We can change the focus point. We can make the focus point real close, center the focus point, or make the focus point farther away Which just kind of changes the look. I think I like far.
Speaker 1
00:11:07 - 00:11:24
Now I can click share and save it as an mp4 And we'll go ahead and download this with the rest of our video files. Alright, let's go ahead and upload another image here Let's see how it looks with 2 people in the image. Not quite as cool with 2 people. Let's go ahead and do a sort of horizontal animation style. I think that looks cool for this image here.
Speaker 1
00:11:24 - 00:11:44
I think I like the focus point as far on this 1. And I'll go ahead and download this image. I'm gonna go ahead and repeat this process with all the rest of my images. So I have a pool of images using Leopix that I could potentially use in my video as well. And I'm just going to kind of fast forward through that part, but pretty much going to do the same process with each 1 of my images that I have available here.
Speaker 1
00:11:44 - 00:12:14
So I've gone ahead and used Leopix to convert all my images into something with a little bit more depth. And now I want to use a tool called Kyber and run some of my images through here just to give another kind of cool cinematic, colorful effect to some of my still images. So let's go ahead and click on create video here. We'll begin with an image. Let's grab this image of a robot planting some plants here and let's continue to our prompt and we'll type an AI robot in the style of realistic video.
Speaker 1
00:12:15 - 00:12:27
Let's just see what happens when we try that. The duration. Let's go ahead and leave it at 8 seconds. I can always speed it up or slow it down when I edit it. Let's just kind of leave everything at its defaults and see where it gets us and let's click generate.
Speaker 1
00:12:27 - 00:12:41
All right. And here's the video that it generated. You can see It starts with my initial image and it shows various types of robots planting plants. And I think this will look really cool during that scene of my video. So we can go ahead and upscale the video here.
Speaker 1
00:12:41 - 00:12:53
It's going to use 1 credit. So we'll go ahead and do that. And it says it's going to take about 6 minutes. So while this is upscaling, let's go ahead and work on our dialogue. So we already have the entire dialogue generated by chat GPT here.
Speaker 1
00:12:53 - 00:13:20
This is our narrator. Now to record the dialogue, I'm going to use 11 labs, 11 labs, you get a certain amount of free credits every month, and I'm just going to use free credits and it should do the trick for a pretty realistic sounding voice. So I'm going to create a brand new unique voice for this video. I can do that by coming up to voice lab here and clicking on add voice and then clicking on voice design. Let's go ahead and make it a male voice, middle-aged, American accent.
Speaker 1
00:13:20 - 00:13:25
I'm gonna go ahead and set this you know somewhere in the middle and let's generate and see what this voice sounds like.
Speaker 2
00:13:25 - 00:13:31
First we thought the PC was a calculator then we found out how to turn numbers into letters and we thought it was a typewriter.
Speaker 1
00:13:31 - 00:13:37
So let's go ahead and see what happens when I bring the accent strength all the way up and see how it changes the voice.
Speaker 3
00:13:37 - 00:13:43
First we thought the PC was a calculator. Then we found out how to turn numbers into letters and we thought it was a typewriter.
Speaker 1
00:13:43 - 00:13:59
I really like that deep voice. I think that deeper voice is exactly what I'm looking for, for this sort of dystopian video that I'm making here. So let's go ahead and click use voice and let's just call it deep male voice. All right. Now that we've got this new voice, let's go ahead and use it.
Speaker 1
00:13:59 - 00:14:07
And I'm going to grab all of my dialogue here. So we've got the narrator. I'm going to copy and paste this first sentence here. Paste it in here. Let's go ahead and click generate.
Speaker 3
00:14:07 - 00:14:12
In the darkest hour of humanity when all hope seemed lost a beacon of light emerged from the ashes.
Speaker 1
00:14:12 - 00:14:19
All right let's mess with the voice settings that was a little too robotic. Let's bring the stability down quite a bit and let's test this again.
Speaker 3
00:14:19 - 00:14:24
In the darkest hour of humanity when all hope seemed lost a beacon of light emerged from the ashes.
Speaker 1
00:14:24 - 00:14:44
If you add these little ellipses at the end of a text chunk here it will actually add a little bit of a pause. I think that's kind of 1 of the downsides of 11 labs. It's a little bit hard to control the pace of the voice, but it sounds really good. So I'm going to go ahead and add the rest of our narration in here and just have it record all in 1 long chunk. And then I'll sort of break it up inside of the editing a little bit.
Speaker 1
00:14:44 - 00:14:50
So there is going to be some manual work when I edit it together. All right. So I've pasted the entire narration in. So let's take a listen.
Speaker 3
00:14:50 - 00:15:33
In the darkest hour of humanity, when all hope seemed lost, a beacon of light emerged from the ashes desperate to save civilization, a group of scientists and Engineers worked tirelessly to create the ultimate AI robot with advanced intelligence and an unwavering sense of duty. This AI robot, known as the Guardian, was designed to restore balance and rebuild our broken world. The Guardian traversed the wastelands, bringing aid and hope to the survivors who, against all odds, still clung to life. Together with humanity, the Guardian forged a new path, working hand in hand to rekindle the once extinguished flame of progress. Through knowledge and unity, the Guardian nurtured a generation destined to rebuild and reshape the future of our species.
Speaker 3
00:15:33 - 00:15:40
Over time, the once devastated world began to heal as both humans and AI worked in harmony to create a new era of prosperity.
Speaker 1
00:15:40 - 00:15:55
All right, so that sounds pretty good to me. All right, now that I've generated the audio here, If I come up here to history, you can see the 1 that I just created. So I generated it twice. I think the first 1 came out better. So I'm going to go ahead and select this 1, click download selected.
Speaker 1
00:15:55 - 00:16:14
And now I've got the MP3 of the entire audio here. Now, 1 last step, I want to add some background music to the whole thing. So I'm going to use moobert.com. Moobert is 1 of my favorite sites to generate audio tracks from, because you can actually use it for free as long as you give attribution. So I'm going to go ahead and log in and create a new audio with it.
Speaker 1
00:16:14 - 00:16:30
Let's Go ahead and select moods. We want it to be sort of like a dark, let's set like a serious mood here. We can set our duration. We want it to be roughly 80 seconds. Cause if we remember our film should go to about 1 minute and 20 seconds.
Speaker 1
00:16:30 - 00:17:24
So let's go ahead and just save this as 1 minute and 20 seconds and with a tense serious mood let's generate the track and see what we get out of it. That's almost a little too poppy for for. So let's go ahead and download this 1, added it to my downloads folders. Let's go and click view. You can see here it is.
Speaker 1
00:17:24 - 00:17:43
It's still sort of prepping it for download and we're ready. So let's go ahead and download this 1. It asks where you're going to publish it. So I'm going to go ahead and put my YouTube channel and then I will go ahead and also put my Twitter channel because I'll likely put it on both and we'll go ahead and click continue and we'll agree and download. Alright so now I've got a folder with all of my files in it.
Speaker 1
00:17:43 - 00:18:04
We've got our voiceover, we've got our background music, and then we've got all of our little clips to go along with it. Now it's time to just assemble it all together, and to do that I use DaVinci Resolve. You can use DaVinci Resolve completely for free. Just go to blackmagicdesign.com slash products slash DaVinci Resolve to find it, and I'm going to create a new project here. I'm going to pull in all of these assets that I just created here.
Speaker 1
00:18:04 - 00:18:11
Jump over to my edit page. I'm going to start by dropping my audio onto the timeline so we can listen back to it.
Speaker 3
00:18:11 - 00:18:17
In the darkest hour of humanity, when all hope seemed lost, A beacon of light emerged from the ashes.
Speaker 1
00:18:17 - 00:18:25
And then we've got our background music so I can drag this behind it here. I like to set background music at about a negative, let's go negative 20 here.
Speaker 3
00:18:25 - 00:18:35
In the darkest hour of humanity when all hope seemed lost, a beacon of light emerged from the ashes. Desperate to save civilization, a group of scientists and..."
Speaker 1
00:18:35 - 00:18:51
So now I'm just going to start lining up the perfect clips based on what ChatGPT gave me. So from 0 to 5 seconds we have our opening shot, a wide-angle view of a dystopian cityscape which looks like this in my video. We'll go ahead and pull this onto the timeline and you can see what this looks like.
Speaker 3
00:18:51 - 00:18:55
In the darkest hour of humanity when all hope seemed lost.
Speaker 1
00:18:56 - 00:19:20
So that scene is supposed to go all the way through to where it says In the darkest hour of humanity when all hope seemed lost a beacon of light emerged from the ashes. I'm going to go ahead and stretch out this video. I can do that by right-clicking on the video, clicking on retime controls, dragging this all the way to the end of that section, And that'll slow it down to about 60%, but it will cover that entire area. So let's listen back.
Speaker 3
00:19:20 - 00:19:28
In the darkest hour of humanity, when all hope seemed lost, a beacon of light emerged from the ashes. Desperate to save civil.
Speaker 1
00:19:28 - 00:19:43
And now I'm gonna move on to my next scene. Desperate to save civilization, a group of scientists and engineers worked tirelessly to create the ultimate AI robot. And so we have a dimly lit underground laboratory filled with computers, wires and advanced equipment. So I'm going to go ahead and add that in. Here's what I've got for this shot.
Speaker 3
00:19:43 - 00:19:49
From the ashes Desperate to save civilization, a group of scientists and engineers worked.
Speaker 1
00:19:49 - 00:20:01
So now I'm just going to start working through and adding all of these video clips to the timeline and I'll show you the finished result. Once I've added them all in. All right. So I've officially finished the entire video and I'm going to go ahead and play the whole thing for you right now.
Speaker 3
00:20:01 - 00:20:34
In the darkest hour of humanity, when all hope seemed lost, a beacon of light emerged from the ashes. Desperate to save civilization, a group of scientists and engineers worked tirelessly to create the ultimate AI robot. With advanced intelligence and an unwavering sense of duty. This AI robot, known as the Guardian, was designed to restore balance and rebuild our broken world. The Guardian traversed the wastelands, bringing aid and hope to the survivors who, against all odds, still clung to life.
Speaker 3
00:20:34 - 00:20:57
Together with humanity, the Guardian forged a new path, working hand in hand to rekindle the once extinguished flame of progress. Through knowledge and unity, the Guardian nurtured a generation destined to rebuild and reshape the future of our species. Over time, the once devastated world began to heal as both humans and AI worked in harmony to create a new era of prosperity.
Speaker 1
00:21:02 - 00:21:14
Boom, and there's the whole video. That is my entire short sci-fi film. Throughout that video, I used a ton of different tools. I used Gen 2 to generate some of the videos. I used Leopix.
Speaker 1
00:21:15 - 00:21:30
I used Kyber. I used Genmo. I used Moobert for the background music, I used 11labs for the voiceover, I used Midjourney for some of the images, and then I brought it all together inside of DaVinci Resolve. And that was the resulting video. Hopefully you enjoyed it.
Speaker 1
00:21:30 - 00:21:57
This was a really really fun tutorial to make. This is what I've been wanting to make for a while. It took quite some time to put into it. I know this video was probably fairly quick and rapid-pace but the reality is behind the scenes this actually took me hours and hours and hours to make because using tools like Gen 2 you have to keep trying, keep prompting, give it another prompt, give it another prompt, give it another prompt until you finally get that visual that you're looking for. Same with things like Mid Journey.
Speaker 1
00:21:57 - 00:22:52
All of these tools, while AI did generate all of this, it was quite a bit of effort going through and testing something going that didn't work testing something that didn't work testing something that didn't work until finally I had something where I could finally pull all the pieces together and make it look like something decent so hopefully you enjoyed this video Hopefully you have a better idea of how you can go and create short films. This 1 was only a minute long total, but you can use this same process and create an hour long video if you want. It'll probably take you weeks or even months to do it using the current tools that are available, but you can definitely do it. I just love finding interesting ways to take this tool over here and this tool over here and this tool up here and this tool down here and just mash them all together and create something really unique and creative and hopefully I demonstrated that and how you can do something like that with your own work. And if you love nerding out about this AI stuff as much as I do, head on over to futuretools.io.
Speaker 1
00:22:53 - 00:23:09
This is where I curate all of the coolest tools that I come across. I'm adding new tools every single day. I'm also keeping up with all the AI news for you. So if you come to this tab here, you can find all of the latest news in the AI space. If there's a little too much, you're getting overwhelmed by all this stuff, join the free newsletter.
Speaker 1
00:23:09 - 00:23:45
Every Friday, I'll send you the TLDR of just the 5 coolest tools I came across, handful of news articles, handful of YouTube videos, and 1 cool way to make money with AI. It goes out every single Friday, and it's for the people that just want the TLDR of the week of all of the cool stuff that happened in AI. All you gotta do is head on over to futuretools.io and click on join the free newsletter. Thank you so much for tuning into this video and nerding out with me and watching the process of pulling all these cool AI tools together and making something unique and creative as the end result. If you enjoy videos like this, make sure to give this video a thumbs up and click the subscribe button.
Speaker 1
00:23:45 - 00:23:45
That'll make sure you see more AI videos like this 1 in your news feed and once again really appreciate you See you guys in the next video
Omnivision Solutions Ltd