Adding another tag to this image will fill the entire width of the page when viewed on mobile, and that causes one of the character portrait tags to be pushed down to the next line and it looks weird. I see the utility of the tag, I just can't add it instantly, I have to mess with the layout.
Much sooner than 2040. There are already decent short films: Plastic: https://youtu.be/vtPcpWvAEt0 StormtrooperVlogs: https://youtu.be/NY2nnsAplSo
You could argue it's already too easy because both of these sort of keep going after the script runs out. And there's a lot of inconsistency. Though to be fair, they were rushed to market to be first of their kind.
I think we're going to see low budget anime that does CGI characters with AI filtering pretty soon. It'll probably look better than cheap CGI.
---
For this piece I chose shots where her body looked the best and the motion matched up close enough to cross fade seamlessly (you can't render more than 5-8 seconds at a time currently).
I added the grain effect in editing. It helps to cover up some of the color stabilization issues and I thought it suited the look of the scene. And the process is much more like film directing than traditional animation, since I run batches at night then go through the "dailies" until I collect enough good shots to stitch together into a scene.
Indeed, Link gets it twice from Ganon, blissfully unaware that he's the bad guy. Zelda's warning to watch out for a pig monster wasn't specific enough.
There's a bit of artifacting from the video frame generation, then a bit more added from RIFE interpolation which sometimes has confuses small repeating patterns like fingers or toplip/bottom lip.
It's 4 cycles. I've figured out a technique for finding a good overlap point and then crossfading it with RIFE to make a seamless loop. I also made a script that will stabilize the colors on each frame, because the AI throws in a rainbow drift sometimes.
I've run a bunch of different experiments on page 4 and 5 and not had much luck. It doesn't understand the puffed out cheeks, it can't match the opacity of the smoke, it cooks the colors, it generates amateur or gross looking animation, all sorts of issues. It tends to like darker colors and more detail, but I tried running the middle panel through stable diffusion and making it a photo realistic person, and I still couldn't get good movement - something about the angle maybe.
Page 11 and 12 show some promise using sex and jiggle specific LoRAs.
LC started out as just forced fem slavery fun time, but it has become more complicated than that in my head. I'm definitely at the point where I have more stories than drawing power, so I don't want to promise anything then disappoint.
>Nice, which Ai you use? WAN2.1 It seems to be the best model out there currently, runs locally on 24GB, is trainable and has LoRAs
>Did you have twitter? I used to have a twitter bot that posted whenever a new image was released, but when Elon took over the API broke and I never bothered looking into how to fix it. If you log in (in the upper right), you can setup email notifications.
I took a break, just working on refining some story ideas I had. Now I'm working on a longer commission, and it'll be a little bit before I start posting pages from it.
I tried the scenes from Deep Breath and had difficulty getting good results on the transformation. The sex scenes seem to work well with the available LoRAs, so there's definitely potential.
The real problem with LC1 is that there's a lot of panel breaking with characters overlapping and big gutters between the panels. So there's not a lot of places to pull a clean frame to inference on. I thought about trying a vertical video instead of horizontal but the panels kind of alternate between the two.
I made this using Wan2.1GP, a new open weight AI video model that can run locally and supports image-to-video. I provided single frames from the comic and generated 3-5 second clips and edited them together and added sound effects.
Looks like you'll have to reconnect your patreon account from https://satinminions.com/myProfile I think this was caused by a patreon outage a few days ago. I'll go over the code and improve the error handling and logging of this.
On the technical side, this is images from the comic page cropped out and run through cog video, then upscaled and image-to-image filtered in stable diffusion using a LORA and hypernet that I trained on images of the morale officer. (The comic page itself also has backgrounds that are enhanced with stable diffusion - I sketched and shaded a rough background, then filtered them to add greeblies). I then ran it through RIFE to interpolate from 8fps to 16fps.
This took me a couple days of experimenting and messing around, with a large amount of compute time in between. It takes ~10 minutes to render a 6 second clip so it's a matter of figuring out a prompt and parameters, running a batch over night, then cherry picking the best results.
Hand animating this scene would take weeks, but the results and methods aren't really comparable. You're seeing this scene because it worked, where several other scenes did not. While I was generating clips, it was not clear that the walking shot was going to work at all as there were lots of derpy results, and even in the final clip I couldn't use the whole thing. But the final result is very realistic with shading and 3D camera motion that would be very hard to duplicate traditionally. The same with the shot of her walking down the stairs - it looks just like a handheld camera shot and I think you would need to have a motion capture source to duplicate that.
There are lots of things the model just can't do. Like having her walk into the hallway and turn down the stairs - not happening. It has to go forward from the image you provide, so it can't have a can't have a character walk into frame or hit a specific mark. If this model had the 2 keyframe targets that tooncrafter has it would be amazing, but I don't know if that's possible. Also just on some preliminary testing, I don't think it can make the princess from better this way jump up in surprise in the lightning scene.
It favors realistic shading - if you give it a flat shaded character you can get limited animation or lots of distortion. I've had mixed results on some of my characters but it seems to pick up well on this style of the morale officer. The output from cogvideo is very smooth, but it can do some weird things to faces and the overall shading style, thus I ran it through stable diffusion which has it's own problems and adds some flickering.
Re: Jiggle physics, I've gotten some ok results by feeding it photorealistic stable diffusion renders of girls in loose tshirts. Here's some misc results for science: https://satinminions.com/Cog-Misc-Results.html
The model seems to shine at realistic scenes with minimal movement. You get lots of subtle secondary motion and it holds the character's proportions well. I think there might be a hybrid use where you draw a character, generate an animation, then roto over it. Also for backgrounds and establishing shots, you can add rotation and animation to what would otherwise be an ordinary pan. Another case that seems to work with a little fiddling is techno babble displays, like the brain scan and graph stuff that appears on the morale officer's hud. So there are use cases but it's not the end times yet.
Multi-answer, or approval voting, is the superior type of voting system. It avoids the spoiler effect and you can safely vote for third parties without "throwing your vote away"
For the sake of completionism, here is a result from Kling AI image-to-video. It's another web service with the usual problems that come from using someone else's computer.
The motion and consistency is an improvement over previous models. There's distortion in the face but this might be correctable with a style-specific post process.
The main problem is that this animation took multiple *days* to render. The site is just completely overloaded. This would be acceptable if they had a job queue that honestly said "you are number 45000 in line" or whatever, but instead they have a fake progress bar that gets to 99% and stops. This is apparently a persistent problem even for paid accounts.
Right now I think we're getting the worst end of AI, where it's convincing enough on the surface to blow up social media, but not really practical enough for a big improvement in production.
There is a (bright?) future where a single person will be able to make a complete film using these tools. I don't know what it'll be like when there's a thousand or ten thousand new shows released every year but it sounds like we will live in interesting times.
My worry is that the best tools will remain proprietary and lawfare will be used to remove the open source alternatives.
Some people think that the current tech will cap out and never really be useful - I think this is cope. The capabilities are clearly already there, it's just a matter of controlling them.
It is unnerving for sure - especially when I go to deviantArt or rule34 or even google images and the entire page is destroyed with ai garbo. I've said this before but that's why I feel the need to mess with it, because I need to understand the threat and how to use it.
Oh I guess my other big worry is that we'll enter a completely post-fact world and it'll be impossible to find a verifiably real image of anything. Every book will be rewritten and every image repainted, all recently recorded history will be in doubt, etc.
The stability and movement are impressive and it has a good understanding of materials such as the sequins and satin gloves. It tends to favor a photo realistic style, 3D rendered style, or a generic flat anime look depending on how you prompt it. Sometimes it will cut to a completely different scene, like this: https://satinminions.com/LumaLabs-Combat-Mode-Cut.html
We're fast approaching the point where generative video AI can replace stock footage, things like "man using computer" or "drone shot of new york city at night" can be completely convincing. And of course memes where consistency doesn't matter are great. But I don't think the current model architectures will be able to replicate a style or character without specific training.
The current target audience seems to be the zero-skill user, where you only have to provide the bare minimum - because that's what gets the most engagement. As a "professional" though, I would much rather see tools that required more advanced inputs - for instance, input a complete line animation and a single colored frame and it would propagate that coloring through the animation.