-itten by a smart person. This painting has lots of detail? Must be a good artist. But now AI is here that can generate surface level bullshit on demand. An offhand comment can become an essay, a doodle can become a painting. For some people, their entire way of judging things is being disrupted. Now the content and specificity of a work needs to be understood, and that is much more difficult.
I am not above this frustration myself. AI is used to crap flood plausible looking content onto all platforms now. I find myself quickly checking images for the tell-tale signs of AI, listening for verbal ticks in youtube shorts, just to quickly determine if I should ignore them. But it ain't going away. In the future there is going to be more AI in the world, not less. It's better to understand it than be taken by surprise.
This turned into a minor blog post, don't mind me:
Stable diffusion works best on single character scenes in standard poses. Pinups basically. It's also pretty good at backgrounds.
The issue is that it doesn't reliably understand relative conditional statements or adjectives. So if you prompt it "a blue ball on top of a red box", it will give you that... sometimes - but you'll also get random combinations of red, blue, ball and box. It gets worse the more adjectives and conditionals you add in.
In the case of multiple characters it's almost impossible to apply the conditionals correctly. If you say "a woman with tan skin and red hair putting bunny ears on a kneeling woman with short blonde hair wearing a red leather bustier", you're gonna get red hair on both of them a because that's a common thing and you have the "red" token in there twice. It's also going to screw up normal features twice as much because there are more things in the scene. Even getting a single character that looks good often requires rolling the dice dozens of times because random bits will be screwed up.
I have done some multiple character scenes by inpainting each part individually, but the problem is you end up with a slightly different style and lighting on each part. Plus it's very time consuming. The patchwork look is a problem even on single character composite images, it's easy to accidentally stray into something that looks like a collage or that un-tooned homer simpson meme. The recent gold dress pinup kind of strays into that territory.
Dall-E is better at comprehension but it's proprietry and too large of a model to run on consumer hardware at this time. I'm generally not interested in AI models that I can't run locally and I'm especially disinterested in corporate mandated artificial brain damage. Even if you're not trying to make edgy stuff, unfiltered models are just better because they have a more complete understanding of the world. Even stable diffusion has fallen into the "trust and safety" trap and it looks like future developments will have to be underground.
The control net extension, which I used to make the latest Alison pic, is a major improvement, but it still doesn't solve multiple characters or animations. I think the possibility is there. I could see something being added like a segmentation map where each segment could be given different prompts. Temporal stability has been shown to be possible in things like nvidia's styleGAN, and some newer text-to-video models. At some point you will be able to go from a sketch animation to a perfect render. The capability in the AI model is there, it just needs to be activated appropriately. Similar to how chat-gpt is an activation layer on top of gpt-3.
I've done a lot of AI tinkering instead of drawing lately, and some people don't like it - but I hope everyone can appreciate that this an existential crisis in art. Lots of people are in "anger" and "denial" stages of grief. I've had some truly bizarre discussions on other forums where I try to demonstrate SD's ability to generate backgrounds and they will start picking apart some 3 pixel high blob of a bush in the distance because it's not an exact technical drawing. Like, have you ever seen a painting by a person? Bob Ross? The guy just smooshed his brush on the canvas and it looks great. A bunch of people were upset that netflix made an anime short using AI for the backgrounds - but tons of anime have been using crappily filtered stock photographs and 3D models for backgrounds for decades, AI could only improve this situation. Even big budget titles frequently use painted over photographs because even among artists, very few people can generate an accurate scene entirely from their mind.
A big problem is that a lot of people are walking around without any comprehension of what they're looking at or reading or listening to. They just make value judgements based on surface level traits that, in the past, have reliably served as proxies for quality. There's a bunch of big words here? Must be wr
Writing all that makes it seem like some huge process. It's way easier than drawing something. But a breakdown of the drawing process is like:
Draw some lines. Color 'em in.
The main bottleneck for AI is the time it takes to generate images. If my computer was infinitely fast, it would only take minutes. As it is now I usually queue up a batch to run when I'm afk, which isn't often.
The first step is to dial in the prompt and settings. I test different models and denoising strengths to see what looks good. Sometimes it has trouble understanding a pose until you give it the right prompt. In this case it understood it fairly well, but specifying hands_on_hips, short_sleeves, and pencil_skirt helped.
Once it starts looking good, I generate a bunch of images with the those settings. Then I piece together the good ones in photoshop and do a little retouching. It's rare for one image to have good everything so I'll take the face from one, the body from another, etc. The hands are usually messed up and have to have parts redrawn.
Then I export the composite image and upscale it 3x. Then I'll do AI inpainting on key parts to high-res-ify them. In this case, her face, her boobs, and her waist. I tried her skirt a couple times but I ended up keeping the original.
The main thing for this one is that I knew it would work since I had already done the prompt testing part at low resolution on my old video card. In general a simple pinup pose like this is easy since there is lots of training data for it. I have yet to get good results on unusual angles like this https://satinminions.com/Suction-Dildo-Shower-04.html or first person shots.
It is possible to just run an image through and get pretty good results. But if you want a high res output where you struggle to find flaws, more work is required. Here is the raw output for this image: https://satinminions.com/SD-15675-Raw.html
The premise is that she has an organic brain inside a synthetic body. The brain would need food, water, and oxygen to survive, but a greatly reduced amount compared to a full human body. These needs would also be regulated and buffered by the synthetic systems. A chemical oxygen generator and CO2 scrubber can be hundreds more efficient natural hemoglobin, allowing her to hold her breath for hours if necessary - although that would be uncomfortable as the brain is accustomed to breathing.
The brain would also still require sleep, although the requirement for the "slow wave" period where they body repairs itself may be reduced. So she may feel suspiciously un-fatigued at times.
A synthetic body would probably need to repair itself though. Think about it, how many times can you bend a usb cable back and forth before it stops working, how long do your clothes last? Anything soft is going to need to be self-healing or it'll turn into dust in a matter of months. And then there's the matter of expendable lubrication.
So I'd imagine she'd have some kind of digestive system and eat small specially prepared supplements. I had an idea for a little story page about this.
For power, she'd probably have to be nuclear. You just can't get enough power density out of chemical batteries and the charge anxiety would be too much. I had an idea where when she's first installed she's on battery power and feels cold and lethargic, then she goes down to the reactor room to have her permanent power supply installed and immediately feels much warmer and energetic.
AI Smoothing does not work well here because the original animation is kind of jank and the smoothing it just makes the jank more obvious.
For a full remix, stable diffusion does not have temporal consistency, which means you can't really get the same character out of it twice, even with the same settings. It ends up looking like every frame was drawn by a different person.
AI really struggles with hands and feet, especially in unusual orientations like this. I've run about 100 iterations and they mostly turn out as weird blobs.
I've discovered that things can be greatly improved with inpainting and a little touchup in photoshop. I was able to fix her face and hand. If I can manage to find a new graphics card I'll be able to iterate a lot faster.
The top images are determined by Elo score from which thumbnails get clicked on in the engagement widget (the group of 6 random thumbnails that appear next to/under images). It's not strictly random, it's slightly more likely to show already highly ranked images and new images, although I tweaked it to not show too many locked images recently because I thought the frequent upsell might be annoying.
It's a pretty consistent measure of how appealing an image is. Covers don't show up because they're often not as stimulating as an action scene or as curiosity provoking as a thumbnail with tiny text. There's also the bottom 50 https://satinminions.com/bottom50 which shows a pretty clear bias for old uncolored sketches where not much is happening.
I made the cover just recently as part of the website rewrite and reorganization. It's scheduled to go public at the end of the month. Same with the POTG2 cover.
I kept this as patron bonus content for quite a while because it was a random concept cover that wasn't attached to anything. I didn't want to indicate to the public that this was a new upcoming comic when it was not.
After the recent website rewrite, it is much easier to reorganize comics. So I used this as a cover for a collection of the other Wetware concept pieces.
This is a click-through comic with many panels and animation effects. It's effectively a subchapter of the story.
The images directly after this are the static images for the first subchapter. Like a CG set for a h-game. I'm not entirely sure how to organize everything.
[Mistress tying the man's feet] Man: Is this necessary? Mistress: You play in my bed, you play by my rules. Man: Uh.. fair enough I sup-oof [Man is interrupted as Mistress pulls him back, tightening the ropes on his wrists] Mistress: Now then, I *am* going to have to torture you for a bit. Man: t.torture? Mistress: Mmm, yes. If you have been possessed, it'll only come out when you're at your most primal level of consciousness. Oh don't worry, you'll like it...At least at first.
Also, did you type that "+" in your username or did it show up on it's own? I had some trouble with encoding/decoding spaces but I'm not sure how there'd be one space and one plus.
You don't appear to be logged in (your comment does not have a user id). In the upper right corner of the page, do you see "Login" or do you see your username? If you click that link do you see your profile or do you get a login prompt?
Also, your account does not have a patreon id. After logging in here, click the "Connect with Patreon" link from your profile. After signing in on patreon.com and clicking "allow", it should redirect you back here and you should see your tier level listed in your profile. If you already did this, did it seem to work or were there any errors?
Volume 1 and 2 are packed fairly tightly together, it's a continuous series of events. You could slot this in somewhere after the end of Vol.2 and before page 19 of Vol.3 - that's when she decides she's being too easily swept up in things like this.
I use an old Intuos 3 tablet, which is better than a lot of new tablets because of the button cluster. It has several buttons grouped together that can be differentiated by feel along with a scroll strip. The trend with newer tablets is to have a row of buttons down the side which is not as quick or easy to press.
I used Krita this image, and most of the recent images. It is free and open source. It's good for drawing but if I need to do a complex layout with masking, lots of text, or a super smooth gradient, I'll go back to Photoshop CS 9. CS9 is old but it does not require a subscription or internet activation.
For a few of the really smooth vector animations, I used a program that I wrote myself.
The hair changes a bit when the face shift happen because male and female skulls have different shapes. Female typical hairlines are also more rounded than male typical hairlines. I fudged it a bit to look cute during the rest of the animation.
The hair becomes long later because that's what the potion does. I'm not sure why some people get hung up on hair growth as unrealistic when *bones* were changing size moments earlier.
The clothing is actually crucial to showing the changing contours and size of the body. It also flaps around to show motion.
It was mainly an aesthetic choice here. There definitely is a servant hierarchy but it's not so strict. The three of them are basically on the same level with their chunky collars and menial duties.
The two maids act superior to Alison because she is shorter, younger and newer. They're a little taken aback at her ability to "climb the ranks" so quickly, but they take it in stride.
I am not above this frustration myself. AI is used to crap flood plausible looking content onto all platforms now. I find myself quickly checking images for the tell-tale signs of AI, listening for verbal ticks in youtube shorts, just to quickly determine if I should ignore them. But it ain't going away. In the future there is going to be more AI in the world, not less. It's better to understand it than be taken by surprise.