[ Home ] [ all ] [ liz / b / meta ] [ Rules / FAQ / Search / News ]

/b/ -Burrow

Bury yourself in the ground lizzie
Password (For file deletion.)

File: 1692041974555.png (1.43 MB,1412x850,AI crap.png)


Anything related


File: 1692044222848.png (1.36 MB,1000x1298,9RxDYwK.png)

I don't have enough data to make Poppy speak properly.

BUT VA played in many cartoons/anime dubs.
Listened to compilation of Kate Higgins voices, one in Monster High sounded almost the same (and some others)

It doesn't sound like her most of the time, but it did in that comp.
This shit will take so long to make. (6hours series + 10 movies)

Definetely will try to sell it when I am done lol


File: 1692058210582.jpg (24.1 KB,459x480,1551862034715.jpg)

I'm losing my fucking mind trying to get stable diffusion to work.
For some reason venv generates a version of python that is not compatible with pytorch and it's making me insane. I wish I knew how to configure it so that every time a virtual environment is generated it uses the one I got installed (3.10), but I didn't find anything now.
It's also fairly late


File: 1692078877331.jpg (111.33 KB,591x662,animegirls.jpg)

Bahahaha it sounds like your python, pytorch or miniconda crap is not installed on C / root drive.

I've been through this and losing mind part too. That's why I laugh, thought it was just me.

Reinstall everything related to Python then install SD from scratch or delete venv folder and run setup.

You should have no issues after that, but if you do. I will make step by step guide. Just post errors


I wouldn't be too sure. Running whereis python3 shows me that I have it installed in these two folders:
I'm not too sure what it is, but I think there must be something wrong with the ./webui-user.sh file

# Uncomment and change the variables below to your need:#

# Install directory without trailing slash

# Name of the subdirectory

# Commandline arguments for webui.py, for example: export COMMANDLINE_ARGS="–medvram –opt-split-attention"
export COMMANDLINE_ARGS="–recision full –no-half"

# python3 executable

# git executable
#export GIT="git"

# python3 venv without trailing slash (defaults to ${install_dir}/${clone_dir}/venv)

# script to launch to start the app
#export LAUNCH_SCRIPT="launch.py"

# install command for torch
export TORCH_COMMAND="pip3 install torch torchvision torchaudio –extra-index-url https://download.pytorch.org/whl/rocm5.4.2"

# Requirements file to use for stable-diffusion-webui
#export REQS_FILE="requirements_versions.txt"

# Fixed git repos

# Fixed git commits

# Uncomment to enable accelerated launch
#export ACCELERATE="True"

# Uncomment to disable TCMalloc
#export NO_TCMALLOC="True"


Personally I don't think I didn't install it right. But I can't be sure really, the main problem is, what I wrote before really:

This program is tested with 3.10.6 Python, but you have 3.6.15.


File: 1692140026516.jpg (119.44 KB,712x500,6jfms0.jpg)

wait a minute, that looks like nothing Michaelsoft Binbows
*linux flashback intensifies*

Oh hell no, I remember fighting for hours while making bootable Windows iso to runaway from this system for gigabrains and it's constant errors.


Aah, so it's like that, huh. I understand everything now.


Man, that's some rookie stuff. Just change what it should use.
but for real, I have no idea what I'm looking at. Just guessing. You are on your own Lizzie


File: 1692142091064-0.png (2.1 MB,1331x1196,Poppy in drop pod.png)

File: 1692142091064-1.png (1.11 MB,1298x592,WHAT ARE THOOOOSE.png)

I know nothing of linux, just joking.

This is my favorite, Poppy in Drop Pod from Deep Rock Galactic.

I forgot that IMG2IMG is shite and it lacks something.
As effect colors get dull with each generation and on higher resolution black blops.

The only retouch that I made was in paint (like 2minutes for chair and boot).
Rest is Stable Diffusion
I will make 3440X1550 when I fix the issues, will be gorgeous!


What a punchable face


File: 1692292059042.png (1.58 MB,1720x768,00097-862847854.png)

OH YEAH, I figured it all out
now check this out!

I have some things that I want to correct but for today it's enough.
Had many laughs while doing.




File: 1692482190932-0.png (828.62 KB,640x960,00698-3711777755.png)

File: 1692482190932-1.png (850.62 KB,640x960,00697-3711777754.png)

Oh you wanker, I will beat your ass for not reading replies on the slowest imgboard ever.
>>7146 >>7147 >>7148

Don't forget to put .vae into models>vae folder. Otherwise you will have shitty colors and artifacts like I did.

I don't think that you need training that much.
Unless you are some freako with weird waifu.


Easily achievable, you are going to have loads of fun.
It took me 4 minutes to set prompt and settings. No LORA, just model.


File: 1692484248583.png (749.05 KB,640x960,00006-3711777757.png)

Didn't write it clear, I meant writing prompt, on abyssorangemix3 model and hitting generate.
So as you can see, it already looks top tier. You can retouch weird ai fails in inpaint function.
Honestly it shocked me how well it looks. Generated 6 and all stunning. First time tried to make anime lol

I was curious how my LORA would act. unfortunately the quality of yordle art is nowhere near, so my dataset can't make it that sharp and natural. Lack of blue skin ppl in ai models makes huge difference too :d

I ran it with the same prompt.
Don't like it at all, it's better at cute art.
(Except headpats, shit's always dark and terrific, like before-crime scene)

Also LIIIIIZMIN, spoiler image doesn't work.


So I got it to work, manage to make some stuff. I actually found a character model for the two I wanna work with, but now I got some questions.
How do I set the character models? Can I increase the resolution of the art (I got a 6800 XT for reference)? It seems like it takes too much VRAM, but can't I allocate it somewhere else? What are the best settings to have?
Sorry if this is a bit much, I just feel a lot like a kid in a videogame store.


Note: It takes too much VRAM when increasing the resolution of an image to generate


File: 1692495003415.png (1.45 MB,1024x1024,first of many.png)

Apparently I was able to make it work with a square resolution.
But most of the images I'd rather make are different than that.
Take this as a token of gratification


File: 1692495930289-0.png (386.45 KB,768x336,00402-1195893998.png)

File: 1692495930289-1.png (391.14 KB,768x336,00363-3145949797.png)

File: 1692495930289-2.png (333.78 KB,768x336,00364-3145949798.png)

you go into models folder, create new named Lora and put these files there.

Now to make them work:
1. U need a word that calls your Lora, author should write somewhere what is it.
2. Additonally you need activate it, below generate button there is icon with sun and landscape(in the middle), click it, go to Lora.
If you don't see them then just refresh. Click choosen lora. It should look like this <lora:whatevertheycalledit:1> in your prompt now.
:1 stands for strenght of it, author should write which value works best.
That's all


When making image, you don't aim for high resolution at first. You go for 512 pixels at max, you can go up to 764 and 1024 but it's better to first find prompt and seed that you like on lower ones because it's quicker.
Now when you find decent one and want to have it in higher resolution.
(that's why it's worth turning on option to make txt file to each image generated, it's in webui)
There are two ways:
Upscaling with hires fix (most often used) - don't know much yet, seek guides

>not enough vram

Also your batch size may be too big.
You need to test how much your vram can handle. Batch means how many imgs are made at the same time (it's faster than generating one at the time)
After getting not enough vram info, IT OFTEN STAYS INSIDE after.
Then you need to dump the memory from with webui afaik or restart console.
I just close it and run again every time.

>allocate vram somewhere else

I wanted to do that too.
You can't, vram is fast af and nothing comes close - in theory U can but it's so impractical that nobody did it.

>Sorry if this is a bit much,

Nah, all good. I was literally eagerly waiting for you to reply hahaha

I think that covers the basics. Rest you can find on reddit most often although info gets outdated pretty fast. Training used to take so much time and resources, currently it's in a pretty good spot.
also if you bought your gpu recently I would return it and go for some used nvidia as they are better optimized for AI.

Remember it's best to use tags from danbooru.
If my great prompt give out shit I just start over. Tries and many tries.

>best settings

Well it's matter of prefernce. I just run 512 by 512orLower, you adapt steps to resolution.
I use DPM++2M Karras on 30 steps.
Examples in the pic: 768x336. I will upscale them in the future


File: 1692498792305.png (5.09 MB,3680x1600,98C 208F on memory holy sh….png)

AWWW I'm proud of you, thanks!
Well just change the height, or width and no more squares :DDDD

Also you motivated me to try upscaling.
So I just read and best method is what I described previously by IMG2IMG and "resize by".
You keep trying as high as possible till you run out of vram.
Then scroll down and at the bottom click script->SD upscale->choose your scale factor->R-ESREGAN 4X+ Anime6B

My room got so fucking hot reached 98*C /208F. Got kinda scared opened window wide and turned up fans to vacuum cleaner noise levels.
Turned out these temps are fine but I will get additonal fan.

THESE DETAILS THO, except weird hand and arm going into dress. But that can be easily fixed with inpainting.
info: used 3090 24gb, scale factor 2, 7min 4sec
My psu has gimmick of showing how much it draws: 460W


File: 1692499649363.png (1010.44 KB,1024x1024,00000-2961104887.png)

>also if you bought your gpu recently I would return it and go for some used nvidia as they are better optimized for AI.
Ehh I trust that the technology will eventually reach diminishing returns and AMD and Nvidia will be at the same ground eventually. Bought it more than 3 months ago and it was from a refurbish anyway.

The main issue that I seem to be getting is this one:

torch.cuda.OutOfMemoryError: HIP out of memory. Tried to allocate 4.50 GiB (GPU 0; 15.98 GiB total capacity; 6.82 GiB already allocated; 4.53 GiB free; 11.36 GiB reserved in total by PyTorch) If reserved memory is >> allocated memory try setting max_split_size_mb to avoid fragmentation. See documentation for Memory Management and PYTORCH_HIP_ALLOC_CONF


I did eventually follow a youtube guide that helped me set up Loras so that was good.

Did try what you said about upscaling, but now my girl has no teeth :cold_face_emoji:

Also, how would I make the legs from here?


Also, to make this image, I had to use a Lora for an artist that I don't have, but because I didn't have it, it didn't use the artstyle this artist would use. I can use multiple loras without conflict?


>AMD and Nvidia will be at the same ground
I hope for that too

Had that issue.
If I remember correctly u need to edit webui.py
Third answer:

>granny smile

Oh haha, when stuff like that happens you can go to img2img inpainting and regenerate certain part.
If it's fighting it then cut smile from previous good image and img2img resize to right scale with low denoising, paste it. There might be lines, just deal with them by inpainting again.

>making legs

Outpainting for that
There you go: https://stable-diffusion-art.com/outpainting/

>multiple loras

They work on the same canvas so they will mix in together. So your artstyle will change.
You could play around lora strength to see if your character and style stays consistent.
If not then generate first with char. lora then send to img2img and run with style.

Oh you butterfingers, I can see it
>Don't forget to put .vae into models>vae folder
Go get it, colors won't be dull


oh dhis is givving me he cemichals in the brain i can't tipe


File: 1692552810266.png (25.23 KB,192x166,5f2439a6f9e2cb758be415d8d8….png)

Man, how do I use an artstyle Lora? I found an artist that I followed a lot recently and this is making me shiver with how good the results can possibly be.


File: 1692569141875.jpg (67.86 KB,621x642,3a6edf06e36331d26099fb79cf….jpg)

The same way you use normal lora.
I wrote it previously too.

Ouch my feelings


Sorry I have an attention defecit disorder.


Oh then no need to be sorry, everything is fine :D
I just assumed something else.

you go into models folder, create new named Lora and put these files there.

Now to make them work:
1. U need a word that calls your Lora, author should write somewhere what is it.
2. Additonally you need activate it, below generate button there is icon with sun and landscape(in the middle), click it, go to Lora.
If you don't see them then just refresh. Click choosen lora. It should look like this <lora:whatevertheycalledit:1> in your prompt now.
:1 stands for strenght of it, author should write which value works best.
That's all


File: 1692617684522.png (257.48 KB,512x512,00000-2836307600.png)

when using inpainting, should I try to generate the whole thing again? or just an attribute? Not this but in another image for example the face came out realistic and I wanted to get rid of that, besides another minor quirks. How would I go about doing that? I highlight what I want to paint over and then just write "smile, face"?
In the end I fixed the issue by unticking "Restore Faces"


Just attribute.
Exactly as you write.

Your image is 512x512 and face takes small part of an image.
So inpainting may struggle to make it right with low pixel count.

Restore faces just "saves" previous face/expression when running through img2img.
It's not for fixing face, I thought that too at first.
I only use it when the resolution is big and face seems perfect.

Did you try other images ratios?
Like horizontal and vertical.
You just change size in settings, it often makes generation more lively and interesting when compared to 1:1


File: 1692686277062.jpg (39.42 KB,567x557,1665331353960388.jpg)

I did eventually configure my GPU so that it can run these kind of images now.
I am also figuring out that the Noise Volume modifies images look when using img2img. Which makes making hybrids of loras easier to understand, but still not much easy to apply.
But having tons of fun with this, I just installed an app on my phone that let's me run SD remotely from my PC to my phone, although the options you get in them are quite limited. Can't upload the results I got for some reason


File: 1692814770270.webm (1.67 MB,720x720,b95737345c749ea75d080dbbb….webm)

imade my firstntuabledrawingUugwaaaaaaaa


File: 1693971700483-0.png (1.54 MB,1056x1113,asdgfdhgdhf.png)

File: 1693971700483-1.png (1.59 MB,1440x1048,00080-3770623158.png)

Feel no more need to fiddle with settings for LORA it's satisfactory for me.
Now I feel quite proud but also empty.
Idk what to do now


How about making a doujin?


Great idea! Thank you

Going to be a huge pain in the ass to make AI do exactly what I want.
I don't think that creating new one from scratch without any drawings or 3d modeling will work.
Will test how it goes on basing on some existing one.


File: 1694082015708.jpg (265.37 KB,1116x967,lol.jpg)

hahaha that loss of detail in last panel
Most likely due to low pixel rate, it was 1424x528 but not high enough for face.

It took quite a bit to make, most time of to recolor skin as it doesn't work as it needs a correct base color lmao

Using 3d model and editing software would make it look pretty nice and for sure faster


That looks phenomenal


File: 1696528332195.png (567.39 KB,720x512,01075-2597429106.png)



File: 1696546308352-0.png (899.66 KB,760x536,106603629.70500003_image.png)

File: 1696546308352-1.png (968.09 KB,760x536,106671115.89000002_image.png)

File: 1696546308352-2.png (490.66 KB,720x512,01191-469877468.png)

File: 1696546308352-3.png (791.83 KB,513x663,85764927.08_image.png)

ahahah so thick, I don't know what's up with it.
It either makes them flat or gives huge milkers.
Even if I blacklist words


You tried ComfyUI?


Perfectly thick. It is as things should be.


Hmm I falsely thought that Stable Diffusion understands concepts and activities.
Thought it had some understanding on it's own. But it's just dumb mashing shit up that looks good.

Not yet , but that's on my radar.
Next up are OpenPoses and Dynamic Prompt.
After learning these extensions I'm going to jump on blender, make poses and set light.

Can't wait till POV option gets better and multiple character generation is fully supported.

I tried making some more stuff today but everything looked bad.
BECAUSE I forgot to specify the breast size… and that was lowering overall quality lol

My Poppy's voice project is on halt because I don't feel like watching movies just to cut audio.
Last one gave me so little, damn VA was switching between voices hard af even though it was only one.


File: 1698978098899.png (497.1 KB,512x720,00029-1220187055.png)



I do


File: 1706453307281.png (913.33 KB,699x694,pops runner.PNG)

blade runner Poppy lol


File: 1712241799737-0.png (859.39 KB,832x1216,00044-3814088809.png)

File: 1712241799737-1.png (1004.45 KB,832x1216,00076-814888908.png)

I dunno if Poppyfag is still lurking, but I'd like to know if you can give some tips to making a good gen.
Personally I feel like mine are kinda boring, and the hands also could get an improvement.
Are you still on 1.5 or did you make the jump to XL?


File: 1712242390392.png (3.49 MB,1536x2560,1712107098079416.png)

Take for comparison this gen, it feels much more attractive as in, colours, the lines are more defined, seems like there is more detail too.
Also nothing seems screwy.


File: 1712602502486.png (9.69 KB,270x218,rip.PNG)

>I dunno if Poppyfag is still lurking
I'm here everyday

>some tips

Try different models/LORAs, you can just copy prompts from people that use the same model.
Best site I know is https://civitai.com/images?tags=4


read some inpaint guides

>1.5 or XL

I didn't try XL, I had huge plans to learn Bender, pose and set specific places for the background.
But I spiralled into doing nothing creative and slacking off, then my drive ran out of space so I had to get rid of python and other SD related stuff.

Wow that really does look good, not even giving off ai vibes.


>didn't try XL, I had huge plans to learn Bender, pose and set specific places for the background.
I see. I made those past gens with PonyXL because it's what anons recommended. I think it looks neat for gens that didn't take much hassle, but for some reason I don't think I want big tiddy now :C
>But I spiralled into doing nothing creative and slacking off, then my drive ran out of space so I had to get rid of python and other SD related stuff.
Oh that sucks. Hope you can get a new driver soonishly. And that you find reasons to go back to genning.


File: 1715321262554.png (6.83 KB,522x136,AAAAAA.PNG)

>I made those past gens with PonyXL
Good stuff Liz
>get a new driver soonishly
Look Liz I have new B: drive.
Not really, I'm trying to get koboldai working and screwed something up and it gave me new magical drive! :-DD

Also tried kobold.cpp something that is supposed to work with the most stupid ppl but it doesn't launch for me ahhh


File: 1715359118659.png (29.37 KB,666x455,wtf.PNG)

Reinstalled on another drive so I can have space to do ai stuff again and holy shit what have I done hahah

SOMEHOW I screwed it up lol
Installed on Disk 0 C: BUT it doesn't boot up if the Disk 2 is gone!

And "SSD" particion is lost probably because of lack of letter but I'm not going to risk bricking my pc, since I don't have any flashdrives.

I will be a hackerman one day, you will fear me one day

[Return][Go to top][Catalog][Post a Reply]
Delete Post [ ]
[ Home ] [ all ] [ liz / b / meta ] [ Rules / FAQ / Search / News ]