Cavendish

joined 1 year ago
[–] Cavendish@lemmynsfw.com 1 points 1 week ago

Just got the new 0.4.2 testflight version. That is a very long list of updates and fixes! Thanks for continuing to make improvements.

[–] Cavendish@lemmynsfw.com 0 points 1 year ago (1 children)

In the Expanse books, there's a planet called Auberon that has an 8 hour rotation, so 4 hours of light and 4 hours of dark. They decided that "1 day" would be light-dark-light and "1 night" is dark-light-dark. It's really interesting how they describe the way society adapts to the cycle of having a midnight sun and both a midmorning and evening sunset.

[–] Cavendish@lemmynsfw.com 1 points 1 year ago

I thought i had a good system where each outpost was only exporting 1 solid, 1 liquid, and 1 gas. This allowed me to isolate and sort at the receiving outpost.

The problem occurs when each outposts import & export containers get full. At which point materials flow from the export station, go to the import location where they can’t be unloaded, THEN THEY COME BACK to the original outpost where they get offloaded. You wind up with all the same materials filling both the import and export containers. Now the entire material flow is completely borked, nothing is getting imported, and all you have access to is the stuff thats locally produced.

[–] Cavendish@lemmynsfw.com 2 points 1 year ago

If you look at the ground map when you land on a "coast" biome, you can see where all the elevation dots go flat. That will help direct you to the ocean if you can't see it over the hills/fog.

[–] Cavendish@lemmynsfw.com 2 points 1 year ago

I have a bunch of Akila City security guards that wander around in their underwear in my game. I just assumed it was casual Friday.

[–] Cavendish@lemmynsfw.com 1 points 1 year ago

Oof. Clearly I hadn't tried it.

[–] Cavendish@lemmynsfw.com 1 points 1 year ago (2 children)

Even with that trick, going from 1000 down to 1 is still painful. Probably faster to give them all 1000, then switch and take 999 back.

[–] Cavendish@lemmynsfw.com 1 points 1 year ago

Stable Diffusion always (? I think anyway) puts prompt metadata into the output image. The problem is that it's easy to strip it out converting to jpg or other formats. Even just uploading to Lemmy will strip the metadata, That is why I use catbox.moe which preserves all of that info.

[–] Cavendish@lemmynsfw.com 5 points 1 year ago

Post it again, and screw the downvotes. I thought that image was pretty good and am kicking myself now for not saying so before you took it down. Please don't lurk, this community needs more diversity.

[–] Cavendish@lemmynsfw.com 3 points 1 year ago (2 children)

Normally, you can use a tool like pngchunk.com to read the metadata, but I just realized that I did an SD Upscale on this one and it didn't preserve the prompt. Sorry about that, I'll put the entire metadata dump below in a spoiler. I'm not sure it'll be that helpful though, this image uses a custom LoRA I'm working on that I haven't released yet, and is complicated by the fact that I started with one model (ToonYou beta 6) for the first 40% of the generation, then switch to a realistic checkpoint merge for the last 60% as a refiner.

prompt


{ "parameters": "Photo of a young 21yo woman posing in (cav_rdrguarma:1.5),\nmasterwork, best quality, soft shadow\n(Photograph with film grain, Sony A7 camera, f1.2, shallow depth of field, 85mm lens),\nnight photo of a jungle \n\nADDCOMM\n\n(sun-kissed to honey ombre hair color in a voluminous curls style:1.2),\nADDROW\n\npendant necklace,\n(light khaki smocked bodice sundress with a flowy skirt and puff sleeves:1.1) (top pulled down showing breasts:1.2) droptop \n\n(flat breast, normal_nipples :1.3), \n(tan lines, beauty marks:0.6)\n\n(SkinHairDetail:0.5)\n\nNegative prompt: (child, childlike) BadDream UnrealisticDream Asian-Less-Neg\r\namateur, blurry, logo, watermark, signature, cropped, out of frame, worst quality, low quality, jpeg artifacts, poorly lit, overexposed, underexposed, glitch, error, out of focus, \r\n(semi-realistic, cgi, 3d, render, sketch, cartoon, drawing, digital art, anime, manga:1.3), \r\n(poorly drawn hands, poorly drawn face:1.2), deformed iris, deformed pupils, morbid, duplicate, mutilated, extra fingers, mutated hands, poorly drawn eyes, mutation, deformed, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, incoherent,\r\ngrayscale, jeans, denim\nSteps: 50, Sampler: DPM++ 2S a Karras, CFG scale: 7, Seed: 1046635747, Size: 640x832, Model hash: e8d456c42e, Model: toonyou_beta6, VAE hash: 63aeecb90f, VAE: vae-ft-mse-840000-ema-pruned.safetensors, Clip skip: 2, RP Active: True, RP Divide mode: Matrix, RP Matrix submode: Rows, RP Mask submode: Mask, RP Prompt submode: Prompt, RP Calc Mode: Attention, RP Ratios: "1,1", RP Base Ratios: 0.2, RP Use Base: False, RP Use Common: True, RP Use Ncommon: False, RP Change AND: False, RP LoRA Neg Te Ratios: 0, RP LoRA Neg U Ratios: 0, RP threshold: 0.4, RP LoRA Stop Step: 0, RP LoRA Hires Stop Step: 0, RP Flip: False, Lora hashes: "cav_rdrguarma-v4: 0540b2c6b046, cav_rdrguarma-v4: 0540b2c6b046, droptop: 24494c0ed389, Breasts_Helper_Trail_v2: 470f04826a09", TI hashes: "SkinHairDetail: edf710bf1ea5, BadDream: 758aac443515, UnrealisticDream: a77451e7ea07, Asian-Less-Neg: 22d2f003e76f", Refiner: Cavendish_Hotel [8ead5e5021], Refiner switch at: 0.4, Version: v1.6.0, Hashes: {"vae": "735e4c3a44", "embed:Asian-Less-Neg": "22d2f003e7", "embed:BadDream": "758aac4435", "embed:SkinHairDetail": "edf710bf1e", "embed:UnrealisticDream": "a77451e7ea", "lora:cav_rdrguarma-v4": "aac45a2863", "lora:droptop": "917fcd35a6", "lora:Breasts_Helper_Trail_v2": "124fe77b5d", "model": "e8d456c42e"}\nTemplate: Photo of a young 21yo woman posing in (cav_rdrguarma:1.5),\nmasterwork, best quality, soft shadow\n(Photograph with film grain, Sony A7 camera, f1.2, shallow depth of field, 85mm lens),\nnight photo of a jungle \n\n, \n\n(sun-kissed to honey ombre hair color in a voluminous curls style:1.2),\nBREAK Photo of a young 21yo woman posing in (cav_rdrguarma:1.5),\nmasterwork, best quality, soft shadow\n(Photograph with film grain, Sony A7 camera, f1.2, shallow depth of field, 85mm lens),\nnight photo of a jungle \n\n, \n\npendant necklace,\n(light khaki smocked bodice sundress with a flowy skirt and puff sleeves:1.1) (top pulled down showing breasts:1.2) droptop \n\n(flat breast, normal_nipples :1.3), \n(tan lines, beauty marks:0.6)\n\n(SkinHairDetail:0.5)\n\nNegative Template: (child, childlike) BadDream UnrealisticDream Asian-Less-Neg\r\namateur, blurry, logo, watermark, signature, cropped, out of frame, worst quality, low quality, jpeg artifacts, poorly lit, overexposed, underexposed, glitch, error, out of focus, \r\n(semi-realistic, cgi, 3d, render, sketch, cartoon, drawing, digital art, anime, manga:1.3), \r\n(poorly drawn hands, poorly drawn face:1.2), deformed iris, deformed pupils, morbid, duplicate, mutilated, extra fingers, mutated hands, poorly drawn eyes, mutation, deformed, dehydrated, bad anatomy, bad proportions, extra limbs, cloned face, disfigured, gross proportions, malformed limbs, missing arms, missing legs, extra arms, extra legs, fused fingers, too many fingers, long neck, incoherent,\r\ngrayscale, jeans, denim" }

[–] Cavendish@lemmynsfw.com 1 points 1 year ago

Thank you! I use a bunch of different custom merges. See this very large xy comparison grid i posted earlier: https://files.catbox.moe/1k6mmr.jpg

I'd recommend Absolute Reality or LazyMix for an off-the-shelf model.

[–] Cavendish@lemmynsfw.com 5 points 1 year ago (2 children)

There's not much out there on training LoRAs that aren't anime characters, and that just isn't my thing. I don't know a chibi from a booru, and most of those tutorials sound like gibberish to me. So I'm kind of just pushing buttons and seeing what happens over lots of iterations.

For this, I settled on the class of place. I tried location but it gave me strange results, like lots of pictures of maps, and GPS type screens. I didn't use any regularization images. Like you mentioned, i couldn't think of what to use. I think the regularization would be more useful in face training anyway.

I read that a batch size of one gave more detailed results, so I set it there and never changed it. I also didn't use any repeats since I had 161 images.

I did carefully tag each photo with a caption .txt file using Utilities > BLIP Captioning in Kohya_ss. That improved results over the versions I made with no tags. Results improved again dramatically when I went back and manually cleaned up the captions to be more consistent. For instance, consolidating building, structure, barn, church, house all to just cabin.

Epochs was 150, which gave me 24,150 steps. Is that high or low? I have no idea. They say 2000 steps or so for a face, and a full location is way more complex than a single face... It seems to work, but it took me 8 different versions to get a model I was happy with.

Let me know what ends up working for you. I'd love to have more discussions about this stuff. As a reward for reading this far, here's a sneak peek at my next lora based on RDR2's Guarma island. https://files.catbox.moe/w1jdya.png. Still a work in progress.

3
submitted 1 year ago* (last edited 1 year ago) by Cavendish@lemmynsfw.com to c/bean@lemmy.world
 

Just a quick test.

spoiler


Regular spoiler

Fancy Custom Spoiler


Fancier spoiler

view more: next ›