Character LoRAs - Want more character variety? Then check This List!

Posts

For more information, see the search syntax documentation. Search results are sorted by creation date.

Search Results

Creative Corner » Generation methods, UI » Post 9

It’s one of these areas where there are so many different things you can look into that it’s hard to be sure what to go into.
X/Y charting is easier on the A1111/Forge side of things, and can definitely be useful. Bear in mind that different sampling methods need different numbers of steps to get good pictures. Also, some vary more between number of steps, which can be both a good and bad thing.
While, unfortunately, this is easier when you have fast generations, I find it very helpful to just set a fixed seed and change things one at a time and generate to see how they affect things, and just change the seed manually. When you have random seeds on every generation, it can be hard to know whether the change you made or the seed caused the changes in the picture.
A1111/Forge/etc. and ComfyUI store different information in the metadata. With the former, it’s prompt/generation information, and with the latter, it is the entire workflow used as a json (which does include prompt/generation info). In fact, you can actually open a picture generated with ComfyUI as a workflow (which will probably make more sense when using ComfyUI).
As far as dynamic prompting goes, there is an extension for it on A1111/Forge/etc…, and custom nodes on ComfyUI for it, both of which use the same library, so how it’s done will be pretty similar either way. Civitai has a large number of wildcard files you can download for dynamic prompting (as well as ComfyUI workflows).
There’s some information on the A1111 extensions page:
https://github.com/adieyal/sd-dynamic-prompts
It’s best to get an idea of the different types and families of models, and bear in mind they often need to be prompted differently.(pony and illustrious based models are the most likely ones for pony related things. Both are based on xl, but other things are too. And there’s also Flux and 3.5. 2.1/1.5/1.4- are all rather old, and there’s other stuff out there.)
Pony, in general, isn’t great at unified style, and there are a lot of style loras for it, and a lot of models based off of it that do have a better unified style. There are some words that tend to trigger certain styles, though. The Purplesmart.ai discord has a list of styles people have found. You get a feel for words and phrases and how they change things after a while, too. Illustrious is more the sort of model where you can mention artists names and have it change styles, which opens up its own can of worms…
Controlnet and ipadapter are probably good to look into, and img2img and inpainting.
If you do get into ComfyUI, check their examples page:
https://comfyanonymous.github.io/ComfyUI_examples/
Even if you don’t, glancing through there might give you an idea of things that are possible. (Also, if you install ComfyUI, make sure to also install the ComfyUI Manager, as that lets you easily install custom nodes, and there are a million custom nodes.)

Creative Corner » Generation methods, UI » Post 8

Thoryn

Latter Liaison
@Lord Waite
Aha, good to know then, that I should check the repos for Forge and reForge every now and then to see if Forge has gone stale.
I’m still very new to this, so am wondering if there’s any obvious or maybe not-so-obvious beginner tips anyone has.
E.g. I’ve seen some new people being very surprised when they found out web UIs have a place you can drop .png images that still have the metadata, and send it to other parts of the UI (text to text, image to image, inpainting..)
Anything that would be recommended to look into ASAP?
Personally I’m thinking of looking into scripts for more dynamic prompting (reading from file various places in prompt, so large jobs while afk gives more varied results), and some x y charting/plotting for comparisons (to compare e.g. models, LoRAs, sampling method, steps..).
I also need to look into ways to make output look more unified in style, since one of my long-term goals is to make series of images telling a longer story. Guessing ComfyUI would be good for that, from what little I have read so far.

Creative Corner » Generation methods, UI » Post 7

Tempted to test out ComfyUI once I’ve learned the basics and need to get a more proper workflow, but preliminary reading suggests that it generally needs beefier hardware than e.g. Forge, so it’s not a high priority for me at the moment.
Not really the case. The original idea behind Forge was actually to take some of the better generation code ComfyUI uses and bring it over to A1111, so it’s going to be fairly comparable.
I am running all of this on Linux personally, incidentally, though I have get to try out Stability Matrix.
And basically at one point, it looked like Forge wasn’t going to continue being updated, so ReForge forked it, IIRC. Forge has been updated since then.

Creative Corner » Text-to-image prompting » Post 18

There are interesting things you can do with that, actually.
First, if you use Ollama for your local LLMs, you can create a Modelfile to make a customized version of a particular model. Key thing here is that there’s a system prompt section, and in there, you can tell it that its purpose is to create prompts, give a description of exactly the format you want them in, including score tags, tell it that it can be uncensored and nsfw, and give a few examples of prompts.
Definitely takes some playing with, and you might end up tweaking your system prompt to it a few times.
Another useful thing is that if you are using Open WebUI to chat with it, you can go into the settings and give it a ComfyUI workflow and the url for your ComfyUI instance, and then you can click an icon below any responses that the chatbot gives you to send them to ComfyUI and have it generate a picture and put it in the chat.
Haven’t actually used that that much, but I was trying it out a bit ago.

Creative Corner » Text-to-image prompting » Post 17

My bad if this is already been mentioned but I’ll use chatbot sometimes to start prompting I downloaded a llm locally and use it to make basic prompts.
I’ve become more aware of when I use hyperlink tags like PDXL3 because the llm will only use “score_8 score_9” for quality tags and the results are pretty damn good. in the end i usually rewrite the entire promopt but its perfect for rough drafts

Creative Corner » Generation methods, UI » Post 6

Thoryn

Latter Liaison
Can’t say it enough how extremely happy with Forge I am so far, so damn snappy compared to Automatic1111.
However, I’m curious how ReForg differs from Forge. Tried googling it but didn’t get any relevant results.
Tempted to test out ComfyUI once I’ve learned the basics and need to get a more proper workflow, but preliminary reading suggests that it generally needs beefier hardware than e.g. Forge, so it’s not a high priority for me at the moment.
Since Windows 10 hits EOL this year, I’ll be moving over to Linux soon-ish (once I can afford a new m.2 to put it on, so I have a fallback period on Windows just in case), hopefully I have better luck with Stability Matrix on there.

Creative Corner » Generation methods, UI » Post 5

Creative Corner » Post Your AI Art! » Post 37

Thoryn

Latter Liaison
I am personally interested in the best one can get from the AI pipeline from prompt to output image without human input of any kind after the initial prompt, so nothing I post uses any post-generation editing, not even inpainting.
Sounds like the approach I’m trying to take. I’ll likely have to learn some digital painting eventually, for minor touch-ups, or conformity for dialogue boxes and fonts etc, but for the most part, I want to do it with prompting.
Reading your description of how you do it, and your elaboration of it on your user page, makes me hopeful that you’ll share any knowledge you can in an appropriate thread in this forum.
Remember seeing extensions for Automatic1111 that could do wildcard prompting, but I never got around to test it. I assume it’s possible with others as well.

Creative Corner » Generation methods, UI » Post 4

Thoryn

Latter Liaison
I voted for Automatic1111 as it’s what I used at the time of voting, but today I switched to Forge (very much because of @tyto4tme4l mentioning their speed difference with it) - and in my case as well, it’s much, much faster - gens that took me 10-15 minutes on Automatic takes me 10-25 seconds with Forge.
The UI for the most part also looks the same for me, since I wasn’t familiar with all of the functionality in either yet, so it was an easy switch. ComfyUI’s UI looks very interesting though.
Right now I’m hoping Stability Matrix gets some, you know, stability, and can install and run properly on my system, so it’s easier to handle different UIs, models, LoRAs etc. Will definitely look at it again once I move on to Linux, but for now I am very happy with basic Forge.

Creative Corner » Text-to-image prompting » Post 16

Main thing with ComfyUI is that it involves making workflows where you connect together nodes, and takes a bit to get used to and you end up learning a little more of how stable diffusion actually works. Also, the UI will look out of date in a fair amount of videos and such on it because they overhauled it recently (and have been making things easier).
If you want an idea of what you are in for on that front, the first screenshot in my first post in this thread was a ComfyUI workflow.
Oh, and my main reason for suggesting Uni_pc as a sampler was just that one of the big things it’s good at is giving good images with less steps. It’s entirely possible A1111 might not have it, as it wasn’t around when it was originally made.
If you do go the ComfyUI route here’s their website and github:
https://www.comfy.org/
https://github.com/comfyanonymous/ComfyUI
If you scroll down on the second one, there’s a link to a portable version for windows.

Creative Corner » Text-to-image prompting » Post 15

Thoryn

Latter Liaison
@Lord Waite
I’ve had it run out of GPU memory before, during larger batches as well as too high resolution for upscaling, and in those instances it just stopped when running out during the batch, or failed at the beginning for the upscale, and gave an error saying out of memory in the status area. But it’s still certainly a possibility.
Right now, I’m unsure if I should pursue Stability Matrix some more (it looked really promising).. in the mean time, I’ll look at things like Forge and ComfyUI directly.

Creative Corner » Text-to-image prompting » Post 14

@Thoryn
Even before I went almost totally to ComfyUI, I went to SD Next and Forge and was going to those instead of A1111, so I can’t tell. I’ve been feeling like A1111 is getting out of date, really.
Could be it’s switching over to using the cpu for some reason, because generating on the cpu does take forever. Maybe the GPU was running out of memory…

Creative Corner » Text-to-image prompting » Post 13

Thoryn

Latter Liaison
@tyto4tme4l
Stability Matrix looked very promising, sadly it just freezes (and on one of the four attempts, leaked memory) for me.
With Automatic1111, I get random really long generation times every now and then. Can generate things “quickly” (ie 15 minutes) for hours, then suddenly it takes two hours to even reach 20% completion.. it even happens with only a different seed. It’s as if the model gets stuck, and there’s no step-bro to help it out.
Anyone else seeing this on Automatic? Seen it happen with others?

Creative Corner » Text-to-image prompting » Post 12

@tyto4tme4l
Yeah, IIRC, one of the points behind Forge was that it reworked things on A1111 to use code from ComfyUI for some of the backend, because ComfyUI is faster and better with memory.
I’ve got it, but I got used to ComfyUI, and while it might have a steep learning curve, it’s a lot more flexible for things once you know it. (Though, alright, inpainting is still going to be easier on a different UI.)

Creative Corner » Text-to-image prompting » Post 11

Thoryn

Latter Liaison
@Lord Waite
For the record, I’m using RTX 3070 TI with 8GB of video memory.
Good point on the fact that LoRAs also use some memory.. best to avoid if possible.
Don’t have the sampler named Uni_pc on my setup, so I kept it on DPM++ 2S a paired with Karras - as when I did a test with all the samplers (same prompt, seed etc), that was one of the fastest (Euler A was slightly faster and what I’ve been using when experimenting with Automatic a couple times before, but lately I have seen more instances of DPM++ 2S a in the wild, so figured I’d give it a try).
I copied your prompt exactly, and it actually gave a decent composition this time!
At 1024x1024 and 15 steps, it took 13 minutes… maybe this is passable for getting the composition, then I can use img2img to flesh things out? I should start experimenting with the pipeline soon, like only loading one or two LoRAs at a time (especially for kinks and concepts I know the model can’t do at well or at all), figure out grid-editing/prompting, inpainting etc.
score_9,score_8_up,score_7_up,score_6_up,score_5_up,score_4_up,
source_pony,rating_safe,
Flurry Heart lying under Christmas tree,ceiling,window,Christmas decorations on walls,presents,pony,filly,cute,star on top of tree,
Steps: 15, Sampler: DPM++ 2S a, Schedule type: Karras, CFG scale: 6, Seed: 40318147, Size: 1024x1024, Model hash: 67ab2fd8ec, Model: ponyDiffusionV6XL_v6StartWithThisOne, VAE hash: 95f26a5ab0, VAE: sdxl_vae.safetensors, Version: v1.10.1

Creative Corner » Text-to-image prompting » Post 10

tyto4tme4l

Something of an artist
@Thoryn
If you have a weak GPU, then how about trying out Forge WebUI? It looks almost exactly like A1111, but it should be much faster, especially on a weak GPU. I don’t know about newer versions, but I’m using a release from 02.2024 and it’s working great. I have GeForce 3060Ti with 8GB VRAM and I can generate four 1024x1024 pictures in slightly above one minute. Pretty much no OOM, errors or crashes.
https://github.com/lllyasviel/stable-diffusion-webui-forge/releases
There are also other UIs like ReForge or ComfyUI, I’d recommend testing different options to see what suits you best. Stability Matrix is great for installing and maintaining multiple UIs.
https://github.com/LykosAI/StabilityMatrix

Creative Corner » Text-to-image prompting » Post 9

@Thoryn
No problem.
512x512 was fine for 1.5 based models, just XL and newer has moved beyond that. Pony v5 was 2.1 based, so 768x768, and v4 and earlier was 1.5-, IIRC?
What I’ve got is a 3060 12GB, btw, which is probably about as low as you can go and still have 12GB. Though an 8GB card would work as well…
The lora does also add to the amount of memory used. While I don’t think the Flurry Heart one is needed, I will note that my first post was using the Wholesome MLP lora, which is a rather nice art style lora.
You could try changing the sampler to Uni_pc and normal, and lower the steps to, say, 12-14, and see if that speeds things up a bit for you.

Creative Corner » Text-to-image prompting » Post 8

Thoryn

Latter Liaison
@Lord Waite
Thanks for the tips. Have used 512x512 because I have basically the bare minimum of capable hardware (with that res, it With 75 prompt chunks, it usually takes me almost 5 for ~23-25 steps, and 10 minutes for 32-35.. and you’re saying I need to quadruple the res, oof..
(I promised myself not to throw more money at expensive GPUs as I am broke and have stopped gaming, but a 5090 starts to look more appealing the more I mess around with AI).
You’re correct that I’m using Automatic1111 by the way.
(Have pondered alternatives, as the cmd window spews errors left and right even on a fresh and up-to-date version, but it’s the devil I know right now..)
I will avoid LoRAs, clean out the negative prompts (and add things only as needed), up the res to 1024x1024 and do some testing.
Thanks again for all the input, really appreciate it.

Creative Corner » Text-to-image prompting » Post 7

No problem. And I can definitely see a couple potential issues.
First, 512x512 is not going to get good results with pony v6. It’s an XL based model, so generally speaking, we’re talking 1024x1024. Other good resolutions are 1152 x 896, 896 x 1152, 1216 x 832, 832 x 1216, 1344 x 768, 768 x 1344, 1536 x 640, & 640 x 1536. That’s basically what XL was trained on.
That negative prompt has way too much in it. “e621_p_low” isn’t actually a tag v6 knows. That was for older versions of pony, and was a precursor to the score tags. Usually, I’d say to start out with a very minimal negative prompt, and add things as needed, and also remove them if they aren’t working. Negative prompts that aren’t needed can actually make an image worse.
You do usually want to have a source tag and a rating tag after the scores. source_pony, source_furry, source_anime, and source_cartoon are the big source tags, then rating_safe, rating_questionable, and rating_explicit.
Not sure the lora is needed, either. You could probably remove it and just say “Flurry Heart”. Also, the point of “star on top of tree” was to try and get it to put the star at the top of the tree in the picture, and by extension, the rest of the tree.
I’d personally try something more like:
score_9, score_8_up, score_7_up, score_6_up, score_5_up, score_4_up,
source_pony, rating_safe,
Flurry Heart lying under Christmas tree, ceiling, window, Christmas decorations on walls, presents, pony, filly, cute, star on top of tree,
with no negative prompt and no lora, and go from there. Definitely keep in mind that the longer a prompt is, the less anything in it actually is weighted. The tokenizer can handle 75 (well, 77, but the other two are used internally) tokens, then after that, the prompt gets broken into 75 token chunks, and that’s the point they start meaning less individually.
(Looks like you’re using something A111 based, so there might be a token count at the top of the prompt entry box?)

Creative Corner » Text-to-image prompting » Post 6

Thoryn

Latter Liaison
@Lord Waite
Thanks for the tip.
I basically only get things like this, where the prompt acts like a toddler and doesn’t listen at all.

parameters
score_9,score_8_up,score_7_up,score_6_up,score_5_up,score_4_up,
Panoramic view of a spacious room with high ceiling,large tainted glass windows,Christmas decorations on walls,Flurry Heart_(Mlp),lying under Christmas tree,presents,<lora:Flurry Heart-Mlp-PonyXL:0.7>,pony,filly,cute,
Negative prompt: anthro,closeup,wip,sketch,blurry,disfigured,bad_hands,badly_drawn,bad_anatomy,watercolor,e621_p_low,thicc,thick,wide_hips,chubby,poofy,hyper,watermark,missing_tail,pillow,bed,couch,sofa,mattress,smiling,standing,sitting,scared,afraid,zoomed_in,cropped,signature,
Steps: 23, Sampler: DPM++ 2S a, Schedule type: Karras, CFG scale: 7, Seed: 582442736, Size: 512x512, Model hash: 67ab2fd8ec, Model: ponyDiffusionV6XL_v6StartWithThisOne, VAE hash: 95f26a5ab0, VAE: sdxl_vae.safetensors, Lora hashes: “Flurry Heart-Mlp-PonyXL: e75f8a2d04d3”, Version: v1.10.1

Creative Corner » Text-to-image prompting » Post 5

Creative Corner » Text-to-image prompting » Post 4

Thoryn

Latter Liaison
@Lord Waite
Still not seeing any success, no matter how much I emphasize the look or size of the room.
Is it possible to attach images here? Can only see URLs to enter, but I have nowhere to host them.

Creative Corner » Text-to-image prompting » Post 3

It does seem rather temperamental. I generally find it helps if you mention things the view needs to be bigger to accommodate. I had some success with this prompt on pony v6:
score_8, score_8_up, score_7_up, score_6_up, score_5_up, score_4_up,
source_pony, rating_safe,
cute Pegasus Derpy sleeping under a tree, Christmas, presents, ornaments, star on top of tree, window, ceiling,
god rays,
“star on top of tree” and “ceiling” helped clue it in that it should be a further away shot.
(Generally, the format I use for pony v6 models is score tags, source, rating, a general description, then tags.)

Creative Corner » Text-to-image prompting » Post 2

Thoryn

Latter Liaison
Remember trying those two features a year or two ago, and not making it work. Guess I’ll just have to try again.
Another question though, when prompting do you guys use natural language describing what you envision, or comma,separated,keywords,like,this?
A mixture? Varies by model?
And to “save on tokens”, do you use more concise but lesser used words, or favor using more tokens for more words just in case the model wouldn’t understand the lesser used ones?

Creative Corner » Text-to-image prompting » Post 1

Default search

If you do not specify a field to search over, the search engine will search for posts with a body that is similar to the query's word stems. For example, posts containing the words winged humanization, wings, and spread wings would all be found by a search for wing, but sewing would not be.

Allowed fields

Field SelectorTypeDescriptionExample
authorLiteralMatches the author of this post. Anonymous authors will never match this term.author:Joey
bodyFull TextMatches the body of this post. This is the default field.body:test
created_atDate/Time RangeMatches the creation time of this post.created_at:2015
idNumeric RangeMatches the numeric surrogate key for this post.id:1000000
myMetamy:posts matches posts you have posted if you are signed in. my:posts
subjectFull TextMatches the title of the topic.subject:time wasting thread
topic_idLiteralMatches the numeric surrogate key for the topic this post belongs to.topic_id:7000
topic_positionNumeric RangeMatches the offset from the beginning of the topic of this post. Positions begin at 0.topic_position:0
updated_atDate/Time RangeMatches the creation or last edit time of this post.updated_at.gte:2 weeks ago
user_idLiteralMatches posts with the specified user_id. Anonymous users will never match this term.user_id:211190
forumLiteralMatches the short name for the forum this post belongs to.forum:meta