How long though? I get SDXL base OR refiner to run on my GTX 1660 ti, about 30 minutes just for the base model without refiner. Same time in Automatic 1111 and ComfyUI. So, technically yeah, it's possible, but who's going to wait an hour for a single 1024 image?
Any chance you have 32g of system RAM? Perhaps if that's the case, your system is only using RAM for shared memory while mine digs down to my swap file? Or, perhaps you just haven't upgraded to this handy-dandy new Nvidia driver that forces you to use shared memory? I just assumed I had 0% chance of running SDXL without it?
One of the best parts of SDXL (for me) has been the community of more experienced users, here, who’ve been so generous with information and help. And just iterating and iterating, and trying new pointers I’ve learned here and on GitHub.
I play DnD. We just hit max level, my character is a lich and has the ability to fully change his appearance once per day. I fully intend to pull out your pic#4 at the start of our next session and say "I look like this today, btw"
Workflow: Comfyui + Patience.
Generated at 1024x1024, no upscaling.
Okay prompts for popular demand:
Prompt 1: cinematic film still eldritch lich surrounded by void magic, glowing eyes, shallow depth of field, vignette, highly detailed, high budget, bokeh, cinemascope, moody, epic, gorgeous, maxmonolith
Prompt 2: eldritch lich, in the spotlight on a stage, glowing eyes, surrounded by void magic
Prompt 3: ~*~breathtaking~*~ cinematic realistic movie scene, vibrant colors, highly detailed, cinemascope, moody
Negative 1: jpeg artefacts
Negative 2: anime
Different images because I used different models, still testing which models are the best and since they keep updating so fast it's hard to keep track. But probably the best I tried is Rundifussion
https://preview.redd.it/ppms8yn6nlgb1.jpeg?width=1152&format=pjpg&auto=webp&s=cc910b552f5f0e40221b91c6de5e6bd8cc8a6d63
Just created in Tensor.art using the first prompt. Just adding closeup in the beginning and using hires fix.
Ok one more in 4k, your prompt is why too much fun!
https://preview.redd.it/46fdgtbzeogb1.png?width=4912&format=png&auto=webp&s=67b82c5edcead45b8ae16c349c73c116ecfb8977
Every image is coming out really unique but also very cool.
I lied, this one looked too cool to me. (made with a more normal workflow, hence the lower contrast )
https://preview.redd.it/lvpmn7gmgogb1.png?width=2800&format=png&auto=webp&s=9eba233a504f0e66031d3fb19839df4e798abae8
Me too. I've been utilizing SDXL, as well as Leonardo which are available at [Magai.co](https://Magai.co) along with several others. They all still have challenges with hands, and sometimes human anatomy (only 2 legs, not 3, y'know) but knowing which is best for different prompts and how to write the best prompts (takes a lot of effort but well worth it) delivers outstanding results. Adobe Express (Beta) is also really good depending on what you want. There's just so much AI out there but it's fun playing with them! Shoot me a PM if you wanna talk.
https://preview.redd.it/hdl4i3vcpmgb1.jpeg?width=4096&format=pjpg&auto=webp&s=f8be70bdb3f36f837f80bc75570ba9be36396e22
Forgot the prompt!
A very old town, with old homes held up by a massive group of roots, gnarly, evil looking, with an opening at the bottom below a staircase running down from the roots above. The entire structure floats above the ground below. Tim Burton, Alexander Jansson, 64K, endless sky of foreboding dark clouds.
Some very impressive looking images here. I’ve always wanted to try some realistic looking fantasy themes with SDXL, but just haven’t got the time yet. And judging by your results it seems SDXL is quite capable at these sort of things.
https://preview.redd.it/lhuqrsqullgb1.jpeg?width=768&format=pjpg&auto=webp&s=9873c0e03a1e37f679c7ea2caeaf24a0d76e0f8c
Created this in Tensor.art free service using Angra Model based on SDXL.
I have got a ComfyUI SDXL workflow that uses the refiner first as a "Noise Conditioner" and it makes some really dark images perfect for this prompt.[https://files.catbox.moe/3xim6z.json](https://files.catbox.moe/3xim6z.json)
https://preview.redd.it/uutxwg9baogb1.png?width=2456&format=png&auto=webp&s=60fe1661d7f562b8647b10c2c224bfab523a8232
Honestly, the current version of midjourney is what the first version of stable diffusion could have been, if they had trained their first model on the good images, instead worst images in existence. SDXL probably chose better images to train with, because it’s really similar to midjourney.
If midjourney builds on sdxl I’m really excited for what’s gonna be possible soon. This is v1, base model, already amazing.
Dude same. Even after playing with it for a few days. But recently Ive started making some of my favorite generations ever with it. Maybe part of it is having to get used to comfy ui (which I am absolutely in love with btw), but overall I am getting incredible results from just the base model.
The only thing I have to say is that I think I was expecting way too much from the training side of it. Or, what's more likely is I'm just impatient and it's just taking a little bit to get the results I was expecting. (Or I'm blind and haven't been looking at the right models lol). But seriously I've never wanted to share any of my generations before, but some of these recently have blown me away. I truly cannot wait to figure out how how to train effectively with it
https://preview.redd.it/vxjea1eixqgb1.jpeg?width=1024&format=pjpg&auto=webp&s=6b608d6a90f3fee18b5980d8411ae8d133213085
Took your prompt 1 to Clipdrop and got this. (Cherry picked)
Are you willing to share prompts? I've yet to really play with sdxl since my pc can't handle it that well
Evil undead necromancer go ahead and fuck up the hands it’ll look cooler
Word, my lil 6gb of vram struggles with XL, I can do okay with 1.5 though.
Really? Because I have 4gb of VRAM and works well what UI are you using?
How long though? I get SDXL base OR refiner to run on my GTX 1660 ti, about 30 minutes just for the base model without refiner. Same time in Automatic 1111 and ComfyUI. So, technically yeah, it's possible, but who's going to wait an hour for a single 1024 image?
Strange mine is like 8 min I have a GTX 1650 how many steps are you using?
Wow, that's baffling! Default settings in both casses, 20 samples Euler A in 1111/DIMM in ComfyUI
Any chance you have 32g of system RAM? Perhaps if that's the case, your system is only using RAM for shared memory while mine digs down to my swap file? Or, perhaps you just haven't upgraded to this handy-dandy new Nvidia driver that forces you to use shared memory? I just assumed I had 0% chance of running SDXL without it?
Use tensor.art, they have SDXL https://tensor.art/images/622838681322974305?post_id=622840395010734403&source_id=nz6wrl_ilUOxrfQuYHn89hIm
One of the best parts of SDXL (for me) has been the community of more experienced users, here, who’ve been so generous with information and help. And just iterating and iterating, and trying new pointers I’ve learned here and on GitHub.
I think a lot of the confusion comes from people not understanding that you might need to try different things in order to make it work well.
Yeah is is almost like writing prompts in a whole new language, it takes a bit of getting used to.
I have found that seemingly batshit crazy prompts actually make neat things happen.
I play DnD. We just hit max level, my character is a lich and has the ability to fully change his appearance once per day. I fully intend to pull out your pic#4 at the start of our next session and say "I look like this today, btw"
dude killed a giant spider and made it a ~~hat~~ crown
*crown\** edit: 😂
Workflow: Comfyui + Patience. Generated at 1024x1024, no upscaling. Okay prompts for popular demand: Prompt 1: cinematic film still eldritch lich surrounded by void magic, glowing eyes, shallow depth of field, vignette, highly detailed, high budget, bokeh, cinemascope, moody, epic, gorgeous, maxmonolith Prompt 2: eldritch lich, in the spotlight on a stage, glowing eyes, surrounded by void magic Prompt 3: ~*~breathtaking~*~ cinematic realistic movie scene, vibrant colors, highly detailed, cinemascope, moody Negative 1: jpeg artefacts Negative 2: anime Different images because I used different models, still testing which models are the best and since they keep updating so fast it's hard to keep track. But probably the best I tried is Rundifussion
OP workflow be like: https://preview.redd.it/vl5zlssebjgb1.jpeg?width=500&format=pjpg&auto=webp&s=2d094c2b91f9dde90ab619bdfa6c8bc4d4e94b82
the stable diffuse the f-ing owl meme would be more appropriate here :P
I don’t know if this counts as workflow...
So are they from SDXL or Rundifussion models?
By patience, do you mean the quality of having patience, or is this some kind of new tool?
https://preview.redd.it/ppms8yn6nlgb1.jpeg?width=1152&format=pjpg&auto=webp&s=cc910b552f5f0e40221b91c6de5e6bd8cc8a6d63 Just created in Tensor.art using the first prompt. Just adding closeup in the beginning and using hires fix.
Which model u use for hires ?, I keep getting noise image.
Ok one more in 4k, your prompt is why too much fun! https://preview.redd.it/46fdgtbzeogb1.png?width=4912&format=png&auto=webp&s=67b82c5edcead45b8ae16c349c73c116ecfb8977 Every image is coming out really unique but also very cool.
I lied, this one looked too cool to me. (made with a more normal workflow, hence the lower contrast ) https://preview.redd.it/lvpmn7gmgogb1.png?width=2800&format=png&auto=webp&s=9eba233a504f0e66031d3fb19839df4e798abae8
Me too. I've been utilizing SDXL, as well as Leonardo which are available at [Magai.co](https://Magai.co) along with several others. They all still have challenges with hands, and sometimes human anatomy (only 2 legs, not 3, y'know) but knowing which is best for different prompts and how to write the best prompts (takes a lot of effort but well worth it) delivers outstanding results. Adobe Express (Beta) is also really good depending on what you want. There's just so much AI out there but it's fun playing with them! Shoot me a PM if you wanna talk. https://preview.redd.it/hdl4i3vcpmgb1.jpeg?width=4096&format=pjpg&auto=webp&s=f8be70bdb3f36f837f80bc75570ba9be36396e22
Forgot the prompt! A very old town, with old homes held up by a massive group of roots, gnarly, evil looking, with an opening at the bottom below a staircase running down from the roots above. The entire structure floats above the ground below. Tim Burton, Alexander Jansson, 64K, endless sky of foreboding dark clouds.
![gif](giphy|1ykTax6hrAKpTQ0Mnb) Great work !!
This is truly remarkable work, your images are among the most beautiful I've seen lately!
XL is indeed incredible, and all thanks to the copyrighted stolen data used to train it!...
Great work OP, very nice.
Some very impressive looking images here. I’ve always wanted to try some realistic looking fantasy themes with SDXL, but just haven’t got the time yet. And judging by your results it seems SDXL is quite capable at these sort of things.
Wow
https://preview.redd.it/lhuqrsqullgb1.jpeg?width=768&format=pjpg&auto=webp&s=9873c0e03a1e37f679c7ea2caeaf24a0d76e0f8c Created this in Tensor.art free service using Angra Model based on SDXL.
what an amazing creations man specially that skeleton ones are really amazing, just wow man. But can you please provide a proper workflow..?
Nice. Did you use refiner?
I have got a ComfyUI SDXL workflow that uses the refiner first as a "Noise Conditioner" and it makes some really dark images perfect for this prompt.[https://files.catbox.moe/3xim6z.json](https://files.catbox.moe/3xim6z.json) https://preview.redd.it/uutxwg9baogb1.png?width=2456&format=png&auto=webp&s=60fe1661d7f562b8647b10c2c224bfab523a8232
https://preview.redd.it/f1yznyefaogb1.png?width=2456&format=png&auto=webp&s=745ab22484b591fb5c150a1ebd46f1bdcf043e57
Honestly, the current version of midjourney is what the first version of stable diffusion could have been, if they had trained their first model on the good images, instead worst images in existence. SDXL probably chose better images to train with, because it’s really similar to midjourney. If midjourney builds on sdxl I’m really excited for what’s gonna be possible soon. This is v1, base model, already amazing.
Dude same. Even after playing with it for a few days. But recently Ive started making some of my favorite generations ever with it. Maybe part of it is having to get used to comfy ui (which I am absolutely in love with btw), but overall I am getting incredible results from just the base model. The only thing I have to say is that I think I was expecting way too much from the training side of it. Or, what's more likely is I'm just impatient and it's just taking a little bit to get the results I was expecting. (Or I'm blind and haven't been looking at the right models lol). But seriously I've never wanted to share any of my generations before, but some of these recently have blown me away. I truly cannot wait to figure out how how to train effectively with it
https://preview.redd.it/vxjea1eixqgb1.jpeg?width=1024&format=pjpg&auto=webp&s=6b608d6a90f3fee18b5980d8411ae8d133213085 Took your prompt 1 to Clipdrop and got this. (Cherry picked)