-
Notifications
You must be signed in to change notification settings - Fork 132
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[Feature Request]: Webui implementation? #2
Comments
Unfortunately the automatic1111 webui is a nightmare to integrate with; between modifications that have to be made to the model, sampling process, and attention implementation, I suspect I'd end up wanting to claw my eyes out even more than I usually do when working with that codebase. PRs are welcome (if they have decent code quality or people don't mind me rewriting half of it before merging) but I am absolutely not going to integrate with auto1111 myself. That said, I am going to add a little webui to this, and I have someone lined up to make a colab notebook for it once that's done 😄 |
well not using Automatics webui... but use the webui from the same original repo |
The original repo has a webui? That's news to me... |
Oh hey, so it does! Neat. well, that was already on my list of things to do, so I'll try to get to that in the next day or two depending on how busy I am 😄 |
Thank you :), because not gonna lie i reached the final part of the installation and generating... and i didn't know how to do that i kept pointing to the yaml file and nothing was happening... I'm pretty sure i was doing something wrong, but a webUI would be dope |
Ah, yeah, this CLI takes .json files rather than .yamls (im not a huge fan of yaml for this stuff); there's some examples in I'll probably be able to throw something together in Gradio tomorrow (i'm in Australia so it's pretty late here 😉) but no promises! |
i mean .json file yeah, i got to this point in the installation process, animatediff --help.. tried to point to my json file... couldn't figure it out |
{
"name": "miku",
"base": "",
"path": "models/sd/SomethingV2_2.safetensors",
"motion_module": "models/motion-module/mm_sd_v14.ckpt",
"seed": [-1, -1],
"scheduler": "ddim",
"steps": 20,
"guidance_scale": 8.25,
"prompt": [
"hatsune miku, 1girl, upper body, clouds, twintails, happy, smile, hand on own cheek, :d, looking at viewer, from side, best quality",
"hatsune miku, 1girl, full body, blue hair, twintails, happy, smile, looking at viewer, dramatic, masterpiece, best quality",
],
"n_prompt": [
"simple colors, simple background, low quality, worst quality, bad quality"
]
} If you save that as
and that should Just Work™ (you'll also need to have downloaded the motion module from the original repo's gdrive links and put them in [edit]: Side note: Use PowerShell, not cmd.exe :) |
alright well idk whats wrong atm... but the first time the installation process worked, now arent working, py -m venv .venv i think that does the job because it doesn't bring a error that python isn't recognized i tired both Powershell and cmd |
I've been working on integrating it into my GUI for Stable Diffusion Deluxe and almost got it all working. Started with the originals AnimateDiff, but quickly realized that all the dependencies are "ancient" and just wouldn't work with the rest of my project that requires latest Diffusers and Torch.. So I appreciate you refactoring things. Here's what it looks like so far....
I was originally using absolute path to point to the motion_module and lora because wanted it working with Windows paths too, but I modified it to relative path like the example, but still gives that error. So close, any suggestions? Thanks. |
i now have the same error as you^ |
@Skquark i got it to work, Basically i threw the error message to Claude AI and the solution is simple, just follow this. "It looks like the issue is still occurring where a string is being passed to path_from_cwd() instead of a Path object. Based on the traceback, in src/animatediff/utils/model.py, this line: python Copy code logger.debug(f"Downloaded {path_from_cwd(result)}") To fix, you'll need to update that section like we discussed earlier: python Copy code result = get_motion_modules() Let me know if this helps resolve the issue! We just need to make sure a Path object is being passed rather than the raw string result." and that made it work |
Good stuff, easy enough fix, been meaning to try Claude AI for those extra long prompt requests. Got my UI working now, loving it... Submitted PR #11 with that little update, plus downgraded minimum versions of a few requirements so it runs smoothly on Colab. My GUI is at DiffusionDeluxe.com for Google Colab or local desktop, but haven't tested if AnimateDiff is working locally since my GPU isn't good enough, so if anyone wants to try that I'd appreciate. Minor feature request for it to work better in the GUI is to save the image frames to output folder as they're being generated rather than at the end of run so you can preview the progress as they're being created. |
Might need someone to confirm it but I believe the frames are generated in parallel. It may be possible to display it live by running a seperate process on the cpu to create a GIF of the animation while it is in process. BTW that's a boat load of features you've got in your UI haha. How long did it take for you to implement them all? I'm kind of surprised there isn't more interest in animateDiff though I think the lack of control over the output is keeping people away. |
You might be right about them generated in parallel, at least to the number of Context frames to condition. No biggie if that can't be done. |
A part of me wonders if training an LLM to use traditional CGI software to generate movies would allow better control and coherency over the final output than using tools like animateDiff and Pikalabs. These models for text2mov aren't able to store the physical information of the pixels they're generating. It works well generating a scene from the POV of a car moving forward but once the car reverses it'll be very difficult for these models to generate images that is consistent with what the viewer has previously seen. I think these models would be very useful in generating proof of concept/trailers though. However in hindsight. Once LLMs can do that there's probably not much left that we humans actually need to do. |
Great, looking forward to using it on webui |
We have AnimateDiff working on twisty.ai as the default generator with 4 of the style models. It's the original distribution with minor tweaks, running on replicate. Currently we are moving this package over to AWS so we can keep it warm and render more quickly. We would like to integrate your version for performance and longer video run length. Any suggestions on how we support your version? Do we just start from scratch and wrap your system in replicates COG wrapper? We use a gem economy for rendering, currently free. Eventually it will be priced to support the rendering expense and provide something back to the model, LoRa and generator creators. |
pls
The text was updated successfully, but these errors were encountered: