Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Question about sign-language-processing/spoken-to-signed-translation/issues/26 #2

Closed
AmitMY opened this issue Mar 15, 2024 · 1 comment

Comments

@AmitMY
Copy link
Contributor

AmitMY commented Mar 15, 2024

          Hi, nice work. 

If I may ask, which scripts of the pose-to-video for the diffusion models do you use?
Have they the holistic bones as input? Or how do know what the expected output of the diffusion model should be?

Originally posted by @florianbaer in sign-language-processing/spoken-to-signed-translation#26 (comment)

@AmitMY
Copy link
Contributor Author

AmitMY commented Mar 15, 2024

The command used to generate the video outputs there is

pose_to_video --type=controlnet --model=sign/sd-controlnet-mediapipe --pose=assets/testing-reduced.pose --video=assets/outputs/controlnet-animatediff.mp4 --processors animatediff

It is based on holistic, yes

@AmitMY AmitMY changed the title Hi, nice work. Question about sign-language-processing/spoken-to-signed-translation/issues/26 Mar 15, 2024
@AmitMY AmitMY closed this as completed Mar 22, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant