Coder Social home page Coder Social logo

animateanyone's Introduction

AnimateAnyone

Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation

Li Hu, Xin Gao, Peng Zhang, Ke Sun, Bang Zhang, Liefeng Bo

YouTube

Teaser Image

Updates

Thank you all for your incredible support and interest in our project. We've received lots of inquiries regarding a demo or the source code. We want to assure you that we are actively working on preparing the demo and code for public release. Although we cannot commit to a specific release date at this very moment, please be certain that the intention to provide access to both the demo and our source code is firm.

Our goal is to not only share the code but also ensure that it is robust and user-friendly, transitioning it from an academic prototype to a more polished version that provides a seamless experience. We appreciate your patience as we take the necessary steps to clean, document, and test the code to meet these standards.

Thank you for your understanding and continuous support.

Citation

@article{hu2023animateanyone,
  title={Animate Anyone: Consistent and Controllable Image-to-Video Synthesis for Character Animation},
  author={Li Hu and Xin Gao and Peng Zhang and Ke Sun and Bang Zhang and Liefeng Bo},
  journal={arXiv preprint arXiv:2311.17117},
  website={https://humanaigc.github.io/animate-anyone/},
  year={2023}
}

animateanyone's People

Contributors

humanaigc avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

animateanyone's Issues

An empty project

An empty project, with just a readme, getting so many stars, is truly unbelievable.

What is the least requires GPU memory?

I reproduced the code and trained it on V100 32GB, but the OOM still occured even when the batch_size=1, image_resolution=128x128, and fp16 amp training.

Some Questions About the ReferenceNet

The ReferenceNet takes as input the VAE encode image. Does it add noise to it?

If you are not adding noise to the ReferenceNet image latents, do you call the ReferenceNet U-Net multiple times with the same timesteps as the denoising network, or are just calling it with a single timestep?

project is from the Alibaba development team?

When you hear that a project is from the Alibaba development team, you should understand the following points:

  1. The possibility of it being open-source is low.
  2. There's more talk than action.
  3. Many projects are abandoned halfway.

Request for Consideration: Inclusive Image Usage in Repository

I hope this finds you well. I would like to bring to your attention a concern regarding the usage of images, particularly pictures of women, in your GitHub repository. There's a possibility that the individuals in these images may not have consented to their use.

In the spirit of promoting inclusivity and respecting individuals' privacy, I kindly request a review and consideration for alternative, more inclusive imagery in your repository. It's essential to ensure that the visual content aligns with ethical practices and fosters a welcoming environment for all contributors and users.

Many journals also have be speaking out about using images without consent, so you can get problems with publishing this. Source: 1 Journal of Modern Optics, 2 Nature Nanotechnology 3 Optical Engineering

Furthermore here are some resources for alternative pictures or just use a picture of yourself, that would be a great showcase on how quickly it is possible to make anyone dance with your application! :)

I appreciate your understanding and attention to this matter. If you have any questions or would like further clarification, please feel free to reach out!

Questions about the training data

Excellent work!
I’m surprised that it animates both real and cartoon characters very well. Does the training dataset contain cartoon characters? And how to ensure the pose sequences are applicable to both real and cartoon?

发布的时间越早越好

既然论文已经发布,代码求求了,不然现在的内卷情况外一那天抖音就出现这个功能了怎么办,抢占先机最好

When will the source code be released?

Hello! I am very interested in your impressive and amazing work after watching your demo!
I wonder when will you release your source code because it seems amazing!

Thank you!

I've just completed a simplified version of AnimateAnyone and invite everyone to give it a try! Currently, the training code has been made available, and soon we'll be releasing our pre-trained models as well. You can access the project via this link: [https://github.com/guoqincode/AnimateAnyone-unofficial](https://github.com/guoqincode/AnimateAnyone-unofficial). Looking forward to your feedback and support!

I've just completed a simplified version of AnimateAnyone and invite everyone to give it a try! Currently, the training code has been made available, and soon we'll be releasing our pre-trained models as well. You can access the project via this link: https://github.com/guoqincode/AnimateAnyone-unofficial. Looking forward to your feedback and support!

Collaborators for an Implementation (Awaiting Code Release...)

the project is interesting in the context of security, allowing for the replication of dangerous scenes.

Therefore, I am proceeding with the implementation using the few details provided in the article.

I have already set up a draft for the temporal part and am testing various trainings, adapting ControlLDM as I inferred from the images in the article. I am seeing the first results, and most of the details are maintained, but the other network, the one that preserves the details, is missing.

It is clear from the videos that it is done in steps, on groups of frames because the jump is noticeable when looking at, for example, the background.

Is anyone working on sketching out the detail preservation part?

[Official Updates] Follow-up plans for the project

Thank you all for your incredible support and interest in our project. We've received lots of inquiries regarding a demo or the source code. We want to assure you that we are actively working on preparing the demo and code for public release. Although we cannot commit to a specific release date at this very moment, please be certain that the intention to provide access to both the demo and our source code is firm.

Our goal is to not only share the code but also ensure that it is robust and user-friendly, transitioning it from an academic prototype to a more polished version that provides a seamless experience. We appreciate your patience as we take the necessary steps to clean, document, and test the code to meet these standards.

Thank you for your understanding and continuous support.

I love this!

I would love this very much as a composer, director and animation guy! I want to participate!

建个群抱团取暖。

图片到视频的探索群。

主要解决图片到视频过程中遇到的一系列问题:包括环境安装,图片到动作的提取,视频崩坏,显存爆满,多卡运行,底模更换等一些列问题,收集一些优秀作品案例。提高沟通效率,降低学习成本,推动项目发展。imageimage

Men

Men wearing a compression shirt

Some problems of my unofficial implementation

Hi,

I have unofficially reproduced the code for 'Animate Anyone' based on the description in your paper. However, I encountered two issues during the training process:

Currently, with a single GPU and a batch size of 2, I have trained for 8k iterations. The generated images show a significant difference in the background compared to the target images, which are pure white. The third row in the following figure.

The faces reconstructed by the VAE decoding exhibit distortion. I'm wondering if it's possible to utilize the latent diffusion model to capture the information lost by the VAE and correct the distorted faces. In your video demo, the faces appear clear, and I'm unsure how to address this issue.

image

I love this!

I would love this very much as a composer, director and animation guy! I want to participate!

Considerations for Keeping Your Code Closed Source

You should consider not releasing your code as open source if your business model relies on maintaining a competitive edge and generating revenue from software sales. Following OpenAI's example, keeping your code closed source allows you to control distribution and directly monetize your innovations. It also protects your intellectual property and ensures that you can provide high-quality support and services to your customers, which often serves as an additional revenue stream. This decision is not solely about security and ethics; it's also about establishing a strong financial foundation for sustainable growth.

Unable to Locate Downloadable Code

Hello,

I am eagerly looking forward to your amazing tool's demo. I attempted to download the code and understood the general explanation, but I could not locate the actual source files. If the source has already been published, I would be grateful if you could direct me to it. I am looking forward to future updates.

Thank you.

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.