Code Monkey home page Code Monkey logo

music-dance-video-synthesis's Introduction

Self-supervised Dance Video Synthesis Conditioned on Music

Pytorch implementation for this paper by Xuanchi Ren, Haoran Li, Zijian Huang, Qifeng Chen

To appear in ACM MM 2020

[Paper] [Paper_MM]

The demo video is shown at: https://youtu.be/UNHv7uOUExU

The dataset and the code for training and test is released.

A notebook for demo and quick start will be provided soon.

Some Demo:

More samples can be seen in demo video.

Requirement:

python 3.5 + pytorch 1.0

For the Testing part, you should install ffmpeg for music video.

We use tensorboardX for logging. If you don't install it, you can just comment the line in train.py.

Training:

This training process is intended for the clean part dataset, which could be downloaded here.

  1. Download the dataset and put it under ./dataset

  2. Run

python train.py

training script will load config of config.py. If you want to train the model on other datasets, you should change the config in config.py.

Testing:

If you want to use the pretrained model, you can firstly download it from here, put it under "pretrain_model" and change the path of get_demo.py to "./pretrain_model/generator_0400.pth".

  1. Run
python get_demo.py --output the_output_path
  1. Make the output skeleton sequence to music video
cd Demo
./frame2vid.sh

Note that you should change the paths and the "max" variable in frame2vid.sh.

Pose2Vid:

For this part, we adapt the method of the paper "Everybody dance now".

And We use this pytorch implementation.

Metrics:

For the proposed cross-modal metric in our paper, we re-implement the paper: Human Motion Analysis with Deep Metric Learning (ECCV 2018).

The implementation of this paper can be seen at: https://github.com/xrenaa/Human-Motion-Analysis-with-Deep-Metric-Learning

Dataset:

To use the dataset, please refer the notebook "dataset/usage_dataset.ipynb"

As state in the paper, we collect 60 videos in total, and divide them into 2 part according to the cleaness of the skeletons.

The clean part(40 videos): https://drive.google.com/file/d/1o79F2F7-dZ7Cvpzf6hsVMwvfNg9LM3_K/view?usp=sharing

The noisy part(20 videos): https://drive.google.com/file/d/1pZ3JszX7393dQwm6x6bxxbiKb0wLIJGE/view?usp=sharing

To support further study, we also provide other collected data:

Ballet: https://drive.google.com/open?id=1NR6S20EI1C37fsDhaNkRI_P1MLT9Ox7u

Popping: https://drive.google.com/file/d/1oLIxtczDZBvPdCAk8wuiI9b4FsnCItMG/view?usp=sharing

Boy_kpop: https://drive.google.com/file/d/14-kEdudvaGLSapAr4prp4D67wzyACVQt/view?usp=sharing

Besides, we also provide the BaiduNetDisk version: https://pan.baidu.com/s/15wLkdPnlZiCxgnPv51hgpg (includes all the dataset)

Questions

If you have questions for our work, please email to [email protected].

Citation

If you use this code for your research, please cite our paper.

@InProceedings{ren_mm_dance,
author  = {Ren, Xuanchi and Li, Haoran and Huang, Zijian and Chen, Qifeng},
title = {Self-supervised Dance Video Synthesis Conditioned on Music},
booktitle = {ACM MM},
year = {2020}
}

music-dance-video-synthesis's People

Contributors

xrenaa avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

music-dance-video-synthesis's Issues

Demo question

Hello,

I have few questions about demo

  1. how can i use my own music for demo?

  2. which get_demo.py is correct one? there are multiple of them on the root folder as well as under Demo folder

  3. I get similar bug as in #3 . I added audio folder in demo folder, but i get same error for output folder now

Thank you

demo question

你好,
我跑起來出現了如下的異常
image
另外有些疑問:
1,如果要輸入原始視頻,重新訓練該怎麼做,就是那些json文件是怎麼生成的
2,怎麼輸入音頻
3,這個項目和https://github.com/NVlabs/Dancing2Music 哪個更好呢?

Cross-modal Evaluation code

Hi,

I found that you have provided the second part of the cross-modal evaluation in another github page.

I wonder if it is possible for you to further provide the followings:

  1. the first part of the cross-modal evaluation. (The part that create a dictionary mapping music embedding to dance sequence)
  2. For now, there is only training code on github for the second part of the cross-modal evaluation. How can I use this trained model to get the score reported in the paper? Also, can you provide the pretrain model? (the one you used for generating the result in paper)

Thanks!

Cannot Get Reported Result With Default Setting

Thanks for the great work. I've tried to run the code as instructed using the clean part data. The results are far from impressive as shown in the demo video (like this). Here is the loss curve:

img

I wonder if I have to change some settings to make it work. Any help is appreciated!

music feature extraction

Hi author,thank you for your great work!
I have read your paper and I have a question that how to extrat the music feature ? If i want to use my music as input , how should I extract the feature from music?

Details about your paper?

Hi,

  1. what's the difference between the clean part and the noisy part?
  2. which training and test data did you use to obtain the results in your Table 1?
  3. How to map the generated skeletons to the images? Are you retrain the model? If so, where are the ground truth images for training?

Thanks a lot.

FileNotFoundError: [Errno 2] No such file or directory: './dataset/clean_revised_pose_pairs.json' & RuntimeError: DataLoader worker (pid(s) 9784, 7708, 11984, 11008, 3636, 15428, 10724, 1548, 2312, 12376, 1792) exited unexpectedly

Run time

python train.py

I get the following error
RuntimeError: DataLoader worker (pid(s) 9784, 7708, 11984, 11008, 3636, 15428, 10724, 1548, 2312, 12376, 1792) exited unexpectedly
Help to fix the error

image

&

Run time

python get_demo.py --output the_output_path

I get the following error
FileNotFoundError: [Errno 2] No such file or directory: './dataset/clean_revised_pose_pairs.json'
Help to fix the error

image

Thank

How can I use my own dataset to train network?

Hello. First of all, thank you for the great work. I'm so impressed by this project and trying to use my own dataset to train this network. However, I have no idea about making my dataset to json file that 'train.py' can read. How should I process my dataset? Thank you.

Demo bugs

Hi, I re-trained the model, but I have bugs in get_demo.py.
image

And how could I generate a gif? Thanks a lot.

Testing error

Hi, I am trying to train and test on other datasets you provided such as Ballet and Popping. However, the training is good but I am meeting the following error when testing, can you help me to fix it? Thanks!

Traceback (most recent call last):
File "get_demo.py", line 73, in
data=DanceDataset(args.data)
File "/data0/htang/projects/Music-Dance-Video-Synthesis/dataset/lisa_dataset_test.py", line 22, in init
sub_keys=sorted(pose_dict_boy[str(key)].keys())
KeyError: '047'

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.