Coder Social home page Coder Social logo

pose-group / dcpose Goto Github PK

View Code? Open in Web Editor NEW
365.0 10.0 64.0 81.55 MB

This is an official implementation of our CVPR 2021 paper "Deep Dual Consecutive Network for Human Pose Estimation" (https://openaccess.thecvf.com/content/CVPR2021/papers/Liu_Deep_Dual_Consecutive_Network_for_Human_Pose_Estimation_CVPR_2021_paper.pdf)

Python 99.80% MATLAB 0.20%
human-pose-estimation video-pose-estimation

dcpose's Introduction

Deep Dual Consecutive Network for Human Pose Estimation (CVPR2021)

Introduction

This is the official code of Deep Dual Consecutive Network for Human Pose Estimation.

Multi-frame human pose estimation in complicated situations is challenging. Although state-of-the-art human joints detectors have demonstrated remarkable results for static images, their performances come short when we apply these models to video sequences. Prevalent shortcomings include the failure to handle motion blur, video defocus, or pose occlusions, arising from the inability in capturing the temporal dependency among video frames. On the other hand, directly employing conventional recurrent neural networks incurs empirical difficulties in modeling spatial contexts, especially for dealing with pose occlusions. In this paper, we propose a novel multi-frame human pose estimation framework, leveraging abundant temporal cues between video frames to facilitate keypoint detection. Three modular components are designed in our framework. A Pose Temporal Merger encodes keypoint spatiotemporal context to generate effective searching scopes while a Pose Residual Fusion module computes weighted pose residuals in dual directions. These are then processed via our Pose Correction Network for efficient refining of pose estimations. Our method ranks No.1 in the Multi-frame Person Pose Estimation Challenge on the large-scale benchmark datasets PoseTrack2017 and PoseTrack2018. We have released our code, hoping to inspire future research.

Visual Results

On PoseTrack

Comparison with SOTA method

Experiments

Results on PoseTrack 2017 validation set

Method Head Shoulder Elbow Wrist Hip Knee Ankle Mean
PoseFlow 66.7 73.3 68.3 61.1 67.5 67.0 61.3 66.5
JointFlow - - - - - - - 69.3
FastPose 80.0 80.3 69.5 59.1 71.4 67.5 59.4 70.3
SimpleBaseline(2018 ECCV) 81.7 83.4 80.0 72.4 75.3 74.8 67.1 76.7
STEmbedding 83.8 81.6 77.1 70.0 77.4 74.5 70.8 77.0
HRNet(2019 CVPR) 82.1 83.6 80.4 73.3 75.5 75.3 68.5 77.3
MDPN 85.2 88.8 83.9 77.5 79.0 77.0 71.4 80.7
PoseWarper(2019 NIPS) 81.4 88.3 83.9 78.0 82.4 80.5 73.6 81.2
DCPose 88.0 88.7 84.1 78.4 83.0 81.4 74.2 82.8

Results on PoseTrack 2017 test set(https://posetrack.net/leaderboard.php)

Method Head Shoulder Elbow Wrist Hip Knee Ankle Total
PoseFlow 64.9 67.5 65.0 59.0 62.5 62.8 57.9 63.0
JointFlow - - - 53.1 - - 50.4 63.4
KeyTrack - - - 71.9 - - 65.0 74.0
DetTrack - - - 69.8 - - 65.9 74.1
SimpleBaseline 80.1 80.2 76.9 71.5 72.5 72.4 65.7 74.6
HRNet 80.0 80.2 76.9 72.0 73.4 72.5 67.0 74.9
PoseWarper 79.5 84.3 80.1 75.8 77.6 76.8 70.8 77.9
DCPose 84.3 84.9 80.5 76.1 77.9 77.1 71.2 79.2

Results on PoseTrack 2018 validation set

Method Head Shoulder Elbow Wrist Hip Knee Ankle Mean
AlphaPose 63.9 78.7 77.4 71.0 73.7 73.0 69.7 71.9
MDPN 75.4 81.2 79.0 74.1 72.4 73.0 69.9 75.0
PoseWarper 79.9 86.3 82.4 77.5 79.8 78.8 73.2 79.7
DCPose 84.0 86.6 82.7 78.0 80.4 79.3 73.8 80.9

Results on PoseTrack 2018 test set

Method Head Shoulder Elbow Wrist Hip Knee Ankle Mean
AlphaPose++ - - - 66.2 - - 65.0 67.6
DetTrack - - - 69.8 - - 67.1 73.5
MDPN - - - 74.5 - - 69.0 76.4
PoseWarper 78.9 84.4 80.9 76.8 75.6 77.5 71.8 78.0
DCPose 82.8 84.0 80.8 77.2 76.1 77.6 72.3 79.0

Installation & Quick Start

Check docs/installation.md for instructions on how to build DCPose from source.

Citation

@InProceedings{Liu_2021_CVPR,
    author    = {Liu, Zhenguang and Chen, Haoming and Feng, Runyang and Wu, Shuang and Ji, Shouling and Yang, Bailin and Wang, Xun},
    title     = {Deep Dual Consecutive Network for Human Pose Estimation},
    booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
    month     = {June},
    year      = {2021},
    pages     = {525-534}
}

dcpose's People

Contributors

chenhaomingbob avatar

Stargazers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

Watchers

 avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar  avatar

dcpose's Issues

A question about model output

Why not let the model directly output 15 joints corresponding to Posetrack dataset, but output 17 joints corresponding to COCO dataset, and then convert to posetrack?

Change the model

Hello, what should I do if I want to make changes based on the model you published.

Can this run on the windows system?

I tried to run on the windows system, but the problem of VC toolkit parsing was always prompted, and the bug still appeared after referring to the solution.How can i solve??
build\lib.win-amd64-3.6\deform_conv_cuda.cp36-win_amd64.pyd : fatal error LNK1120: 8 个无法解析的外部命令
error: command 'C:\Program Files (x86)\Microsoft Visual Studio\2019\BuildTools\VC\Tools\MSVC\14.29.30037\bin\HostX86\x64\link.exe' failed with exit status 1120

Licensing

Dear DCPose, I couldn't find any licensing information .Could you please add licensing information about that repository, so that it becomes clear how to safely use this repository.
Many thanks,
Wasif

Filtering still / moving persons

Hi,

I am looking for a way to get pose estimations during a video.
However, I have to filter static people and gives attention to only moving persons (actors).
Is this api able to do that ?
Regards

两个问题

  1. 在测试阶段,使用的行人检测框架是?
  2. 从当前帧crop出的人体,扩大25%用于截取前一帧和后一帧的同一个人体,这是假设时序连贯。但你们如何能将整个视频的同一个人物的相同节点连贯起来? 这才是tracking的定义,论文中没有提到。

谢谢。

cannot import name 'deform_conv_cuda' from partially initialized module 'thirdparty.deform_conv' (most likely due to a circular import)

大佬您好,我想运行试一下效果,运行 video.py 的时候报了这个错误,不知道要怎么解决,谢谢啦

cannot import name 'deform_conv_cuda' from partially initialized module 'thirdparty.deform_conv' (most likely due to a circular import)

报错位置:
File "xxx / DCPose/thirdparty/deform_conv/functions/deform_conv.py", line 5, in
from .. import deform_conv_cuda

DCPose/datasets/zoo/posetrack/posetrack_utils/poseval/py/convert.py", line 200, in from_new for person_info in track_data["annotations"]: KeyError: 'annotations'

hello, I have met a problem,,when I reproduced the excellent research, DCPose
I run command
python run.py
--cfg ../configs/posetimation/DcPose/posetrack18/model_RSN.yaml --test
the yaml about val and test

VAL:
ANNOT_DIR: './DcPose_supp_files/posetrack18_annotation_dirs/val/'
COCO_BBOX_FILE: './DcPose_supp_files/posetrack18_precomputed_boxes/val_boxes.json'
USE_GT_BBOX: false
BBOX_THRE: 1.0
IMAGE_THRE: 0.2
IN_VIS_THRE: 0.2
NMS_THRE: 1.0
OKS_THRE: 0.9
FLIP_VAL: false
POST_PROCESS: true

TEST:
ANNOT_DIR: './DcPose_supp_files/posetrack18_annotation_dirs/test'
COCO_BBOX_FILE: './DcPose_supp_files/posetrack18_precomputed_boxes/test_boxes.json'
USE_GT_BBOX: false
BBOX_THRE: 1.0
IMAGE_THRE: 0.2
IN_VIS_THRE: 0.2
NMS_THRE: 1.0
OKS_THRE: 0.9
FLIP_TEST: false
POST_PROCESS: true

it appear this error in test but not in val

#---------------------------------------------------------------------------------------------------------#
Traceback (most recent call last):
File "run.py", line 34, in
main()
File "run.py", line 30, in main
runner.launch()
File "/home/xxx/project/DCPose/engine/defaults/runner.py", line 63, in launch
evaluator.exec()
File "/home/xxx/project/DCPose/engine/defaults/evaluator.py", line 21, in exec
self.eval()
File "/home/xxx/project/DCPose/engine/defaults/evaluator.py", line 75, in eval
phase=self.phase)
File "/home/xxx/project/DCPose/engine/core/function.py", line 252, in eval
filenames_map, filenames, imgnums)
File "/home/xxx/project/DCPose/datasets/zoo/posetrack/PoseTrack.py", line 547, in evaluate
AP = evaluate_simple.evaluate(annot_dir, output_dir, eval_track=False)[0]
File "/home/xxx/project/DCPose/datasets/zoo/posetrack/posetrack_utils/poseval/py/evaluate_simple.py", line 16, in evaluate
gtFramesAll, prFramesAll = load_data_dir(['', gtdir, preddir])
File "/home/xxx/project/DCPose/datasets/zoo/posetrack/posetrack_utils/poseval/py/eval_helpers.py", line 399, in load_data_dir
data = convert_videos(data)[0]
File "/home/xxx/project/DCPose/datasets/zoo/posetrack/posetrack_utils/poseval/py/convert.py", line 621, in convert_videos
videos = Video.from_new(track_data)
File "/home/xxx/project/DCPose/datasets/zoo/posetrack/posetrack_utils/poseval/py/convert.py", line 200, in from_new
for person_info in track_data["annotations"]:
KeyError: 'annotations'
#---------------------------------------------------------------------------------------------------------#

Questions about additional data

Thank you for your work!

I noticed the result of your model posted on the leaderboard of PoseTrack17 is with additional COCO data.

Could you give more details of using additional data?

Thanks!

Throws error when I try to train from scratch in posetrack 18

The below error occurs:

022-04-23 06:32:46 [posetimation.zoo.DcPose.dcpose_rsn] ERROR: => please download pre-trained models first!
Traceback (most recent call last):
File "run.py", line 33, in
main()
File "run.py", line 29, in main
runner.launch()
File "/home/awanish/DCPose-main/engine/defaults/runner.py", line 52, in launch
trainer = DefaultTrainer(self.cfg, self.output_path_dict, PE_Name=self.args.PE_Name)
File "/home/awanish/DCPose-main/engine/defaults/trainer.py", line 32, in init
self.model = build_model(cfg, phase='train')
File "/home/awanish/DCPose-main/posetimation/zoo/build.py", line 17, in build_model
model_instance.init_weights()
File "/home/awanish/DCPose-main/posetimation/zoo/DcPose/dcpose_rsn.py", line 285, in init_weights
raise ValueError('{} is not exist!'.format(self.pretrained))
ValueError: /home/awanish/DCPose-main/DcPose_supp_files/pretrained_models/pretrained_.pth is not exist!

there is no file with the name pretrained_.pth given in supp files folder

Questions about the performance

Thanks for your excellent work!

I run the code successfully and I can use your pretrained model to get a result of 82.8 for PoseTrack17 validation dataset, but if I use the provided code and config files to train a new model and the result is only 81.6.

So I want to know that is there any other settings about the training or the config files about the training are not the updated version?

different video frame,same target?

I would like to ask, yolo only can detect target in a single frame .when defining clip i(p,c,n),by expanding the bounding box by about 25%, the same target between frames at different intervals is cropped. How do you determine the different positions of the same target in different frames? How to cut it?

Skip frames

Hello Dear author. I am very pleased with your work. Thank you!

But the thing is it is sometimes skipping frames. Do you know how i can correct that I mean which file should look to correct.

Thank you very much!

hi ,bro.how can i run code using my own video?

cd demo/                   
mkdir input/
# Put your video in the input directory
python video.py

Your method above didnt work.Maybe because the pretrained model (pretrained_coco_model)is not suitable?
thanks a lot !

Training Time

Hi,I saw you said "train our model for a batch size of 32 for 20 epochs with 2 Nvidia GeForce Titan X GPUs" in the paper.
Can I know the sepcific training time you will cost in this situation? Thanks

Pose Temporal Merger

Hi folks,

thanks for your amazing work.

I have a question regarding the results I have obtained using DCPose.

Below you can see three consecutive images (from a longer sequence). It seems that the temporal component does not correct the false association (left/right knee) which it could have if it used the information from the previous and the next frame. Am I using it in a wrong way, or you expected this kind of behaviour.

00000040
00000041
00000042

Cheers!

Speed

I want to know the speed. Anyone could tell me?

ninja: build stopped: subcommand failed

`/opt/conda/lib/python3.6/site-packages/torch/include/ATen/core/TensorBody.h:395:1: note: declared here
T * data() const {
^
ninja: build stopped: subcommand failed.
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1673, in _run_ninja_build
env=env)
File "/opt/conda/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "setup.py", line 42, in
zip_safe=False)
File "/opt/conda/lib/python3.6/site-packages/setuptools/init.py", line 153, in setup
return distutils.core.setup(**attrs)
File "/opt/conda/lib/python3.6/distutils/core.py", line 148, in setup
dist.run_commands()
File "/opt/conda/lib/python3.6/distutils/dist.py", line 955, in run_commands
self.run_command(cmd)
File "/opt/conda/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/develop.py", line 34, in run
self.install_for_development()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/develop.py", line 136, in install_for_development
self.run_command('build_ext')
File "/opt/conda/lib/python3.6/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/opt/conda/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/build_ext.py", line 79, in run
_build_ext.run(self)
File "/opt/conda/lib/python3.6/site-packages/Cython/Distutils/old_build_ext.py", line 186, in run
_build_ext.build_ext.run(self)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 339, in run
self.build_extensions()
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 708, in build_extensions
build_ext.build_extensions(self)
File "/opt/conda/lib/python3.6/site-packages/Cython/Distutils/old_build_ext.py", line 195, in build_extensions
_build_ext.build_ext.build_extensions(self)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 448, in build_extensions
self._build_extensions_serial()
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 473, in _build_extensions_serial
self.build_extension(ext)
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/build_ext.py", line 196, in build_extension
_build_ext.build_extension(self, ext)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 533, in build_extension
depends=ext.depends)
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 538, in unix_wrap_ninja_compile
with_cuda=with_cuda)
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1359, in _write_ninja_file_and_compile_objects
error_prefix='Error compiling objects for extension')
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1683, in _run_ninja_build
raise RuntimeError(message) from e
RuntimeError: Error compiling objects for extension
`

在自定义数据集上重新训练

您好,请问我想在自己的数据集(一个手势姿态数据集,包含21个关节点)上重新训练和测评您的模型,需要将我的数据转换成什么样的格式呢?在训练和测评阶段有什么其他额外的数据处理嘛?

The comparison with PoseWarper might not be fair.

Due to some untold reason, the "neck" joint that PoseWarper used in training is not the ground truth, but the average of left and right shoulder. This brings down the PoseWarper's results about 23 mAP across PoseTrack2017 and PoseTrack2018 valid sets and test sets. In other words, the "true" accuracy of the PoseWarper would be 23 mAP higher than what they report in their paper.

In your paper, you compare your results with the PoseWarper's "faulty" results. From your code, we can see you correctly use the "neck" ground truth in training. That's probably why your "Head" results are much better across PoseTrack valid and test datasets.

I'm curious, you guys really know nothing about the "neck" problem in PoseWarper?

YOLOv3

请问您一下,您使用的 YOLOv3 检测训练代码有吗

Posetrack download

Hello,The Posetrack official website is down and I'm unable to access the dataset. I could see that many have faced the same issue.It would be great if you could resolve the website issue or send us the data personally.Thank you so much for your time.

/DCPose/thirdparty/deform_conv# python setup.py develop

`ninja: build stopped: subcommand failed.
Traceback (most recent call last):
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1673, in _run_ninja_build
env=env)
File "/opt/conda/lib/python3.6/subprocess.py", line 438, in run
output=stdout, stderr=stderr)
subprocess.CalledProcessError: Command '['ninja', '-v']' returned non-zero exit status 1.

The above exception was the direct cause of the following exception:

Traceback (most recent call last):
File "setup.py", line 42, in
zip_safe=False)
File "/opt/conda/lib/python3.6/site-packages/setuptools/init.py", line 153, in setup
return distutils.core.setup(**attrs)
File "/opt/conda/lib/python3.6/distutils/core.py", line 148, in setup
dist.run_commands()
File "/opt/conda/lib/python3.6/distutils/dist.py", line 955, in run_commands
self.run_command(cmd)
File "/opt/conda/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/develop.py", line 34, in run
self.install_for_development()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/develop.py", line 136, in install_for_development
self.run_command('build_ext')
File "/opt/conda/lib/python3.6/distutils/cmd.py", line 313, in run_command
self.distribution.run_command(command)
File "/opt/conda/lib/python3.6/distutils/dist.py", line 974, in run_command
cmd_obj.run()
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/build_ext.py", line 79, in run
_build_ext.run(self)
File "/opt/conda/lib/python3.6/site-packages/Cython/Distutils/old_build_ext.py", line 186, in run
_build_ext.build_ext.run(self)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 339, in run
self.build_extensions()
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 708, in build_extensions
build_ext.build_extensions(self)
File "/opt/conda/lib/python3.6/site-packages/Cython/Distutils/old_build_ext.py", line 195, in build_extensions
_build_ext.build_ext.build_extensions(self)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 448, in build_extensions
self._build_extensions_serial()
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 473, in _build_extensions_serial
self.build_extension(ext)
File "/opt/conda/lib/python3.6/site-packages/setuptools/command/build_ext.py", line 196, in build_extension
_build_ext.build_extension(self, ext)
File "/opt/conda/lib/python3.6/distutils/command/build_ext.py", line 533, in build_extension
depends=ext.depends)
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 538, in unix_wrap_ninja_compile
with_cuda=with_cuda)
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1359, in _write_ninja_file_and_compile_objects
error_prefix='Error compiling objects for extension')
File "/opt/conda/lib/python3.6/site-packages/torch/utils/cpp_extension.py", line 1683, in _run_ninja_build
raise RuntimeError(message) from e
RuntimeError: Error compiling objects for extension
`

Datasets

Thank you very much for your work.
I am in China now and I cannot download the posetrack2017 and posetrack2018 dataset. Could you please share it with me?
Or can I train directly with the COCO dataset? What should I do?(Do I need to change the format in COCO?)
Looking forward to your reply ! Thanks!!

error in modulated_deformable_im2col_cuda": too many resources requested for launch

Got this error while running the demo on Nvidia TX2:
Solution is to change both .cu files in thirdparty/deform_conv/src to reduce the number of cuda threads.

Namely: from const int CUDA_NUM_THREADS = 1024; to const int CUDA_NUM_THREADS = 512; did the trick. Jetson TX2 has 8gb of vram.

NB: after doing so, rewmember to run again python setup.py develop to compile the deform_conv module

error in setup.py

src/deform_conv_cuda.cpp:559:3: error: ‘AT_CHECK’ was not declared in this scope; did you mean ‘DCHECK’?

  559 |   AT_CHECK(input.is_contiguous(), "input tensor has to be contiguous");
      |   ^~~~~~~~
      |   DCHECK

error: command '/usr/bin/gcc' failed with exit code 1

performance

I ran training new model on posetrack17 and test it on validation dataset for three times. I only got 80.9mAP for three times where there are large gap. 2 * 2080 and other config is not changed. I used 0, 1000, 8000 random seed for different initialization. So where's the problem.

DCPose support status for other dataset

Hi guys, I am working on a pose estimation project with DCPose right now and thanks for the work!

My project requires to know more about the foot keypoints detection since it focus on the running motion. However, PoseTrack dataset doesn't have the required feet annotation? I was wondering how can I set up the COCO-full-body dataset to train with DCPose and do full body keypoint detection with it?

The link to coco: https://github.com/jin-s13/COCO-WholeBody

Any documents will be helpful

复现

有两个问题想问一下作者,万分感谢!1.数据集下载的时候,比如posetrack18要全部下载完吗?2.posetrack17的脚本怎么用?

Performance questions

Thank you for your good research. And thank you for revealing the code quickly.
I tried to implement your code in the same environment.
However, using a single GPU environment, the batch have become 1/2 size. And I got the following results.
edit config line 4 : GPUS: (1,) (in my environment using 0:3080, 1:2080TI)

Posetrack 2017 val
Model Head Shoulder Elbow Wrist Hip Knee Ankle Mean
DcPose_RSN 86.3907 87.718 83.2292 76.2394 80.1681 79.1894 71.2038 80.9779
Posetrack 2018 val
Model Head Shoulder Elbow Wrist Hip Knee Ankle Mean
DcPose_RSN 83.8176 86.2703 81.4414 75.3437 77.1077 77.9727 72.2061 79.4758

The result of not achieving the performance suggested is because the batch size is small?

I would like to change the deform conv module to torchvision to run your code on CUDA11.1.
https://pytorch.org/vision/stable/_modules/torchvision/ops/deform_conv.html#deform_conv2d

I also encountered an error in the posetrack 2017 test dataset.

2021-04-02 14:23:13 [engine.core.function] INFO: test: [3100/5462]      Time 1.659 (1.713)      Data 0.027s (0.083s)    Accuracy 0.000 (0.006)
2021-04-02 14:25:59 [engine.core.function] INFO: test: [3200/5462]      Time 1.659 (1.712)      Data 0.027s (0.081s)    Accuracy 0.000 (0.006)
Traceback (most recent call last):
  File "run.py", line 33, in <module>
    main()
  File "run.py", line 29, in main
    runner.launch()
  File "/DCPose/engine/defaults/runner.py", line 63, in launch
    evaluator.exec()
  File "/DCPose/engine/defaults/evaluator.py", line 20, in exec
    self.eval()
  File "/DCPose/engine/defaults/evaluator.py", line 73, in eval
    phase=self.phase)
  File "/DCPose/engine/core/function.py", line 165, in eval
    input_x, input_sup_A, input_sup_B, target_heatmaps, target_heatmaps_weight, meta = next(self.dataloader_iter)
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 345, in __next__
    data = self._next_data()
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 856, in _next_data
    return self._process_data(data)
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/dataloader.py", line 881, in _process_data
    data.reraise()
  File "/opt/conda/lib/python3.7/site-packages/torch/_utils.py", line 394, in reraise
    raise self.exc_type(msg)
AttributeError: Caught AttributeError in DataLoader worker process 0.
Original Traceback (most recent call last):
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/_utils/worker.py", line 178, in _worker_loop
    data = fetcher.fetch(index)
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/_utils/fetch.py", line 44, in fetch
    data = [self.dataset[idx] for idx in possibly_batched_index]
  File "/opt/conda/lib/python3.7/site-packages/torch/utils/data/_utils/fetch.py", line 44, in <listcomp>
    data = [self.dataset[idx] for idx in possibly_batched_index]
  File "/DCPose/datasets/zoo/posetrack/PoseTrack.py", line 100, in __getitem__
    return self._get_spatiotemporal_window(data_item)
  File "/DCPose/datasets/zoo/posetrack/PoseTrack.py", line 166, in _get_spatiotemporal_window
    self.logger.error(error_msg)
AttributeError: 'PoseTrack' object has no attribute 'logger'

Val issues

when i train this net , it work acc is about 0.75-0.8
but while i val it ,acc is 0.000 who know why? thanks

ValueError: Unknown CUDA arch (8.6) or GPU not supported

Thank you for the amazing work! I am having an issue while installing DCN

File "C:\Users\92336\anaconda3\envs\DCPose\lib\site-packages\torch\utils\cpp_extension.py", line 1027, in _get_cuda_arch_flags
   raise ValueError("Unknown CUDA arch ({}) or GPU not supported".format(arch))
ValueError: Unknown CUDA arch (8.6) or GPU not supported

I am using python 3.6.12, pytorch-1.4.0, and CUDA 10.0

My GPU Specifications: GeForce RTX 3080

Windows System

Any solutions for this?

when the pretrained model will release?

Hi, thanks for your great job, I'm working on gait recognition, and I want to extract human skeletons accurately. I run on the demo, but it doesn't work because of the lack of the yolov3.weight this file. Thanks again!

posetrack download

您好,PoseTrack官网目前已无法访问,方便分享下PoseTrack17、PoseTrack18的数据集及其标注文件吗?
谢谢~

模型的速度问题

在论文中没有找到速度指标的描述,有没有前辈,可以提供下这方面的数据,感谢!

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.