Next Article in Journal
Advances in Condition Monitoring of Railway Infrastructure
Previous Article in Journal
Real-Time Myocardial Infarction Detection Approaches with a Microcontroller-Based Edge-AI Device
Previous Article in Special Issue
Dynamic Japanese Sign Language Recognition Throw Hand Pose Estimation Using Effective Feature Extraction and Classification Approach
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

HDPose: Post-Hierarchical Diffusion with Conditioning for 3D Human Pose Estimation

1
Department of Information and Communications Engineering, Sejong University, Seoul 05006, Republic of Korea
2
Department of Electrical Engineering, Sejong University, Seoul 05006, Republic of Korea
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(3), 829; https://doi.org/10.3390/s24030829
Submission received: 12 December 2023 / Revised: 16 January 2024 / Accepted: 22 January 2024 / Published: 26 January 2024
(This article belongs to the Special Issue Sensing and Vision Technologies for Human Activity Recognition)

Abstract

:
Recently, monocular 3D human pose estimation (HPE) methods were used to accurately predict 3D pose by solving the ill-pose problem caused by 3D-2D projection. However, monocular 3D HPE still remains challenging owing to the inherent depth ambiguity and occlusions. To address this issue, previous studies have proposed diffusion model-based approaches (DDPM) that learn to reconstruct a correct 3D pose from a noisy initial 3D pose. In addition, these approaches use 2D keypoints or context encoders that encode spatial and temporal information to inform the model. However, they often fall short of achieving peak performance, or require an extended period to converge to the target pose. In this paper, we proposed HDPose, which can converge rapidly and predict 3D poses accurately. Our approach aggregated spatial and temporal information from the condition into a denoising model in a hierarchical structure. We observed that the post-hierarchical structure achieved the best performance among various condition structures. Further, we evaluated our model on the widely used Human3.6M and MPI-INF-3DHP datasets. The proposed model demonstrated competitive performance with state-of-the-art models, achieving high accuracy with faster convergence while being considerably more lightweight.

1. Introduction

The goal of monocular 3D human pose estimation (HPE) is to localize 3D body joints from 2D images or video. This task is crucial in computer vision applications, such as human–robot interaction [1], autonomous driving [2], the metaverse [3], and VR [4]. In recent years, coupled with the success of deep learning, the performance of monocular 3D HPE has made notable progress [5,6,7,8,9]. The lifting-based approach [10] involves mapping from a 2D pose to a 3D pose. This approach does not estimate the 3D pose directly from the image, thereby achieving high performance with less influence from background and lighting conditions. The dilated temporal-based approach [11] uses a fully convolutional model dependent on dilated temporal convolutions over 2D keypoints to effectively estimate 3D poses in videos. By learning the trajectory of a 2D keypoint’s movement over time, this approach facilitates the forecasting of the sequential movement of each joint within the human anatomy across successive frames of a video. Furthermore, the transformer-based approach [7] was proposed to learn the temporal trajectories by capturing the surrounding sequences and the long-range associations of the input sequence. These approaches have made significant contributions to improving the performance of 3D HPE. By leveraging the capabilities of these methods, the accuracy and reliability of pose estimation have been considerably improved. However, despite advancements in 3D HPE, monocular 3D HPE faces challenges in terms of depth ambiguity and various occlusion scenarios, which result in the reconstruction of incorrect 3D poses.
To overcome these limitations, previous studies have proposed methods utilizing diffusion models. Diffusion models have facilitated the realization of imposing performance. Representatively, denoising diffusion probabilistic models (DDPMs) [12] have rendered image generation plausible by gradually denoising sampled data from a Gaussian distribution. Intuitively, the DDPM mechanism entails treating the incorrect data in 3D poses, which are caused by occlusion, as noise. Based on the gradual denoising of this noise via small iterative steps, the correct 3D pose can ultimately be converged. However, reconstructing the correct 3D pose using only the incorrect 3D pose sampled from the pure a Gaussian distribution is challenging owing to the lack of sufficient evidence to predict the correct 3D pose from an incorrect one. Moreover, the learning process for reconstructing the correct 3D pose is time-consuming owing to the numerous steps involved in the denoising process. Therefore, in the denoising process, the conditions are vital for providing guidance for the transition from an incorrect 3D pose towards the correct 3D pose. This study focused on manipulating conditions to guide 3D poses more accurately and quickly.
Currently, 3D HPE using diffusion involves two distinct methods for providing conditions. The first approach [13] is the provision of a simple 2D pose. This method provides evidence through a 2D pose, thus enabling the 3D pose sampled from pure Gaussian noise to be guided towards the correct 3D pose. However, a simple 2D pose does not yield high performance, as it cannot capture the temporal information between frames in a 2D pose sequence. The second approach [14] is a context encoder that includes spatial and temporal information and can outperform simple 2D poses. This method can outperform the first approach because it captures correlations in 2D pose joints and frames. However, this method requires more training steps to converge to optimal performance, resulting in longer training times.
To address the issues highlighted, this study proposed a hierarchical diffusion 3D human pose estimation (HDPose). We aimed to learn to guide the denoising model to clean the 3D pose in a precise and rapid manner by passing detailed spatial and temporal information to it in a hierarchical structure. We were inspired by the method proposed by Lu et al. [15], wherein the HDAE [15] exploited the low-level-to-high-level feature hierarchy. This method can efficiently and deterministically transform a semantic image from Gaussian noise. Therefore, we designed a hierarchical conditioning model inspired by the above literature. The proposed method utilized the context encoder described previously. This effectively encapsulated the context encoder of the final layer output, which integrated skip connections and spatio-temporal features, thereby seamlessly capturing both spatial and temporal features at diverse scales. Consequently, we hierarchically aggregated the last features extracted from the conditional model to each encoder of the denoising model. This enabled the neural network to learn more efficiently and converge rapidly. We compared the performance of various methods that feed features from the condition model to the denoising model. Empirically, we found that the most effective approach was to hierarchically aggregate the holistic representation produced by the final encoder of the condition. In addition, compared to previous models [13,14], we observed that we achieve significantly better performance by using a pre-trained condition model.
We experimented with different conditional model structures, including simple 2D pose sequences, non-hierarchical structure, pre-hierarchical structure, and post-hierarchical structure. Our model was easily trainable and could more accurately predict 3D poses that matched 2D poses with high-quality poses. Moreover, it quickly converged to the optimal 3D pose during the learning process and was a more lightweight model than the previous approach [7,16]. In summary, the primary contributions of this study are as follows:
  • We propose a novel hierarchical diffusion-based (HDPose) method that can converged to a fast and accurate 3D pose by aggregating spatio-temporal information to all layers of the denoising model.
  • We performed experiments with various conditioning methods, including simple 2D pose, non-hierarchical, pre-hierarchical, and post-hierarchical structure. Through empirical observation, we identified the model structure that yielded the best performance.
  • Our proposed hierarchical model, when compared to state-of-the-art methods, demonstrated competitive results. It maintained a lightweight model, showcasing its effectiveness on the Human3.6M and MPI-INF-3DHP datasets.

2. Related Work

2.1. 3D Human Pose Estimation

Monocular 3D human pose estimation methods can typically be divided into two categories: frame-based approaches and sequence-based approaches. Many prior approaches to frame-based methods have employed either a CNN to directly regress the 3D pose from the image [17,18,19] or an off-the-shelf 2D pose detector to predict the 2D pose, followed by a lifting process to obtain the 3D pose, e.g., [9,10,20,21,22,23,24]. Pavlakos et al. [25] directly regressed the 3D coordinates of each joint in a pose by transforming the pose into a 3D volume and training a CNN to predict the likelihood of each joint being located within each voxel of the volume. Martinez et al. [10] proposed a method to detect 2D poses and then lifted them to 3D poses to ensure that they were less sensitive to changes in background and lighting.
The sequence-based method predicts a consistent 3D pose by using temporal information from a sequence of 2D human poses, e.g., [7,8,26,27,28,29]. Pavllo et al. [11] proposed a fully convolutional neural network architecture that utilized dilated temporal convolutions over 2D keypoints to estimate the 3D pose in a video sequence. Zheng et al. [7] proposed a transformer-based approach for accurate 3D human pose estimation from videos.
In contrast to the frame-based approach, the sequence-based approach utilizes temporal information to infer consistent 3D poses even when body joints are occluded in individual frames. The current state-of-the-art model by Zhang et al. [16] further enhances performance through a sequence-to-sequence (seq2seq) approach built upon the Transformer architecture. This advancement addresses limitations found in the previously proposed seq2frame approach. The seq2frame method, which predicts individual frames based on long-range temporal information processed by the transformer, necessitates the repetitive input of 2D keypoint sequences with substantial overlap to deduce the 3D pose for all frames. This leads to the issue of redundant computations. To resolve this challenge, Zhang et al. [16] introduced a novel transformer-based 3D human pose estimation method utilizing the seq2seq approach, enabling the prediction of consecutive frames more efficiently. We were inspired by the work of Zhang et al. [16]. Our approach integrates transformer-based 3D human pose estimation to effectively harness long-range temporal information within videos. We adopt a seq2seq approach, building upon the foundations laid by Zhang et al. [16]. This allows us to more accurately capture the nuances of human motion over extended periods, enhancing the overall effectiveness of our pose estimation framework.

2.2. Generative 3D Human Pose Estimation

Generative methods have emerged as a promising approach for addressing the challenges of occlusion and depth ambiguity in 3D human pose estimation. Recently, methods using many generative models, such as GAN [30], CVAE [31], and Normalizing Flows [32], have been proposed. Barsoum et al. [33] proposed a novel sequence-to-sequence model for probabilistic human motion prediction. Sharma et al. [34] addressed the ambiguity of 2D-3D lifting by generating multiple 3D posture possibilities. Wehrbein et al. [35] explored ambiguous 2D-3D inverse problems using a regularized flow-based approach with deterministic 3D-2D mapping and uncertainty modeling from 2D detectors.
Recently, methods utilizing diffusion have been proposed. Diffpose [36] considered the cross-correlation between the joints, which was not considered as a condition in previous studies. Using an embedding transformer as a condition, it is provided as a condition to the diffusion model through a joint-wise embedding vector. The study by Choi et al. [13] uses GCN [37] as a denoising model that captures the spatial anatomy of the person well, and utilizes 2D keypoints as a condition. However, this work is only optimized for frame-by-frame 3D HPE operations, which has the limitation of not exploiting important temporal trajectories in video sequences and thus does not achieve high performance. To improve on this, the work of Rommel et al. [14] utilizes spatial and temporal context by using pre-trained models as conditionals, but this method also requires significant time for convergence to the desired pose. Therefore, we provide a hierarchical aggregation of spatial and temporal context using pre-trained conditionals to enable faster convergence.

3. Proposed Method: Hierarchical Diffusion 3D Human Pose Estimation (HDPose)

3.1. Diffusion Model

The diffusion model learns to gradually denoise a sampled 3D pose starting from pure noise. A diffusion process can be divided into two processes: forward and reverse processes.
  • Forward Process can be modeled as a Markov chain [38] wherein Gaussian noise is gradually added to the ground truth 3D pose x 0 at each subsequent step t until the state attains a Gaussian distribution. It is denoted as N ( 0 , I ) . To train the diffusion model to denoise a 3D pose in a progressive manner, it must be provided with supervisory signals in the form of ground truth distributions. We can generate samples from these distributions using the forward diffusion process iteration, starting with the ground truth 3D pose distribution and gradually adding noise. This process can predefine q ( x 1 : T | x 0 ) through variance noise scheduler β t and step t as follows:
    q ( x 1 : T | x 0 ) = t = 1 T q ( x t | x t 1 )
    q ( x t | x t 1 ) : = N ( x t ; 1 β t x t 1 , β t I )
    We used the cosine noise variance schedule [39] to control the amount of noise added to the 3D pose at each step of the diffusion process. We enabled a reparameterization trick to make the diffusion process more efficient by enabling direct sampling from the noise distribution. Following DDPM [12], this process can be expressed as:
    x t : = α ¯ t x 0 + 1 α ¯ t ϵ
    where α t : = 1 β t , α ¯ t : = s = 1 t α s and ϵ N ( 0 , I ) Gaussian noise ϵ . We can optimize L by randomly sampling t during training, thereby exploiting these properties.
  • Reverse Process is a process of reconstruction of the correct 3D pose from an incorrect 3D pose. The task of accurately reconstructing a 3D pose from a random distribution remains a significant challenge. To address this, we adapted the diffusion process based on the context information derived from the 2D sequence. This approach ensured the attainment of a deterministic 3D pose that aligns with the spatial and temporal embedding vectors. Reverse processes can also be expressed as a joint distribution p θ ( x t 1 | x t ) , which describes the probability of observing a 3D pose x t 1 at timestep t.
    p θ ( x t 1 | x t ) : = N ( x t 1 ; μ θ ( x t , c , t ) , Σ θ ( x t , c , t ) )
    In the DDPM [12], Σ θ ( x t , c , t ) was fixed as constant. Considering the mean parameter μ θ ( x t , c , t ) , we can compute the distribution of the previous timestep x t 1 using the μ θ function, which is defined as follows.
    μ θ ( x t , c , t ) : = 1 α t ( x t β t 1 α ¯ t ϵ θ ( x t , c , t ) )
    Therefore, the only remaining task is to predict ϵ θ . However, in the above method, timestep t is typically set to a value greater than 100 to ensure that the model can accurately learn the diffusion process. As a result, this can make the reverse diffusion process computationally expensive. Instead of predicting the noise, we approximated the reverse diffusion process using DDIM [40] to reduce the computational cost, which required fewer iterations. Therefore, we directly predicted the correct 3D pose x ˜ 0 from the trained network.
    x ˜ 0 = f θ ( x t , c , t )

3.2. Training and Sampling Process

Training process. Initially, we randomly selected a timestep t U ( 1 , , T ) and sample noise ϵ N ( 0 , I ) . x t was obtained by gradually adding noise to the ground truth 3D pose, with the noise level being dependent on the specific timestep t. We processed the pose denoising by inputting x t , condition c, and timestep t into the denoising model f θ , as defined in Equation (6). This model was responsible for reconstructing the predicted 3D pose x ˜ 0 . Subsequently, we applied gradient descent steps to Equation (7) until convergence to the correct 3D pose was achieved:
L = E t [ 1 , T ] , x 0 , c [ x 0 x ˜ 0 2 ]
Throughout the training, the entire diffusion process is supervised. We optimize the denoising model using a mean squared error (MSE) between the ground truth 3D pose and the predicted 3D pose.
Sampling process involves estimating the correct 3D pose using the trained denoising model f θ . We initiated the process by sampling the initial 3D pose x 0 , T from a Gaussian noise distribution corresponding to timestep T. The pose x 0 , 0 was directly predicted from x 0 , t and then fed into the denoising model to produce the incorrect 3D pose x 0 , t 1 for the subsequent timestep. This procedure is described by the following equation, which outlines the DDIM [40] process:
x 0 , t 1 = α ¯ t 1 · x ˜ 0 , 0 + 1 α ¯ t 1 σ t 2 · ϵ t + σ t ϵ
ϵ t = x 0 , t α ¯ t · x ˜ 0 , 0 1 α ¯ t
In accordance with Equation (3), ϵ t is derived. As σ t nears zero, its determinate nature intensifies. We commence at timestep T with x 0 , T and recursively predict the 3D pose for the next timestep using the denoising model defined in Equation (6). At each timestep, the predicted 3D pose serves as the input for the denoising model, facilitating the prediction of the subsequent 3D pose. This process is iteratively conducted N times, where N belongs to the range [1, T] and satisfies N < T .

3.3. Pre-Trained Model of Conditioning

In general, obtaining a correct 3D pose solely from the incorrect 3D data derived from a Gaussian distribution is challenging. To address this, we integrated additional conditions to more precisely steer the 3D pose reconstruction process. Nevertheless, the denoising process, when implemented using only basic 2D pose conditions, cannot achieve optimal results. The reasons for the suboptimal results when using only basic 2D pose conditions in the denoising process are as follows. First, utilizing simple 2D poses fails to effectively capture temporal information, which leads to an inability to accurately predict the appropriate trajectories for each joint, thus not achieving the best possible performance. Second, even when context information containing spatial–temporal data is provided to enhance the process, there is an issue with the prolonged time required for convergence to the optimal pose. When using a transformer as the backbone of a condition model, the self-attention mechanism has the ability to effectively capture long sequences. While this is an advantageous property for modeling complex sequences, it does require significant computational resources. This computational intensity comes from computing the interactions between all the elements in the sequence, and for the 3D HPE task, using long-range associations of the input sequence to capture temporal information between sequences improves accuracy, but requires computational complexity and significant memory to store the attention weights.
Therefore, we use a pre-trained [41] spatio-temporal transformer as the condition model. This method helps save learning time and resources and plays an important guiding role for more accurate 3D pose estimation. We pass the spatial and temporal context to the diffusion model in a hierarchical manner, which allows the diffusion model to converge faster.

3.4. Hierarchical Conditioning Diffusion for 3D Human Pose Estimation

This section presents an overview of conditional diffusion applied to 3D HPE, as illustrated in Figure 1.
  • Pre-Hierarchical Structure
    In Figure 1c, this architecture is inspired by the work of Sun et al. [42]. Their work showed that connecting feature maps of varying depths allows networks to integrate and utilize multiscale information, which can lead to a more nuanced understanding of the input data. Instead of using a simple 2D pose as a condition, we use an E 2 D consisting of a spatial encoder E S and a temporal encoder E T . The spatial encoder E S learns the spatial correlations between all joints in the i t h frame. This approach allows the model to gain a more accurate understanding of the actual structure of the body and the natural connections between joints, leading to more precise and realistic pose estimation. Initially, the 2D pose x 2 D is transformed into a higher-dimensional embedding vector X R F × J × D via linear projection. This vector X is combined with the learnable spatial location sign E s R F × J × D and then input to E S . The output of E S is processed by E T , which captures the joint correlation of each frame. As observed in [16], there are significant differences in the motion trajectories of the joints from frame to frame, so it is essential to learn a distinct trajectory pattern for each joint in each frame.
    E F 2 D = C o n c a t ( E 1 S , E 1 T , E L S , E L T ) W F
    The depth of the encoder is denoted by L. In this framework, the fusion module uses a linear projection to align the dimensions of the concatenated features ( E i S , E i T ) with the dimensions of the spatial encoder used for denoising. In the denoising model f θ ( E F 2 D , x t 3 D , t ) , the input features include both the condition E F 2 D and the incorrect 3D pose x t 3 D and the time interval t. The incorrect 3D pose is then merged with the associated condition feature E F 2 D and jointly trained. The training process described in Section 3.2 is then performed.
  • Post-Hierarchical Structure
    As shown in Figure 1d, we introduce a post-hierarchical structure as an efficient way to accurately and quickly guide the construction of the correct 3D pose. From our observations, we found that spreading the final extracted features in a hierarchical structure to each encoder layer of the denoising model yields the most effective results. Similar to (b), we extract E 2 D from the final layer of the condition model. We then pass this feature, which covers the entire spatial and temporal information, to the denoising model f θ ( E 2 D , x t 3 D , t ) . In traditional 3D HPE methods using diffusion, it is common to associate the condition with x 3 D only in the initial encoder, whereas our approach incorporates it in all encoder layers. The incorrect 3D pose x t 3 D is transformed into a high-dimensional embedding vector, which is then added along with a spatial position embedding E s and a time interval embedding t. At each encoder step, this embedding vector is further aggregated with the global features of the condition model E 2 D to produce global condition information. The output of the spatial encoder E S is combined with the temporal position embedding E t . The combined features are then reshaped in the R J × F × D dimension and provided as input to the temporal encoder E T . This procedure is repeated on all encoder layers to the final depth to extract the final 3D pose x 0 : t . The x t 3 D is utilized to generate a noisy 3D pose x 0 : t 1 to be input to the denoising model as the next step, which is input via DDIM [40]. This procedure is repeated N times. The goal is to progressively refine the pose to an accurate 3D structure. This process is repeated N times, progressively refining the pose to an accurate 3D reconstruction. Figure 2 shows the detailed architecture of the Post-Hierarchical Structure.

4. Experiment Results

4.1. Datasets and Evaluation Metrics

We conducted our experiments using datasets commonly adopted in previous studies. To perform a comprehensive evaluation of our proposed method, we used well-known benchmark datasets for human pose estimation: Human3.6M [43,44] and MPI-INF-3DHP [45].
Human3.6M is a comprehensive and challenging dataset for 3D human pose estimation in indoor environments. The dataset was captured using four synchronized high-resolution cameras operating at 50 Hz, providing high-quality data for a wide range of human poses and activities. Our model is trained on five subjects (S1, S5, S6, S7, S8) and evaluated on two subjects (S9, S11).
MPI-INF-3DHP is a widely used large-scale dataset for 3D human pose estimation, involving both indoor and complex outdoor scenarios. It features eight actors performing eight distinct activities captured from 14 synchronized cameras. The dataset comprises over 1.3 million frames meticulously collected from these cameras. We split a training set containing eight subjects and a test set containing seven subjects, the same as in the previous study [7,8,16].
Evaluation Metrics. The performance of the proposed method was evaluated using the same metrics as in previous methods. First, the proposed models were evaluated on Human3.6M using standard protocols (i.e., Protocol 1, Protocol 2). Protocol 1 used mean per-joint position error (MPJPE), which is the average of the Euclidean distance in millimeters between the ground truth pose and the predicted pose. Protocol 2 used P-MPJPE, applied to the alignment between the ground truth pose and the predicted pose. The MPI-INF-3DHP reports the percentage of correct keypoints (PCK), Area Under the Curve (AUC), and MPJPE as evaluation metrics.

4.2. Implementation Details

All experiments were conducted on individual NVIDIA GeForce RTX 3090 GPUs (Geforce RTX 3090 GPUs is developed by NVIDIA, based in Santa Clara, CA, USA). We employed a batch size of 4 and performed 300 epochs in training. The initial learning rate was set to 0.001 and weight decay was 7 × 10−6 per epoch. The Adam optimizer was used for optimization. Data augmentation was performed by applying flipping and horizontal transformations. The experiments used the PyTorch framework and the method proposed by Zhang et al. [16] as the backbone for both the condition and denoising models. The spatial and temporal encoders each had a depth of 8 and multi-head attention of 8. The embedding dimension was set to 256. As a hyperparameter for diffusion, the number of hypotheses H was set to 1, β started at 0.99 and decreased to 0.01, and cosine noise variance schedule [39] was used. We experimented with a timestep of T 1000.

4.3. Quantitative Results

  • Results on benchmark dataset. Table 1 demonstrates that our method yielded nearly identical results to the previous SOTA models in terms of average MPJPE for Protocol 1 at 41.0 mm and for Protocol 2 at 32.8 mm, while outperforming other models. Specifically, on the Human3.6M, our model exhibited an improvement of approximately 4.21% (1.8 mm) for Protocol 1 and 4.65% (1.6 mm) for Protocol 2, compared to the method proposed by Shan et al. [46]. On the MPI-INF-3DHP, it exhibited a significant enhancement of 38.45% (from 58.0 mm to 35.7 mm) compared to the method proposed by Liu et al. [8]. For detailed information, refer to Table 2.
  • Results on computational complexity. To evaluate the computational complexity of our model, we compared the number of trainable parameters with those of previous models. Despite its lightweight design, our model matched the performance of SOTA models, with a reasonable number of floating-point operations per second (FLOPs). The training process performed on a single GeForce 3090 GPU completes 100 epochs in about 24 h. Further emphasizing its efficiency, a comparison of frames per second (FPS) during the inference process revealed real-time capabilities. These detailed results are elucidated in Table 3.
  • Results on comparison of convergence. We compared the convergence speed of our proposed method HDPose and the state-of-the-art model [7,16], and the result is reported in Figure 3. Comparing the optimal MPJPE performance over 100 epochs, Zheng et al. [7] achieved 45.1 mm in 100 epochs and Zhang et al. [16] achieved 42.2 mm in 96 epochs. On the other hand, we can see that our proposed model converges to 42.1 mm already in 60 epochs with a faster learning process. By comparing the convergence speed with other state-of-the-art (SOTA) models, we found that our proposed model outperforms others by up to 26% at peak performance. We found that our model has a faster learning convergence speed than other models using pre-trained conditioning models. The model was trained to recognize weights that were optimized to recognize features that were already useful in a spatial and temporal context.
  • Results on visualization. Figure 4 presents a comparison of the state-of-the-art (SOTA) methods [7,16,28] and HDPose by visualizing their performance across three actions: Sitting, Greeting, and WalkingDog in Subject S11 of Human3.6M. We found that our proposed method generated more plausible poses than previous works and closely resembled the ground truth 3D pose. We also presented visualization results on MPI-INF-3DHP and on 3DPW [47], an “in-the-wild” dataset that reflects a real-world environment with varying lighting, backgrounds, and camera angles. More details on this can be found in the Appendix A.

4.4. Ablation Study

Using pytorch profiler [54], we found that the total time taken to perform a CUDA operation or task was 0.028 ms. We also analyzed the computational complexity according to the embedding dimension, as shown in Table 4. We fixed the embedding dimension of the conditioning model ( d c ) to 256 as the embedding dimension of the pre-trained model, and compared the number of parameters and FLOPs with MPJPE while changing the embedding dimension of the denoising model ( d d ), and found the most optimized performance at 256. Comparing the FPS and CUDA computation speed of our model with the 5.0 M of memory that can run on resource-intensive embedded devices, we believe that our model is suitable for use on embedded devices such as robots. This evidence strongly indicated that the proposed method was lightweight and efficient and rapidly converged to accurate 3D poses, thereby offering competitive performance in the realm of 3D HPE.
Performance differences in condition. Conditioning to demonstrate that the post-hierarchical structure achieves the highest performance and efficiency, we compared four different condition models for 100 epochs on the Human3.6M dataset with the same hyperparameter settings, including the encoder depth of 8, and one hypothesis. The post-hierarchical structure exhibited a significant improvement in MPJPE compared to the simple 2D pose (43.29%, 72.3 mm → 41.0 mm), pre-hierarchical structure (9.69%, 45.4 mm → 41.0 mm), and non-hierarchical structure (3.53%, 42.5 mm → 41.0 mm). The detailed results are presented in Figure 5.
Visualization of self-attention matrix among joints and frames. We visualize the spatial and temporal attention of the denoising model. As shown in Figure 6, the left side is the attention matrix to understand the correlation between each joint, demonstrating the capability of the model to distinguish between the left and right sides of a joint, and the right side is the attention matrix between frames, normalized to a value in the range of [0, 1]. It is easy to observe that our model has learned the connectivity between joints, even when they are physically far apart, and is also good at learning long-range associations of the input sequence.
Analysis depending on hierarchical structure. We compared the effect of hierarchical methods on accuracy to assess the performance of the proposed model at various depths of conditioning models. We observed a decline in the performance of the pre-hierarchical structure with increasing depth. Conversely, the performance of the post-hierarchical structure improved at deeper levels. These findings are detailed in Figure 7. Considering that the model’s input was a 2D pose, the integration of low-level and high-level conditioning models did not significantly enhance the 3D pose reconstruction. Consequently, we opted for a hierarchical broadcasting approach for denoising, which was centered on the final comprehensive representation of the 2D pose. This method was found to surpass existing techniques in terms of effectiveness.
Spatio-temporal encoder of each component. In our approach to hierarchically integrate the final comprehensive representation into a denoising model, we analyzed the performance variances when aggregating solely the spatial encoder, the temporal encoder, or a combination of both. These results are detailed in Table 5. It was observed that aggregating all spatial encoders reduced the performance compared to the use of concatenation methods. Aggregation of only the temporal encoder exhibited performance on par with employing only the spatial encoder. However, the highest performance was achieved when both the spatial and temporal encoders were aggregated together. Ultimately, the model attained its peak performance through the aggregation of all spatial and temporal encoders.

4.5. Limitations and Discussion

Our method still has some unsolved problems, and an example of some failure cases is shown in Figure 8. We perform 3D human pose estimation for a single person. However, when many people pass by, they may overlap and be recognized as one person, making the keypoints indistinguishable from each other. We designed a model that contains spatial and temporal information as a condition model, which is passed to the diffusion model to make the model more robust to occlusion. Here, the condition acts as a guide for the diffusion model to restore the correct pose. Conversely, if it estimates the wrong pose, it will reconstruct the wrong pose. Due to the nature of the camera, the image size is limited, so if a part of the person is cropped out of the image, the temporal information is not available and the model cannot handle severe occlusion.

5. Conclusions

In this paper, we proposed HDpose, a new framework designed for hierarchical conditioning in diffusion-based 3D human pose estimation. When performing 3D human pose estimation with a diffusion model, the simple use of a 2D pose as the condition necessitates several steps to converge to the correct 3D pose, often resulting in suboptimal performance. Therefore, we emphasized the need for a method that rapidly converged to the correct 3D pose. HDpose leveraged a condition model to generate a holistic representation, which was then aggregated across all layers of the denoising model to ultimately converge to the correct 3D pose. By comparing the convergence speed with other state-of-the-art (SOTA) models, we found that our proposed model outperforms others by up to 26% at peak performance. When evaluated against two widely used benchmark datasets in comparison with state-of-the-art (SOTA) methods, our model demonstrated equivalent performance while significantly reducing the model size by approximately 85.12%. Thus, a more lightweight model was realized.

Author Contributions

D.L. proposed the idea and defined the concept for this study, and also prepared the data and scenario design. The first draft of the manuscript was written by D.L. The review process, supervision, and approval of the final manuscript was done by J.K. All authors have read and agreed to the published version of the manuscript.

Funding

This research was partly supported by the Ministry of Science and ICT (MSIT), Korea, under the Information Technology Research Center (ITRC) support program (IITP-2023-2021-0-01816) supervised by the Institute for Information Communications Technology Planning Evaluation (IITP), by Institute of Information communications Technology Planning Evaluation (IITP) grant funded by the Korea government (MSIT) (No. 2023-0-00545, Development of Intelligent digital twin object federation and data processing technology), and by Institute of Information communications Technology Planning Evaluation (IITP) under the metaverse support program to nurture the best talents (IITP-2023-RS-2023-00254529) grant funded by the Korea government (MSIT).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The Human3.6M dataset is available upon request from the author [43]. The MPI-INF-3DHP dataset: https://vcai.mpi-inf.mpg.de/3dhp-dataset/. The 3DPW dataset: https://virtualhumans.mpi-inf.mpg.de/3DPW/.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Visualization of the HDPose in Various Environments

Figure A1. Qualitative comparison of the proposed HDPose method using the MPI-INF-3DHP.
Figure A1. Qualitative comparison of the proposed HDPose method using the MPI-INF-3DHP.
Sensors 24 00829 g0a1
Figure A2. Qualitative result of the proposed HDPose method with “in the wild” using the 3DPW.
Figure A2. Qualitative result of the proposed HDPose method with “in the wild” using the 3DPW.
Sensors 24 00829 g0a2

References

  1. Ehlers, K.; Brama, K. A human-robot interaction interface for mobile and stationary robots based on real-time 3D human body and hand-finger pose estimation. In Proceedings of the 2016 IEEE 21st International Conference on Emerging Technologies and Factory Automation (ETFA), Berlin, Germany, 6–9 September 2016; IEEE: Piscataway, NJ, USA, 2016; pp. 1–6. [Google Scholar]
  2. Zheng, J.; Shi, X.; Gorban, A.; Mao, J.; Song, Y.; Qi, C.R.; Liu, T.; Chari, V.; Cornman, A.; Zhou, Y.; et al. Multi-modal 3D human pose estimation with 2D weak supervision in autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 4478–4487. [Google Scholar]
  3. Zhou, Y.; Huang, H.; Yuan, S.; Zou, H.; Xie, L.; Yang, J. MetaFi++: WiFi-enabled Transformer-based Human Pose Estimation for Metaverse Avatar Simulation. IEEE Internet Things J. 2023, 10, 14128–14136. [Google Scholar] [CrossRef]
  4. Mehta, D.; Sridhar, S.; Sotnychenko, O.; Rhodin, H.; Shafiei, M.; Seidel, H.P.; Xu, W.; Casas, D.; Theobalt, C. Vnect: Real-time 3d human pose estimation with a single rgb camera. Acm Trans. Graph. (Tog) 2017, 36, 44. [Google Scholar] [CrossRef]
  5. Chen, C.H.; Ramanan, D. 3d human pose estimation = 2d pose estimation+ matching. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7035–7043. [Google Scholar]
  6. Hossain, M.R.I.; Little, J.J. Exploiting temporal information for 3d human pose estimation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 68–84. [Google Scholar]
  7. Zheng, C.; Zhu, S.; Mendieta, M.; Yang, T.; Chen, C.; Ding, Z. 3d human pose estimation with spatial and temporal transformers. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Virtual, 11–17 October 2021; pp. 11656–11665. [Google Scholar]
  8. Li, W.; Liu, H.; Tang, H.; Wang, P.; Van Gool, L. Mhformer: Multi-hypothesis transformer for 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 13147–13156. [Google Scholar]
  9. Xu, T.; Takano, W. Graph stacked hourglass networks for 3d human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 16105–16114. [Google Scholar]
  10. Martinez, J.; Hossain, R.; Romero, J.; Little, J.J. A simple yet effective baseline for 3d human pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2640–2649. [Google Scholar]
  11. Pavllo, D.; Feichtenhofer, C.; Grangier, D.; Auli, M. 3d human pose estimation in video with temporal convolutions and semi-supervised training. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 7753–7762. [Google Scholar]
  12. Ho, J.; Jain, A.; Abbeel, P. Denoising diffusion probabilistic models. Adv. Neural Inf. Process. Syst. 2020, 33, 6840–6851. [Google Scholar]
  13. Choi, J.; Shim, D.; Kim, H.J. DiffuPose: Monocular 3D Human Pose Estimation via Denoising Diffusion Probabilistic Model. arXiv 2023, arXiv:2212.02796. [Google Scholar]
  14. Rommel, C.; Valle, E.; Chen, M.; Khalfaoui, S.; Marlet, R.; Cord, M.; Perez, P. DiffHPE: Robust, Coherent 3D Human Pose Lifting with Diffusion. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV) Workshops, Paris, France, 2–3 October 2023; pp. 3220–3229. [Google Scholar]
  15. Lu, Z.; Wu, C.; Chen, X.; Wang, Y.; Qiao, Y.; Liu, X. Hierarchical Diffusion Autoencoders and Disentangled Image Manipulation. arXiv 2023, arXiv:2304.11829. [Google Scholar]
  16. Zhang, J.; Tu, Z.; Yang, J.; Chen, Y.; Yuan, J. MixSTE: Seq2seq Mixed Spatio-Temporal Encoder for 3D Human Pose Estimation in Video. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 13232–13242. [Google Scholar]
  17. Tekin, B.; Rozantsev, A.; Lepetit, V.; Fua, P. Direct Prediction of 3D Body Poses from Motion Compensated Sequences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  18. Chen, Y.; Tu, Z.; Kang, D.; Chen, R.; Bao, L.; Zhang, Z.; Yuan, J. Joint Hand-object 3D Reconstruction from a Single Image with Cross-branch Feature Fusion. arXiv 2020, arXiv:2006.15561. [Google Scholar] [CrossRef] [PubMed]
  19. Kocabas, M.; Athanasiou, N.; Black, M.J. VIBE: Video Inference for Human Body Pose and Shape Estimation. arXiv 2019, arXiv:1912.05656. [Google Scholar]
  20. Habibie, I.; Xu, W.; Mehta, D.; Pons-Moll, G.; Theobalt, C. In the wild human pose estimation using explicit 2d features and intermediate 3d representations. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 10905–10914. [Google Scholar]
  21. Tekin, B.; Márquez-Neila, P.; Salzmann, M.; Fua, P. Learning to fuse 2d and 3d image cues for monocular body pose estimation. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 3941–3950. [Google Scholar]
  22. Zhao, L.; Peng, X.; Tian, Y.; Kapadia, M.; Metaxas, D.N. Semantic graph convolutional networks for 3d human pose regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3425–3435. [Google Scholar]
  23. Zhou, X.; Huang, Q.; Sun, X.; Xue, X.; Wei, Y. Towards 3d human pose estimation in the wild: A weakly-supervised approach. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 398–407. [Google Scholar]
  24. Tome, D.; Russell, C.; Agapito, L. Lifting from the deep: Convolutional 3d pose estimation from a single image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2500–2509. [Google Scholar]
  25. Pavlakos, G.; Zhou, X.; Derpanis, K.G.; Daniilidis, K. Coarse-to-fine volumetric prediction for single-image 3D human pose. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 7025–7034. [Google Scholar]
  26. Dabral, R.; Mundhada, A.; Kusupati, U.; Afaque, S.; Sharma, A.; Jain, A. Learning 3d human pose from structure and motion. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 668–683. [Google Scholar]
  27. Zeng, A.; Sun, X.; Huang, F.; Liu, M.; Xu, Q.; Lin, S. Srnet: Improving generalization in 3d human pose estimation with a split-and-recombine approach. In Proceedings of the Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020; Proceedings, Part XIV 16. Springer: Berlin/Heidelberg, Germany, 2020; pp. 507–523. [Google Scholar]
  28. Liu, R.; Shen, J.; Wang, H.; Chen, C.; Cheung, S.c.; Asari, V. Attention mechanism exploits temporal contexts: Real-time 3d human pose reconstruction. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5064–5073. [Google Scholar]
  29. Zhou, L.; Chen, Y.; Wang, J. Dual-Path Transformer for 3D Human Pose Estimation. IEEE Trans. Circuits Syst. Video Technol. 2023, 1. [Google Scholar] [CrossRef]
  30. Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
  31. Sohn, K.; Yan, X.; Lee, H. Learning Structured Output Representation Using Deep Conditional Generative Models. In Proceedings of the 28th International Conference on Neural Information Processing Systems, Sanur, Indonesia, 8–12 December 2021; MIT Press: Cambridge, MA, USA, 2015. NIPS’15. Volume 2, pp. 3483–3491. [Google Scholar]
  32. Rezende, D.J.; Mohamed, S. Variational Inference with Normalizing Flows. arXiv 2016, arXiv:1505.05770. [Google Scholar]
  33. Barsoum, E.; Kender, J.R.; Liu, Z. HP-GAN: Probabilistic 3D human motion prediction via GAN. arXiv 2017, arXiv:1711.09561. [Google Scholar]
  34. Sharma, S.; Varigonda, P.T.; Bindal, P.; Sharma, A.; Jain, A. Monocular 3D Human Pose Estimation by Generation and Ordinal Ranking. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  35. Wehrbein, T.; Rudolph, M.; Rosenhahn, B.; Wandt, B. Probabilistic Monocular 3D Human Pose Estimation With Normalizing Flows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Virtual, 11–17 October 2021; pp. 11199–11208. [Google Scholar]
  36. Holmquist, K.; Wandt, B. Diffpose: Multi-hypothesis human pose estimation using diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Paris, France, 2–3 October 2023; pp. 15977–15987. [Google Scholar]
  37. Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. arXiv 2016, arXiv:1609.02907. [Google Scholar]
  38. Norris, J.R. Markov Chains; Number 2; Cambridge University Press: Cambridge, UK, 1998. [Google Scholar]
  39. Nichol, A.; Dhariwal, P. Improved Denoising Diffusion Probabilistic Models. arXiv 2021, arXiv:2102.09672. [Google Scholar]
  40. Song, J.; Meng, C.; Ermon, S. Denoising Diffusion Implicit Models. In Proceedings of the International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
  41. Han, X.; Zhang, Z.; Ding, N.; Gu, Y.; Liu, X.; Huo, Y.; Qiu, J.; Yao, Y.; Zhang, A.; Zhang, L.; et al. Pre-trained models: Past, present and future. AI Open 2021, 2, 225–250. [Google Scholar] [CrossRef]
  42. Sun, K.; Xiao, B.; Liu, D.; Wang, J. Deep high-resolution representation learning for human pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, 15–20 June 2019; pp. 5693–5703. [Google Scholar]
  43. Ionescu, C.; Papava, D.; Olaru, V.; Sminchisescu, C. Human3.6M: Large Scale Datasets and Predictive Methods for 3D Human Sensing in Natural Environments. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 1325–1339. [Google Scholar] [CrossRef] [PubMed]
  44. Ionescu, C.; Li, F.; Sminchisescu, C. Latent Structured Models for Human Pose Estimation. In Proceedings of the International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011. [Google Scholar]
  45. Mehta, D.; Rhodin, H.; Casas, D.; Fua, P.; Sotnychenko, O.; Xu, W.; Theobalt, C. Monocular 3D Human Pose Estimation In The Wild Using Improved CNN Supervision. In Proceedings of the 3D Vision (3DV), 2017 Fifth International Conference on 3D Vision (3DV), Qingdao, China, 10–12 October 2017; IEEE: Piscataway, NJ, USA, 2017. [Google Scholar] [CrossRef]
  46. Shan, W.; Liu, Z.; Zhang, X.; Wang, S.; Ma, S.; Gao, W. P-stmo: Pre-trained spatial temporal many-to-one model for 3d human 545 pose estimation. In Proceedings of the European Conference on Computer Vision, Tel Aviv, Israel, 23–27 October 2022; Springer: Berlin/Heidelberg, Germany, 2022; pp. 461–478. [Google Scholar]
  47. Von Marcard, T.; Henschel, R.; Black, M.; Rosenhahn, B.; Pons-Moll, G. Recovering Accurate 3D Human Pose in The Wild Using IMUs and a Moving Camera. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  48. Qian, X.; Tang, Y.; Zhang, N.; Han, M.; Xiao, J.; Huang, M.C.; Lin, R.S. HSTFormer: Hierarchical Spatial-Temporal Transformers for 3D Human Pose Estimation. arXiv 2023, arXiv:2301.07322. [Google Scholar]
  49. Zhao, W.; Wang, W.; Tian, Y. Graformer: Graph-oriented transformer for 3d pose estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 20438–20447. [Google Scholar]
  50. Cai, Y.; Ge, L.; Liu, J.; Cai, J.; Cham, T.J.; Yuan, J.; Thalmann, N.M. Exploiting spatial-temporal relationships for 3d pose estimation 537 via graph convolutional networks. In Proceedings of the IEEE/CVF International Conference on Computer 538 Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 2272–2281. [Google Scholar]
  51. Shan, W.; Lu, H.; Wang, S.; Zhang, X.; Gao, W. Improving robustness and accuracy via relative information encoding in 3d 540 human pose estimation. In Proceedings of the 29th ACM International Conference on Multimedia, Virtual Event, China, 20–24 October 2021; pp. 3446–3454. [Google Scholar]
  52. Chen, T.; Fang, C.; Shen, X.; Zhu, Y.; Chen, Z.; Luo, J. Anatomy-aware 3d human pose estimation with bone-based pose decomposition. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 198–209. [Google Scholar] [CrossRef]
  53. Wang, J.; Yan, S.; Xiong, Y.; Lin, D. Motion Guided 3D Pose Estimation from Videos. arXiv 2020, arXiv:2004.13985. [Google Scholar]
  54. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library. In Advances in Neural Information Processing Systems 32; Neural Information Processing Systems Foundation, Inc.: La Jolla, CA, USA, 2019. [Google Scholar]
Figure 1. An overview of various conditioning structures. This framework comprises 4 distinct conditioning structures. (a) In the Simple 2D Pose Structure, prior to the diffusion process, x 3 D sampled from the Gaussian distribution and simple x 2 D are concatenated and used as the input. (b) The Non-Hierarchical Structure aims to improve upon the limited performance of simple x 2 D by extracting spatial-temporal context information. This information is then concatenated with x 3 D for use as input. (c) The Pre-Hierarchical structure aggregates x 3 D and the output E F 2 D from each layer, incorporating both low-level and high-level information via a fusion module. (d) The Post-Hierarchical structure is a denoising model that utilizes the holistic representation E 2 D . In this structure, x 3 D is projected to a higher dimension via linear projection. Next, It feeds spatial and temporal encoders with hierarchical conditioning aggregation. This process is repeated N times, ultimately converging to a refined 3D pose when t = 0.
Figure 1. An overview of various conditioning structures. This framework comprises 4 distinct conditioning structures. (a) In the Simple 2D Pose Structure, prior to the diffusion process, x 3 D sampled from the Gaussian distribution and simple x 2 D are concatenated and used as the input. (b) The Non-Hierarchical Structure aims to improve upon the limited performance of simple x 2 D by extracting spatial-temporal context information. This information is then concatenated with x 3 D for use as input. (c) The Pre-Hierarchical structure aggregates x 3 D and the output E F 2 D from each layer, incorporating both low-level and high-level information via a fusion module. (d) The Post-Hierarchical structure is a denoising model that utilizes the holistic representation E 2 D . In this structure, x 3 D is projected to a higher dimension via linear projection. Next, It feeds spatial and temporal encoders with hierarchical conditioning aggregation. This process is repeated N times, ultimately converging to a refined 3D pose when t = 0.
Sensors 24 00829 g001
Figure 2. Detailed architecture of the post-hierarchical structure method. Beginning with the condition model, the 2D pose with dimensions (243, 17, 2) undergoes a linear projection, transforming its dimensions to (243, 17, 512). After being processed multiple times through the spatial and temporal encoders, a dimensional transformation to (243, 17, 256) is carried out. This allows the resulting features to be integrated with each corresponding spatial and temporal encoder within the denoising model.
Figure 2. Detailed architecture of the post-hierarchical structure method. Beginning with the condition model, the 2D pose with dimensions (243, 17, 2) undergoes a linear projection, transforming its dimensions to (243, 17, 512). After being processed multiple times through the spatial and temporal encoders, a dimensional transformation to (243, 17, 256) is carried out. This allows the resulting features to be integrated with each corresponding spatial and temporal encoder within the denoising model.
Sensors 24 00829 g002
Figure 3. Comparison of convergence speeds across state-of-the-art models.
Figure 3. Comparison of convergence speeds across state-of-the-art models.
Sensors 24 00829 g003
Figure 4. Qualitative comparison of the proposed HDPose method with other SOTA approaches using the Human3.6M.
Figure 4. Qualitative comparison of the proposed HDPose method with other SOTA approaches using the Human3.6M.
Sensors 24 00829 g004
Figure 5. Ablation experiments based on different conditioning methods.
Figure 5. Ablation experiments based on different conditioning methods.
Sensors 24 00829 g005
Figure 6. Visualization of spatial and temporal self-attention matrix. Each row and column shows the index of the joint and frames. The left side indicates the number of joints. On the right side, the figure denotes the number of frames.
Figure 6. Visualization of spatial and temporal self-attention matrix. Each row and column shows the index of the joint and frames. The left side indicates the number of joints. On the right side, the figure denotes the number of frames.
Sensors 24 00829 g006
Figure 7. Ablation study of conditioning depth on a hierarchical structure.
Figure 7. Ablation study of conditioning depth on a hierarchical structure.
Sensors 24 00829 g007
Figure 8. Failure cases caused by multi-person overlapping and image cropping.
Figure 8. Failure cases caused by multi-person overlapping and image cropping.
Sensors 24 00829 g008
Table 1. Quantitative evaluation of 3D human pose estimation methods using the standard evaluation metrics MPJPE (mm) and PA-MPJPE(mm) on the Human3.6M dataset. Part of the data in Table 1 was referenced from the respective papers [13,16,36,48]. (‡)—using the diffusion method. Bold: best.
Table 1. Quantitative evaluation of 3D human pose estimation methods using the standard evaluation metrics MPJPE (mm) and PA-MPJPE(mm) on the Human3.6M dataset. Part of the data in Table 1 was referenced from the respective papers [13,16,36,48]. (‡)—using the diffusion method. Bold: best.
Protocol 1 (MPJPE)Dir.DiscEatGreetPhonePhotoPosePurch.SitSitD.SmokeWaitWalkD.WalkWalkT.Avg.
Zhao et al. [49]45.250.848.050.054.965.048.247.160.270.051.648.754.139.743.151.8
Cai et al. [50] (N = 7)44.647.445.648.850.859.047.243.957.961.949.746.651.337.139.448.8
Pavllo et al. [11] (N = 243)45.246.743.345.648.155.144.644.357.365.847.144.049.032.833.946.8
Liu et al. [28] (N = 243)41.343.944.042.248.057.142.243.257.361.347.043.547.032.631.845.1
Zeng [27]46.647.143.941.645.849.646.540.053.461.146.142.643.131.532.644.8
Shan et al. [51] (N = 243)40.844.541.442.746.355.641.841.953.760.845.041.544.830.831.944.3
Zheng et al. [7] (N = 81)41.544.839.842.546.551.642.142.053.360.745.543.346.131.832.244.3
Chen et al. [52] (N = 243)41.443.240.142.946.651.941.742.353.960.245.441.746.031.532.744.1
Li et al. [8] (N = 351)39.243.140.140.944.951.240.641.353.560.343.741.143.829.830.643.0
Shan et al. [46] (N = 243)38.942.740.441.145.649.740.939.955.559.444.942.242.729.429.442.8
Zhang et al. [16] (N = 243)37.940.737.839.642.350.239.939.951.655.642.139.940.827.928.040.9
Choi et al. [13] (H = 10) ‡43.450.745.450.249.653.448.645.056.970.747.848.251.343.143.449.4
Holmquist et al. [36] (H = 200) ‡38.143.135.343.146.648.239.037.651.959.341.747.645.437.436.043.3
Ours (N = 243, H = 1)37.840.737.739.642.450.239.840.251.855.842.239.841.027.928.141.0
Protocol 2 (PA-MPJPE)Dir.DiscEatGreetPhonePhotoPosePurch.SitSitD.SmokeWaitWalkD.WalkWalkT.Avg.
Cai et al. [50] (N = 7)35.737.836.940.739.645.237.434.546.950.140.536.141.029.633.239.0
Liu et al. [28] (N = 243)32.335.233.335.835.941.533.232.744.650.937.032.437.025.227.235.6
Zheng et al. [7] (N = 81)32.534.832.634.635.339.532.132.042.848.534.832.435.324.526.034.6
Chen et al. [52] (N = 243)32.635.132.835.436.340.432.432.342.749.036.832.436.024.926.535.0
Li et al. [8]31.534.932.833.635.339.632.032.243.548.736.432.634.323.925.134.4
Shan et al. [46] (N = 243)31.335.232.933.935.439.332.531.544.648.236.332.934.423.823.934.4
Zhang et al. [16] (N = 243)30.833.130.331.833.139.131.130.542.544.534.030.832.722.122.932.6
Choi et al. [13] (H = 10) ‡35.940.336.741.439.843.437.135.546.259.739.938.041.932.934.239.9
Holmquist et al. [36] (H = 200) ‡27.931.429.730.234.937.127.328.239.046.134.232.333.626.127.532.4
Ours (N = 243, H = 1)31.033.230.631.933.239.231.130.742.545.034.130.732.522.023.032.8
Table 2. Quantitative evaluation of 3D human pose estimation methods using the evaluation metrics PCK, AUC, MPJPE (mm) on the MPI-INF-3DHP dataset. Part of the data in Table 2 was referenced from the respective papers [7,46,48]. Bold: best.
Table 2. Quantitative evaluation of 3D human pose estimation methods using the evaluation metrics PCK, AUC, MPJPE (mm) on the MPI-INF-3DHP dataset. Part of the data in Table 2 was referenced from the respective papers [7,46,48]. Bold: best.
Method PCK↑AUC↑MPJPE↓
Pavllo et al. [11] (T = 243)CVPR’1985.551.584.8
Wang et al. [53] (T = 96)ECCV’2086.962.168.1
Chen et al. [52] (T = 25)TCSVT’2187.954.079.1
Liu et al. [8] (T = 9)CVPR’2293.863.358.0
Zhang et al. [16] (T = 243)CVPR’2296.975.835.4
Ours (T = 243)Ours96.575.635.7
Table 3. Analysis on computational complexity. Part of the data in Table 3 was referenced from the respective papers [16,46].
Table 3. Analysis on computational complexity. Part of the data in Table 3 was referenced from the respective papers [16,46].
MethodMPJPEParams (M)FLOPsFPS
Zheng et al. [7]44.49.51358269
Shan et al. [46]42.86.717373040
Zhang et al. [16]40.933.66454547
Ours41.05.078.54054
Table 4. Ablation study computational complexity according to embedding dimension. Embedding dimension of condition ( d c ), embedding dimension of post-hierarchical model ( d d ).
Table 4. Ablation study computational complexity according to embedding dimension. Embedding dimension of condition ( d c ), embedding dimension of post-hierarchical model ( d d ).
d c d d MPJPEParams (M)FLOPs
2566447.60.35.0
25612844.41.219.7
25625641.05.078.5
25651241.719.9313.0
Table 5. Analysis of performance based on hierarchical aggregation of denoising models.
Table 5. Analysis of performance based on hierarchical aggregation of denoising models.
Hiearchical AggregationSpatialTemporalMPJPE
Spatial only×45.9
Temporal only×45.6
Spatial and temporal41.0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, D.; Kim, J. HDPose: Post-Hierarchical Diffusion with Conditioning for 3D Human Pose Estimation. Sensors 2024, 24, 829. https://doi.org/10.3390/s24030829

AMA Style

Lee D, Kim J. HDPose: Post-Hierarchical Diffusion with Conditioning for 3D Human Pose Estimation. Sensors. 2024; 24(3):829. https://doi.org/10.3390/s24030829

Chicago/Turabian Style

Lee, Donghoon, and Jaeho Kim. 2024. "HDPose: Post-Hierarchical Diffusion with Conditioning for 3D Human Pose Estimation" Sensors 24, no. 3: 829. https://doi.org/10.3390/s24030829

APA Style

Lee, D., & Kim, J. (2024). HDPose: Post-Hierarchical Diffusion with Conditioning for 3D Human Pose Estimation. Sensors, 24(3), 829. https://doi.org/10.3390/s24030829

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop