Neural Mesh Refiner for 6-DoF Pose Estimation
How can we effectively utilise the 2D monocular image information for recovering the 6D pose (6-DoF) of the visual objects? Deep learning has shown to be effective for robust and real-time monocular pose estimation. Oftentimes, the network learns to regress the 6-DoF pose using a naive loss function. However, due to a lack of geometrical scene understanding from the directly regressed pose estimation, there are misalignments between the rendered mesh from the 3D object and the 2D instance segmentation result, e.g., bounding boxes and masks prediction. This paper bridges the gap between 2D mask generation and 3D location prediction via a differentiable neural mesh renderer. We utilise the overlay between the accurate mask prediction and less accurate mesh prediction to iteratively optimise the direct regressed 6D pose information with a focus on translation estimation. By leveraging geometry, we demonstrate that our technique significantly improves direct regression performance on the difficult task of translation estimation and achieve the state of the art results on Peking University/Baidu - Autonomous Driving dataset and the ApolloScape 3D Car Instance dataset.
Di Wu, Yihao Chen, Xianbiao Qi, Yongjian Jian, Weixuan Chen, Rong Xiao

arXiv:2003.07561, 2020


6D-VNet: End-to-end 6-DoF Vehicle Pose Estimation from Monocular RGB Images
We present a conceptually simple framework for 6DoF object pose estimation, especially for autonomous driving scenario. Our approach efficiently detects traffic participants in a monocular RGB image while simultaneously regressing their 3D translation and rotation vectors. The method, called 6D-VNet, extends Mask R-CNN by adding customised heads for predicting vehicle's finer class, rotation and translation. The proposed 6D-VNet is trained end-to-end compared to previous methods. Furthermore, we show that the inclusion of translational regression in the joint losses is crucial for the 6DoF pose estimation task, where object translation distance along longitudinal axis varies significantly, e.g., in autonomous driving scenarios. Additionally, we incorporate the mutual information between traffic participants via a modified non-local block. As opposed to the original non-local block implementation, the proposed weighting modification takes the spatial neighbouring information into consideration whilst counteracting the effect of extreme gradient values.
Di Wu, Zhaoyong Zhuang, Xiangqun Can, Wenbin Zou, Xia Li

IEEE Conference on Computer Vision and Pattern Recognition workshop, 2019


Kernalised Multi-resolution Convnet for Visual Tracking
Kernelized Correlation Filters (KCF) have demonstrated excellent performance for high-speed visual object tracking. The key to their success is by observing that the resulting data matrix of translated patches is circulant, the cyclic shifts could be diagonalised with the Discrete Fourier Transform, reducing both storage and computation to obtain the next frame response map. Built upon their seminar work, there has been a plethora of recent improvements relying on convolutional neural network (CNN) pretrained on ImageNet as a feature extractor for visual tracking. However, most of their works replying on ad hoc analysis to design the weights for different layers either using boosting or hedging techniques as an ensemble tracker. In this paper, we go beyond the conventional KCF framework and introduce a novel formulation for training from the hierarchical response maps to directly output the target movement. We interpret the response maps from the hierarchies of convolutional layers as a nonlinear counter part of an image pyramid representation and employ a learning framework for mapping the hierarchical outputs to target translation. Our proposed formulation enables efficient integration of multi-resolution deep feature maps trained from massive online image data and demonstrates its competitiveness over the corresponding counterpart baselines.
Di Wu, Wenbin Zou, Xia Li

IEEE Conference on Computer Vision and Pattern Recognition workshop, 2017


Deep Dynamic Neural Networks for Multimodal Gesture Segmentation and Recognition
This paper describes a novel method called Deep Dynamic Neural Networks (DDNN) for multimodal gesture recognition. A semi-supervised hierarchical dynamic framework based on a Hidden Markov Model (HMM) is proposed for simultaneous gesture segmentation and recognition where skeleton joint information, depth and RGB images, are the multimodal input observations. Unlike most traditional approaches that rely on the construction of complex handcrafted features, our approach learns high-level spatio-temporal representations using deep neural networks suited to the input modality: a Gaussian-Bernouilli Deep Belief Network (DBN) to handle skeletal dynamics, and a 3D Convolutional Neural Network (3DCNN) to manage and fuse batches of depth and RGB images. This is achieved through the modeling and learning of the emission probabilities of the HMM required to infer the gesture sequence. This purely data driven approach achieves a Jaccard index score of 0.81 in the ChaLearn LAP gesture spotting challenge. The performance is on par with a variety of state-of-the-art hand-tuned feature-based approaches and other learning-based methods, therefore opening the door to the use of deep learning techniques in order to further explore multimodal time series data.
Di Wu, Lionel Pigou, Pieter-Jan Kindermans, Nam Le, Ling Shao, Joni Dambre, and Jean-Marc Odobez

The IEEE Transactions on Pattern Analysis and Machine Intelligence (TPAMI) 2016

DOI: 10.1109/TPAMI.2016.2537340


EUMSSI team at the MediaEval Person Discovery Challenge
We present the results of the EUMSSI team's participation in the Multimodal Person Discovery task at the MediaEval challenge 2015. Our system ranks the first amongst other 9 international research teams. The goal is to identify all people who simultaneously appear and speak in a video corpus, which implicitly involves both audio stream and visual stream. We emphasize on improving each modality separately and benchmarking them to analyze their pros and cons.
Nam Le, Di Wu, Sylvain Meignier and Jean-Marc Odobze

Proceedings of the MediaEval 2015 Workshop, Wurzen, Germany

Challenge Webpage


Deep Dynamic Neural Networks for Gesture Segmentation and Recognition
The purpose of this paper is to describe a novel Deep Dynamic Neural Networks(DDNN) for the track 3 of the Chalearn Looking at People 2014 challenge. A generalised semi-supervised hierarchical dynamic framework is proposed for simultaneous gesture segmentation and recognition taking both skeleton and depth images as input modules. First, Deep Belief Networks(DBN) and 3D Convolutional Neural Networks (3DCNN) are adopted for skeletal and depth module accordingly to extract high level spatio-temporal features. Then the learned representations are used for estimating emission probabilities of the Hidden Markov Models to infer action sequence. The framework can be easily extended by including an ergodic state to segment and recognise video sequence in a frame-to-frame mechanism, rendering it possible for online segmentation and recognition for diverse input modules. Some normalisation details pertaining to preprocessing raw features are also discussed. This purely data-driven approach achieves 0.8162 score in this gesture spotting challenge. The performance is on par with a variety of the state-of-the-art hand-tuned-feature approaches and other learning-based methods, opening the doors for using deep learning techniques to explore time series multimodal data.
D. Wu and L. Shao in ECCV, workshop 2014, Zurich

Github Project , Challenge Webpage


Multimodal Dynamic Networks for Gesture Recognition
Multimodal input is a real-world situation in gesture recognition applications such as sign language recognition. In this paper, we propose a novel bi-modal (audio and skeleton joints) dynamic network for gesture recognition. First, state-of-the-art dynamic Deep Belief Networks are deployed to extract high level audio and skeletal joints representations. Then, instead of traditional late fusion, we adopt another layer of perceptron for cross modality learning taking the input from each individual net's penultimate layer. Finally, to account for temporal dynamics, the learned shared representations are used for estimating the emission probability to infer action sequences. In particular, we demonstrate that multimodal feature learning will extract semantically meaningful shared representations, outperforming individual modalities, and the early fusion scheme's efficacy against the traditional method of late fusion.
D. Wu and L. Shao in ACM Multimedia 2014, Orlando

Leveraging Hierarchical Parametric Network for Skeletal Joints Action Segmentation and Recognition
Over the last few years, with the immense popularity of the Kinect, there has been renewed interest in developing methods for human gesture and action recognition from 3D skeletal data. A number of approaches have been proposed to extract representative features from 3D skeletal data, most commonly hard wired geometric or bio-inspired shape context features. We propose a hierarchial dynamic framework that first extracts high level skeletal joints features and then uses the learned representation for estimating emission probability to infer action sequence. Currently gaussian mixture models are the dominant technique for modeling the emission distribution of hidden Markov models. We show that better action recognition using skeletal features can be achieved by replacing gaussian mixture models by deep neural networks that contain many layers of features to predict probability distributions over states of hidden Markov models. The framework can be easily extended to include a ergodic state to segment and recognize actions simultaneously.
D. Wu and L. Shao in CVPR 2014

PDF, Video Spotlight, Github Project


Learning Deep and Wide: A Spectral Method for Learning Deep Networks
Building intelligent systems that are capable of extracting high-level representations from high-dimensional sensory data lies at the core of solving many computer vision related tasks. We propose the Multi-Spectral Neural Networks (MSNN) to learn features from multi-column Deep Neural Networks and embed the penultimate hierarchical discriminative manifolds into a compact representation. The low dimensional embedding explores the complementary property of different views wherein the distribution of each view is sufficiently smooth and hence achieves robustness given few labeled training data. Our experiments show that spectrally embedding several deep neural networks can explore the optimum output from the multi-column networks and consistently decrease the error rate compared to a single deep network.
L. Shao, D. Wu and X. Li in IEEE Transactions on Neural Networks and Learning Systems 2014 (formerly known as IEEE Transactions on Neural Networks)

Multi-Max-Margin Support Vector Machine for Multi-Source Human Action Recognition
We propose a new ensemble-based classifier for multi-source human action recognition called Multi-Max-Margin Support Vector Machine (MMM-SVM). This ensemble method incorporates the decision values of multiple sources and makes an informed final prediction by merging multi-source feature’s intrinsic decision strength. Experiments performed on the benchmark IXMAS multi-view dataset demonstrate that the performance of our multiview system can further improve the accuracy over single view by 3%-13% and consistently outperform the direct-concatenation method. We further apply this ensemble technique for combining the decision values of contextual and motion information in the UCF Sports dataset and the results are comparable to the state-of-the-art, which exhibits our algorithm’s potential for further extension in other areas of feature fusion problems.
D. Wu, L. Shao in Neurocomputing 2013

One Shot Learning Gesture Recognition with Kinect Sensor
Gestures are both natural and intuitive for Human-Computer-Interaction (HCI) and the one-shot learning scenario is one of the real world situations in terms of gesture recognition problems. In this demo, we present a hand gesture recognition system using the Kinect sensor, which addresses the problem of one-shot learning gesture recognition with a user defined training and testing system. Such a system can behave like a remote control where the user can allocate a specific function using a prefered gesture by performing it only once. To adopt the gesture recognition framework, the system first automatically segments an action sequence into atomic tokens, and then adopts the Extended-Motion-History-Image (Extended-MHI) for motion feature representation. We evaluate the performance of our system quantitatively in Chalearn Gesture Challenge, and apply it to a virtual one shot learning gesture recognition system.
D. Wu, F. Zhu, L. Shao, H. Zhang in In ACM Multimedia 2012

One shot learning gesture recognition from RGBD images
We present a system to classify the gesture from only one learning example. The inputs are duo-modality, i.e. RGB and depth sensor from Kinect. Our system performs morphological denoising on depth images and automatically segments the temporal boundaries. Features are extracted based on Extended-Motion-History-Image(Extended-MHI) and the Multi-view Spectral Embedding (MSE) algorithm is used to fuse duo modalities in a physically meaningful manner. Our approach achieves less than 0.3 in Levenshtein distance in CHALEARN Gesture Challenge validation batches.
D. Wu, F. Zhu, and L. Shao in In CVPR workshop on gesture recognition 2012

Silhouette Analysis Based Action Recognition via Exploiting Human Poses
In this paper, we propose a novel scheme for human action recognition that combines the advantages of both local and global representations. We explore human silhouettes for human action representation by taking into account the correlation between sequential poses in an action. A modified bag-of-words model, named bag-of-correlated-poses (BoCP), is introduced to encode temporally local features of actions. To utilize the property of visual word ambiguity, we adopt the soft assignment strategy to reduce the dimensionality of our model and circumvent the penalty of computational complexity and quantization error. To compensate for the loss of structural information, we propose an extended motion template, i.e., extensions of the motion history image (extended-MHI) to capture the holistic structural features. The proposed scheme takes advantages of local and global features and therefore provides a discriminative representation for human actions. Experimental results prove the viability of the complimentary properties of two descriptors and the proposed approach outperforms the state-of-the-art methods on the IXMAS action recognition dataset.
D. Wu and L.Shao in In IEEE Transactions on Circuits and Systems for Video Technology 2013

Action Recognition using Correlogram of Body Poses and Spectral Regression
Human action recognition is an important topic in computer vision with its applications range from robotics, video surveillance, human-computer interaction, user interface design, multi-media video retrieval and among others. This paper proposed a novel representation for human action using Correlogram of Body Poses (CBP) which is a descendant of our previous work on Histogram of Body Poses (HBP). In addition, the paper explores a recently proposed subspace learning method called Spectral Regression Discriminant Analysis (SRDA) which aims to speed up the traditional subspace learning methods. Results show that this new representation of human action achieves a rather satisfactory rate and be compared to most state-of-art methods.
L Shao, D Wu, X Chen in ICIP2010

This document was last modified on: