no code implementations • ECCV 2020 • Yanda Meng, Wei Meng, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Xiaowei Huang, Yalin Zheng
In particular, thanks to the proposed aggregation GCN, our network benefits from direct feature learning of the instances’ boundary locations and the spatial information propagation across the image.
1 code implementation • 27 Apr 2024 • Damith Chamalke Senadeera, Xiaoyun Yang, Dimitrios Kollias, Gregory Slabaugh
In this paper we introduce CUE-Net, a novel architecture designed for automated violence detection in video surveillance.
2 code implementations • CVPR 2022 • Hongrun Zhang, Yanda Meng, Yitian Zhao, Yihong Qiao, Xiaoyun Yang, Sarah E. Coupland, Yalin Zheng
Multiple instance learning (MIL) has been increasingly used in the classification of histopathology whole slide images (WSIs).
no code implementations • 9 Mar 2022 • Yanda Meng, Xu Chen, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Yihong Qiao, Xiaowei Huang, Yalin Zheng
In this paper, we propose a novel multi-level aggregation network to regress the coordinates of the vertices of a 3D face from a single 2D image in an end-to-end manner.
1 code implementation • 8 Mar 2022 • Yanda Meng, Joshua Bridge, Meng Wei, Yitian Zhao, Yihong Qiao, Xiaoyun Yang, Xiaowei Huang, Yalin Zheng
This paper proposes an adaptive auxiliary task learning based approach for object counting problems.
1 code implementation • 27 Oct 2021 • Yanda Meng, Hongrun Zhang, Dongxu Gao, Yitian Zhao, Xiaoyun Yang, Xuesheng Qian, Xiaowei Huang, Yalin Zheng
Our model is well-suited to obtain global semantic region information while also accommodates local spatial boundary characteristics simultaneously.
1 code implementation • ICCV 2021 • Kenan Dai, Jie Zhao, Lijun Wang, Dong Wang, Jianhua Li, Huchuan Lu, Xuesheng Qian, Xiaoyun Yang
Deep learning based visual trackers entail offline pre-training on large volumes of video datasets with accurate bounding box annotations that are labor-expensive to achieve.
1 code implementation • ICCV 2021 • Yanda Meng, Hongrun Zhang, Yitian Zhao, Xiaoyun Yang, Xuesheng Qian, Xiaowei Huang, Yalin Zheng
Semi-supervised approaches for crowd counting attract attention, as the fully supervised paradigm is expensive and laborious due to its request for a large number of images of dense crowd scenarios and their annotations.
no code implementations • 5 Apr 2021 • Xuehu Liu, Pingping Zhang, Chenyang Yu, Huchuan Lu, Xuesheng Qian, Xiaoyun Yang
To capture richer perceptions and extract more comprehensive video representations, in this paper we propose a novel framework named Trigeminal Transformers (TMT) for video-based person Re-ID.
1 code implementation • CVPR 2021 • Xin Chen, Bin Yan, Jiawen Zhu, Dong Wang, Xiaoyun Yang, Huchuan Lu
The correlation operation is a simple fusion manner to consider the similarity between the template and the search region.
Ranked #5 on Visual Tracking on TNL2K
1 code implementation • CVPR 2021 • Xuehu Liu, Pingping Zhang, Chenyang Yu, Huchuan Lu, Xiaoyun Yang
Specifically, we first propose a Global-guided Correlation Estimation (GCE) to generate feature correlation maps of local features and global features, which help to localize the high- and low-correlation regions for identifying the same person.
1 code implementation • CVPR 2021 • Bin Yan, Xinyu Zhang, Dong Wang, Huchuan Lu, Xiaoyun Yang
Many recent trackers adopt the multiple-stage tracking strategy to improve the quality of bounding box estimation.
Ranked #15 on Semi-Supervised Video Object Segmentation on VOT2020
Semi-Supervised Video Object Segmentation Visual Object Tracking
no code implementations • 4 Jul 2020 • Pengyu Zhang, Jie Zhao, Dong Wang, Huchuan Lu, Xiaoyun Yang
In this study, we propose a novel RGB-T tracking framework by jointly modeling both appearance and motion cues.
Ranked #6 on Rgb-T Tracking on GTOT
1 code implementation • 4 Jul 2020 • Bin Yan, Dong Wang, Huchuan Lu, Xiaoyun Yang
In recent years, the multiple-stage strategy has become a popular trend for visual tracking.
2 code implementations • CVPR 2020 • Kenan Dai, Yunhua Zhang, Dong Wang, Jianhua Li, Huchuan Lu, Xiaoyun Yang
Most top-ranked long-term trackers adopt the offline-trained Siamese architectures, thus, they cannot benefit from great progress of short-term trackers with online update.
Ranked #11 on Visual Object Tracking on LaSOT-ext
1 code implementation • CVPR 2020 • Bin Yan, Dong Wang, Huchuan Lu, Xiaoyun Yang
An effective and efficient perturbation generator is trained with a carefully designed adversarial loss, which can simultaneously cool hot regions where the target exists on the heatmaps and force the predicted bounding box to shrink, making the tracked target invisible to trackers.
2 code implementations • ICCV 2019 • Peixia Li, Bo-Yu Chen, Wanli Ouyang, Dong Wang, Xiaoyun Yang, Huchuan Lu
In this work, we propose a novel gradient-guided network to exploit the discriminative information in gradients and update the template in the siamese network through feed-forward and backward operations.
Ranked #3 on Visual Object Tracking on OTB-2015 (Precision metric)
1 code implementation • ICCV 2019 • Bin Yan, Haojie Zhao, Dong Wang, Huchuan Lu, Xiaoyun Yang
In this work, we present a novel robust and real-time long-term tracking framework based on the proposed skimming and perusal modules.
no code implementations • ICCV 2019 • Pingping Zhang, Wei Liu, Yinjie Lei, Huchuan Lu, Xiaoyun Yang
To address these issues, in this work we propose a novel deep learning framework, named Cascaded Context Pyramid Network (CCPNet), to jointly infer the occupancy and semantic labels of a volumetric 3D scene from a single depth image.
Ranked #5 on 3D Semantic Scene Completion on NYUv2 (using extra training data)