TSU-net: Two-stage multi-scale cascade and multi-field fusion U-net for right ventricular segmentation

Author(s):  
Xiuquan Du ◽  
Xiaofei Xu ◽  
Heng Liu ◽  
Shuo Li
Author(s):  
Yutong Yan ◽  
Pierre-Henri Conze ◽  
Gwenolé Quellec ◽  
Mathieu Lamard ◽  
Beatrice Cochener ◽  
...  

2021 ◽  
Vol 12 (4) ◽  
pp. 2204
Author(s):  
Jiaxuan Li ◽  
Peiyao Jin ◽  
Jianfeng Zhu ◽  
Haidong Zou ◽  
Xun Xu ◽  
...  
Keyword(s):  

2020 ◽  
Vol 34 (07) ◽  
pp. 11037-11044
Author(s):  
Lianghua Huang ◽  
Xin Zhao ◽  
Kaiqi Huang

A key capability of a long-term tracker is to search for targets in very large areas (typically the entire image) to handle possible target absences or tracking failures. However, currently there is a lack of such a strong baseline for global instance search. In this work, we aim to bridge this gap. Specifically, we propose GlobalTrack, a pure global instance search based tracker that makes no assumption on the temporal consistency of the target's positions and scales. GlobalTrack is developed based on two-stage object detectors, and it is able to perform full-image and multi-scale search of arbitrary instances with only a single query as the guide. We further propose a cross-query loss to improve the robustness of our approach against distractors. With no online learning, no punishment on position or scale changes, no scale smoothing and no trajectory refinement, our pure global instance search based tracker achieves comparable, sometimes much better performance on four large-scale tracking benchmarks (i.e., 52.1% AUC on LaSOT, 63.8% success rate on TLP, 60.3% MaxGM on OxUvA and 75.4% normalized precision on TrackingNet), compared to state-of-the-art approaches that typically require complex post-processing. More importantly, our tracker runs without cumulative errors, i.e., any type of temporary tracking failures will not affect its performance on future frames, making it ideal for long-term tracking. We hope this work will be a strong baseline for long-term tracking and will stimulate future works in this area.


2022 ◽  
Vol 14 (1) ◽  
pp. 27
Author(s):  
Junda Li ◽  
Chunxu Zhang ◽  
Bo Yang

Current two-stage object detectors extract the local visual features of Regions of Interest (RoIs) for object recognition and bounding-box regression. However, only using local visual features will lose global contextual dependencies, which are helpful to recognize objects with featureless appearances and restrain false detections. To tackle the problem, a simple framework, named Global Contextual Dependency Network (GCDN), is presented to enhance the classification ability of two-stage detectors. Our GCDN mainly consists of two components, Context Representation Module (CRM) and Context Dependency Module (CDM). Specifically, a CRM is proposed to construct multi-scale context representations. With CRM, contextual information can be fully explored at different scales. Moreover, the CDM is designed to capture global contextual dependencies. Our GCDN includes multiple CDMs. Each CDM utilizes local Region of Interest (RoI) features and single-scale context representation to generate single-scale contextual RoI features via the attention mechanism. Finally, the contextual RoI features generated by parallel CDMs independently are combined with the original RoI features to help classification. Experiments on MS-COCO 2017 benchmark dataset show that our approach brings continuous improvements for two-stage detectors.


Author(s):  
Y. Yan ◽  
P.-H. Conze ◽  
G. Quellec ◽  
M. Lamard ◽  
B. Cochener ◽  
...  

Micromachines ◽  
2021 ◽  
Vol 12 (12) ◽  
pp. 1458
Author(s):  
Yanpeng Sun ◽  
Zhanyou Chang ◽  
Yong Zhao ◽  
Zhengxu Hua ◽  
Sirui Li

At night, visual quality is reduced due to insufficient illumination so that it is difficult to conduct high-level visual tasks effectively. Existing image enhancement methods only focus on brightness improvement, however, improving image quality in low-light environments still remains a challenging task. In order to overcome the limitations of existing enhancement algorithms with insufficient enhancement, a progressive two-stage image enhancement network is proposed in this paper. The low-light image enhancement problem is innovatively divided into two stages. The first stage of the network extracts the multi-scale features of the image through an encoder and decoder structure. The second stage of the network refines the results after enhancement to further improve output brightness. Experimental results and data analysis show that our method can achieve state-of-the-art performance on synthetic and real data sets, with both subjective and objective capability superior to other approaches.


Electronics ◽  
2021 ◽  
Vol 10 (19) ◽  
pp. 2434
Author(s):  
Jucheng Yang ◽  
Feng Wei ◽  
Yaxin Bai ◽  
Meiran Zuo ◽  
Xiao Sun ◽  
...  

Convolutional neural networks and the per-pixel loss function have shown their potential to be the best combination for super-resolving severely degraded images. However, there are still challenges, such as the massive number of parameters requiring prohibitive memory and vast computing and storage resources as well as time-consuming training and testing. What is more, the per-pixel loss measured by L2 and the Peak Signal-to-Noise Ratio do not correlate well with human perception of image quality, since L2 simply does not capture the intricate characteristics of human visual systems. To address these issues, we propose an effective two-stage hourglass network with multi-task co-optimization, which enables the entire network to focus on training and testing time and inherent image patterns such as local luminance, contrast, structure and data distribution. Moreover, to avoid overwhelming memory overheads, our model is capable of performing real-time single image multi-scale super-resolution, so it is memory-friendly, meaning that memory space is utilized efficiently. In addition, in order to best use the underlying structure and perception of image quality and the intermediate estimates during the inference process, we introduce a cross-scale training strategy with 2×, 3× and 4× image super-resolution. This effective multi-task two-stage network with the cross-scale strategy for multi-scale image super-resolution is named EMTCM. Quantitative and qualitative experiment results show that the proposed EMTCM network outperforms state-of-the-art methods in recovering high-quality images.


2021 ◽  
Vol 422 ◽  
pp. 34-50
Author(s):  
Houzhang Fang ◽  
Mingjiang Xia ◽  
Hehui Liu ◽  
Yi Chang ◽  
Liming Wang ◽  
...  

Sign in / Sign up

Export Citation Format

Share Document