Listvedel7026
5% higher than that of Region Proposal Network (RPN), surpassing all the existing region proposal approaches. We also integrate SC-RPN into Fast R-CNN and Faster R-CNN to test its effectiveness on object detection task, the experimental results achieve a gain of 3.2% and 3.8% in terms of mAP compared to the original ones.Human attention is an interactive activity between our visual system and our brain, using both low-level visual stimulus and high-level semantic information. Previous image salient object detection (SOD) studies conduct their saliency predictions via a multitask methodology in which pixelwise saliency regression and segmentation-like saliency refinement are conducted simultaneously. However, this multitask methodology has one critical limitation the semantic information embedded in feature backbones might be degenerated during the training process. Our visual attention is determined mainly by semantic information, which is evidenced by our tendency to pay more attention to semantically salient regions even if these regions are not the most perceptually salient at first glance. This fact clearly contradicts the widely used multitask methodology mentioned above. To address this issue, this paper divides the SOD problem into two sequential steps. First, we devise a lightweight, weakly supervised deep network to coarsely locate the semantically salient regions. Next, as a postprocessing refinement, we selectively fuse multiple off-the-shelf deep models on the semantically salient regions identified by the previous step to formulate a pixelwise saliency map. Compared with the state-of-the-art (SOTA) models that focus on learning the pixelwise saliency in single images using only perceptual clues, our method aims at investigating the object-level semantic ranks between multiple images, of which the methodology is more consistent with the human attention mechanism. Our method is simple yet effective, and it is the first attempt to consider salient object detection as mainly an object-level semantic reranking problem.This paper describes a high-resolution 3D navigation and tracking system using magnetic field gradients, that can replace X-Ray fluoroscopy in high-precision surgeries. Monotonically varying magnetic fields in X, Y and Z directions are created in the field-of-view (FOV) to produce magnetic field gradients, which encode each spatial point uniquely. Highly miniaturized, wireless and battery-less devices, capable of measuring their local magnetic field, are designed to sense the gradient field. One such device can be attached to an implant inside the body and another to a surgical tool, such that both can simultaneously measure and communicate the magnetic field at their respective locations to an external receiver. The relative location of the two devices on a real-time display can enable precise surgical navigation without using X-Rays. A prototype device is designed consisting of a micro-chip fabricated in 65nm CMOS technology, a 3D magnetic sensor and an inductor-coil. Planar electromagnetic coils are designed for creating the 3D magnetic field gradients in a 20x20x10cm3 of scalable FOV. Unambiguous and orientation-independent spatial encoding is achieved by (i) using the gradient in the total field magnitude instead of only the Z-component; and (ii) using a combination of the gradient fields to correct for the non-linearity and non-monotonicity in X and Y gradients. The resultant X and Y FOV yield ≥90% utilization of their respective coil-span. The system is tested in vitro to demonstrate a localization accuracy of less then 100μm in 3D, the highest reported to the best of our knowledge.Instance level detection and segmentation of thoracic diseases or abnormalities are crucial for automatic diagnosis in chest X-ray images. selleck products Leveraging on constant structure and disease relations extracted from domain knowledge, we propose a structure-aware relation network (SAR-Net) extending Mask R-CNN. The SAR-Net consists of three relation modules 1. the anatomical structure relation module encoding spatial relations between diseases and anatomical parts. 2. the contextual relation module aggregating clues based on query-key pair of disease RoI and lung fields. 3. the disease relation module propagating co-occurrence and causal relations into disease proposals. Towards making a practical system, we also provide ChestX-Det, a chest X-Ray dataset with instance-level annotations (boxes and masks). ChestX-Det is a subset of the public dataset NIH ChestX-ray14. It contains ~ 3500 images of 13 common disease categories labeled by three board-certified radiologists. We evaluate our SAR-Net on it and another dataset DR-Private. Experimental results show that it can enhance the strong baseline of Mask R-CNN with significant improvements. The ChestX-Det is released at https//github.com/Deepwise-AILab/ChestX-Det-Dataset.Landmark correspondences are a widely used type of gold standard in image registration. However, the manual placement of corresponding points is subject to high inter-user variability in the chosen annotated locations and in the interpretation of visual ambiguities. In this paper, we introduce a principled strategy for the construction of a gold standard in deformable registration. Our framework (i) iteratively suggests the most informative location to annotate next, taking into account its redundancy with previous annotations; (ii) extends traditional pointwise annotations by accounting for the spatial uncertainty of each annotation, which can either be directly specified by the user, or aggregated from pointwise annotations from multiple experts; and (iii) naturally provides a new strategy for the evaluation of deformable registration algorithms. Our approach is validated on four different registration tasks. The experimental results show the efficacy of suggesting annotations according to their informativeness, and an improved capacity to assess the quality of the outputs of registration algorithms. In addition, our approach yields, from sparse annotations only, a dense visualization of the errors made by a registration method. The source code of our approach supporting both 2D and 3D data is publicly available at https//github.com/LoicPeter/evaluation-deformable-registration.