This paper presents a high-fidelity inhomogeneous ground clutter simulation method for airborne phased array Pulse Doppler (PD) radar aided by a digital elevation model (DEM) and digital land classification data (DLCD). The method starts by extracting the basic geographic information of the Earth's surface scattering points from the DEM data, then reads the Earth's surface classification codes of Earth's surface scattering points according to the DLCD. After determining the landform types, different backscattering coefficient models are selected to calculate the backscattering coefficient of each Earth surface scattering point. Finally, the high-fidelity inhomogeneous ground clutter simulation of airborne phased array PD radar is realized based on the Ward model. The simulation results show that the classifications of landform types obtained by the proposed method are more abundant, and the ground clutter simulated by different backscattering coefficient models is more real and effective.
Detecting novel anomalies in medical imaging is challenging due to the limited availability of labeled data for rare abnormalities, which often display high variability and subtlety. This challenge is further compounded when small abnormal regions are embedded within larger normal areas, as whole-image predictions frequently overlook these subtle deviations. To address these issues, we propose an unsupervised Patch-GAN framework designed to detect and localize anomalies by capturing both local detail and global structure. Our framework first reconstructs masked images to learn fine-grained, normal-specific features, allowing for enhanced sensitivity to minor deviations from normality. By dividing these reconstructed images into patches and assessing the authenticity of each patch, our approach identifies anomalies at a more granular level, overcoming the limitations of whole-image evaluation. Additionally, a patch-ranking mechanism prioritizes regions with higher abnormal scores, reinforcing the alignment between local patch discrepancies and the global image context. Experimental results on the ISIC 2016 skin lesion and BraTS 2019 brain tumor datasets validate our framework's effectiveness, achieving AUCs of 95.79% and 96.05%, respectively, and outperforming three state-of-the-art baselines.
Self-supervised learning (SSL) has been successfully applied to remote sensing image classification by designing pretext tasks to extract valuable feature representations of targets. However, existing SSL methodologies overlook the edge information integral to ground objects, culminating in frequent misclassifications at target boundaries. Additionally, the scarcity of training samples often restricts the full utilization of the knowledge encapsulated in the pre-training model. To address these issues, we propose a novel self-supervised edge perception learning framework (SEPLF) to improve the classification performance of high-resolution remote sensing images (HRSI). The framework comprises self-supervised edge perception learning (SEPL) and training sample augmentation (TSA) algorithms. On the one hand, the SEPL approach leverages morphological data enhancement strategies to render the extracted invariant features more robust. It also effectively mines the potential information concealed at target edges, augmenting ground objects's edge separability. On the other hand, the TSA algorithm not only obtains a large number of training samples but also enhances the intra-class diversity of the samples by considering different spectral features of the same category of ground objects. Experimental results validate that our proposed method outperforms state-of-the-art algorithms, particularly with limited labeled samples.
Despite the great success achieved by prevailing binary local descriptors, they are still suffering from two problems: 1) vulnerable to the geometric transformations; 2) lack of an effective treatment to the highly-correlated bits that are generated by directly applying the scheme of image hashing. To tackle both limitations, we propose an unsupervised Transformation-invariant Binary Local Descriptor learning method (TBLD). Specifically, the transformation invariance of binary local descriptors is ensured by projecting the original patches and their transformed counterparts into an identical high-dimensional feature space and an identical low-dimensional descriptor space simultaneously. Meanwhile, it enforces the dissimilar image patches to have distinctive binary local descriptors. Moreover, to reduce high correlations between bits, we propose a bottom-up learning strategy, termed Adversarial Constraint Module, where low-coupling binary codes are introduced externally to guide the learning of binary local descriptors. With the aid of the Wasserstein loss, the framework is optimized to encourage the distribution of the generated binary local descriptors to mimic that of the introduced low-coupling binary codes, eventually making the former more low-coupling. Experimental results on three benchmark datasets well demonstrate the superiority of the proposed method over the state-of-the-art methods. The project page is available at https://github.com/yoqim/TBLD.
In recent years considerable research efforts have been devoted to compression techniques of convolutional neural networks (CNNs). Many works so far have focused on CNN connection pruning methods which produce sparse parameter tensors in convolutional or fully-connected layers. It has been demonstrated in several studies that even simple methods can effectively eliminate connections of a CNN. However, since these methods make parameter tensors just sparser but no smaller, the compression may not transfer directly to acceleration without support from specially designed hardware. In this paper, we propose an iterative approach named Auto-balanced Filter Pruning, where we pre-train the network in an innovative auto-balanced way to transfer the representational capacity of its convolutional layers to a fraction of the filters, prune the redundant ones, then re-train it to restore the accuracy. In this way, a smaller version of the original network is learned and the floating-point operations (FLOPs) are reduced. By applying this method on several common CNNs, we show that a large portion of the filters can be discarded without obvious accuracy drop, leading to significant reduction of computational burdens. Concretely, we reduce the inference cost of LeNet-5 on MNIST, VGG-16 and ResNet-56 on CIFAR-10 by 95.1%, 79.7% and 60.9%, respectively.
Background Individual differences have been detected in individuals with opioid use disorders (OUD) in rehabilitation following protracted abstinence. Recent studies suggested that prediction models were effective for individual-level prognosis based on neuroimage data in substance use disorders (SUD). Aims This prospective cohort study aimed to assess neuroimaging biomarkers for individual response to protracted abstinence in opioid users using connectome-based predictive modelling (CPM). Methods One hundred and eight inpatients with OUD underwent structural and functional magnetic resonance imaging (fMRI) scans at baseline. The Heroin Craving Questionnaire (HCQ) was used to assess craving levels at baseline and at the 8-month follow-up of abstinence. CPM with leave-one-out cross-validation was used to identify baseline networks that could predict follow-up HCQ scores and changes in HCQ (HCQ follow-up −HCQ baseline) . Then, the predictive ability of identified networks was tested in a separate, heterogeneous sample of methamphetamine individuals who underwent MRI scanning before abstinence for SUD. Results CPM could predict craving changes induced by long-term abstinence, as shown by a significant correlation between predicted and actual HCQ follow-up (r=0.417, p<0.001) and changes in HCQ (negative: r=0.334, p=0.002;positive: r=0.233, p=0.038). Identified craving-related prediction networks included the somato-motor network (SMN), salience network (SALN), default mode network (DMN), medial frontal network, visual network and auditory network. In addition, decreased connectivity of frontal-parietal network (FPN)-SMN, FPN-DMN and FPN-SALN and increased connectivity of subcortical network (SCN)-DMN, SCN-SALN and SCN-SMN were positively correlated with craving levels. Conclusions These findings highlight the potential applications of CPM to predict the craving level of individuals after protracted abstinence, as well as the generalisation ability; the identified brain networks might be the focus of innovative therapies in the future.
Despite the thrilling success achieved by existing binary descriptors, most of them are still in the mire of three limitations: 1) vulnerable to the geometric transformations; 2) incapable of preserving the manifold structure when learning binary codes; 3) NO guarantee to find the true match if multiple candidates happen to have the same Hamming distance to a given query. All these together make the binary descriptor less effective, given large-scale visual recognition tasks. In this paper, we propose a novel learning-based feature descriptor, namely Unsupervised Deep Binary Descriptor (UDBD), which learns transformation invariant binary descriptors via projecting the original data and their transformed sets into a joint binary space. Moreover, we involve a ℓ 2,1 -norm loss term in the binary embedding process to gain simultaneously the robustness against data noises and less probability of mistakenly flipping bits of the binary descriptor, on top of it, a graph constraint is used to preserve the original manifold structure in the binary space. Furthermore, a weak bit mechanism is adopted to find the real match from candidates sharing the same minimum Hamming distance, thus enhancing matching performance. Extensive experimental results on public datasets show the superiority of UDBD in terms of matching and retrieval accuracy over state-of-the-arts.