Deformable image registration plays a crucial role in medical image analysis. Despite various advanced registration models having been proposed, achieving accurate and efficient deformable registration remains challenging. Leveraging the recent outstanding performance of Mamba in computer vision, we introduced a novel model called MCRDP-Net. MCRDP-Net adapted a dual-stream network architecture that combined Mamba blocks and convolutional blocks to simultaneously extract global and local information from fixed and moving images. In the decoding stage, we employed a pyramid network structure to obtain high-resolution deformation fields, achieving efficient and precise registration. The effectiveness of MCRDP-Net was validated on public brain registration datasets, OASIS and IXI. Experimental results demonstrated significant advantages of MCRDP-Net in medical image registration, with DSC, HD95, and ASD reaching 0.815, 8.123, and 0.521 on the OASIS dataset and 0.773, 7.786, and 0.871 on the IXI dataset. In summary, MCRDP-Net demonstrates superior performance in deformable image registration, proving its potential in medical image analysis. It effectively enhances the accuracy and efficiency of registration, providing strong support for subsequent medical research and applications.
Atrial fibrillation (AF) is a life-threatening heart condition, and its early detection and treatment have garnered significant attention from physicians in recent years. Traditional methods of detecting AF heavily rely on doctor’s diagnosis based on electrocardiograms (ECGs), but prolonged analysis of ECG signals is very time-consuming. This paper designs an AF detection model based on the Inception module, constructing multi-branch detection channels to process raw ECG signals, gradient signals, and frequency signals during AF. The model efficiently extracted QRS complex and RR interval features using gradient signals, extracted P-wave and f-wave features using frequency signals, and used raw signals to supplement missing information. The multi-scale convolutional kernels in the Inception module provided various receptive fields and performed comprehensive analysis of the multi-branch results, enabling early AF detection. Compared to current machine learning algorithms that use only RR interval and heart rate variability features, the proposed algorithm additionally employed frequency features, making fuller use of the information within the signals. For deep learning methods using raw and frequency signals, this paper introduced an enhanced method for the QRS complex, allowing the network to extract features more effectively. By using a multi-branch input mode, the model comprehensively considered irregular RR intervals and P-wave and f-wave features in AF. Testing on the MIT-BIH AF database showed that the inter-patient detection accuracy was 96.89%, sensitivity was 97.72%, and specificity was 95.88%. The proposed model demonstrates excellent performance and can achieve automatic AF detection.
In clinical, manually scoring by technician is the major method for sleep arousal detection. This method is time-consuming and subjective. This study aimed to achieve an end-to-end sleep-arousal events detection by constructing a convolutional neural network based on multi-scale convolutional layers and self-attention mechanism, and using 1 min single-channel electroencephalogram (EEG) signals as its input. Compared with the performance of the baseline model, the results of the proposed method showed that the mean area under the precision-recall curve and area under the receiver operating characteristic were both improved by 7%. Furthermore, we also compared the effects of single modality and multi-modality on the performance of the proposed model. The results revealed the power of single-channel EEG signals in automatic sleep arousal detection. However, the simple combination of multi-modality signals may be counterproductive to the improvement of model performance. Finally, we also explored the scalability of the proposed model and transferred the model into the automated sleep staging task in the same dataset. The average accuracy of 73% also suggested the power of the proposed method in task transferring. This study provides a potential solution for the development of portable sleep monitoring and paves a way for the automatic sleep data analysis using the transfer learning method.
In order to realize the quantitative assessment of muscle strength in hand function rehabilitation and then formulate scientific and effective rehabilitation training strategies, this paper constructs a multi-scale convolutional neural network (MSCNN) - convolutional block attention module (CBAM) - bidirectional long short-term memory network (BiLSTM) muscle strength prediction model to fully explore the spatial and temporal features of the data and simultaneously suppress useless features, and finally achieve the improvement of the accuracy of the muscle strength prediction model. To verify the effectiveness of the model proposed in this paper, the model in this paper is compared with traditional models such as support vector machine (SVM), random forest (RF), convolutional neural network (CNN), CNN - squeeze excitation network (SENet), MSCNN-CBAM and MSCNN-BiLSTM, and the effect of muscle strength prediction by each model is investigated when the hand force application changes from 40% of the maximum voluntary contraction force (MVC) to 60% of the MVC. The research results show that as the hand force application increases, the effect of the muscle strength prediction model becomes worse. Then the ablation experiment is used to analyze the influence degree of each module on the muscle strength prediction result, and it is found that the CBAM module plays a key role in the model. Therefore, by using the model in this article, the accuracy of muscle strength prediction can be effectively improved, and the characteristics and laws of hand muscle activities can be deeply understood, providing assistance for further exploring the mechanism of hand functions.
In order to meet the need of autonomous control of patients with severe limb disorders, this paper designs a nursing bed control system based on motor imagery-brain computer interface (MI-BCI). In view of the low decoding performance of cross-subjects and the dynamic fluctuation of cognitive state in the existing MI-BCI technology, the neural network structure optimization and user interaction feedback enhancement are improved. Firstly, the optimized dual-branch graph convolution multi-scale neural network integrates dynamic graph convolution and multi-scale convolution. The average classification accuracy is higher than that of multi-scale attention temporal convolution network, Gram angle field combined with convolution long short term memory hybrid network, Transformer-based graph convolution network and other existing methods. Secondly, a dual visual feedback mechanism is constructed, in which electroencephalogram (EEG) topographic map feedback can improve the discrimination of spatial patterns, and attention state feedback can enhance the temporal stability of signals. Compared with the single EEG topographic map feedback and non-feedback system, the average classification accuracy of the proposed method is also greatly improved. Finally, in the four classification control task of nursing bed, the average control accuracy of the system is 90.84%, and the information transmission rate is 84.78 bits/min. In summary, this paper provides a reliable technical solution for improving the autonomous interaction ability of patients with severe limb disorders, which has important theoretical significance and application value.
Glioma is a primary brain tumor with high incidence rate. High-grade gliomas (HGG) are those with the highest degree of malignancy and the lowest degree of survival. Surgical resection and postoperative adjuvant chemoradiotherapy are often used in clinical treatment, so accurate segmentation of tumor-related areas is of great significance for the treatment of patients. In order to improve the segmentation accuracy of HGG, this paper proposes a multi-modal glioma semantic segmentation network with multi-scale feature extraction and multi-attention fusion mechanism. The main contributions are, (1) Multi-scale residual structures were used to extract features from multi-modal gliomas magnetic resonance imaging (MRI); (2) Two types of attention modules were used for features aggregating in channel and spatial; (3) In order to improve the segmentation performance of the whole network, the branch classifier was constructed using ensemble learning strategy to adjust and correct the classification results of the backbone classifier. The experimental results showed that the Dice coefficient values of the proposed segmentation method in this article were 0.909 7, 0.877 3 and 0.839 6 for whole tumor, tumor core and enhanced tumor respectively, and the segmentation results had good boundary continuity in the three-dimensional direction. Therefore, the proposed semantic segmentation network has good segmentation performance for high-grade gliomas lesions.
In response to the issues of single-scale information loss and large model parameter size during the sampling process in U-Net and its variants for medical image segmentation, this paper proposes a multi-scale medical image segmentation method based on pixel encoding and spatial attention. Firstly, by redesigning the input strategy of the Transformer structure, a pixel encoding module is introduced to enable the model to extract global semantic information from multi-scale image features, obtaining richer feature information. Additionally, deformable convolutions are incorporated into the Transformer module to accelerate convergence speed and improve module performance. Secondly, a spatial attention module with residual connections is introduced to allow the model to focus on the foreground information of the fused feature maps. Finally, through ablation experiments, the network is lightweighted to enhance segmentation accuracy and accelerate model convergence. The proposed algorithm achieves satisfactory results on the Synapse dataset, an official public dataset for multi-organ segmentation provided by the International Conference on Medical Image Computing and Computer Assisted Intervention (MICCAI), with Dice similarity coefficient (DSC) and 95% Hausdorff distance (HD95) scores of 77.65 and 18.34, respectively. The experimental results demonstrate that the proposed algorithm can enhance multi-organ segmentation performance, potentially filling the gap in multi-scale medical image segmentation algorithms, and providing assistance for professional physicians in diagnosis.
Chromatin three-dimensional genome structure plays a key role in cell function and gene regulation. Single-cell Hi-C techniques can capture genomic structure information at the cellular level, which provides an opportunity to study changes in genomic structure between different cell types. Recently, some excellent computational methods have been developed for single-cell Hi-C data analysis. In this paper, the available methods for single-cell Hi-C data analysis were first reviewed, including preprocessing of single-cell Hi-C data, multi-scale structure recognition based on single-cell Hi-C data, bulk-like Hi-C contact matrix generation based on single-cell Hi-C data sets, pseudo-time series analysis, and cell classification. Then the application of single-cell Hi-C data in cell differentiation and structural variation was described. Finally, the future development direction of single-cell Hi-C data analysis was also prospected.
Alzheimer’s disease (AD) classification models usually segment the entire brain image into voxel blocks and assign them labels consistent with the entire image, but not every voxel block is closely related to the disease. To this end, an AD auxiliary diagnosis framework based on weakly supervised multi-instance learning (MIL) and multi-scale feature fusion is proposed, and the framework is designed from three aspects: within the voxel block, between voxel blocks, and high-confidence voxel blocks. First, a three-dimensional convolutional neural network was used to extract deep features within the voxel block; then the spatial correlation information between voxel blocks was captured through position encoding and attention mechanism; finally, high-confidence voxel blocks were selected and combined with multi-scale information fusion strategy to integrate key features for classification decision. The performance of the model was evaluated on the Alzheimer’s Disease Neuroimaging Initiative (ADNI) and Open Access Series of Imaging Studies (OASIS) datasets. Experimental results showed that the proposed framework improved ACC and AUC by 3% and 4% on average compared with other mainstream frameworks in the two tasks of AD classification and mild cognitive impairment conversion classification, and could find the key voxel blocks that trigger the disease, providing an effective basis for AD auxiliary diagnosis.
To address issues such as loss of detailed information, blurred target boundaries, and unclear structural hierarchy in medical image fusion, this paper proposes an adaptive feature medical image fusion network based on a full-scale diffusion model. First, a region-level feature map is generated using a kernel-based saliency map to enhance local features and boundary details. Then, a full-scale diffusion feature extraction network is employed for global feature extraction, alongside a multi-scale denoising U-shaped network designed to fully capture cross-layer information. A multi-scale feature integration module is introduced to reinforce texture details and structural information extracted by the encoder. Finally, an adaptive fusion scheme is applied to progressively fuse region-level features, global features, and source images layer by layer, enhancing the preservation of detail information. To validate the effectiveness of the proposed method, this paper validates the proposed model on the publicly available Harvard dataset and an abdominal dataset. By comparing with nine other representative image fusion methods, the proposed approach achieved improvements across seven evaluation metrics. The results demonstrate that the proposed method effectively extracts both global and local features of medical images, enhances texture details and target boundary clarity, and generates fusion image with high contrast and rich information, providing more reliable support for subsequent clinical diagnosis.