west china medical publishers
Keyword
  • Title
  • Author
  • Keyword
  • Abstract
Advance search
Advance search

Search

find Keyword "Attention mechanism" 19 results
  • Predicting epileptic seizures based on a multi-convolution fusion network

    Current epilepsy prediction methods are not effective in characterizing the multi-domain features of complex long-term electroencephalogram (EEG) data, leading to suboptimal prediction performance. Therefore, this paper proposes a novel multi-scale sparse adaptive convolutional network based on multi-head attention mechanism (MS-SACN-MM) model to effectively characterize the multi-domain features. The model first preprocesses the EEG data, constructs multiple convolutional layers to effectively avoid information overload, and uses a multi-layer perceptron and multi-head attention mechanism to focus the network on critical pre-seizure features. Then, it adopts a focal loss training strategy to alleviate class imbalance and enhance the model's robustness. Experimental results show that on the publicly created dataset (CHB-MIT) by MIT and Boston Children's Hospital, the MS-SACN-MM model achieves a maximum accuracy of 0.999 for seizure prediction 10 ~ 15 minutes in advance. This demonstrates good predictive performance and holds significant importance for early intervention and intelligent clinical management of epilepsy patients.

    Release date:2025-10-21 03:48 Export PDF Favorites Scan
  • Research on fault diagnosis of patient monitor based on text mining

    The conventional fault diagnosis of patient monitors heavily relies on manual experience, resulting in low diagnostic efficiency and ineffective utilization of fault maintenance text data. To address these issues, this paper proposes an intelligent fault diagnosis method for patient monitors based on multi-feature text representation, improved bidirectional gate recurrent unit (BiGRU) and attention mechanism. Firstly, the fault text data was preprocessed, and the word vectors containing multiple linguistic features was generated by linguistically-motivated bidirectional encoder representation from Transformer. Then, the bidirectional fault features were extracted and weighted by the improved BiGRU and attention mechanism respectively. Finally, the weighted loss function is used to reduce the impact of class imbalance on the model. To validate the effectiveness of the proposed method, this paper uses the patient monitor fault dataset for verification, and the macro F1 value has achieved 91.11%. The results show that the model built in this study can realize the automatic classification of fault text, and may provide assistant decision support for the intelligent fault diagnosis of the patient monitor in the future.

    Release date:2024-04-24 09:40 Export PDF Favorites Scan
  • A lightweight convolutional neural network for myositis classification from muscle ultrasound images

    Existing classification methods for myositis ultrasound images have problems of poor classification performance or high computational cost. Motivated by this difficulty, a lightweight neural network based on a soft threshold attention mechanism is proposed to cater for a better IIMs classification. The proposed network was constructed by alternately using depthwise separable convolution (DSC) and conventional convolution (CConv). Moreover, a soft threshold attention mechanism was leveraged to enhance the extraction capabilities of key features. Compared with the current dual-branch feature fusion myositis classification network with the highest classification accuracy, the classification accuracy of the network proposed in this paper increased by 5.9%, reaching 96.1%, and its computational complexity was only 0.25% of the existing method. The obtained results support that the proposed method can provide physicians with more accurate classification results at a lower computational cost, thereby greatly assisting them in their clinical diagnosis.

    Release date:2024-10-22 02:39 Export PDF Favorites Scan
  • Multi-scale feature extraction and classification of motor imagery electroencephalography based on time series data enhancement

    The brain-computer interface (BCI) based on motor imagery electroencephalography (MI-EEG) enables direct information interaction between the human brain and external devices. In this paper, a multi-scale EEG feature extraction convolutional neural network model based on time series data enhancement is proposed for decoding MI-EEG signals. First, an EEG signals augmentation method was proposed that could increase the information content of training samples without changing the length of the time series, while retaining its original features completely. Then, multiple holistic and detailed features of the EEG data were adaptively extracted by multi-scale convolution module, and the features were fused and filtered by parallel residual module and channel attention. Finally, classification results were output by a fully connected network. The application experimental results on the BCI Competition IV 2a and 2b datasets showed that the proposed model achieved an average classification accuracy of 91.87% and 87.85% for the motor imagery task, respectively, which had high accuracy and strong robustness compared with existing baseline models. The proposed model does not require complex signals pre-processing operations and has the advantage of multi-scale feature extraction, which has high practical application value.

    Release date:2023-08-23 02:45 Export PDF Favorites Scan
  • Deep overparameterized blood cell detection algorithm utilizing hybrid attention mechanisms

    To address the challenges in blood cell recognition caused by diverse morphology, dense distribution, and the abundance of small target information, this paper proposes a blood cell detection algorithm - the "You Only Look Once" model based on hybrid mixing attention and deep over-parameters (HADO-YOLO). First, a hybrid attention mechanism is introduced into the backbone network to enhance the model's sensitivity to detailed features. Second, the standard convolution layers with downsampling in the neck network are replaced with deep over-parameterized convolutions to expand the receptive field and improve feature representation. Finally, the detection head is decoupled to enhance the model's robustness for detecting abnormal cells. Experimental results on the Blood Cell Counting Dataset (BCCD) demonstrate that the HADO-YOLO algorithm achieves a mean average precision of 90.2% and a precision of 93.8%, outperforming the baseline YOLO model. Compared with existing blood cell detection methods, the proposed algorithm achieves state-of-the-art detection performance. In conclusion, HADO-YOLO offers a more efficient and accurate solution for identifying various types of blood cells, providing valuable technical support for future clinical diagnostic applications.

    Release date:2025-10-21 03:48 Export PDF Favorites Scan
  • A joint distillation model for the tumor segmentation using breast ultrasound images

    The accurate segmentation of breast ultrasound images is an important precondition for the lesion determination. The existing segmentation approaches embrace massive parameters, sluggish inference speed, and huge memory consumption. To tackle this problem, we propose T2KD Attention U-Net (dual-Teacher Knowledge Distillation Attention U-Net), a lightweight semantic segmentation method combined double-path joint distillation in breast ultrasound images. Primarily, we designed two teacher models to learn the fine-grained features from each class of images according to different feature representation and semantic information of benign and malignant breast lesions. Then we leveraged the joint distillation to train a lightweight student model. Finally, we constructed a novel weight balance loss to focus on the semantic feature of small objection, solving the unbalance problem of tumor and background. Specifically, the extensive experiments conducted on Dataset BUSI and Dataset B demonstrated that the T2KD Attention U-Net outperformed various knowledge distillation counterparts. Concretely, the accuracy, recall, precision, Dice, and mIoU of proposed method were 95.26%, 86.23%, 85.09%, 83.59%and 77.78% on Dataset BUSI, respectively. And these performance indexes were 97.95%, 92.80%, 88.33%, 88.40% and 82.42% on Dataset B, respectively. Compared with other models, the performance of this model was significantly improved. Meanwhile, compared with the teacher model, the number, size, and complexity of student model were significantly reduced (2.2×106 vs. 106.1×106, 8.4 MB vs. 414 MB, 16.59 GFLOPs vs. 205.98 GFLOPs, respectively). Indeedy, the proposed model guarantees the performances while greatly decreasing the amount of computation, which provides a new method for the deployment of clinical medical scenarios.

    Release date:2025-02-21 03:20 Export PDF Favorites Scan
  • Research on emotion recognition in electroencephalogram based on independent component analysis-recurrence plot and improved EfficientNet

    To accurately capture and effectively integrate the spatiotemporal features of electroencephalogram (EEG) signals for the purpose of improving the accuracy of EEG-based emotion recognition, this paper proposes a new method combining independent component analysis-recurrence plot with an improved EfficientNet version 2 (EfficientNetV2). First, independent component analysis is used to extract independent components containing spatial information from key channels of the EEG signals. These components are then converted into two-dimensional images using recurrence plot to better extract emotional features from the temporal information. Finally, the two-dimensional images are input into an improved EfficientNetV2, which incorporates a global attention mechanism and a triplet attention mechanism, and the emotion classification is output by the fully connected layer. To validate the effectiveness of the proposed method, this study conducts comparative experiments, channel selection experiments and ablation experiments based on the Shanghai Jiao Tong University Emotion Electroencephalogram Dataset (SEED). The results demonstrate that the average recognition accuracy of our method is 96.77%, which is significantly superior to existing methods, offering a novel perspective for research on EEG-based emotion recognition.

    Release date:2024-12-27 03:50 Export PDF Favorites Scan
  • Image segmentation of skin lesions based on dense atrous spatial pyramid pooling and attention mechanism

    The skin is the largest organ of the human body, and many visceral diseases will be directly reflected on the skin, so it is of great clinical significance to accurately segment the skin lesion images. To address the characteristics of complex color, blurred boundaries, and uneven scale information, a skin lesion image segmentation method based on dense atrous spatial pyramid pooling (DenseASPP) and attention mechanism is proposed. The method is based on the U-shaped network (U-Net). Firstly, a new encoder is redesigned to replace the ordinary convolutional stacking with a large number of residual connections, which can effectively retain key features even after expanding the network depth. Secondly, channel attention is fused with spatial attention, and residual connections are added so that the network can adaptively learn channel and spatial features of images. Finally, the DenseASPP module is introduced and redesigned to expand the perceptual field size and obtain multi-scale feature information. The algorithm proposed in this paper has obtained satisfactory results in the official public dataset of the International Skin Imaging Collaboration (ISIC 2016). The mean Intersection over Union (mIOU), sensitivity (SE), precision (PC), accuracy (ACC), and Dice coefficient (Dice) are 0.901 8, 0.945 9, 0.948 7, 0.968 1, 0.947 3, respectively. The experimental results demonstrate that the method in this paper can improve the segmentation effect of skin lesion images, and is expected to provide an auxiliary diagnosis for professional dermatologists.

    Release date:2023-02-24 06:14 Export PDF Favorites Scan
  • Small-scale cross-layer fusion network for classification of diabetic retinopathy

    Deep learning-based automatic classification of diabetic retinopathy (DR) helps to enhance the accuracy and efficiency of auxiliary diagnosis. This paper presents an improved residual network model for classifying DR into five different severity levels. First, the convolution in the first layer of the residual network was replaced with three smaller convolutions to reduce the computational load of the network. Second, to address the issue of inaccurate classification due to minimal differences between different severity levels, a mixed attention mechanism was introduced to make the model focus more on the crucial features of the lesions. Finally, to better extract the morphological features of the lesions in DR images, cross-layer fusion convolutions were used instead of the conventional residual structure. To validate the effectiveness of the improved model, it was applied to the Kaggle Blindness Detection competition dataset APTOS2019. The experimental results demonstrated that the proposed model achieved a classification accuracy of 97.75% and a Kappa value of 0.971 7 for the five DR severity levels. Compared to some existing models, this approach shows significant advantages in classification accuracy and performance.

    Release date: Export PDF Favorites Scan
  • Multi-tissue segmentation model of whole slide image of pancreatic cancer based on multi task and attention mechanism

    Accurate segmentation of whole slide images is of great significance for the diagnosis of pancreatic cancer. However, developing an automatic model is challenging due to the complex content, limited samples, and high sample heterogeneity of pathological images. This paper presented a multi-tissue segmentation model for whole slide images of pancreatic cancer. We introduced an attention mechanism in building blocks, and designed a multi-task learning framework as well as proper auxiliary tasks to enhance model performance. The model was trained and tested with the pancreatic cancer pathological image dataset from Shanghai Changhai Hospital. And the data of TCGA, as an external independent validation cohort, was used for external validation. The F1 scores of the model exceeded 0.97 and 0.92 in the internal dataset and external dataset, respectively. Moreover, the generalization performance was also better than the baseline method significantly. These results demonstrate that the proposed model can accurately segment eight kinds of tissue regions in whole slide images of pancreatic cancer, which can provide reliable basis for clinical diagnosis.

    Release date:2023-02-24 06:14 Export PDF Favorites Scan
2 pages Previous 1 2 Next

Format

Content