SwePub
Sök i LIBRIS databas

  Extended search

id:"swepub:oai:DiVA.org:bth-26459"
 

Search: id:"swepub:oai:DiVA.org:bth-26459" > LAD-Net :

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

LAD-Net : A lightweight welding defect surface non-destructive detection algorithm based on the attention mechanism

Liang, Feng (author)
Kunming University of Science and Technology, China
Zhao, Lun (author)
Shenzhen Polytechnic University, China
Ren, Yu (author)
Shenzhen Polytechnic University, China
show more...
Wang, Sen (author)
Kunming University of Science and Technology, China
To, Sandy (author)
The Hong Kong Polytechnic University, Hong Kong
Abbas, Zeshan (author)
Shenzhen Polytechnic University, China
Islam, Md. Shafiqul, 1984- (author)
Blekinge Tekniska Högskola,Institutionen för maskinteknik
show less...
 (creator_code:org_t)
Elsevier, 2024
2024
English.
In: Computers in industry (Print). - : Elsevier. - 0166-3615 .- 1872-6194. ; 161
  • Journal article (peer-reviewed)
Abstract Subject headings
Close  
  • Ultrasound welding technology is widely applied in the field of industrial manufacturing. In complex working conditions, various factors such as welding parameters, equipment conditions and operational techniques contribute to the formation of diverse and unpredictable line defects during the welding process. These defects exhibit characteristics such as varied shapes, random positions, and diverse types. Consequently, traditional defect surface detection methods face challenges in achieving efficient and accurate non-destructive testing. To achieve real-time detection of ultrasound welding defects efficiently, we have developed a lightweight network called the Lightweight Attention Detection Network (LAD-Net) based on an attention mechanism. Firstly, this work proposes a Deformable Convolution Feature Extraction Module (DCFE-Module) aimed at addressing the challenge of extracting features from welding defects characterized by variable shapes, random positions, and complex defect types. Additionally, to prevent the loss of critical defect features and enhance the network's capability for feature extraction and integration, this study designs a Lightweight Step Attention Mechanism Module (LSAM-Module) based on the proposed Step Attention Mechanism Convolution (SAM-Conv). Finally, by integrating the Efficient Multi-scale Attention (EMA) module and the Explicit Visual Center (EVC) module into the network, we address the issue of imbalance between global and local information processing, and promote the integration of key defect features. Qualitative and quantitative experimental results conducted on both ultrasound welding defect data and the publicly available NEU-DET dataset demonstrate that the proposed LAD-Net method achieves high performance. On our custom dataset, the F1 score and mAP@0.5 reached 0.954 and 94.2%, respectively. Furthermore, the method exhibits superior detection performance on the public dataset. © 2024 Elsevier B.V.

Subject headings

TEKNIK OCH TEKNOLOGIER  -- Materialteknik -- Bearbetnings-, yt- och fogningsteknik (hsv//swe)
ENGINEERING AND TECHNOLOGY  -- Materials Engineering -- Manufacturing, Surface and Joining Technology (hsv//eng)

Keyword

Attention mechanism
Defect detection
LAD-Net
Ultrasonic welding
Complex networks
Defects
Extraction
Feature extraction
Nondestructive examination
Signal detection
Ultrasonic testing
Welding
Attention detection
Attention mechanisms
Detection networks
Features extraction
Lightweight attention detection network
Nondestructive detection
Random position
Ultrasonic weldings
Welding defects
Convolution

Publication and Content Type

ref (subject category)
art (subject category)

Find in a library

To the university's database

  • 1 of 1
  • Previous record
  • Next record
  •    To hitlist

Search outside SwePub

Kungliga biblioteket hanterar dina personuppgifter i enlighet med EU:s dataskyddsförordning (2018), GDPR. Läs mer om hur det funkar här.
Så här hanterar KB dina uppgifter vid användning av denna tjänst.

 
pil uppåt Close

Copy and save the link in order to return to this view