Steel rail surface defect image expansion method based on attention mechanism and DCGAN
An attention and defect technology, applied in the field of rail surface defect detection, can solve the problems of poor performance, low sample diversity, and it is difficult for the algorithm to detect the untrained, so as to increase the diversity and improve the application effect.
- Summary
- Abstract
- Description
- Claims
- Application Information
AI Technical Summary
Problems solved by technology
Method used
Image
Examples
Embodiment 1
[0059] An image augmentation method for rail surface defects based on attention mechanism and DCGAN, characterized in that it includes the following steps:
[0060] S1. Create a track surface defect dataset, including the following steps:
[0061] S11. Select images: select no less than 200 photos of rail surface defects;
[0062] S12. Crop the picture: crop out the surface area of the rail containing the defect, and adjust the size to 300×130.
[0063] 2. Build a self-attention module, including the following steps:
[0064] S21. Perform linear transformation and channel adjustment on the feature map x in the hidden layer through 1×1 convolution and convert it into 3 feature maps f(x), g(x), h(x), where f(x), g (x) used to calculate attention;
[0065] S22. Transpose the output of f(x) and multiply it with the output of g(x), and normalize by softmax to obtain the attention map M(x);
[0066] S23. Multiply the attention map M(x) and the linearly transformed original fea...
Embodiment 2
[0098] An image augmentation method for rail surface defects based on attention mechanism and DCGAN, characterized in that it includes the following steps:
[0099] 1) Make a rail surface defect data set, select no less than 200 rail surface defect photos, cut out the rail surface area containing the defects, and adjust the size to 300×130.
[0100] 2) Build a self-attention module, calculate the relationship between any two pixels in the image, and capture the internal correlation of features, such as figure 1 shown. The specific process of building a self-attention module is as follows:
[0101] 21) Perform linear transformation and channel adjustment on the feature map x in the hidden layer by 1×1 convolution and convert it into 3 feature maps f(x), g(x), h(x), where f(x), g (x) used to calculate attention;
[0102] 22) Transpose the output of f(x) and multiply it with the output of g(x), and normalize by softmax to get the attention map M(x);
[0103] 23) Multiply the ...
PUM
Login to View More Abstract
Description
Claims
Application Information
Login to View More 


