CPC G06V 10/82 (2022.01) [G06V 10/764 (2022.01)] | 6 Claims |
1. An activity recognition method of an LRF (large receptive field) large kernel attention convolution network based on a large receptive field, comprising:
collecting an action signal, carrying out a preprocessing and a data partition on the action signal to obtain a data set; and
training an LRF large-kernel attention convolution network model based on the data set, and introducing a trained LRF large-kernel attention convolution network model into a mobile wearable recognition device for a human posture recognition;
wherein the LRF large-kernel attention convolution network model comprises:
a n LRF large-kernel attention convolution network with three layers and a fully connected classification output layer, wherein the LRF large-kernel attention convolution network comprises a local depth convolution layer, a long-distance depth expansion convolution layer and a 1×1 ordinary convolution layer for a feature extraction; and the fully connected classification output layer is used for an action classification;
a calculation method of the LRF large-kernel attention convolution network model comprises:
wherein X represents an input matrix, t represents a time step of the input matrix, and s represents a sensor mode the input matrix;
compressing the input matrix X into one-dimensional data, and introducing into a self-attention module, outputting a weighted sum of all value vectors, and using a Softmax function for a normalization;
wherein Q, K, and V represent a query value, a key value and a vector value respectively; and dk represents a scaling factor;
proposing an LRF attention mechanism to capture time information and modal information in sensor activity images:
X′=ReLU(BN(Conv2d(X))) (3)
wherein X′ represents a node output matrix in four dimensions, ReLU represents an activation function, and Conv2d represents a two-dimensional convolution operation;
obtaining a normalized output result of the node output matrix X′ by a layer normalization function, and further strengthening a network anti-degradation ability by a shortcut link:
X″=X′+LRF(LN(X′)) (4)
wherein a symbol LRF and a symbol LN represent a leak-kernel receptive field attention mechanism and the layer normalization function respectively;
outputting a feedforward network comprising a multilayer perceptron and a normalization layer by formula (4):
X″′=X″+MLP(L N(X′)) (5)
wherein symbol MLP and LN represent the multilayer perceptron and a layer normalization respectively.
|