CPC G06F 40/284 (2020.01) [G06F 40/30 (2020.01); G06N 3/047 (2023.01)] | 20 Claims |
1. A computer-implemented method comprising:
generating, by one or more processors and using a multi-context convolutional self-attention machine learning framework, a cross-context token representation based at least in part on an input text token of an input text sequence wherein:
the multi-context convolutional self-attention machine learning framework comprises a shared token embedding machine learning model, a plurality of context-specific self-attention machine learning models, and a cross-context representation inference machine learning model,
the shared token embedding machine learning model is configured to generate an initial token embedding for the input text token,
a context-specific self-attention machine learning model of the plurality of context-specific self-attention machine learning models is (a) associated with a distinct context window size of a plurality of distinct context window sizes, and (b) configured to generate a context-specific token representation for the input text token based at least in part on the initial token embedding, and
the cross-context representation inference machine learning model is configured to generate the cross-context token representation based at least in part on the context-specific token representation;
generating, by the one or more processors and a natural language processing machine learning model, a natural language processing output for the input text sequence based at least in part on the cross-context token representation; and
initiating, by the one or more processors, the performance of one or more prediction-based actions based at least in part on the natural language processing output.
|