US 12,321,706 B2
Soft knowledge prompts for language models
Siamak Shakeri, New York City, NY (US); Cicero Nogueira dos Santos, Glen Ridge, NJ (US); Daniel Matthew Cer, Santa Clara, CA (US); Zhe Dong, Zurich (CH); Jianmo Ni, Santa Clara, CA (US); Yun-Hsuan Sung, San Francisco, CA (US); and John Nham, Fremont, CA (US)
Assigned to Google LLC, Mountain View, CA (US)
Filed by Google LLC, Mountain View, CA (US)
Filed on Feb. 9, 2023, as Appl. No. 18/166,806.
Prior Publication US 2024/0273294 A1, Aug. 15, 2024
Int. Cl. G06F 40/295 (2020.01)
CPC G06F 40/295 (2020.01) 21 Claims
OG exemplary drawing
 
1. A computer-implemented method, comprising:
identifying, by one or more processors of a computing system, a soft knowledge prompt in response to a received input text, wherein the received input text masks an object entity;
concatenating, by the one or more processors, the identified soft knowledge prompt to a sequence of word embeddings of the input text;
applying, by the one or more processors, the concatenated soft knowledge prompt and the sequence of word embeddings to a trained language model;
predicting, by the one or more processors, an object entity name of the object entity;
computing, by the one or more processors, a cross-entropy loss according to the predicted object entity name;
updating the identified soft knowledge prompt based on the computed cross-entropy loss, wherein the updated soft knowledge prompt is configured to function as a part of an auxiliary memory of a language model that is activated when solving a specific task; and
disambiguating a named entity that appears in the received input text.