US 12,204,847 B2
Systems and methods for text summarization
Alexander R. Fabbri, New York, NY (US); Prafulla Kumar Choubey, San Jose, CA (US); Jesse Vig, Los Altos, CA (US); Chien-Sheng Wu, Mountain View, CA (US); and Caiming Xiong, Menlo Park, CA (US)
Assigned to Salesforce, Inc., San Francisco, CA (US)
Filed by Salesforce, Inc., San Francisco, CA (US)
Filed on Oct. 6, 2022, as Appl. No. 17/938,572.
Claims priority of provisional application 63/355,323, filed on Jun. 24, 2022.
Prior Publication US 2023/0419017 A1, Dec. 28, 2023
Int. Cl. G06F 17/00 (2019.01); G06F 40/166 (2020.01); G06F 40/284 (2020.01); G06N 20/00 (2019.01)
CPC G06F 40/166 (2020.01) [G06F 40/284 (2020.01); G06N 20/00 (2019.01)] 23 Claims
OG exemplary drawing
 
1. A method for text summarization, the method comprising:
receiving, via a communication interface, a training dataset comprising at least an uncompressed text, a compressed text, and one or more information entities that are not mentioned in the compressed text but included in the uncompressed text;
generating, by a perturber model, a perturbed text based on inserting at least a first information entity from the one or more information entities into the compressed text;
training the perturber model based on a first training objective comparing the perturbed text generated from the compressed text and the at least first information entity, and the uncompressed text including the one or more information entities;
generating, by the trained perturber model, a perturbed summary in response to an input of a reference summary and at least a second information entity randomly selected from a corresponding source document of the reference summary;
generating, via an editor model, a predicted summary by removing information from the perturbed summary conditioned on the source document of the reference summary; and
training the editor model based on a second training objective of a cross-entropy loss between a predicted token distribution of the predicted summary conditioned on the source document and a token distribution of the reference summary.