US 12,223,430 B1
Empirical game theoretic system and method for adversarial decision analysis
Sean Soleyman, Calabasas, CA (US); Deepak Khosla, Camarillo, CA (US); Fan H. Hung, Woodland Hills, CA (US); and Samuel D. Johnson, Santa Monica, CA (US)
Assigned to HRL LABORATORIES, LLC, Malibu, CA (US)
Filed by HRL Laboratories, LLC, Malibu, CA (US)
Filed on Jul. 8, 2021, as Appl. No. 17/370,825.
Application 17/370,825 is a continuation in part of application No. 17/167,001, filed on Feb. 3, 2021, granted, now 12,061,673.
Application 17/167,001 is a continuation in part of application No. 17/064,566, filed on Oct. 6, 2020.
Application 17/064,566 is a continuation in part of application No. 16/792,869, filed on Feb. 17, 2020.
Claims priority of provisional application 63/107,199, filed on Oct. 29, 2020.
Claims priority of provisional application 63/027,829, filed on May 20, 2020.
Claims priority of provisional application 62/953,008, filed on Dec. 23, 2019.
Claims priority of provisional application 62/814,133, filed on Mar. 5, 2019.
Int. Cl. G06N 3/086 (2023.01); G05B 13/02 (2006.01); G05B 13/04 (2006.01)
CPC G06N 3/086 (2013.01) [G05B 13/027 (2013.01); G05B 13/04 (2013.01)] 22 Claims
OG exemplary drawing
 
1. A system for performing a simulated vehicle control task based on adversarial decision analysis, the system comprising:
one or more processors and a non-transitory computer-readable medium having executable instructions encoded thereon such that when executed, the one or more processors perform operations of:
performing a plurality of empirical game theoretic analyses between an evolving population of neural network strategies and an adversary population of neural network strategies, wherein each empirical game theoretic analysis comprises:
using a neuroevolution procedure, performing a fitness-based selection of a strategy in the evolving population of neural network strategies that out-performs the adversary population of neural network strategies; and
using an empirical game theory procedure, iteratively running the neuroevolution procedure and adding the selected strategy to the adversary population of neural network strategies with each iteration,
wherein addition of the selected strategy to the adversary population of neural network strategies results in monotonic strategy improvement with each iteration;
following the plurality of empirical game theoretic analyses, selecting a final strategy for the evolving population of neural network strategies and the adversary population of neural network strategies using a tournament selection procedure;
using the final strategy, training a neural network; and
using the trained neural network to perform a simulated vehicle control task.