US 11,811,794 B2
Privacy interface for data loss prevention via artificial intelligence models
Gabriel Gabra Zaccak, Cambridge, MA (US); William Hartman, Nantucket, MA (US); Andrés Rodriguez Esmeral, Vancouver (CA); Devin Daniel Reich, Olympia, WA (US); Marina Titova, Menlo Park, CA (US); Brett Robert Redinger, Oakland, CA (US); Philip Joseph Dow, South Lake Tahoe, CA (US); Satish Srinivasan Bhat, Fremont, CA (US); Walter Adolf De Brouwer, Los Altos, CA (US); and Scott Michael Kirk, Belmont, CA (US)
Assigned to Sharecare AI, Inc., Palo Alto, CA (US)
Filed by Sharecare AI, Inc., Palo Alto, CA (US)
Filed on May 12, 2021, as Appl. No. 17/319,025.
Claims priority of provisional application 63/023,854, filed on May 12, 2020.
Prior Publication US 2021/0360010 A1, Nov. 18, 2021
Int. Cl. H04L 9/40 (2022.01); G06N 20/20 (2019.01)
CPC H04L 63/1416 (2013.01) [G06N 20/20 (2019.01)] 19 Claims
OG exemplary drawing
 
1. A system for preventing exfiltration of training data by feature reconstruction attacks on model instances trained on the training data during a training job, comprising: a privacy interface that presents a plurality of modulators for a plurality of training parameters; modulators in the plurality of modulators configured to respond to selection commands via the privacy interface to trigger procedural calls that modify corresponding training parameters in the plurality of training parameters for respective training cycles in the training job; a trainer configured to execute the training cycles in dependence on the modified training parameters, and determine a performance accuracy of the model instances for each of the executed training cycles; a differential privacy estimator configured to estimate a privacy guarantee for each of the executed training cycles in dependence on the modified training parameters; a feedback provider configured to visualize, on the privacy interface, the privacy guarantee, the performance accuracy, and the modified training parameters for each of the executed training cycles; and a susceptibility predictor that determines, in dependence on the modified training parameters, susceptibility of the model instances to the feature reconstruction attacks, including model inversion attacks, member inference attacks, and gradient leakage attacks.