VISUALIZING FEATURE VARIATION EFFECTS ON COMPUTER MODEL PREDICTION

    公开(公告)号:US20220405299A1

    公开(公告)日:2022-12-22

    申请号:US17743173

    申请日:2022-05-12

    Abstract: A model visualization system analyzes model behavior to identify clusters of data instances with similar behavior. For a selected feature, data instances are modified to set the selected feature to different values evaluated by a model to determine corresponding model outputs. The feature values and outputs may be visualized in an instance-feature variation plot. The instance-feature variation plots for the different data instances may be clustered to identify latent differences in behavior of the model with respect to different data instances when varying the selected feature. The number of clusters for the clustering may be automatically determined, and the clusters may be further explored by identifying another feature which may explain the different behavior of the model for the clusters, or by identifying outlier data instances in the clusters.

    IDENTIFYING AND MITIGATING DISPARATE GROUP IMPACT IN DIFFERENTIAL-PRIVACY MACHINE-LEARNED MODELS

    公开(公告)号:US20230385443A1

    公开(公告)日:2023-11-30

    申请号:US18202435

    申请日:2023-05-26

    CPC classification number: G06F21/6245

    Abstract: A model evaluation system evaluates the extent to which privacy-aware training processes affect the direction of training gradients for groups. A modified differential-privacy (“DP”) training process provides per-sample gradient adjustments with parameters that may be adaptively modified for different data batches. Per-sample gradients are modified with respect to a reference bound and a clipping bound. A scaling factor may be determined for each per-sample gradient based on the higher of the reference bound or a magnitude of the per-sample gradient. Per-sample gradients may then be adjusted based on a ratio of the clipping bound to the scaling factor. A relative privacy cost between groups may be determined as excess training risk based on a difference in group gradient direction relative to an unadjusted batch gradient and the adjusted batch gradient according to the privacy-aware training.

    IDENTIFYING AND MITIGATING DISPARATE GROUP IMPACT IN DIFFERENTIAL-PRIVACY MACHINE-LEARNED MODELS

    公开(公告)号:US20230385444A1

    公开(公告)日:2023-11-30

    申请号:US18202440

    申请日:2023-05-26

    CPC classification number: G06F21/6245

    Abstract: A model evaluation system evaluates the extent to which privacy-aware training processes affect the direction of training gradients for groups. A modified differential-privacy (“DP”) training process provides per-sample gradient adjustments with parameters that may be adaptively modified for different data batches. Per-sample gradients are modified with respect to a reference bound and a clipping bound. A scaling factor may be determined for each per-sample gradient based on the higher of the reference bound or a magnitude of the per-sample gradient. Per-sample gradients may then be adjusted based on a ratio of the clipping bound to the scaling factor. A relative privacy cost between groups may be determined as excess training risk based on a difference in group gradient direction relative to an unadjusted batch gradient and the adjusted batch gradient according to the privacy-aware training.

Patent Agency Ranking