BOOSTING CLASSIFICATION AND REGRESSION TREE PERFORMANCE WITH DIMENSION REDUCTION

    公开(公告)号:US20230186107A1

    公开(公告)日:2023-06-15

    申请号:US17550551

    申请日:2021-12-14

    IPC分类号: G06N5/00 G06K9/62

    CPC分类号: G06N5/003 G06K9/6257

    摘要: A system and method can be provided for constructing and training a decision tree for machine learning. A training set can be received. The decision tree can be initialized by constructing a root node and a root solver can be trained with the training set. A processor can grow the decision tree by iteratively splitting nodes of the decision tree, where at a node of the decision tree, dimension reduction is performed on features of data of the training set received at the node, and the data having reduced dimension is split based on a routing function, for routing to another node of the decision tree. The dimension reduction and the split can be performed together at the node based on solving a nonlinear optimization problem.

    MULTI-POLYTOPE MACHINE FOR CLASSIFICATION

    公开(公告)号:US20230128821A1

    公开(公告)日:2023-04-27

    申请号:US17491494

    申请日:2021-09-30

    IPC分类号: G06N20/10

    摘要: A computer implemented method of generating a classifier engine for machine learning includes receiving a set of data points. A semi-supervised k-means process is applied to the set of data points from each class. The set of data points in a class is clustered into multiple clusters of data points, using the semi-supervised k-means process. Multi-polytopes are constructed for one or more of the clusters from all classes. A support vector machine (SVM) process is run on every pair of clusters from all classes. Separation hyperplanes are determined for the clustered classes. Labels are determined for each cluster based on the separation by hyperplanes.

    TRAINING NEURAL NETWORKS WITH CONVERGENCE TO A GLOBAL MINIMUM

    公开(公告)号:US20240119274A1

    公开(公告)日:2024-04-11

    申请号:US17951587

    申请日:2022-09-23

    发明人: Lam Minh Nguyen

    IPC分类号: G06N3/08

    CPC分类号: G06N3/08

    摘要: Select an initial weight vector for a convex optimization sub-problem associated with a neural network having a non-convex network architecture loss surface. With at least one processor, approximate a solution to the convex optimization sub-problem that obtains a search direction, to learn a common classifier from training data. With the at least one processor, update the initial weight vector by subtracting the approximate solution to the convex optimization sub-problem times a first learning rate. With the at least one processor, repeat the approximating and updating steps, for a plurality of iterations, with the updated weight vector from a given one of the iterations taken as the initial weight vector for a next one of the iterations, to obtain a final weight vector for the neural network, until convergence to a global minimum is achieved, to implement the common classifier.

    INTERPRETABLE CLUSTERING VIA MULTI-POLYTOPE MACHINES

    公开(公告)号:US20230267339A1

    公开(公告)日:2023-08-24

    申请号:US17675202

    申请日:2022-02-18

    IPC分类号: G06N5/02

    CPC分类号: G06N5/022

    摘要: In unsupervised interpretable machine learning, one or more datasets having multiple features can be received. A machine can be trained to jointly cluster and interpret resulting clusters of the dataset by at least jointly clustering the dataset into clusters and generating hyperplanes in a multi-dimensional feature space of the dataset, where the hyperplanes separate pairs of the clusters, where a hyperplane separates a pair of clusters. Jointly clustering the dataset into clusters and generating hyperplanes can repeat until convergence, where the clustering in a subsequent iteration uses the generated hyperplanes from a previous iteration to optimize performance of the clustering. The hyperplanes can be adjusted to further improve the performance of the clustering. The clusters and interpretation of the clusters can be provided, where a cluster's interpretation is provided based on hyperplanes that construct a polytope containing the cluster.

    SHUFFLING-TYPE GRADIENT METHOD FOR TRAINING MACHINE LEARNING MODELS WITH BIG DATA

    公开(公告)号:US20220171996A1

    公开(公告)日:2022-06-02

    申请号:US17109112

    申请日:2020-12-01

    IPC分类号: G06K9/62 G06N3/02 G06F17/16

    摘要: A computer-implemented method for a shuffling-type gradient for training a machine learning model using a stochastic gradient descent (SGD) includes the operations of uniformly randomly distributing data samples or coordinate updates of a training data, and calculating the learning rates for a no-shuffling scheme and a shuffling scheme. A combined operation of the no-shuffling scheme and the shuffling scheme of the training data is performed using a stochastic gradient descent (SGD) algorithm. The combined operation is switched to performing only the shuffling scheme from the no-shuffling scheme based on one or more predetermined criterion; and training the machine learning models with the training data based on the combined no-shuffling scheme and shuffling scheme.

    INTELLIGENT DYNAMIC CONDITION-BASED INFRASTRUCTURE MAINTENANCE SCHEDULING

    公开(公告)号:US20240103959A1

    公开(公告)日:2024-03-28

    申请号:US17949731

    申请日:2022-09-21

    IPC分类号: G06F11/07 G06F11/00

    CPC分类号: G06F11/0793 G06F11/008

    摘要: In example aspects of this disclosure, a method includes generating, by one or more computing devices, a parametric model that expresses condition states for each of a plurality of assets, and the probability of the assets transitioning between the condition states; generating, by the one or more computing devices, stochastic degradation predictions of a group of the assets, based on the condition states and the probability of transitioning between the condition states for at least some of the assets; and generating, by the one or more computing devices, a maintenance schedule based on: the stochastic degradation predictions of the group of the assets, costs of corrective maintenance for assets in a failed state, and costs of scheduled maintenance for the assets.