Transformer failure diagnosis method and system based on integrated deep belief network

    公开(公告)号:US12131247B2

    公开(公告)日:2024-10-29

    申请号:US17126067

    申请日:2020-12-18

    申请人: WUHAN UNIVERSITY

    IPC分类号: G06N3/047 G06F17/14 G06N3/08

    CPC分类号: G06N3/047 G06F17/14 G06N3/08

    摘要: A transformer failure diagnosis method and system based on an integrated deep belief network are provided. The disclosure relates to the fields of electronic circuit engineering and computer vision. The method includes the following: obtaining a plurality of vibration signals of transformers of various types exhibiting different failure types, retrieving a feature of each of the vibration signals, and establishing training data through the retrieved features; training a plurality of deep belief networks exhibiting different learning rates through the training data and obtaining a failure diagnosis correct rate of each of the deep belief networks; and keeping target deep belief networks corresponding to the failure diagnosis correct rates that satisfy requirements, building an integrated deep belief network through each of the target deep belief networks, and performing a failure diagnosis on the transformers through the integrated deep belief network.

    Methods and Systems for Quantifying Uncertainty in Neural Link Predictors for Knowledge Graphs

    公开(公告)号:US20240354595A1

    公开(公告)日:2024-10-24

    申请号:US18136463

    申请日:2023-04-19

    IPC分类号: G06N5/02 G06N3/047

    CPC分类号: G06N5/02 G06N3/047

    摘要: The present disclosure describes methods and systems for quantifying certainty for a prediction based on a knowledge graph. The method includes receiving a target triple and a knowledge graph comprising a set of structured data and a set of certainty scores for the structured data; converting the target triple to an embeddings space according to neighborhood sampling by a neural network, wherein the embeddings space includes a set of point coordinates; generating a plausibility prediction for the target triple using a scoring function; repeating converting the target triple to the embedding space and generating another plausibility prediction for the target triple N times with dropouts to obtain N plausibility scores for the target triple, wherein N is an integer larger than one; generating a predicted plausibility score and a certainty score for the target triple; and outputting the predicted plausibility score and the certainty score.

    ACTIVATION FUNCTION FOR HOMOMORPHICALLY-ENCRYPTED NEURAL NETWORKS

    公开(公告)号:US20240303471A1

    公开(公告)日:2024-09-12

    申请号:US18178684

    申请日:2023-03-06

    申请人: Intel Corporation

    IPC分类号: G06N3/047

    CPC分类号: G06N3/047

    摘要: Implementations herein disclose an activation function for homomorphically-encrypted neural networks. A data-agnostic activation technique is provided that collects information about the distribution of the most-dominant activated locations in the feature maps of the trained model and maintains a map of those locations. This map, along with a defined percent of random locations, decides which neurons in the model are activated using an activation function. Advantages of implementations herein include allowing for efficient activation function computations in encrypted computations of neural networks, yet no data-dependent computation is done during inference time (e.g., data-agnostic). Implementations utilize negligible overhead in model storage, while preserving the same accuracy as with general activation functions and runs in orders of magnitude faster than approximation-based activation functions. Furthermore, implementations herein can be applied post-hoc to already-trained models and, as such, do not utilize fine-tuning.

    Probabilistic neural network architecture generation

    公开(公告)号:US12079726B2

    公开(公告)日:2024-09-03

    申请号:US18107612

    申请日:2023-02-09

    摘要: Examples of the present disclosure describe systems and methods for probabilistic neural network architecture generation. In an example, an underlying distribution over neural network architectures based on various parameters is sampled using probabilistic modeling. Training data is evaluated in order to iteratively update the underlying distribution, thereby generating a probability distribution over the neural network architectures. The distribution is iteratively trained until the parameters associated with the neural network architecture converge. Once it is determined that the parameters have converged, the resulting probability distribution may be used to generate a resulting neural network architecture. As a result, intermediate architectures need not be fully trained, which dramatically reduces memory usage and/or processing time. Further, in some instances, it is possible to evaluate bigger architectures and/or larger batch sizes while also reducing neural network architecture generation time and maintaining or improving neural network accuracy.

    Method and apparatus for training image caption model, and storage medium

    公开(公告)号:US12073321B2

    公开(公告)日:2024-08-27

    申请号:US17075618

    申请日:2020-10-20

    IPC分类号: G06N3/08 G06N3/045 G06N3/047

    CPC分类号: G06N3/08 G06N3/045 G06N3/047

    摘要: Embodiments of this application disclose a method for training an image caption model, the image caption model including an encoding convolutional neural network (CNN) and a decoding recurrent neural network (RNN). The method includes: obtaining an image eigenvector of an image sample by using the encoding CNN; decoding the image eigenvector by using the decoding RNN, to obtain a sentence used for describing the image sample; determining a matching degree between the sentence obtained through decoding and the image sample and a smoothness degree of the sentence obtained through decoding, respectively; and adjusting the decoding RNN according to the matching degree and the smoothness degree.

    NEURAL GRAPH REVEALERS
    10.
    发明公开

    公开(公告)号:US20240281643A1

    公开(公告)日:2024-08-22

    申请号:US18313907

    申请日:2023-05-08

    IPC分类号: G06N3/047

    CPC分类号: G06N3/047

    摘要: The present disclosure relates to recovering a sparse feature graph based on input data having a collection of samples and associated features. In particular, the systems described herein utilize a fully connected neural network to learn a regression of the input data and determine direct connections between features of the input data while the neural network satisfies one or more sparsity constraints. This regression may be used to recover a feature graph indicating direct connections between the features of the input data. In addition, the feature graph may be presented via an interactive presentation that enables a user to navigate nodes and edges of the graph to gain insights of the input data and associated features.