Abstract:
A method includes fusing multi-modal sensor data from a plurality of sensors having different modalities. At least one region of interest is detected in the multi-modal sensor data. One or more patches of interest are detected in the multi-modal sensor data based on detecting the at least one region of interest. A model that uses a deep convolutional neural network is applied to the one or more patches of interest. Post-processing of a result of applying the model is performed to produce a post-processing result for the one or more patches of interest. A perception indication of the post-processing result is output.
Abstract:
A method includes detecting at least one region of interest in a frame of image data. One or more patches of interest are detected in the frame of image data based on detecting the at least one region of interest. A model including a deep convolutional neural network is applied to the one or more patches of interest. Post-processing of a result of applying the model is performed to produce a post-processing result for the one or more patches of interest. A visual indication of a classification of defects in a structure is output based on the result of the post-processing.
Abstract:
A method includes converting time-series data from a plurality of prognostic and health monitoring (PHM) sensors into frequency domain data. One or more portions of the frequency domain data are labeled as indicative of one or more target modes to form labeled target data. A model including a deep neural network is applied to the labeled target data. A result of applying the model is classified as one or more discretized PHM training indicators associated with the one or more target modes. The one or more discretized PHM training indicators are output.