Abstract:
In some aspects, a system may receive, from a first one-dimensional radar array, first information based at least in part on first reflections associated with an azimuthal plane. The system may further receive, from a second one-dimensional radar array, second information based at least in part on second reflections associated with an elevation plane. Accordingly, the system may detect an object based at least in part on the first information and may determine an elevation associated with the object based at least in part on the second information. Numerous other aspects are described.
Abstract:
Lempel-Ziv (LZ)-based data compression employing implicit variable-length distance coding is disclosed. Distances in LZ-based data compression length and distance blocks are implicit variable-length encoded during data compression to avoid padding encoded distances with extra bits (e.g., trailing 0's) that require fewer bits for storage than the number of bits needed to store maximum distance length. This reduces distance bit lengths in compressed output data to further reduce data size. During data compression, a distance table is generated that contains entries each having an assigned base and a number of extra bits to be read in compressed data during data decompression. In this manner, during data decompression, the entries in the distance table can be consulted to determine the number of bits in the variable-encoded distance in the compressed data to be read since the encoded distances can be encoded in the compressed data in fewer bits and without bit padding.
Abstract:
A device for processing image data is disclosed. The device can obtain a radar point cloud and one or more frames of camera data. The device can determine depth estimates of one or more pixels of the one or more frames of camera data. The device can generate a pseudo lidar point cloud using the depth estimates of the one or more pixels of the one or more frames of camera data, wherein the pseudo lidar point cloud comprises a three-dimensional representation of at least one frame of the one or more frames of camera data. The device can determine one or more object bounding boxes based on the radar point cloud and the pseudo lidar point cloud.
Abstract:
A method of exploiting activation sparsity in deep neural networks is described. The method includes retrieving an activation tensor and a weight tensor where the activation tensor is a sparse activation tensor. The method also includes generating a compressed activation tensor comprising non-zero activations of the activation tensor, where the compressed activation tensor has fewer columns than the activation tensor. The method further includes processing the compressed activation tensor and the weight tensor to generate an output tensor.
Abstract:
This disclosure provides systems, methods, and devices for image signal processing that support multi-source pose merging for depth estimation. In a first aspect, a method of image processing includes generating, in accordance with first image data of a first image frame and second image data of a second image frame, a first mask indicating one or more pixels determined not to change position between the first image frame and the second image frame, generating, in accordance with the first image data and the second image data, a second mask indicating one or more pixels determined not to change position between the first image frame and the second image frame, and combining the first mask with the second mask to generate a third mask. Other aspects and features are also claimed and described.
Abstract:
Aspects for generating compressed data streams with lookback pre-fetch instructions are disclosed. A data compression system is provided and configured to receive and compress an uncompressed data stream as part of a lookback-based compression scheme. The data compression system determines if a current data block was previously compressed. If so, the data compression system is configured to insert a lookback instruction corresponding to the current data block into the compressed data stream. Each lookback instruction includes a lookback buffer index that points to an entry in a lookback buffer where decompressed data corresponding to the data block will be stored during a separate decompression scheme. Once the data blocks have been compressed, the data compression system is configured to move a lookback buffer index of each lookback instruction in the compressed data stream into a lookback pre-fetch instruction located earlier than the corresponding lookback instruction in the compressed data stream.
Abstract:
Multi-processor core three-dimensional (3D) integrated circuits (ICs) (3DICs) and related methods are disclosed. In aspects disclosed herein, ICs are provided that include a central processing unit (CPU) having multiple processor cores (“cores”) to improve performance. To further improve CPU performance, the multiple cores can also be designed to communicate with each other to offload workloads and/or share resources for parallel processing, but at a communication overhead associated with passing data through interconnects which have an associated latency. To mitigate this communication overhead inefficiency, aspects disclosed herein provide the CPU with its multiple cores in a 3DIC. Because 3DICs can overlap different IC tiers and/or align similar components in the same IC tier, the cores can be designed and located between or within different IC tiers in a 3DIC to reduce communication distance associated with processor core communication to share workload and/or resources, thus improving performance of the multi-processor CPU design.
Abstract:
Aspects include computing devices, systems, and methods for implementing generating a cache memory configuration. A server may apply machine learning to context data. The server may determine a cache memory configuration relating to the context data for a cache memory of a computing device and predict execution of an application on the computing device. Aspects include computing devices, systems, and methods for implementing configuring a cache memory of the computing device. The computing device may classify a plurality of cache memory configurations, related to a predicted application execution, based on at least a hardware data threshold and a first hardware data. The computing device may select a first cache memory configuration from the plurality of cache memory configurations in response to the first cache memory configuration being classified for the first hardware data, and configuring the cache memory at runtime based on the first cache memory configuration.
Abstract:
Example systems and techniques are described for training a machine learning model. A system includes memory configured to store image data captured by a plurality of cameras and one or more processors communicatively coupled to the memory. The one or more processors are configured to execute a machine learning model on the image data, the machine learning model including a plurality of layers. The one or more processors are configured to apply a non-linear mapping function to output of one layer of the plurality of layers to generate depth data. The one or more processors are configured to train the machine learning model based on the depth data to generate a trained machine learning model.
Abstract:
The present disclosure describes methods, computer-readable media, and apparatuses for operating neural networks. For example, a first apparatus may receive a set of sparse weight vectors. The first apparatus may compress the set of sparse weight vectors to produce a compressed set of sparse weight vectors. The first apparatus may operate a neural network based on the compressed set of sparse weight vectors. In another example, a second apparatus may receive a set of sparse weight vectors. The second apparatus may perform a sparse computation based on the set of sparse weight vectors, and the performance of the sparse computation may produce one or more partial sums. The second apparatus may operate a neural network based at least in part on the one or more partial sums.