摘要:
There is disclosed, a high speed comparator with bit-wise masking takes advantage of early availability of the reference word and mask word to generate conditional select signals, thereby minimizing the time required to generate a comparator output once the input word to be compared is available.
摘要:
Apparatus for arbitrating the selection of an interrupt for servicing from a plurality of interrupts in which a priority level for each of the plurality of interrupts is programmed in a first register and each of the interrupts which is to be evaluated for selection for servicing is set as pending in a second register. Only a pending interrupt having a priority level above a pre-set current interrupt priority level is selected for servicing and where multiple pending interrupts of the same priority level occur, the one with the highest order bit position in the second register is used.
摘要:
A data processing system and method of communicating data in a data processing system are described. The data processing system includes a communication network to which a plurality of devices are coupled. At least one device among the plurality of devices coupled to the communication network includes mastering circuitry and snooping circuitry. According to the method, a first timing signal having a first frequency and a second timing signal having a second frequency different from the first frequency are generated. Communication transactions on the communication network are initiated utilizing the mastering circuitry, which operates in response to the first timing signal, and are monitored utilizing the snooping circuitry, which operates in response to the second timing signal.
摘要:
A method and system for front-end gathering of store instructions within a processor is disclosed. In accordance with the method and system of the present invention, a store queue within a data-processing system includes a front-end queue and a back-end queue. Multiple entries are provided in the back-end queue, and each entry includes an address field, a byte-count field, and a data field. A determination is first made as to whether or not a data field of a first entry of the front-end queue is filled completely. In response to a determination that the data field of the first entry of the front-end queue is not filled completely, another determination is made as to whether or not an address for a store instruction in a subsequent second entry is equal to an address for the store instruction in the first entry plus a byte count in the first entry. In response to a determination that the address for the store instruction in the subsequent second entry is equal to the address for the store instruction in the first entry plus the byte count in the first entry, the store instruction in the subsequent second entry is collapsed into the store instruction in the first entry.
摘要:
A method and system for controlling access to a shared resource in a data processing system are described. According to the method, a number of requests for access to the resource are generated by a number of requesters that share the resource. Each of the requesters is associated with a priority weight that indicates a probability that the associated requester will be assigned a highest current priority. Each requester is then assigned a current priority that is determined substantially randomly with respect to previous priorities of the requesters. In response to the current priorities of the requesters, a request for access to the resource is granted. In one embodiment, a requester corresponding to a granted request is signaled that its request has been granted, and a requester corresponding to a rejected request is signaled that its request was not granted.
摘要:
A method and system for allocating data among cache memories within a symmetric multiprocessor data-processing system are disclosed. The symmetric multiprocessor data-processing system includes a system memory and multiple processing units, wherein each of the processing units has a cache memory. The system memory is divided into a number of segments, wherein the number of segments is equal to the total number of cache memories. Each of these segments is represented by one of the cache memories such that a cache memory is responsible to cache data from its associated segment within the system memory.
摘要:
There is disclosed a converter for summing inputs includes first, second, third and fourth adders. Each of the adders is adapted to receive a carry-in and to provide as outputs a carry-out and a sum output. Each adder has an associated partial product generated in proximity thereto. The adders are interconnected such that the associated partial product of the first and second adders provide additional inputs to the first adder. The associated partial products of the third and fourth adder provide additional inputs to the second adder. The sum output of the first and second adders provide additional input to the third adder. A sum input to the converter and the sum output of the third adder provide additional inputs to the fourth adder. The output of the fourth adder is the output of the converter.
摘要:
A process for the low pressure chemical vapor deposition of silicon nitride from ammonia and a silane of the formula: (t-C.sub.4 H.sub.9 NH).sub.2 SiH.sub.2 provides improved properties of the resulting film for use in the semiconductor industry.
摘要翻译:用于氨的氮化硅的低压化学气相沉积和式(t-C 4 H 9 N H)2 SiH 2的硅烷的方法提供了用于半导体工业的所得膜的改进的性能。
摘要:
A periodic wireless data transmission has improved access latency obtaining information regarding users' interest in the information and by arranging the information on an transmission in order of descending popularity. In one embodiment, each adjacent pair of topics on the transmission. The topics' positions on the transmission are exchanged if the exchange decreases the average latency for all users. This may be repeated for all of the topics on the transmission. The transmission structure may also be arranged so that the transmission combines a number of transmission channels to obtain greater aggregate capacity. This may be done using, for example, an FDMA transmission structured to be theoretically comparable to a plurality of "striped" disks known as a RAID (redundant array of inexpensive disks).
摘要:
A data processing system includes a processor, a system memory, one or more input/output channel controllers (IOCC), and a system bus connecting the processor, the memory and the IOCCs together for communicating instructions, address and data between the various elements of a system. The IOCC includes a paged cache storage having a number of lines wherein each line of the page may be, for example, 32 bytes. Each page in the cache also has several attribute bits for that page including the so called WIM and attribute bits. The W bit is for controlling write through operations; the I bit controls cache inhibit; and the M bit controls memory coherency. Since the IOCC is unaware of these page table attribute bits for the cache lines being DMAed to system memory, IOCC must maintain memory consistency and cache coherency without sacrificing performance. For DMA write data to system memory, new cache attributes called global, cachable and demand based write through are created. Individual writes within a cache line are gathered by the IOCC and only written to system memory when the I/O bus master accesses a different cache line or relinquishes the I/O bus.