摘要:
Described are techniques for processing a request to access global memory. For a first processor included on a first of a plurality of boards connected by a fabric, a logical address is determined for a global memory location in a system global memory. A first physical address for the logical address is determined. It is determined whether the first physical address is included in a first global partition of the first board. If so, first processing is performed including updating a memory map to map a window of the first processor's logical address space to a physical memory segment located within the first global partition. Otherwise, if the first physical address is included in a second of the plurality of global partitions physically located on one of the plurality of boards other than said first board, second processing is performed to issue the request over the fabric.
摘要:
Described are techniques for partitioning memory. A plurality of boards is provided. Each of the plurality of boards includes a physical memory portion and a set of one or more processor. The physical memory portion in each of said plurality of boards is partitioned into a plurality of logical partitions including a global memory partition accessible by any processor on any of the plurality of boards and one or more other memory partitions configured for use by one or more processors of said each board. Each of the one or more other memory partitions not being accessible to a processor on a board other than said each board.
摘要:
Detecting data errors in connection with a data transfer process includes performing an XOR operation on a plurality of data blocks to obtain a data block XOR result. An XOR operation may be performed on a plurality of cyclic redundancy check (CRC) codes associated with the plurality of data blocks to obtain a CRC XOR result. The data block XOR result and the CRC XOR result may be used to determine whether an error exists in the plurality of data blocks. The system may be used in connection with local IO transfers and in connection with local CPU XOR operations for a RAID system in which data may be mirrored, striped or otherwise distributed across multiple storage devices.
摘要:
An apparatus and method are provided for connecting a host Enterprise System Connection Architecture (ESCON) Input/Output (I/O) interface to a cache of a data storage system. The apparatus includes (a) a set of at least 4 pipelines, each pipeline being coupled on a first end to the host ESCON I/O interface and being coupled on a second end to the cache, (b) a plurality of line processors, each line processor controlling one or more of the pipelines of the set of pipelines, and (c) in each pipeline, a protocol engine, the protocol engine configured to distinguish user data from frame header data and separate the user data from the frame header data for transport over the pipeline.
摘要翻译:提供了一种用于将主机企业系统连接体系结构(ESCON)输入/输出(I / O)接口连接到数据存储系统的缓存的装置和方法。 该装置包括(a)一组至少4条流水线,每条流水线在第一端耦合到主机ESCON I / O接口,并在第二端耦合到高速缓存,(b)多个线路处理器, 每行处理器控制所述一组管道中的一个或多个管道,以及(c)在每个流水线中,协议引擎,所述协议引擎被配置为将用户数据与帧头数据区分开并将所述用户数据与所述帧头数据分开 用于运输管道。
摘要:
A shared resources service processor facilitates messaging between line processors and provides a single point of contact for a user interfacing with line processor(s), for example in a storage system interface. Shared memory is divided into “mailboxes” that are used to communicate between the line processors and the service processor. The service processor issues a system management interrupt to any or all of the line processors. This interrupt indicates to the line processor(s) that it should go out to the shared memory and read its respective mailbox. In operation, the service processor can deliver a message, i.e. command, to a line processor's mailbox, for example to tell a line processor to go off-line or on-line. The service processor will write the command into the mailbox and then assert the system management interrupt on the appropriate line processor that it wants to read the mailbox. The line processor receiving the interrupt will take the interrupt vector, read the mailbox, interpret the command and deliver the appropriate response to the mailbox. The line processor(s) conserves backplane bandwidth by selectively consolidating selected tasks onto the service processor to reduce the number of accesses to the backplane.
摘要:
An apparatus and method are provided for connecting a host Enterprise System Connection Architecture (ESCON) Input/Output (I/O) interface to a cache of a data storage system. The apparatus includes (a) a set of at least 4 pipelines, each pipeline being coupled on a first end to the host ESCON I/O interface and being coupled on a second end to the cache, (b) a plurality of line processors, each line processor controlling one or more of the pipelines of the set of pipelines, and (c) in each pipeline, a protocol engine, the protocol engine configured to distinguish user data from frame header data and separate the user data from the frame header data for transport over the pipeline.
摘要翻译:提供了一种用于将主机企业系统连接体系结构(ESCON)输入/输出(I / O)接口连接到数据存储系统的缓存的装置和方法。 该装置包括(a)一组至少4条流水线,每条流水线在第一端耦合到主机ESCON I / O接口,并在第二端耦合到高速缓存,(b)多个线路处理器, 每行处理器控制所述一组管道中的一个或多个管道,以及(c)在每个流水线中,协议引擎,所述协议引擎被配置为将用户数据与帧头数据区分开并将所述用户数据与所述帧头数据分开 用于运输管道。