US7676604B2 - Task context direct indexing in a protocol engine - Google Patents

Task context direct indexing in a protocol engine Download PDF

Info

Publication number
US7676604B2
US7676604B2 US11/285,825 US28582505A US7676604B2 US 7676604 B2 US7676604 B2 US 7676604B2 US 28582505 A US28582505 A US 28582505A US 7676604 B2 US7676604 B2 US 7676604B2
Authority
US
United States
Prior art keywords
task
context
request
index
memory
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Expired - Fee Related, expires
Application number
US11/285,825
Other versions
US20070118835A1 (en
Inventor
William Halleck
Victor Lau
Pak-Lung Seto
Naichih (Neil) Chang
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Intel Corp
Original Assignee
Intel Corp
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Intel Corp filed Critical Intel Corp
Priority to US11/285,825 priority Critical patent/US7676604B2/en
Assigned to INTEL CORPORATION reassignment INTEL CORPORATION ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: CHANG, NAICHIH (NEIL), HALLECK, WILLIAM, LAU, VICTOR, SETO, PAK-LUNG
Publication of US20070118835A1 publication Critical patent/US20070118835A1/en
Application granted granted Critical
Publication of US7676604B2 publication Critical patent/US7676604B2/en
Expired - Fee Related legal-status Critical Current
Adjusted expiration legal-status Critical

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F9/00Arrangements for program control, e.g. control units
    • G06F9/06Arrangements for program control, e.g. control units using stored programs, i.e. using an internal store of processing equipment to receive or retain programs
    • G06F9/46Multiprogramming arrangements
    • G06F9/48Program initiating; Program switching, e.g. by interrupt
    • G06F9/4806Task transfer initiation or dispatching
    • G06F9/4843Task transfer initiation or dispatching by program, e.g. task dispatcher, supervisor, operating system

Definitions

  • This disclosure relates to management of task context in a computer system.
  • SAS Serial Attached Small Computer Systems Interface
  • FCAL Fibre Channel Arbitrated Loop
  • a task represents the work associated with a command or group of linked commands.
  • a context associated with the task is typically stored in a data structure and includes information required to process the task. The information may include the type of command to be issued and an identifier identifying the storage device to which the command is to be issued.
  • One example of a task may be a command to write a file to a disk drive and the parameters of the command include a number of sectors to write on the disk drive.
  • each task request is typically assigned both an Input/Output (I/O) tag and a context identifier (ID) which are used to track the tasks.
  • the context ID is typically assigned by hardware in the storage controller and used by the hardware to associate tasks with specific resources in the storage controller.
  • the I/O tag is typically assigned by firmware in the storage controller and used by the firmware to identify the task.
  • the context ID may be unrelated to the I/O tag.
  • a large task request queue is required in the storage controller to store the task requests and their associated context for all task requests issued to the storage controller. Furthermore, a request to abort or suspend a task issued by firmware in the storage controller requires finding a context ID associated with the I/O tag prior to processing the request. A list of tasks in process must be searched in order to find one with a matching I/O tag. This requires additional logic which is dependent on the number of tasks in progress.
  • An alternative to searching through a list of tasks is to have the hardware in the storage controller return the context ID assigned to each task when it accepts a task from firmware.
  • the returned context ID is stored in a mapping table with the associated I/O tag. Instead of having to search through the entire list, the mapping table can be searched for a matching context ID.
  • the return of a context ID for each task increases the number of accesses to the hardware that are performed for each task. Instead of two accesses, one to issue the request and the second to accept completion status, a third access is required by firmware to retrieve the context ID and store it in the mapping table with the I/O tag.
  • the context ID is rarely used this third access is required for each task in case the context ID is needed by the firmware to abort or suspend a task.
  • FIG. 1 is a block diagram of a system including an input output processor (IOP) which manages tasks to be processed by at least one storage device according to an embodiment of the present invention
  • IOP input output processor
  • FIG. 2 is a block diagram of an embodiment of the IOP shown in FIG. 1 ;
  • FIG. 3 is a flow chart illustrating a method for issuing task requests implemented in the IOP shown in FIG. 2 ;
  • FIG. 4 is a flow chart illustrating a method for processing completed task requests implemented in the IOP shown in FIG. 2 ;
  • FIG. 5 is a flow chart illustrating a method for aborting issued tasks implemented in the IOP shown in FIG. 2 ;
  • FIG. 6 is a block diagram of an embodiment of a task context for a frame transmission task stored in a work queue in the work queue 216 shown in FIG. 2 ;
  • FIG. 7 is a block diagram of an embodiment of a context entry in context memory for storing context for the frame transmission task shown in FIG. 6 .
  • FIG. 1 is a block diagram of a system 100 including an input output processor (IOP) 104 which manages tasks to be processed by at least one storage device according to an embodiment of the present invention.
  • the system 100 includes a Host System 102 which may be coupled to the IOP 104 using a Peripheral Component Interconnect (PCI) interconnect 108 , for example, an interconnect such as PCI Express (PCIe).
  • PCIe Peripheral Component Interconnect
  • PCI Express PCI Express
  • the Host system 102 may include a Central Processing Unit (CPU), a Host Chip set (MCH) and memory.
  • CPU Central Processing Unit
  • MCH Host Chip set
  • the IOP 104 receives task requests from a device driver in an operating system that executes in memory in a Host System 102 for a storage device 106 coupled to the IOP 104 .
  • a task represents the work associated with a command or group of linked commands.
  • the IOP 104 processes each task request that it receives, stores a context associated with the task request together with the task request in a queue of tasks and returns status of the task request when completed by the storage device 106 .
  • the IOP 104 may communicate with the storage device 106 over a storage protocol interconnect— 110 using a serial attached storage protocol.
  • a serial attached storage protocol communication is between an initiator (originator) and a target (responder).
  • the initiator may be a storage protocol controller such as a Host Bus Adapter (HBA) and the target may be a storage device, for example, a disk drive, Digital Video Disk (DVD) drive, compact disk (CD) drive, Redundant Array of Independent Disks (RAID), or tape drive.
  • HBA Host Bus Adapter
  • Commands, data and status information encapsulated in frames are exchanged between the initiator and the target, over storage protocol interconnect 110 using standard serial attached storage protocol suites.
  • Fibre Channel protocol FCP
  • SAS Serial Attached Small Computer System Interface
  • SATA Serial Advanced Technology Attachment
  • SATA Serial ATA: High Speed Serialized AT Attachment
  • SATA Serial ATA Working Group
  • SAS Standard Serial Attached Technology Attachment
  • STP Serial Attached Technology Attachment
  • SMP Serial Management Protocol
  • SSP Serial SCSI Protocol
  • FIG. 2 is a block diagram of an embodiment of the IOP 104 shown in FIG. 1 .
  • the IOP 104 includes a processor 200 , storage protocol engine 202 , context memory 204 , a host interface 208 and a memory subsystem.
  • the memory subsystem includes a memory controller 206 and memory 212 that is accessible by the processor.
  • the host interface 208 allows the host system 102 ( FIG. 1 ) and the processor 200 to communicate through message passing and interrupt generation.
  • the memory 212 in the memory subsystem may be Double Data Rate Synchronous Dynamic Random Access Ram (DDR SDRAM) and the processor 200 may be an Intel XScale® core.
  • DDR SDRAM Double Data Rate Synchronous Dynamic Random Access Ram
  • An engine is a processor or portion of a program that manages and manipulates data.
  • the storage protocol engine 202 manages one of a plurality of serial storage protocols, for example, SSP, FCP, Internet SCSI (iSCSI) or SATA.
  • the IOP 104 may have multiple storage protocol engines with each supporting a different storage protocol. All of the storage protocol engines may be enabled for processing frames at the same time or only one of the plurality of protocol processing engines may be enabled. In the embodiment shown in FIG. 2 , there is only one protocol processing engine 202 .
  • a work queue 216 in the storage protocol engine 202 and a context memory 204 are used to manage task context according to an embodiment of the present invention.
  • the task context may be used to identify where the results of a completed task are to be returned and may also include information needed to execute the task such as, the type of task to execute, information identifying a storage device to which the task is directed and other parameters of the task.
  • An embodiment of a task context is described later in conjunction with FIG. 6 .
  • the work queue 216 in the storage protocol engine 202 stores tasks to be processed by the storage protocol engine 202 .
  • the work queue is a First In First Out (FIFO) that can be accessed by the processor 200 via the internal bus. Tasks are added to the work queue 216 by the processor 200 and removed from the work queue by the storage protocol engine 202 .
  • Each entry in the work queue 216 includes a task request.
  • a task entry includes a task context index (TCI) or task identifier that is used by the storage protocol engine 202 for processing tasks on the task queue.
  • TCI task context index
  • An embodiment of the work queue entry stored on the work queue will be described later in conjunction with FIG. 6
  • the storage protocol engine 202 determines the number of context resources that are available to handle tasks. This number is stored in a context resources register 218 in the storage protocol engine 202 that is accessible to the processor 200 over the internal bus 210 .
  • the context resources register 218 may be one of a plurality of registers in the storage protocol engine that are accessible by the processor 200 .
  • the processor 200 executes firmware, that is, software routines that are stored in a memory 212 accessible by the processor 200 . While executing the firmware, the processor 200 accesses the context resources register 218 in the storage protocol engine 202 to determine the number of context entries in context memory 204 before adding a first task request to the work queue 216 in the storage protocol engine 202 . A free list of available TCIs is created within the range defined by the context resources register and this TCI free list 214 is stored in memory 212 .
  • a TCI is obtained from the free list of available TCIs stored in memory 212 and the task request that includes the associated TCI is added to an entry in the work queue 216 .
  • a scheduler 220 in the storage protocol engine 202 accesses the entries in the work queue 216 to schedule tasks to be processed.
  • the firmware may be referred to as the issuer of the task request because it adds the task request to the work queue and the storage protocol engine may be referred to as the processor of the task request.
  • the TCI is an index to a context entry 224 in context memory 204 that stores context associated with the task.
  • the context memory 204 is used exclusively for storing task context. In alternate embodiments, the context memory could also be used for storing other data or structures shared by the processor 200 and the storage protocol engine 202 .
  • the context resources register 218 in the storage protocol engine 202 stores the total number of context entries in the context memory 204 .
  • the context memory 204 is accessible by both the processor 200 over the internal bus 210 and by the storage protocol engine 202 over a separate bus 226 . Requests for access to the context memory from the processor 200 and the storage protocol engine 202 are arbitrated by a context memory controller that controls access to the context memory.
  • the TCI ranges from 0-(N ⁇ 1).
  • TCIs can be assigned to tasks in sequential order from 0-N.
  • tasks are added to the work queue 216 by the processor in FIFO order, tasks can be scheduled by the scheduler 220 in the storage protocol engine 202 in an order determined by a scheduler.
  • the scheduler 220 has an associated scheduler memory in the storage protocol engine that is neither visible nor directly accessible by the processor 200 .
  • the queued tasks will not necessarily complete in the same order that they were added to the work queue.
  • the TCI associated with the context entry of the completed task is added to the TCI free list 214 managed by the processor 200 so that the processor 200 can assign free context entries in the context memory 204 to new tasks.
  • task requests can be stored in a temporary buffer in memory 212 until one of the context entries in the context memory is freed and the TCI associated with the freed context entry is added to the TCI free list. This allows more than N tasks to be queued in the IOP although only N tasks can be pending in the storage protocol engine 202 .
  • a task flow for issuing a task request to the storage protocol engine 202 is described in conjunction with the flow diagram shown in FIG. 3 .
  • FIG. 3 will be described in conjunction with FIG. 2 .
  • the IOP waits for a task request.
  • the task request may be a new request or a waiting request that is stored in the IOP waiting for an empty context entry.
  • processing continues with block 302 .
  • the TCI free list 214 stored in memory 212 in the IOP 104 is checked to determine if there is an empty context entry in context memory.
  • a TCI in the TCI free list 214 indicates that there is an empty context entry indexed by the TCI that is available in the context memory 204 for storing context associated with a new task. If so, processing continues with block 304 . If not, in an embodiment, at block 308 , the task request is stored in a queue in memory 212 until a context entry in the context memory 204 is freed by a completed task.
  • a TCI is selected for the task from the TCI free list 214 stored in memory 212 in the IOP.
  • the TCI free list 214 is implemented as a First In First Out list of available TCIs.
  • the context associated with the task is written in a context entry identified by the provided TCI in the context memory 204 .
  • the task is issued to the storage protocol engine 202 with the associated TCI.
  • the task request and associated TCI is written to an entry in the work queue 216 in the storage protocol engine 202 and the work queue entry is marked valid.
  • the TCI is also stored in memory 212 together with information regarding the task in a list of tasks in progress. Processing continues with block 300 to wait for a new task request.
  • a scheduler 220 in the storage protocol engine 202 takes a task request from the work queue 216 and uses the TCI in the work queue entry to reference the context memory 204 to extract parameters required for the task from the context associated with the task that is identified by the TCI.
  • the scheduler 220 schedules the task to lower layers of the storage protocol, that is, the transport, link and physical layers that are handled by the storage protocol engine 202 .
  • the task Upon completion of the task, the task is placed on a status queue 222 with the associated TCI. For each completed task, the processor 200 returns the TCI to the TCI free list 214 when it no longer requires the context associated with the TCI.
  • FIG. 4 A task flow for handling completion of a task request is described in conjunction with the flow diagram shown in FIG. 4 .
  • FIG. 4 will be described in conjunction with FIG. 2 .
  • a status queue 222 in the storage protocol engine 202 is checked to determine if any tasks have completed. If so, processing continues with block 402 to handle the completed task.
  • the TCI is stored on the status queue 222 with the status of the completed task.
  • the status of the completed task may be successful, unsuccessful or aborted.
  • the TCI is returned to the TCI free list 214 to be used when adding a new task to the work queue 216 in the storage protocol engine.
  • FIG. 5 A task flow for handling a task abort is described in conjunction with the flow diagram shown in FIG. 5 .
  • FIG. 5 will be described in conjunction with FIG. 2 .
  • the TCI assigned to the task to be aborted prior to adding the task to the work queue 216 is obtained from a list of tasks in progress that are stored in memory 212 .
  • a command to abort the task along with the TCI associated with the task is added to the work queue 216 .
  • the status of the task abort command is added to the status queue and handled as described in conjunction with FIG. 4 .
  • FIG. 6 is a block diagram of an embodiment of a task context for a frame transmission task 600 stored in a work queue entry in the work queue 216 shown in FIG. 2 .
  • the frame transmission task includes a task type field 602 for storing a task type that indicates the type of task which in this case is a frame transmission task.
  • the task type may be a link specific task type, context management task type, link control task type, protocol engine control task type or other task type.
  • the task timeout field 603 stores a relative timeout value associated with the frame transmission task.
  • the frame transmission task also includes a protocol field 604 which identifies the serial storage protocol to be used by the storage protocol engine for the task.
  • the protocol may be SSP, STP/SATA, FC or other serial storage protocol.
  • the priority field 606 stores an indication of the priority, for example, high or low, of the task.
  • the TCI field 608 stores the TCI associated with the task that identifies the context entry in context memory for storing context for the task.
  • the Remote Node Index (RNI) field 610 and the Local Port Index (LPI) field 612 store indices to where information for the remote node and local port are stored.
  • the task control field 614 stores control information about the task, for example, whether the task is an initiator task or a target task and whether the type of operations that it involves, for example, read, write or read-write operations.
  • FIG. 7 is a block diagram of an embodiment of a context entry 700 in context memory for storing context for the frame transmission task shown in FIG. 6 .
  • the context includes information for task management that is stored in a task management portion 702 of the context entry.
  • the task management information includes information provided in the work queue entry, for example, task type 602 ( FIG. 6 ), protocol 604 ( FIG. 6 ) and TCI 608 ( FIG. 6 ).
  • the 10 description portion 704 of the context entry stores information about the frame transmission task such as the IO class (read, write or read/write), the block size, staffing Logical Block Address (LBA), transfer size and endian control (big or little).
  • the task completion status portion 706 of the context entry stores completion status for the task.
  • the hardware reserved processing portion 708 of the context entry stores hardware states.
  • the processor 200 can learn the available resources in the storage protocol engine 202 without requiring knowledge of the actual implementation. Furthermore, the use of the TCI reduces the handshaking commands required between the processor and the storage protocol engine for assigning a task over other methods which require passing a context identifier back to the processor for use in identifying the context associated with the task. The handshaking is reduced because the TCI is used by both the storage protocol engine and the processor to keep track of tasks that are in process.
  • the task context may be built into a context memory buffer associated with a TCI prior to issuing the task request to the storage protocol engine 202 .
  • a context memory buffer associated with a TCI prior to issuing the task request to the storage protocol engine 202 .
  • An embodiment of the invention has been described for a task request issued to a storage device. However, embodiments of the present invention are not limited in this respect. Embodiments of the invention may be used for other applications that require the management of a context associated with tasks being performed.
  • a computer usable medium may consist of a read only memory device, such as a CD ROM disk or conventional ROM devices, or a computer diskette, having a computer readable program code stored thereon.

Abstract

A method and apparatus for managing task context are provided. Upon initialization, a protocol engine provides context resources available for processing tasks to a task issuer. Based on available context resources, the task issuer creates and manages a free list of available task context indices and assigns an index to a task prior to storing task context in a context memory accessible to both the task issuer and the protocol engine and issuing the task to the protocol engine.

Description

FIELD OF THE INVENTION
This disclosure relates to management of task context in a computer system.
BACKGROUND
Storage communication protocols such as the Serial Attached Small Computer Systems Interface (SAS) and Fibre Channel Arbitrated Loop (FCAL) provide a connection-oriented service with acknowledged delivery. A connection is established between devices prior to transfer of data between them. In the event that the connection is terminated, the connection must be re-established prior to resuming transfer of the data. The process for establishing the connection may require an exchange of frames between the devices.
There may be many simultaneous active connections requiring the management of thousands of task requests that are typically issued by a device driver in the operating system to a storage controller. A task represents the work associated with a command or group of linked commands. A context associated with the task is typically stored in a data structure and includes information required to process the task. The information may include the type of command to be issued and an identifier identifying the storage device to which the command is to be issued. One example of a task may be a command to write a file to a disk drive and the parameters of the command include a number of sectors to write on the disk drive.
As the issued task requests may be queued prior to being processed, each task request is typically assigned both an Input/Output (I/O) tag and a context identifier (ID) which are used to track the tasks. The context ID is typically assigned by hardware in the storage controller and used by the hardware to associate tasks with specific resources in the storage controller. The I/O tag is typically assigned by firmware in the storage controller and used by the firmware to identify the task. The context ID may be unrelated to the I/O tag.
A large task request queue is required in the storage controller to store the task requests and their associated context for all task requests issued to the storage controller. Furthermore, a request to abort or suspend a task issued by firmware in the storage controller requires finding a context ID associated with the I/O tag prior to processing the request. A list of tasks in process must be searched in order to find one with a matching I/O tag. This requires additional logic which is dependent on the number of tasks in progress.
An alternative to searching through a list of tasks is to have the hardware in the storage controller return the context ID assigned to each task when it accepts a task from firmware. The returned context ID is stored in a mapping table with the associated I/O tag. Instead of having to search through the entire list, the mapping table can be searched for a matching context ID. However, the return of a context ID for each task increases the number of accesses to the hardware that are performed for each task. Instead of two accesses, one to issue the request and the second to accept completion status, a third access is required by firmware to retrieve the context ID and store it in the mapping table with the I/O tag. Furthermore, although the context ID is rarely used this third access is required for each task in case the context ID is needed by the firmware to abort or suspend a task.
BRIEF DESCRIPTION OF THE DRAWINGS
Features of embodiments of the claimed subject matter will become apparent as the following detailed description proceeds, and upon reference to the drawings, in which like numerals depict like parts, and in which:
FIG. 1 is a block diagram of a system including an input output processor (IOP) which manages tasks to be processed by at least one storage device according to an embodiment of the present invention;
FIG. 2 is a block diagram of an embodiment of the IOP shown in FIG. 1;
FIG. 3 is a flow chart illustrating a method for issuing task requests implemented in the IOP shown in FIG. 2;
FIG. 4 is a flow chart illustrating a method for processing completed task requests implemented in the IOP shown in FIG. 2;
FIG. 5 is a flow chart illustrating a method for aborting issued tasks implemented in the IOP shown in FIG. 2;
FIG. 6 is a block diagram of an embodiment of a task context for a frame transmission task stored in a work queue in the work queue 216 shown in FIG. 2; and
FIG. 7 is a block diagram of an embodiment of a context entry in context memory for storing context for the frame transmission task shown in FIG. 6.
Although the following Detailed Description will proceed with reference being made to illustrative embodiments of the claimed subject matter, many alternatives, modifications, and variations thereof will be apparent to those skilled in the art. Accordingly, it is intended that the claimed subject matter be viewed broadly, and be defined only as set forth in the accompanying claims.
DETAILED DESCRIPTION
FIG. 1 is a block diagram of a system 100 including an input output processor (IOP) 104 which manages tasks to be processed by at least one storage device according to an embodiment of the present invention. The system 100 includes a Host System 102 which may be coupled to the IOP 104 using a Peripheral Component Interconnect (PCI) interconnect 108, for example, an interconnect such as PCI Express (PCIe). A version of PCI Express is described in the PCI Express Specification, Revision 1.0a.—PCI Special Interest Group. The Host system 102 may include a Central Processing Unit (CPU), a Host Chip set (MCH) and memory.
The IOP 104 receives task requests from a device driver in an operating system that executes in memory in a Host System 102 for a storage device 106 coupled to the IOP 104. A task represents the work associated with a command or group of linked commands. The IOP 104 processes each task request that it receives, stores a context associated with the task request together with the task request in a queue of tasks and returns status of the task request when completed by the storage device 106.
The IOP 104 may communicate with the storage device 106 over a storage protocol interconnect—110 using a serial attached storage protocol. Typically, in a serial attached storage protocol, communication is between an initiator (originator) and a target (responder). The initiator may be a storage protocol controller such as a Host Bus Adapter (HBA) and the target may be a storage device, for example, a disk drive, Digital Video Disk (DVD) drive, compact disk (CD) drive, Redundant Array of Independent Disks (RAID), or tape drive.
Commands, data and status information encapsulated in frames are exchanged between the initiator and the target, over storage protocol interconnect 110 using standard serial attached storage protocol suites.
There are many standard serial attached storage protocol suites such as, Fibre Channel protocol (FCP), Serial Attached Small Computer System Interface (SAS) and Serial Advanced Technology Attachment (SATA). A version of the Fibre Channel (FC) standard is described in the American National Standards Institute (ANSI) Standard Fibre Channel Physical and Signaling Interface-2 (FC-FS-2) Aug. 9, 2005 Specification. A version of the Fibre Channel Protocol (FCP-3) standard which defines a mapping protocol for applying the Small Computer System Interface (SCSI) command set to Fibre Channel is described in Information technology—Fibre Channel Protocol for SCSI, Third Version (FCP-3) Revision 4, Sep. 13, 2005 American National Standards Institute (ANSI) (hereinafter termed the “FCP standard”). A version of the SATA protocol is described in “Serial ATA: High Speed Serialized AT Attachment,” Revision 1.0a, published on Jan. 7, 2003 by the Serial ATA Working Group (hereinafter termed the “SATA standard”). A version of the SAS protocol is described in “Information Technology—Serial Attached SCSI—1.1,” Working Draft American National Standard of International Committee For Information Technology Standards (INCITS) T10 Technical Committee, Project T10/1562-D, Revision 1, published Sep. 18, 2003, by ANSI (hereinafter termed the “SAS Standard”). The SAS protocol may comprise Serial Attached Technology Attachment (SATA) Tunneled Protocol (STP), Serial Management Protocol (SMP) and Serial SCSI Protocol (SSP).
FIG. 2 is a block diagram of an embodiment of the IOP 104 shown in FIG. 1. The IOP 104 includes a processor 200, storage protocol engine 202, context memory 204, a host interface 208 and a memory subsystem. The memory subsystem includes a memory controller 206 and memory 212 that is accessible by the processor. The host interface 208 allows the host system 102 (FIG. 1) and the processor 200 to communicate through message passing and interrupt generation. In an embodiment, the memory 212 in the memory subsystem may be Double Data Rate Synchronous Dynamic Random Access Ram (DDR SDRAM) and the processor 200 may be an Intel XScale® core.
An engine is a processor or portion of a program that manages and manipulates data. The storage protocol engine 202 manages one of a plurality of serial storage protocols, for example, SSP, FCP, Internet SCSI (iSCSI) or SATA. The IOP 104 may have multiple storage protocol engines with each supporting a different storage protocol. All of the storage protocol engines may be enabled for processing frames at the same time or only one of the plurality of protocol processing engines may be enabled. In the embodiment shown in FIG. 2, there is only one protocol processing engine 202.
A work queue 216 in the storage protocol engine 202 and a context memory 204 are used to manage task context according to an embodiment of the present invention. The task context may be used to identify where the results of a completed task are to be returned and may also include information needed to execute the task such as, the type of task to execute, information identifying a storage device to which the task is directed and other parameters of the task. An embodiment of a task context is described later in conjunction with FIG. 6.
The work queue 216 in the storage protocol engine 202 stores tasks to be processed by the storage protocol engine 202. In an embodiment, the work queue is a First In First Out (FIFO) that can be accessed by the processor 200 via the internal bus. Tasks are added to the work queue 216 by the processor 200 and removed from the work queue by the storage protocol engine 202. Each entry in the work queue 216 includes a task request. In addition to the task request, a task entry includes a task context index (TCI) or task identifier that is used by the storage protocol engine 202 for processing tasks on the task queue. An embodiment of the work queue entry stored on the work queue will be described later in conjunction with FIG. 6
Upon initialization, based on available context memory 204, the storage protocol engine 202 determines the number of context resources that are available to handle tasks. This number is stored in a context resources register 218 in the storage protocol engine 202 that is accessible to the processor 200 over the internal bus 210. The context resources register 218 may be one of a plurality of registers in the storage protocol engine that are accessible by the processor 200.
The processor 200 executes firmware, that is, software routines that are stored in a memory 212 accessible by the processor 200. While executing the firmware, the processor 200 accesses the context resources register 218 in the storage protocol engine 202 to determine the number of context entries in context memory 204 before adding a first task request to the work queue 216 in the storage protocol engine 202. A free list of available TCIs is created within the range defined by the context resources register and this TCI free list 214 is stored in memory 212.
To issue a task to the storage protocol engine 202, a TCI is obtained from the free list of available TCIs stored in memory 212 and the task request that includes the associated TCI is added to an entry in the work queue 216. A scheduler 220 in the storage protocol engine 202 accesses the entries in the work queue 216 to schedule tasks to be processed. In this embodiment, the firmware may be referred to as the issuer of the task request because it adds the task request to the work queue and the storage protocol engine may be referred to as the processor of the task request.
In an embodiment of the invention, the TCI is an index to a context entry 224 in context memory 204 that stores context associated with the task. In the embodiment shown, the context memory 204 is used exclusively for storing task context. In alternate embodiments, the context memory could also be used for storing other data or structures shared by the processor 200 and the storage protocol engine 202. The context resources register 218 in the storage protocol engine 202 stores the total number of context entries in the context memory 204. The context memory 204 is accessible by both the processor 200 over the internal bus 210 and by the storage protocol engine 202 over a separate bus 226. Requests for access to the context memory from the processor 200 and the storage protocol engine 202 are arbitrated by a context memory controller that controls access to the context memory.
In an embodiment in which the maximum number of context entries is N, the TCI ranges from 0-(N−1). In an embodiment, starting with all context entries in the context memory available, TCIs can be assigned to tasks in sequential order from 0-N. Although tasks are added to the work queue 216 by the processor in FIFO order, tasks can be scheduled by the scheduler 220 in the storage protocol engine 202 in an order determined by a scheduler. The scheduler 220 has an associated scheduler memory in the storage protocol engine that is neither visible nor directly accessible by the processor 200. Thus, the queued tasks will not necessarily complete in the same order that they were added to the work queue. As tasks complete, the TCI associated with the context entry of the completed task is added to the TCI free list 214 managed by the processor 200 so that the processor 200 can assign free context entries in the context memory 204 to new tasks.
When there are no available context entries, task requests can be stored in a temporary buffer in memory 212 until one of the context entries in the context memory is freed and the TCI associated with the freed context entry is added to the TCI free list. This allows more than N tasks to be queued in the IOP although only N tasks can be pending in the storage protocol engine 202.
A task flow for issuing a task request to the storage protocol engine 202 is described in conjunction with the flow diagram shown in FIG. 3. FIG. 3 will be described in conjunction with FIG. 2.
At block 300, the IOP waits for a task request. The task request may be a new request or a waiting request that is stored in the IOP waiting for an empty context entry. Upon receiving a task request, processing continues with block 302.
At block 302, the TCI free list 214 stored in memory 212 in the IOP 104 is checked to determine if there is an empty context entry in context memory. A TCI in the TCI free list 214 indicates that there is an empty context entry indexed by the TCI that is available in the context memory 204 for storing context associated with a new task. If so, processing continues with block 304. If not, in an embodiment, at block 308, the task request is stored in a queue in memory 212 until a context entry in the context memory 204 is freed by a completed task.
At block 304, a TCI is selected for the task from the TCI free list 214 stored in memory 212 in the IOP. In an embodiment the TCI free list 214 is implemented as a First In First Out list of available TCIs. The context associated with the task is written in a context entry identified by the provided TCI in the context memory 204.
At block 306, the task is issued to the storage protocol engine 202 with the associated TCI. To issue the task to the storage protocol engine, the task request and associated TCI is written to an entry in the work queue 216 in the storage protocol engine 202 and the work queue entry is marked valid. The TCI is also stored in memory 212 together with information regarding the task in a list of tasks in progress. Processing continues with block 300 to wait for a new task request.
A scheduler 220 in the storage protocol engine 202 takes a task request from the work queue 216 and uses the TCI in the work queue entry to reference the context memory 204 to extract parameters required for the task from the context associated with the task that is identified by the TCI. The scheduler 220 schedules the task to lower layers of the storage protocol, that is, the transport, link and physical layers that are handled by the storage protocol engine 202.
Upon completion of the task, the task is placed on a status queue 222 with the associated TCI. For each completed task, the processor 200 returns the TCI to the TCI free list 214 when it no longer requires the context associated with the TCI.
A task flow for handling completion of a task request is described in conjunction with the flow diagram shown in FIG. 4. FIG. 4 will be described in conjunction with FIG. 2.
At block 400, a status queue 222 in the storage protocol engine 202 is checked to determine if any tasks have completed. If so, processing continues with block 402 to handle the completed task.
At block 402, the TCI is stored on the status queue 222 with the status of the completed task. The status of the completed task may be successful, unsuccessful or aborted. The TCI is returned to the TCI free list 214 to be used when adding a new task to the work queue 216 in the storage protocol engine.
A task flow for handling a task abort is described in conjunction with the flow diagram shown in FIG. 5. FIG. 5 will be described in conjunction with FIG. 2.
At block 500, the TCI assigned to the task to be aborted prior to adding the task to the work queue 216 is obtained from a list of tasks in progress that are stored in memory 212.
At block 502, a command to abort the task along with the TCI associated with the task is added to the work queue 216. When the abort command is complete, the status of the task abort command is added to the status queue and handled as described in conjunction with FIG. 4.
FIG. 6 is a block diagram of an embodiment of a task context for a frame transmission task 600 stored in a work queue entry in the work queue 216 shown in FIG. 2. The frame transmission task includes a task type field 602 for storing a task type that indicates the type of task which in this case is a frame transmission task. The task type may be a link specific task type, context management task type, link control task type, protocol engine control task type or other task type. The task timeout field 603 stores a relative timeout value associated with the frame transmission task.
The frame transmission task also includes a protocol field 604 which identifies the serial storage protocol to be used by the storage protocol engine for the task. The protocol may be SSP, STP/SATA, FC or other serial storage protocol. The priority field 606 stores an indication of the priority, for example, high or low, of the task. The TCI field 608 stores the TCI associated with the task that identifies the context entry in context memory for storing context for the task. The Remote Node Index (RNI) field 610 and the Local Port Index (LPI) field 612 store indices to where information for the remote node and local port are stored. The task control field 614 stores control information about the task, for example, whether the task is an initiator task or a target task and whether the type of operations that it involves, for example, read, write or read-write operations.
FIG. 7 is a block diagram of an embodiment of a context entry 700 in context memory for storing context for the frame transmission task shown in FIG. 6. The context includes information for task management that is stored in a task management portion 702 of the context entry. The task management information includes information provided in the work queue entry, for example, task type 602 (FIG. 6), protocol 604 (FIG. 6) and TCI 608 (FIG. 6). The 10 description portion 704 of the context entry stores information about the frame transmission task such as the IO class (read, write or read/write), the block size, staffing Logical Block Address (LBA), transfer size and endian control (big or little). The task completion status portion 706 of the context entry stores completion status for the task. The hardware reserved processing portion 708 of the context entry stores hardware states.
Through the context resources register, the processor 200 can learn the available resources in the storage protocol engine 202 without requiring knowledge of the actual implementation. Furthermore, the use of the TCI reduces the handshaking commands required between the processor and the storage protocol engine for assigning a task over other methods which require passing a context identifier back to the processor for use in identifying the context associated with the task. The handshaking is reduced because the TCI is used by both the storage protocol engine and the processor to keep track of tasks that are in process.
Additionally, by sharing the same tag, that is, the TCI between the processor and the storage protocol engine, there is no need for a search mechanism (mapping table) as in prior art methods in order to match a context identifier assigned by the storage protocol engine with an I/O tag.
With the processor assigning context IDs prior to issuing the task to the storage protocol engine, the task context may be built into a context memory buffer associated with a TCI prior to issuing the task request to the storage protocol engine 202. Thus, there is no longer any need to include the entire task context along with the task request on the work queue, reducing the size of the work queue buffer. An embodiment of the invention has been described for a task request issued to a storage device. However, embodiments of the present invention are not limited in this respect. Embodiments of the invention may be used for other applications that require the management of a context associated with tasks being performed.
It will be apparent to those of ordinary skill in the art that methods involved in embodiments of the present invention may be embodied in a computer program product that includes a computer usable medium. For example, such a computer usable medium may consist of a read only memory device, such as a CD ROM disk or conventional ROM devices, or a computer diskette, having a computer readable program code stored thereon.
While embodiments of the invention have been particularly shown and described with references to embodiments thereof, it will be understood by those skilled in the art that various changes in form and details may be made therein without departing from the scope of embodiments of the invention encompassed by the appended claims.

Claims (13)

1. An apparatus comprising: a storage protocol engine that comprises a queue for storing in an entry in the queue both a task request and a task context index, the task request being to request a task; and a context memory coupled via a bus to the storage protocol engine and for storing in a respective entry in the context memory a context associated with the task request, the task context index identifying both the task and a context associated with the task to both an issuer of the task request and a processor of the task request; the issuer of the task request to write to the context memory; the context associated with the task request being stored in the context memory and including a protocol field that identifies a selected storage protocol from a plurality of storage protocols, the selected storage protocol to be used by the storage protocol engine for the task, the storage protocol engine and the context memory being comprised in an input/output (I/O) processor coupled to a host system, the host system including a central processing unit, the context associated with the task request also including the task context index, completion status for the task, and a task type field, the task type field identifying a selected task type for the task from a plurality of possible task types, the plurality of possible task types including: a frame transmission task type, a context management task type, and a protocol engine control task type; the task request, as stored in the queue, including a task timeout field, the task type field, and the protocol field, the task timeout field storing a relative timeout value for the task; the task context index being an index to the respective entry in the context memory that stores the context associated with the task request, the context memory being used exclusively for task context storage; a free list of task identifiers for identifying available context entries in the context memory, wherein the free list is managed by the issuer of the task request; and wherein the issuer of the task request writes the context directly to the context memory in the entry identified by the task context index prior to adding the task request to the queue.
2. The apparatus of claim 1, further comprising:
a register writable by the processor of the task and readable by the issuer of the task request for storing a maximum number of contexts that can be stored in the context memory.
3. The apparatus of claim 1, wherein the issuer of the task request assigns the task context index.
4. The apparatus of claim 1, wherein the task request includes the task to be processed.
5. A method for managing task context comprising: storing in an entry in a queue in a storage protocol engine both a task request and a task context index, the task request being to request a task; storing in a respective entry in a context memory a context associated with the task request, the context memory being coupled via a bus to the storage protocol engine; and the index identifying both the task request and the context associated with the task request to both an issuer of the task request and a processor of the task request; the issuer of the task request to write to the context memory; the context associated with the task request being stored in the context memory and including a protocol field that identifies a selected storage protocol from a plurality of storage protocols, the selected storage protocol to be used by the storage protocol engine for the task, the storage protocol engine and the context memory being comprised in an input/output (I/O) processor coupled to a host system, the host system including a central processing unit, the context associated with the task request also including the task context index, completion status for the task, and a task type field, the task type field identifying a selected task type for the task from a plurality of possible task types, the plurality of possible task types including: a frame transmission task type, a context management task type, and a protocol engine control task type; the task request, as stored in the queue, including a task timeout field, the task type field, and the protocol field, the task timeout field storing a relative timeout value for the task; the task context index being an index to the respective entry in the context memory that stores the context associated with the task request, the context memory being used exclusively for task context storage; wherein the issuer of the task request writes the context directly to the context memory in the entry identified by the task context index prior to adding the task request to the queue; and managing by the issuer of the task, a free list of task identifiers identifying available context entries in the context memory.
6. The method of claim 5, further comprising:
storing a maximum number of contexts that can be stored in the context memory in a register writable by the processor of the task and readable by the issuer of the task request.
7. The method of claim 5, wherein the issuer of the task request assigns the task context index.
8. The method of claim 5, wherein the task request includes the task to be processed.
9. A system comprising: a disk drive; a storage protocol engine and a context memory comprised in an input/output (I/O) processor coupled to the disk drive, the storage protocol engine being coupled via a bus to the context memory and comprising a queue for storing a task request, the task request including a command to be processed by the disk drive and a task context index, the I/O processor being coupled to a host system, the host system including a central processing unit; and a context memory for storing at an entry identified by the index a context associated with the task request, the index identifying both the task request and the context associated with the task request to both an issuer of the task request and the storage protocol engine; the issuer of the task request to write to the context memory; the context associated with the task request being stored in the context memory and including a protocol field that identifies a selected storage protocol from a plurality of storage protocols, the selected storage protocol to be used by the storage protocol engine for the task, the context associated with the task request also including the index, completion status for the task, and a task type field that identifies a selected task type for the task from a plurality of possible task types, the plurality of possible task types including: a flame transmission task type, a context management task type, and a protocol engine control task type; the task request, as stored in the queue, including a task timeout field, the task type field, and the protocol field, the task timeout field storing a relative timeout value for the task; the task context index being an index to the respective entry in the context memory that stores the context associated with the task request, the context memory being used exclusively for task context storage; wherein the issuer of the task request writes the context directly to the context memory in the entry identified by the task context index prior to adding the task request to the queue; and a free list of task identifiers for identifying available context entries in the context memory, wherein the free list is managed by the issuer of the task request.
10. The system of claim 9, wherein the storage protocol engine further comprises:
a register written by the processor of the task and readable by the issuer of the task request for storing a maximum number of contexts that can be stored in the context memory.
11. The system of claim 9, wherein the issuer of the task request assigns the task context index.
12. An article including a machine-accessible medium having associated information, wherein the associated information, when accessed, results in a machine performing: storing in a queue in a storage protocol engine both a task request and a task context index, the task request being to request a task; storing in an entry in a context memory identified by the index a context associated with the task request, the context memory being coupled via a bus to the storage protocol engine; and the index identifying both the task request and the context associated with the task request to both an issuer of the task request and a processor of the task request; the issuer of the task request to write to the context memory; the context associated with the task request being stored in the context memory and including a protocol field that identifies a selected storage protocol from a plurality of storage protocols, the selected storage protocol to be used by the storage protocol engine for the task, the storage protocol engine and the context memory being comprised in an input/output (I/O) processor coupled to a host system, the host system including a central processing unit, the context associated with the task request also including the task context index, completion status for the task, and a task type field that identifies a selected task type for the task from a plurality of possible task types, the plurality of possible task types including: a frame transmission task type, a context management task type, and a protocol engine control task type; the task request, as stored in the queue, including a task timeout field, the task type field, and the protocol field, the task timeout field storing a relative timeout value for the task; the task context index being an index to the respective entry in the context memory that stores the context associated with the task request, the context memory being used exclusively for task context storage; wherein the issuer of the task request writes the context directly to the context memory in the entry identified by the index prior to adding the task request to the queue; and a free list of task identifiers for identifying available context entries in the context memory, wherein the free list is managed by the issuer of the task request.
13. The article of claim 12, wherein the issuer of the task request assigns the task context index.
US11/285,825 2005-11-22 2005-11-22 Task context direct indexing in a protocol engine Expired - Fee Related US7676604B2 (en)

Priority Applications (1)

Application Number Priority Date Filing Date Title
US11/285,825 US7676604B2 (en) 2005-11-22 2005-11-22 Task context direct indexing in a protocol engine

Applications Claiming Priority (1)

Application Number Priority Date Filing Date Title
US11/285,825 US7676604B2 (en) 2005-11-22 2005-11-22 Task context direct indexing in a protocol engine

Publications (2)

Publication Number Publication Date
US20070118835A1 US20070118835A1 (en) 2007-05-24
US7676604B2 true US7676604B2 (en) 2010-03-09

Family

ID=38054896

Family Applications (1)

Application Number Title Priority Date Filing Date
US11/285,825 Expired - Fee Related US7676604B2 (en) 2005-11-22 2005-11-22 Task context direct indexing in a protocol engine

Country Status (1)

Country Link
US (1) US7676604B2 (en)

Cited By (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
CN107992537A (en) * 2017-11-24 2018-05-04 金蝶软件(中国)有限公司 Service attribute transmission method, device, computer equipment and storage medium
US10489188B2 (en) 2017-01-19 2019-11-26 Samsung Electronics Co., Ltd. Method for reducing interrupt latency in embedded systems

Families Citing this family (5)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP4622474B2 (en) * 2004-11-17 2011-02-02 横河電機株式会社 Field device and system using the same
US7809068B2 (en) * 2005-12-28 2010-10-05 Intel Corporation Integrated circuit capable of independently operating a plurality of communication channels
JP2008152470A (en) * 2006-12-15 2008-07-03 Hitachi Ltd Data processing system and semiconductor integrated circuit
US20120265962A1 (en) * 2011-04-17 2012-10-18 Anobit Technologies Ltd. High-performance sas target
CN107515795A (en) * 2017-09-08 2017-12-26 北京京东尚科信息技术有限公司 Multi-task parallel data processing method, device, medium and equipment based on queue

Citations (48)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4805107A (en) * 1987-04-15 1989-02-14 Allied-Signal Inc. Task scheduler for a fault tolerant multiple node processing system
US5555402A (en) 1988-07-26 1996-09-10 Database Excelleration Systems, Inc. A disk storage subsystem for interfacing with a parallel path, a nonvolatile media and a volatile storage medium
US5577224A (en) 1994-12-13 1996-11-19 Microsoft Corporation Method and system for caching data
US5727211A (en) * 1995-11-09 1998-03-10 Chromatic Research, Inc. System and method for fast context switching between tasks
US5781926A (en) 1996-05-20 1998-07-14 Integrated Device Technology, Inc. Method and apparatus for sub cache line access and storage allowing access to sub cache lines before completion of line fill
US5802340A (en) 1995-08-22 1998-09-01 International Business Machines Corporation Method and system of executing speculative store instructions in a parallel processing computer system
US5862406A (en) 1991-04-11 1999-01-19 Mitsubishi Denki Kabushiki Kaisha Array recording system reporting completion of writing data operation prior to the completion of writing redundancy data in alternative recording units
US5944816A (en) * 1996-05-17 1999-08-31 Advanced Micro Devices, Inc. Microprocessor configured to execute multiple threads including interrupt service routines
US6205467B1 (en) * 1995-11-14 2001-03-20 Advanced Micro Devices, Inc. Microprocessor having a context save unit for saving context independent from interrupt requests
US6418489B1 (en) * 1999-10-25 2002-07-09 Motorola, Inc. Direct memory access controller and method therefor
US6430593B1 (en) * 1998-03-10 2002-08-06 Motorola Inc. Method, device and article of manufacture for efficient task scheduling in a multi-tasking preemptive priority-based real-time operating system
US20030021239A1 (en) 2001-07-26 2003-01-30 Mullendore Rodney N. Method and apparatus for scheduling packet flow on a fibre channel arbitrated loop
US20030110325A1 (en) 2001-12-10 2003-06-12 Bradley Roach Direct memory access (DMA) transfer buffer processor
US20030236952A1 (en) 2002-06-21 2003-12-25 Grieff Thomas W. System and method for providing multi-initiator capability to an ATA drive
US20040019718A1 (en) 2002-07-25 2004-01-29 Lsi Logic Corporation Method for receiving user defined frame information structure (FIS) types in a serial-ATA (SATA) system
US20040073719A1 (en) 2002-09-30 2004-04-15 Grimsrud Knut S. Method and apparatus for efficient command queuing within a serial ATA environment
US20040100944A1 (en) 2002-11-27 2004-05-27 Scott Richmond Serial ATA frame structure routing circuitry and protocols
US6790075B1 (en) 2003-07-18 2004-09-14 Yun-Ching Sung Serial ATA interface connector
US20040205259A1 (en) 2003-03-26 2004-10-14 Brea Technologies, Inc. Initiator connection tag for simple table lookup
US20040252716A1 (en) 2003-06-11 2004-12-16 Sam Nemazie Serial advanced technology attachment (SATA) switch
US20040268169A1 (en) 2003-06-25 2004-12-30 Bashford Patrick R. Method and apparatus of automatic power management control for native command queuing Serial ATA device
US6843674B1 (en) 2002-06-03 2005-01-18 Silicon Image, Inc. Method and apparatus for connecting serial ATA storage components
US20050015532A1 (en) 2003-05-20 2005-01-20 Beckett Richard C. Integrated circuit capable of communicating using different communication protocols
US20050102474A1 (en) 2003-11-06 2005-05-12 Sridhar Lakshmanamurthy Dynamically caching engine instructions
US20050138191A1 (en) 2003-12-18 2005-06-23 Pak-Lung Seto Adaptor supporting different protocols
US20050138202A1 (en) 2003-12-18 2005-06-23 Intel Corporation Addresses assignment for adaptor interfaces
US20050149656A1 (en) 2003-12-31 2005-07-07 Pak-Lung Seto Intermediate device capable of communicating using different communication protocols
US20050235072A1 (en) * 2004-04-17 2005-10-20 Smith Wilfred A Data storage controller
US20060004935A1 (en) 2004-06-30 2006-01-05 Pak-Lung Seto Multi-protocol bridge
US20060041691A1 (en) 2004-08-20 2006-02-23 Bashford Patrick R Circuit and method to provide configuration of serial ATA queue depth versus number of devices
US20060047904A1 (en) 2004-09-01 2006-03-02 Derek Rohde Method and system for processing markers, data integrity fields and digests
US20060064568A1 (en) 2004-09-21 2006-03-23 Pak-Lung Seto Integrated circuit capable of mapping logical block address data across multiple domains
US7050940B2 (en) * 2004-03-17 2006-05-23 International Business Machines Corporation Method and system for maintaining and examining timers for network connections
US20060155944A1 (en) 2005-01-13 2006-07-13 Hitachi, Ltd. System and method for data migration and shredding
US20060294286A1 (en) 2005-06-24 2006-12-28 Duerk Vicky P Remote node index mapping mechanism for serial attached storage devices
US20060294344A1 (en) 2005-06-28 2006-12-28 Universal Network Machines, Inc. Computer processor pipeline with shadow registers for context switching, and method
US20070005898A1 (en) 2005-06-30 2007-01-04 William Halleck Method, apparatus and system for task context cache replacement
US20070005850A1 (en) 2005-06-29 2007-01-04 Intel Corporation Port multiplier mapping apparatus, systems, and methods
US20070006235A1 (en) 2005-06-30 2007-01-04 Intel Corporation Task scheduling to devices with same connection address
US20070005896A1 (en) 2005-06-30 2007-01-04 Naichih Chang Hardware oriented host-side native command queuing tag management
US20070002827A1 (en) 2005-06-30 2007-01-04 Victor Lau Automated serial protocol target port transport layer retry mechanism
WO2007002804A2 (en) 2005-06-29 2007-01-04 Intel Corporation Wide-port context cache apparatus, systems, and methods
US20070005838A1 (en) 2005-06-30 2007-01-04 Naichih Chang Serial ATA port addressing
US20070011333A1 (en) 2005-06-30 2007-01-11 Victor Lau Automated serial protocol initiator port transport layer retry mechanism
US20070011360A1 (en) 2005-06-30 2007-01-11 Naichih Chang Hardware oriented target-side native command queuing tag management
US7185266B2 (en) 2003-02-12 2007-02-27 Alacritech, Inc. Network interface device for error detection using partial CRCS of variable length message portions
US20070067504A1 (en) 2005-09-16 2007-03-22 Victor Lau Parallel processing of frame based data transfers
US20070073923A1 (en) 2005-09-29 2007-03-29 Kiran Vemula DMA descriptor management mechanism

Family Cites Families (2)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US7754230B2 (en) * 2004-05-19 2010-07-13 The Regents Of The University Of California Methods and related compositions for reduction of fat
DK2422789T3 (en) * 2004-05-19 2018-02-26 Los Angeles Biomedical Res Inst Harbor Ucla Medical Ct INJECTABLE COMPOSITION CONTAINING SODIUM DEOXYCHOLATE

Patent Citations (49)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4805107A (en) * 1987-04-15 1989-02-14 Allied-Signal Inc. Task scheduler for a fault tolerant multiple node processing system
US5555402A (en) 1988-07-26 1996-09-10 Database Excelleration Systems, Inc. A disk storage subsystem for interfacing with a parallel path, a nonvolatile media and a volatile storage medium
US5862406A (en) 1991-04-11 1999-01-19 Mitsubishi Denki Kabushiki Kaisha Array recording system reporting completion of writing data operation prior to the completion of writing redundancy data in alternative recording units
US5577224A (en) 1994-12-13 1996-11-19 Microsoft Corporation Method and system for caching data
US5802340A (en) 1995-08-22 1998-09-01 International Business Machines Corporation Method and system of executing speculative store instructions in a parallel processing computer system
US5727211A (en) * 1995-11-09 1998-03-10 Chromatic Research, Inc. System and method for fast context switching between tasks
US6205467B1 (en) * 1995-11-14 2001-03-20 Advanced Micro Devices, Inc. Microprocessor having a context save unit for saving context independent from interrupt requests
US5944816A (en) * 1996-05-17 1999-08-31 Advanced Micro Devices, Inc. Microprocessor configured to execute multiple threads including interrupt service routines
US5781926A (en) 1996-05-20 1998-07-14 Integrated Device Technology, Inc. Method and apparatus for sub cache line access and storage allowing access to sub cache lines before completion of line fill
US6430593B1 (en) * 1998-03-10 2002-08-06 Motorola Inc. Method, device and article of manufacture for efficient task scheduling in a multi-tasking preemptive priority-based real-time operating system
US6418489B1 (en) * 1999-10-25 2002-07-09 Motorola, Inc. Direct memory access controller and method therefor
US20030021239A1 (en) 2001-07-26 2003-01-30 Mullendore Rodney N. Method and apparatus for scheduling packet flow on a fibre channel arbitrated loop
US20030110325A1 (en) 2001-12-10 2003-06-12 Bradley Roach Direct memory access (DMA) transfer buffer processor
US6843674B1 (en) 2002-06-03 2005-01-18 Silicon Image, Inc. Method and apparatus for connecting serial ATA storage components
US20030236952A1 (en) 2002-06-21 2003-12-25 Grieff Thomas W. System and method for providing multi-initiator capability to an ATA drive
US20040019718A1 (en) 2002-07-25 2004-01-29 Lsi Logic Corporation Method for receiving user defined frame information structure (FIS) types in a serial-ATA (SATA) system
US20040073719A1 (en) 2002-09-30 2004-04-15 Grimsrud Knut S. Method and apparatus for efficient command queuing within a serial ATA environment
US20040100944A1 (en) 2002-11-27 2004-05-27 Scott Richmond Serial ATA frame structure routing circuitry and protocols
US7185266B2 (en) 2003-02-12 2007-02-27 Alacritech, Inc. Network interface device for error detection using partial CRCS of variable length message portions
US20040205259A1 (en) 2003-03-26 2004-10-14 Brea Technologies, Inc. Initiator connection tag for simple table lookup
US20050015532A1 (en) 2003-05-20 2005-01-20 Beckett Richard C. Integrated circuit capable of communicating using different communication protocols
US7093033B2 (en) 2003-05-20 2006-08-15 Intel Corporation Integrated circuit capable of communicating using different communication protocols
US20040252716A1 (en) 2003-06-11 2004-12-16 Sam Nemazie Serial advanced technology attachment (SATA) switch
US20040268169A1 (en) 2003-06-25 2004-12-30 Bashford Patrick R. Method and apparatus of automatic power management control for native command queuing Serial ATA device
US6790075B1 (en) 2003-07-18 2004-09-14 Yun-Ching Sung Serial ATA interface connector
US20050102474A1 (en) 2003-11-06 2005-05-12 Sridhar Lakshmanamurthy Dynamically caching engine instructions
US20050138191A1 (en) 2003-12-18 2005-06-23 Pak-Lung Seto Adaptor supporting different protocols
US20050138202A1 (en) 2003-12-18 2005-06-23 Intel Corporation Addresses assignment for adaptor interfaces
US20050149656A1 (en) 2003-12-31 2005-07-07 Pak-Lung Seto Intermediate device capable of communicating using different communication protocols
US7050940B2 (en) * 2004-03-17 2006-05-23 International Business Machines Corporation Method and system for maintaining and examining timers for network connections
US20050235072A1 (en) * 2004-04-17 2005-10-20 Smith Wilfred A Data storage controller
US20060004935A1 (en) 2004-06-30 2006-01-05 Pak-Lung Seto Multi-protocol bridge
US20060041691A1 (en) 2004-08-20 2006-02-23 Bashford Patrick R Circuit and method to provide configuration of serial ATA queue depth versus number of devices
US20060047904A1 (en) 2004-09-01 2006-03-02 Derek Rohde Method and system for processing markers, data integrity fields and digests
US20060064568A1 (en) 2004-09-21 2006-03-23 Pak-Lung Seto Integrated circuit capable of mapping logical block address data across multiple domains
US20060155944A1 (en) 2005-01-13 2006-07-13 Hitachi, Ltd. System and method for data migration and shredding
US20060294286A1 (en) 2005-06-24 2006-12-28 Duerk Vicky P Remote node index mapping mechanism for serial attached storage devices
US20060294344A1 (en) 2005-06-28 2006-12-28 Universal Network Machines, Inc. Computer processor pipeline with shadow registers for context switching, and method
WO2007002804A2 (en) 2005-06-29 2007-01-04 Intel Corporation Wide-port context cache apparatus, systems, and methods
US20070005850A1 (en) 2005-06-29 2007-01-04 Intel Corporation Port multiplier mapping apparatus, systems, and methods
US20070006235A1 (en) 2005-06-30 2007-01-04 Intel Corporation Task scheduling to devices with same connection address
US20070005896A1 (en) 2005-06-30 2007-01-04 Naichih Chang Hardware oriented host-side native command queuing tag management
US20070002827A1 (en) 2005-06-30 2007-01-04 Victor Lau Automated serial protocol target port transport layer retry mechanism
US20070005838A1 (en) 2005-06-30 2007-01-04 Naichih Chang Serial ATA port addressing
US20070011333A1 (en) 2005-06-30 2007-01-11 Victor Lau Automated serial protocol initiator port transport layer retry mechanism
US20070011360A1 (en) 2005-06-30 2007-01-11 Naichih Chang Hardware oriented target-side native command queuing tag management
US20070005898A1 (en) 2005-06-30 2007-01-04 William Halleck Method, apparatus and system for task context cache replacement
US20070067504A1 (en) 2005-09-16 2007-03-22 Victor Lau Parallel processing of frame based data transfers
US20070073923A1 (en) 2005-09-29 2007-03-29 Kiran Vemula DMA descriptor management mechanism

Non-Patent Citations (23)

* Cited by examiner, † Cited by third party
Title
"Final Office Action for U.S. Appl. No. 11/170,872, mailed on Sep. 28, 2007", 21 Pages.
"Final Office Action for U.S. Appl. No. 11/172,627, mailed on May 1, 2008", 14 Pages.
"Final Office Action for U.S. Appl. No. 11/172,715, mailed on Apr. 3, 2008", 26 Pages.
"Non-Final Office Action for U.S. Appl. No. 11/170,872, mailed on Jun. 22, 2007", 22 Pages.
"Non-Final Office Action for U.S. Appl. No. 11/171,960, mailed on Jul. 2, 2007", 8 Pages.
"Non-Final Office Action for U.S. Appl. No. 11/172,627, mailed on Sep. 20, 2007", (19 Pages).
"Non-Final Office Action for U.S. Appl. No. 11/172,715, mailed on Nov. 13, 2007", 36 Pages.
"Non-Final Office Action for U.S. Appl. No. 11/240,177, mailed on Feb. 6, 2008", 9 Pages.
"Notice of Allowance for U.S.. Appl. No. 11/171,960, mailed on Jan. 7, 2008", 4 Pages.
"Notification of Transmittal of International Preliminary Examination Report, International Application No. PCT/US2006/025352".
"PCT/US2006/025461 International Search Report", 1 page.
"PCT/US2006/025462 Search Report", (Nov. 14, 2006),2 pages.
"PCT/US2006/025750 Search Report", (Apr. 23, 2007),2 pages.
"PCT/US2006/025753 Search Report", 2 pages.
"Response to Non-Final Office Action for U.S. Appl. No. 11/170,872, filed Aug. 10, 2007", 12 Pages.
"Response to Non-Final Office Action for U.S. Appl. No. 11/171,960, filed Oct. 2, 2007", 10 Pages.
"Response to Non-Final Office Action for U.S. Appl. No. 11/172,627, filed Jan. 25, 2008", 20 Pages.
"Response to Non-Final Office Action for U.S. Appl. No. 11/172,715, filed Feb. 13, 2008", 12 Pages.
"Response to Non-Final Office Action for U.S. Appl. No. 11/240,177, filed Feb. 22, 2008", 11 Pages.
APT Technologies, Inc, et al., "Serial ATA: High Speed Serialized AT Attachment", Revision 1.0a, (Jan. 7, 2003),pp. 1-2, 184-236.
Intel Corporation, "Comparing Serial ATA Native Command Queuing (NCQ) and ATA Tagged Command Queuing (TCQ)", The Advantage of Serial ATA Native Command Queuing, (Oct. 2003),pp. 1-4.
Intel Corporation, "Serial ATA II Native Command Queuing Overview, Application Note", www.intel.com/design/storage/papers/252664.htm. retrieved on Nov. 28, 2006, the whole document, (Apr. 30, 2003),21 pages.
Intel Corporation, et al., "Serial ATA Native Command Queuing", XP-002428259, Joint Whitepaper by Intel Corporation & Seagate Technology, (Jul. 2003), the whole document, 12 pages.

Cited By (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10489188B2 (en) 2017-01-19 2019-11-26 Samsung Electronics Co., Ltd. Method for reducing interrupt latency in embedded systems
CN107992537A (en) * 2017-11-24 2018-05-04 金蝶软件(中国)有限公司 Service attribute transmission method, device, computer equipment and storage medium
CN107992537B (en) * 2017-11-24 2020-05-08 金蝶软件(中国)有限公司 Service attribute transmission method, device, computer equipment and storage medium

Also Published As

Publication number Publication date
US20070118835A1 (en) 2007-05-24

Similar Documents

Publication Publication Date Title
US8402171B1 (en) Host bus adapter that interfaces with host computer bus to multiple types of storage devices
US9946670B2 (en) Determining when to throttle interrupts to limit interrupt processing to an interrupt processing time period
CN102171649B (en) Method and system for queuing transfers of multiple non-contiguous address ranges with a single command
US7676604B2 (en) Task context direct indexing in a protocol engine
US7805543B2 (en) Hardware oriented host-side native command queuing tag management
US6985997B2 (en) System and method for storage system
US7500059B2 (en) Inter-processor communication method using a shared cache memory in a storage system
US6839804B2 (en) Disk array storage device with means for enhancing host application performance using task priorities
KR20070024670A (en) An apparatus and method for high performance volatile disk drive memory access using an integrated dma engine
US20070006235A1 (en) Task scheduling to devices with same connection address
US9928004B2 (en) Assigning device adaptors to use to copy source extents to target extents in a copy relationship
JP3617631B2 (en) Storage control apparatus and control method thereof
US7620751B2 (en) Command scheduling and affiliation management for serial attached storage devices
US20070255897A1 (en) Apparatus, system, and method for facilitating physical disk request scheduling
US7600058B1 (en) Bypass method for efficient DMA disk I/O
JP2001306265A (en) Storage controller and method for controlling the same
TW434491B (en) Increasing I/O performance through storage of packetized operational information in local memory
US20050125573A1 (en) Method and mechanism of accessing segments in data storage systems
EP1579336B1 (en) Improving optical storage transfer performance
EP1288774A2 (en) Integrated drive controller for systems with integrated mass storage
US8775687B1 (en) Method to ensure data coherency in a scalable aggregate neighbor-device interface
US8667188B2 (en) Communication between a computer and a data storage device
US20060155888A1 (en) Request conversion

Legal Events

Date Code Title Description
AS Assignment

Owner name: INTEL CORPORATION,CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HALLECK, WILLIAM;LAU, VICTOR;SETO, PAK-LUNG;AND OTHERS;REEL/FRAME:017282/0418

Effective date: 20051121

Owner name: INTEL CORPORATION, CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:HALLECK, WILLIAM;LAU, VICTOR;SETO, PAK-LUNG;AND OTHERS;REEL/FRAME:017282/0418

Effective date: 20051121

CC Certificate of correction
FPAY Fee payment

Year of fee payment: 4

FEPP Fee payment procedure

Free format text: MAINTENANCE FEE REMINDER MAILED (ORIGINAL EVENT CODE: REM.)

LAPS Lapse for failure to pay maintenance fees

Free format text: PATENT EXPIRED FOR FAILURE TO PAY MAINTENANCE FEES (ORIGINAL EVENT CODE: EXP.)

STCH Information on status: patent discontinuation

Free format text: PATENT EXPIRED DUE TO NONPAYMENT OF MAINTENANCE FEES UNDER 37 CFR 1.362

FP Lapsed due to failure to pay maintenance fee

Effective date: 20180309