US20020129146A1 - Highly available database clusters that move client connections between hosts - Google Patents

Highly available database clusters that move client connections between hosts Download PDF

Info

Publication number
US20020129146A1
US20020129146A1 US10/072,317 US7231702A US2002129146A1 US 20020129146 A1 US20020129146 A1 US 20020129146A1 US 7231702 A US7231702 A US 7231702A US 2002129146 A1 US2002129146 A1 US 2002129146A1
Authority
US
United States
Prior art keywords
connection
client
primary
dbms
connection manager
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US10/072,317
Inventor
Eyal Aronoff
Eyal Kalderon
Bill Romine
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Quest Software Inc
Original Assignee
Quest Software Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Quest Software Inc filed Critical Quest Software Inc
Priority to US10/072,317 priority Critical patent/US20020129146A1/en
Assigned to QUEST SOFTWARE, INC. reassignment QUEST SOFTWARE, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: ARONOFF, EYAL, KALDERON, EYAL, ROMINE, BILL
Publication of US20020129146A1 publication Critical patent/US20020129146A1/en
Priority to US10/661,412 priority patent/US7231391B2/en
Priority to US10/662,039 priority patent/US20040162836A1/en
Priority to US11/471,313 priority patent/US7512682B2/en
Priority to US11/754,891 priority patent/US7606839B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1029Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers using data related to the state of servers by a load balancer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/1034Reaction to server failures by a load balancer
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L9/00Cryptographic mechanisms or cryptographic arrangements for secret or secure communications; Network security protocols
    • H04L9/40Network security protocols
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L67/00Network arrangements or protocols for supporting network services or applications
    • H04L67/01Protocols
    • H04L67/10Protocols in which an application is distributed across nodes in the network
    • H04L67/1001Protocols in which an application is distributed across nodes in the network for accessing one among a plurality of replicated servers
    • H04L67/10015Access to distributed or replicated servers, e.g. using brokers
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99953Recoverability
    • YGENERAL TAGGING OF NEW TECHNOLOGICAL DEVELOPMENTS; GENERAL TAGGING OF CROSS-SECTIONAL TECHNOLOGIES SPANNING OVER SEVERAL SECTIONS OF THE IPC; TECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10TECHNICAL SUBJECTS COVERED BY FORMER USPC
    • Y10STECHNICAL SUBJECTS COVERED BY FORMER USPC CROSS-REFERENCE ART COLLECTIONS [XRACs] AND DIGESTS
    • Y10S707/00Data processing: database and file management or data structures
    • Y10S707/99951File or database maintenance
    • Y10S707/99952Coherency, e.g. same view to multiple users
    • Y10S707/99955Archiving or backup

Definitions

  • the present invention relates to the field of highly available database clusters. More specifically, the invention relates to database clusters that transparently move client connections between hosts.
  • a database is generally considered to be a collection of information or data organized in a way that computer programs can quickly access or select desired portions of the collection.
  • a database management system includes the collection of computer programs that enable the quick storage, selection, modification, and extraction of desired portions of data from the database.
  • Exemplary DBMSs include those commercially available from Oracle Corporation, IBM, or the like.
  • Application programs typically include client programs that connect to a DBMS to provide users the ability to interact with the data of the database, such as, for example, to select, modify, organize, delete, or the like, some or all of the foregoing data.
  • Exemplary application programs include payroll or inventory programs, online stores, or the like.
  • the application programs are designed to be continually connected to a DBMS, thereby having substantially continuous access to data stored within the same. Unless specifically coded to recover, these application programs typically fail when their connection to the DBMS fails or is otherwise unavailable, such as during a system failure. For many application program environments, this failure is undesirable.
  • Database clusters can include two or more DBMSs accessing shared data files.
  • the shared data files can include data files having the same set of data from the replication of changes from one DBMS to another.
  • the shared data files can include multiple DBMSs that access the same physical storage. Through the shared data files, system designers allow one DBMS to replace another in the event of a failure.
  • system designers may also employ application servers in order to reduce the effects of losing a connection to a DBMS.
  • system designers often have application programs connect to an application server, where the application server includes the functionality to recover lost client connections to one or more secondary DBMSs within a database cluster.
  • the application server generally includes a proprietary protocol used in communications from the application program to the application server.
  • the proprietary protocol is generally not native to the DBMS and therefore, each connecting application program will first be routed through the application server.
  • the application server solution is not well suited for geographically diverse storage systems.
  • Embodiments of the present invention seek to overcome some or all of these and other problems.
  • aspects of embodiments of the present disclosure include a highly available database cluster that can maintain a connection with potentially geographically remote client application programs, including non-fault tolerant application programs.
  • the database cluster can advantageously move a client connection between a failing, unbalanced ,or overloaded DBMS, to another DBMS within the database cluster.
  • the database cluster includes connection managers which monitor a connection between a client application program and a primary DBMS.
  • connection manager determines that the primary DBMS is unavailable, has an unbalanced share of the workload of the cluster, or the like
  • the connection manager of a secondary DBMS can assume the connection to the client application as if it were the primary DBMS.
  • the connection manager can finish all open transactions, thus avoiding the need to roll back the same.
  • Embodiments of the connection managers can also monitor the connection at the DBMS communication level, such as, for example, the SQL*Net level.
  • the connection managers capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster.
  • an aspect of an embodiment of the invention includes a data processing system comprising a database cluster which can move a connection between a remote client and a first DBMS within the cluster to a second DBMS within the cluster when the database cluster determines that the first DBMS has failed, wherein the movement of the connection is transparent to the remote client and the connection includes communication in a protocol native to the first and second DBMSs, such as, for example, SQL*Net.
  • a protocol native to the first and second DBMSs such as, for example, SQL*Net.
  • Another aspect of an embodiment of the invention includes a data processing system comprising a database cluster which can move a connection between a remote client and a first DBMS to a second DBMS when the database cluster determines that the first DBMS is executing an unbalanced portion of the cluster workload, wherein the movement of the connection is transparent to the remote client.
  • Another aspect of an embodiment of the invention includes a method of moving a client connection from a first DBMS to a second DBMS.
  • the method comprises monitoring a state of a client connection to a first DBMS, wherein the client connection includes communication in a protocol native to the first DBMS.
  • the method also comprises detecting a condition of the connection which indicates the connection should be moved, and moving the client connection to a second DBMS without the client dropping the client connection.
  • FIG. 1 illustrates a block diagram of an exemplary data processing system including a database cluster according to embodiments of the invention.
  • FIG. 2 illustrates a block diagram of exemplary connection managers of the database cluster of FIG. 1, according to embodiments of the invention.
  • FIG. 3 illustrates a flow chart of a fail-over process, according to embodiments of the invention.
  • aspects of embodiments of the present disclosure include a highly available database cluster that can move connections with one or more client program applications from a first host to a second in the event of, for example, a failure of the first host, an unbalanced or overloaded workload present on the first host, or the like. Additionally, the cluster can provide communication in the native protocol of underlying database management systems (DBMSs), thereby providing fault tolerant connections for potentially geographically remote and potentially non-fault tolerant client application programs.
  • DBMSs database management systems
  • the database cluster includes connection managers that monitor a connection between a client application program and a primary DBMS.
  • the connection manager of a secondary DBMS can assume the connection to the client application as if it were the primary DBMS.
  • the assumption of the connection by the secondary connection manager is transparent to the client.
  • the secondary connection manager can replay or finish all open transactions, thus picking up the connection to the client in a state exactly where the primary DBMS failed.
  • Embodiments of the connection managers can also monitor a connection at the DBMS communication level, such as, for example, the SQL*Net level. According to one embodiment, the connection managers capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster.
  • the connection manager monitors a current state of TCP and IP protocols of a TCP/IP connection. When one connection manager determines that it should assume the TCP/IP connection, the connection manager continues the TCP conversation that the client originally started with the other connection manager.
  • the foregoing solution advantageously provides a database cluster offering high availability to its connecting clients, including non-fault tolerant clients, by moving connections between DBMSs within a database cluster.
  • FIG. 1 illustrates a block diagram of an exemplary data processing system 100 , according to an embodiment of the invention.
  • the data processing system 100 includes a client application program 105 (client 105 ) communicating with a highly available database cluster 110 (cluster 110 ) through a communication network 115 .
  • the client 105 comprises computer programs such as payroll or inventory programs, online stores, human resource applications, or the like, executing on one or more remote computer devices or systems.
  • the client 105 can comprise virtually any client program designed to connect with a DBMS to interact with data stored therein, such as, for example, to select, modify, organize, delete, index, or the like, some or all of the foregoing data.
  • the client 105 can execute on a wide variety of computer devices, such as, for example, personal digital assistants, mobile telephones, handheld computer devices, laptop computers, workstations, mainframe computers, combinations of the same, or the like.
  • the cluster 110 can comprise two or more DBMSs, able to access portions of shared, replicated, or otherwise mirrored data.
  • Exemplary DBMSs include those commercially available from Oracle Corporation, IBM, or the like.
  • the DBMSs of the cluster 110 execute on one or more hosts or other computing devices.
  • the communication network 115 comprises some or all of the Internet.
  • the communications network 115 can include a wide range of interactive communications mediums.
  • the communications network 115 can include interactive television networks, telephone networks, wireless data transmission systems, two-way cable systems, customized computer networks, interactive kiosk networks, automatic teller machine networks, direct links, private local or wide area networks, and the like.
  • the client 105 connects to the cluster 110 through the communication network 115 .
  • the client 105 issues instructions or transactions including one or more operational statements to be carried out against data stored in data files accessible by the cluster 110 .
  • the cluster 110 returns an indication of the same to the client 105 .
  • the cluster 110 can move the foregoing connection with the client 105 from a first host to a second in the event of, for example, a failure of the first host, an unbalanced or overloaded workload present on the first host, or the like.
  • the cluster 110 can provide communication in the native protocol of the underlying two or more DBMSs, thereby providing fault tolerant connections for the potentially geographically remote and potentially non-fault tolerant client 105 .
  • the cluster 110 can monitor a connection at the DBMS communication level, such as, for example, a SQL*Net level. The cluster 110 can capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster.
  • FIG. 1 also shows the cluster 110 including a routing device 120 communicating with a primary host 125 (Host A 125 ) to execute transactions against one or more shared data files 130 . Additionally, FIG. 1 shows the routing device 120 having the ability to communicate with a secondary host 135 (Host B 135 ), which in turn also includes the ability to execute transactions against the one or more shared data files 130 . According to one embodiment, Host A 125 includes a primary connection manager 140 and a primary DBMS 145 , while Host B 135 includes a secondary connection manager 150 and a secondary DBMS 155 . FIG. 1 also shows the primary connection manager 140 communicating with the secondary connection manager 150 .
  • Routing device 120 comprises a device, such as, for example, a router, hub, or the like, that connects any number of computing systems or networks. Generally, routing device 120 uses information in data packets, along with a forwarding table to determine where the data packets go. According to one embodiment, the routing device 120 is configured in such as fashion as to forward all packets destined for the database cluster 110 to both the primary connection manager 140 and the secondary connection manager 150 . An artisan will recognize that the function of such routing will be to enable a virtual IP address (VIP) that may be shared between hosts.
  • VIP virtual IP address
  • the routing device 120 sends all data packets from the client 120 to both the primary connection manager 140 and the secondary connection manager 150 .
  • the secondary connection manager monitors statistics related to, for example, the number of clients connected to the primary connection manger.
  • the primary connection manager assumes responsibility for the data packets send from the client 105 to the primary DBMS 145 .
  • the client 105 sends transactions, in the form of data packets, through the communication network 115 to the primary DBMS 145 , the data packets are routed to the primary connection manager 140 , forming a connection between the primary connection manager 140 and the client 105 .
  • the primary connection manager 140 then forwards the data packets to the primary DBMS 145 , forwards a copy of the data packets to the secondary connection manager 150 , and monitors statistics related to, for example, the number of connected clients and the status of the secondary connection manager 150 . Meanwhile, the secondary connection manager 150 receives the copied data packets, holds them in memory, and monitors statistics related to, for example, the number of connected clients and the status of the primary connection manager 140 .
  • the primary DBMS 145 receives the data packets from the primary connection manager 140 , assembles them into operational statements of transactions, and executes the same against the data files 130 .
  • the primary DBMS 145 then returns the requested data and/or acknowledgment of the received data packets back to the primary connection manager 140 , which in turns forwards a copy to the secondary connection manager 150 and a copy to the respective client 105 through the communication network 115 .
  • the secondary connection manager 150 can detect a condition of the connection between the primary connection manager 140 and the client 105 from the statistics being monitored. For example, the secondary connection manager 150 can detect a failure of the connection, an unbalanced or overloaded workload on the primary connection, or the like. In such circumstances, the secondary connection manager assumes control of the connection and replays any rolled back transactions against the data files 130 through the secondary DBMS 155 as follows.
  • the secondary connection manager 150 communicates with the routing device 120 to acknowledge TCP requests from the client 105 to the primary connection manager 140 . These acknowledgements advantageously keep the client TCP connection from timing out and failing. Additionally, the secondary connection manager 150 replays any operational statements of transactions rolled back due to, for example, the failure of the primary connection. As is generally known in the art, upon failure of a DBMS, all operational statements of open transactions (for, example, non-committed transactions) executed against the data files 130 are rolled back as if they never occurred. However, because the operational statements of open transactions are stored in the foregoing memory of the secondary connection manager 150 , these operational statements from open transactions can be reexecuted against the data files 130 through the secondary DBMS 155 . After replaying the foregoing operational statements, the secondary connection manager 150 begins forwarding data packets from the client 105 to the secondary DBMS 155 to be executed against the data files 130 .
  • the database cluster 110 advantageously moves a connection between the primary DBMS 145 and the client 105 to the secondary DBMS 155 in the cluster 110 , when the primary DBMS 145 fails, becomes unbalanced, overloaded, or the like. Additionally, the database cluster 110 advantageously replays any rolled back statements of open transactions during fail-over to the secondary DBMS 155 , thereby providing an assumption of the connection that is transparent to the client 105 . Accordingly, the cluster 110 avoids failure of non-fault tolerant clients by moving the connection rather than allowing it to fail. Additionally, the cluster 110 advantageously provides communication in the native protocol of the underlying two or more DBMSs, thereby providing fault tolerant connections for the potentially geographically remote and potentially non-fault tolerant client 105 .
  • FIG. 2 illustrates a block diagram of embodiments of the primary and secondary connection managers, 140 and 150 , of the cluster 110 , according to embodiments of the invention.
  • FIG. 2 shows the primary connection manager 140 including a primary connection 205 communicating with a memory 210 including statistics 215 , a monitor process 220 also communicating with the memory 210 , and a protocol shadow 225 communicating with the memory 210 and the primary DBMS 145 .
  • FIG. 2 shows the secondary connection manager 150 including a secondary connection 245 communicating with a memory 250 including statistics 255 , a monitor process 260 also communicating with the memory 250 , and a protocol shadow 265 communicating with the memory 250 and the secondary DBMS 155 .
  • the secondary connection manager 150 includes an import process 270 communicating with the primary connection 205 and a queue 275 .
  • the secondary connection manager 150 also includes a replay process 280 communicating with the queue 275 and the protocol shadow 265 .
  • an additional redo monitor can access one or more log files 285 associated with the primary DBMS 145 .
  • the redo monitor also can communicate with the memory 210 and review the statistics 215 .
  • FIG. 2 also shows the protocol shadow 265 accessing the one or more log files 285 .
  • the client 105 When the client 105 begins a transaction by issuing an operational statement to be applied against the data files 130 , the client 105 distributes the statement across one or more data packets.
  • the data packets are forwarded through the communication network 115 to the routing device 120 , where, as disclosed, the routing device 120 forwards the packets to the primary connection 205 and to the secondary connection 245 .
  • the primary connection 205 examines statistics in the statistics 215 generated by the redo monitor. These statistics include, for example, the current location of transaction being stored in the log files 285 .
  • the primary connection transmits a copy of each data packet along with the current log file location, such as a sequence number, to the import process 270 of the secondary connection manager 150 , and places a copy in the memory 210 .
  • the import process 270 stores the data packets in the queue 275 .
  • the protocol shadow 225 accesses the memory 210 and retrieves the data packets.
  • the protocol shadow 225 forwards the packets to the primary DBMS 145 , where the packets are assembled and the operational statement executed against the data files 130 .
  • the DBMS can also keep a record or log of the executed statement, generally in the log file 285 .
  • the DBMS 145 forwards a result of the statement and/or and acknowledgement of receipt of the same, back to the protocol shadow 225 , preferably in one or more acknowledgement data packets.
  • the protocol shadow 225 transfers the data packets back to the memory 210 , where they are picked up by the primary connection 205 .
  • the primary connection 205 forwards a copy of the data packets to the import process 270 and to the client 105 .
  • the client 105 receives the results and/or acknowledgement of the transmitted statement of an open transaction.
  • the client 105 may then desire to finalize, or commit the transaction against the data files 130 .
  • the client 105 issues a commit statement, which is forwarded to the primary DBMS 145 and the import process 270 , along with the subsequent result and/or acknowledgement, in a manner similar to that disclosed.
  • the protocol shadow 225 stores sufficient data from the data packets that it can assemble the statements of a given transaction.
  • the protocol shadow 225 attaches a marker to the result/acknowledgement data packets associated with the primary DBMS 145 acknowledging execution of the commit statement.
  • the marker comprises a location marker, such as, for example, a sequence number from the primary DBMS 145 .
  • the import process 270 recognizes the marker placed on the data packets associated with the commit statement, and recognizes that the entire transaction has been executed by the primary DBMS 145 against the data files 130 . Therefore, the import process 270 deletes the data packets associated with the now finalized transaction from the queue 275 .
  • the protocol shadow 225 and the import process 270 advantageously work together to ensure that only the data packets associated with open transactions remain in the queue 275 .
  • the primary connection 205 also stores the statistics 215 related to the connection with the client 105 in the memory 210 .
  • the statistics include sufficient information for the monitor process 220 to determine whether the primary connection 205 has failed, is processing an unbalanced or overloaded workload, or the like, and whether the secondary connection 245 has failed, is processing an unbalanced or overloaded workload, or the like.
  • the statistics 215 can include the number of clients seen by the primary connection 205 , the number of clients seen by the secondary connection 245 , the status of communication with secondary communication manager 150 , or the like.
  • the primary connection 205 acquires the statistics 215 corresponding to information from the secondary connection manager 150 through the connection between the primary connection 205 and the secondary connection 245 .
  • the foregoing status of the secondary communication manger 150 can be ascertained through straightforward ping or ping-like commands.
  • FIG. 3 illustrates a flow chart of a fail-over process 300 , according to embodiments of the invention.
  • the fail-over process 300 begins with BLOCK 305 where the cluster 110 monitors the statistics of one or more connections with one or more clients.
  • the monitoring corresponds to the monitor processes 220 and 260 .
  • the cluster 110 detects the need to move the connection from one DBMS to another.
  • the monitor 260 may determine that the primary DBMS 145 has failed, become unbalanced, overloaded, or the like, and determine that the secondary connection manager 150 should assume the connection with the client 105 .
  • the fail-over process 300 proceeds to BLOCK 320 , where the cluster 110 moves the connection from one DBMS to another without losing the connection or causing even a non-fault tolerant client to fail.
  • the secondary connection 245 can communicate with the routing device 120 to assume the IP address (or VIP) of the primary DBMS 145 .
  • the secondary connection manager 150 can replay all statements of open transactions which were rolled back in the data files 130 . Accordingly, the move is transparent to the client 105 who does not lose the connection and does not know that a change has been made.
  • BLOCK 320 can include SUBBLOCK 321 , where the cluster 110 instructs the routing device 120 to forward communication from the client to another DBMS.
  • the secondary connection 245 can assume the IP address of the primary DBMS 145 .
  • BLOCK 320 can also include SUBBLOCK 322 , where the cluster 110 can send a keepalive message to one or more clients to ensure against failure of the connection to the same.
  • the client 105 resends data packets which are not responded to or otherwise acknowledged by the cluster 110 .
  • the client 105 When the client 105 resends the same data packets a predetermined amount of times, the client 105 may register a failure of the connection, thereby causing non-fault tolerant clients (such as those clients not programmed to recover) to also fail.
  • the cluster 110 can respond to the client 105 with a message or acknowledgement that keeps the client 105 from resending the same data packets, therefore keeping the client from determining that the connection has failed.
  • the secondary connection 245 sends the foregoing keepalive messages.
  • BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 324 where the cluster 110 replays any statements from open transactions that were rolled back during the failure of the primary DBMS 145 .
  • the replay process 280 can access the queue 275 to retrieve data packets associated with rolled back transactions and to forward them to the protocol shadow 265 .
  • the import process 270 removes the statements associated with all finalized or committed transactions, thereby leaving only rolled back transactions in the queue 275 .
  • BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 326 where the cluster 110 removes any leftover committed transactions that may have slipped through.
  • Host A 125 can fail after the primary DBMS 145 executes a commit statement for a particular transaction, but before the result/acknowledgement of the same can be transmitted to the import process 270 .
  • the secondary connection manager 150 believes the statements associated with the foregoing transaction were rolled back, e.g., because they were left in the queue 275 , and therefore, the replay process 280 will forward the already committed statements to the protocol shadow 265 .
  • the protocol shadow 265 parses the log file 285 of the primary DBMS 145 to ensure a commit statement associated with the open transaction was not received. When the protocol shadow 265 determines that a commit statement was received, the protocol shadow 265 deletes the statements associated therewith before their associated data packets are forwarded to the secondary DBMS 155 to be executed against the data files 130 .
  • BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 328 where the cluster 110 establishes communication between the client and the secondary DBMS.
  • the protocol shadow 265 begins accessing new data packets stored in the memory 250 by the secondary connection 245 after it assumed the connection to the client 105 from the primary connection manager 140 .
  • the secondary connection manager 150 performs operations similar to the normal operations of the primary connection manager 140 as disclosed above.
  • the system administrator of the database cluster 110 can designate whether the secondary connection manager 150 through the monitor process 260 fails-back to the primary connection manager 140 after the cause of failure of the same is repaired, or whether the secondary connection manager 150 simply becomes the primary and vice versa.
  • the data packets captured from the primary connection manager 140 can be replicated to other DBMSs by replaying the same on the other DBMSs.
  • This replication has several advantages over other replication techniques including a potential reduction in the traffic keeping the database cluster synchronized, thereby advantageously providing economical replication of geographically diverse data files.
  • the captured data packets can also be used to assist a transaction log based replication system.
  • the data packets can be directed to the other databases in the cluster prior to committing the transactions. Accordingly, committed transactions on a particular DBMS are not lost when the DBMS fails, as these transactions may advantageously be replayed on the other DBMSs in the cluster.
  • the captured data packets can also be used to assist a transaction log based replication system when posting replicated modifications. Some modifications (such as a vertical table update or DDL operation) may be difficult to replicate via a log-based replication. When the original data packets are available, posting the original SQL rather than the data from the transaction log may be more efficient and straightforward.
  • software may be added just below the client 105 , thereby providing a mechanism to replay incomplete transactions.
  • a typical client application does not access the database directly, but instead uses some type of intermediate layer such as ODBC or JDBC, OCI, or the like.
  • the foregoing added software can advantageously replace this intermediate layer.

Abstract

Aspects of embodiments of the present disclosure include a highly available database cluster that can maintain a connection with potentially geographically remote client application programs, including non-fault tolerant application programs, even in the event of one of the database management systems (DBMS) of the cluster becoming unavailable. For example, the database cluster can advantageously move a client connection between a failing, unbalanced, or overloaded DBMS, to another DBMS within the cluster. The database cluster can include connection managers that monitor a connection between a client application program and a primary DBMS. When one connection manager determines that the primary DBMS is unavailable, the connection manager of a secondary DBMS can assume the connection to the client application as if it were the primary DBMS. The connection manager can finish all open transactions, thereby avoiding the need to roll back the same. Moreover, the connection managers can monitor the connection at the DBMS communication level, such as, for example, the SQL*Net level.

Description

    REFERENCE TO RELATED APPLICATION
  • The present application claims priority benefit under 35 U.S.C. §119([0001] e) from U.S. Provisional Application No. 60/266,908, filed Feb. 6, 2001, entitled “HIGHLY AVAILABLE DATABASE CLUSTERS,” which is incorporated herein by reference.
  • FIELD OF THE INVENTION
  • The present invention relates to the field of highly available database clusters. More specifically, the invention relates to database clusters that transparently move client connections between hosts. [0002]
  • BACKGROUND OF THE INVENTION
  • A database is generally considered to be a collection of information or data organized in a way that computer programs can quickly access or select desired portions of the collection. A database management system (DBMS) includes the collection of computer programs that enable the quick storage, selection, modification, and extraction of desired portions of data from the database. Exemplary DBMSs include those commercially available from Oracle Corporation, IBM, or the like. Application programs, on the other hand, typically include client programs that connect to a DBMS to provide users the ability to interact with the data of the database, such as, for example, to select, modify, organize, delete, or the like, some or all of the foregoing data. Exemplary application programs include payroll or inventory programs, online stores, or the like. [0003]
  • Often, the application programs are designed to be continually connected to a DBMS, thereby having substantially continuous access to data stored within the same. Unless specifically coded to recover, these application programs typically fail when their connection to the DBMS fails or is otherwise unavailable, such as during a system failure. For many application program environments, this failure is undesirable. [0004]
  • System designers have created various solutions to reduce the effects of an application program losing a connection to a DBMS. For example, system designers often employ database clusters to offer backup solutions to failed systems. Database clusters can include two or more DBMSs accessing shared data files. For example, the shared data files can include data files having the same set of data from the replication of changes from one DBMS to another. Also, the shared data files can include multiple DBMSs that access the same physical storage. Through the shared data files, system designers allow one DBMS to replace another in the event of a failure. [0005]
  • There are several drawbacks associated with the foregoing database clustering solution, especially when employed in environments allowing for little or no down time, such as, for example, high availability solutions. For example, when a DBMS fails, the connection from the application program to the DBMS can be lost, thereby potentially losing all open transactions from the same. Additionally, data not replicated from a failing DBMS can be lost. Moreover, during load balancing, simultaneous updates of the same data on different DBMSs can occur in some replication solutions. Also, a large amount of communication traffic among a cluster, and/or hardware limitations of the same, can reduce the cost effectiveness of geographically diverse systems. Moreover, as discussed, the failure of an individual DBMS results in a failure of non-fault tolerant program applications. [0006]
  • On the other hand, system designers may also employ application servers in order to reduce the effects of losing a connection to a DBMS. For example, system designers often have application programs connect to an application server, where the application server includes the functionality to recover lost client connections to one or more secondary DBMSs within a database cluster. However, the application server generally includes a proprietary protocol used in communications from the application program to the application server. The proprietary protocol is generally not native to the DBMS and therefore, each connecting application program will first be routed through the application server. Thus, the application server solution is not well suited for geographically diverse storage systems. [0007]
  • Embodiments of the present invention seek to overcome some or all of these and other problems. [0008]
  • SUMMARY OF THE INVENTION
  • Therefore, a need exists for a database cluster that can maintain a connection with potentially geographically remote client application programs, including non-fault tolerant application programs, even in the event of a failure or other unavailability of the primary DBMS. Accordingly, aspects of embodiments of the present disclosure include a highly available database cluster that can maintain a connection with potentially geographically remote client application programs, including non-fault tolerant application programs. For example, the database cluster can advantageously move a client connection between a failing, unbalanced ,or overloaded DBMS, to another DBMS within the database cluster. [0009]
  • According to one embodiment, the database cluster includes connection managers which monitor a connection between a client application program and a primary DBMS. When one connection manager determines that the primary DBMS is unavailable, has an unbalanced share of the workload of the cluster, or the like, the connection manager of a secondary DBMS can assume the connection to the client application as if it were the primary DBMS. For example, the connection manager can finish all open transactions, thus avoiding the need to roll back the same. Embodiments of the connection managers can also monitor the connection at the DBMS communication level, such as, for example, the SQL*Net level. According to one embodiment, the connection managers capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster. [0010]
  • Based on the foregoing, an aspect of an embodiment of the invention includes a data processing system comprising a database cluster which can move a connection between a remote client and a first DBMS within the cluster to a second DBMS within the cluster when the database cluster determines that the first DBMS has failed, wherein the movement of the connection is transparent to the remote client and the connection includes communication in a protocol native to the first and second DBMSs, such as, for example, SQL*Net. [0011]
  • Another aspect of an embodiment of the invention includes a data processing system comprising a database cluster which can move a connection between a remote client and a first DBMS to a second DBMS when the database cluster determines that the first DBMS is executing an unbalanced portion of the cluster workload, wherein the movement of the connection is transparent to the remote client. [0012]
  • Another aspect of an embodiment of the invention includes a method of moving a client connection from a first DBMS to a second DBMS. The method comprises monitoring a state of a client connection to a first DBMS, wherein the client connection includes communication in a protocol native to the first DBMS. The method also comprises detecting a condition of the connection which indicates the connection should be moved, and moving the client connection to a second DBMS without the client dropping the client connection. [0013]
  • For purposes of summarizing the invention, certain aspects, advantages and novel features of the invention have been described herein. Of course, it is to be understood that not necessarily all such aspects, advantages or features will be embodied in any particular embodiment of the invention.[0014]
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • A general architecture that implements the various features of the invention will now be described with reference to the drawings. The drawings and the associated descriptions are provided to illustrate embodiments of the invention and not to limit the scope of the invention. Throughout the drawings, reference numbers are re-used to indicate correspondence between referenced elements. In addition, the first digit of each reference number indicates the figure in which the element first appears. [0015]
  • FIG. 1 illustrates a block diagram of an exemplary data processing system including a database cluster according to embodiments of the invention. [0016]
  • FIG. 2 illustrates a block diagram of exemplary connection managers of the database cluster of FIG. 1, according to embodiments of the invention. [0017]
  • FIG. 3 illustrates a flow chart of a fail-over process, according to embodiments of the invention. [0018]
  • DETAILED DESCRIPTION OF THE PREFERRED EMBODIMENT
  • Aspects of embodiments of the present disclosure include a highly available database cluster that can move connections with one or more client program applications from a first host to a second in the event of, for example, a failure of the first host, an unbalanced or overloaded workload present on the first host, or the like. Additionally, the cluster can provide communication in the native protocol of underlying database management systems (DBMSs), thereby providing fault tolerant connections for potentially geographically remote and potentially non-fault tolerant client application programs. [0019]
  • According to one embodiment, the database cluster includes connection managers that monitor a connection between a client application program and a primary DBMS. When one connection manager determines that the primary DBMS is unavailable, has an unbalanced share of the workload of the cluster, or the like, the connection manager of a secondary DBMS can assume the connection to the client application as if it were the primary DBMS. In one embodiment, the assumption of the connection by the secondary connection manager is transparent to the client. Moreover, the secondary connection manager can replay or finish all open transactions, thus picking up the connection to the client in a state exactly where the primary DBMS failed. Embodiments of the connection managers can also monitor a connection at the DBMS communication level, such as, for example, the SQL*Net level. According to one embodiment, the connection managers capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster. [0020]
  • According to one embodiment, the connection manager monitors a current state of TCP and IP protocols of a TCP/IP connection. When one connection manager determines that it should assume the TCP/IP connection, the connection manager continues the TCP conversation that the client originally started with the other connection manager. [0021]
  • The foregoing solution advantageously provides a database cluster offering high availability to its connecting clients, including non-fault tolerant clients, by moving connections between DBMSs within a database cluster. [0022]
  • To facilitate a complete understanding of the invention, the remainder of the detailed description describes the invention with reference to the drawings, wherein like reference numbers are referenced with like numerals throughout. [0023]
  • FIG. 1 illustrates a block diagram of an exemplary [0024] data processing system 100, according to an embodiment of the invention. As shown in FIG. 1, the data processing system 100 includes a client application program 105 (client 105) communicating with a highly available database cluster 110 (cluster 110) through a communication network 115. According to one embodiment, the client 105 comprises computer programs such as payroll or inventory programs, online stores, human resource applications, or the like, executing on one or more remote computer devices or systems.
  • An artisan will recognize from the disclosure herein that the [0025] client 105 can comprise virtually any client program designed to connect with a DBMS to interact with data stored therein, such as, for example, to select, modify, organize, delete, index, or the like, some or all of the foregoing data. Moreover, the artisan will recognize from the disclosure herein that the client 105 can execute on a wide variety of computer devices, such as, for example, personal digital assistants, mobile telephones, handheld computer devices, laptop computers, workstations, mainframe computers, combinations of the same, or the like.
  • The [0026] cluster 110 can comprise two or more DBMSs, able to access portions of shared, replicated, or otherwise mirrored data. Exemplary DBMSs include those commercially available from Oracle Corporation, IBM, or the like. According to one embodiment, the DBMSs of the cluster 110 execute on one or more hosts or other computing devices.
  • The [0027] communication network 115 comprises some or all of the Internet. However, an artisan will recognize from the disclosure herein that the communications network 115 can include a wide range of interactive communications mediums. For example, the communications network 115 can include interactive television networks, telephone networks, wireless data transmission systems, two-way cable systems, customized computer networks, interactive kiosk networks, automatic teller machine networks, direct links, private local or wide area networks, and the like.
  • In one embodiment, the [0028] client 105 connects to the cluster 110 through the communication network 115. The client 105 issues instructions or transactions including one or more operational statements to be carried out against data stored in data files accessible by the cluster 110. When the cluster 110 has executed the instructions or transactions, the cluster 110 returns an indication of the same to the client 105. Moreover, the cluster 110 can move the foregoing connection with the client 105 from a first host to a second in the event of, for example, a failure of the first host, an unbalanced or overloaded workload present on the first host, or the like. Additionally, the cluster 110 can provide communication in the native protocol of the underlying two or more DBMSs, thereby providing fault tolerant connections for the potentially geographically remote and potentially non-fault tolerant client 105. For example, the cluster 110 can monitor a connection at the DBMS communication level, such as, for example, a SQL*Net level. The cluster 110 can capture enough information about the connection to restore the connection to its current state on another DBMS in the cluster.
  • FIG. 1 also shows the [0029] cluster 110 including a routing device 120 communicating with a primary host 125 (Host A 125) to execute transactions against one or more shared data files 130. Additionally, FIG. 1 shows the routing device 120 having the ability to communicate with a secondary host 135 (Host B 135), which in turn also includes the ability to execute transactions against the one or more shared data files 130. According to one embodiment, Host A 125 includes a primary connection manager 140 and a primary DBMS 145, while Host B 135 includes a secondary connection manager 150 and a secondary DBMS 155. FIG. 1 also shows the primary connection manager 140 communicating with the secondary connection manager 150.
  • [0030] Routing device 120 comprises a device, such as, for example, a router, hub, or the like, that connects any number of computing systems or networks. Generally, routing device 120 uses information in data packets, along with a forwarding table to determine where the data packets go. According to one embodiment, the routing device 120 is configured in such as fashion as to forward all packets destined for the database cluster 110 to both the primary connection manager 140 and the secondary connection manager 150. An artisan will recognize that the function of such routing will be to enable a virtual IP address (VIP) that may be shared between hosts.
  • In one embodiment, the [0031] routing device 120 sends all data packets from the client 120 to both the primary connection manager 140 and the secondary connection manager 150. The secondary connection manager monitors statistics related to, for example, the number of clients connected to the primary connection manger. The primary connection manager assumes responsibility for the data packets send from the client 105 to the primary DBMS 145. Thus, when the client 105 sends transactions, in the form of data packets, through the communication network 115 to the primary DBMS 145, the data packets are routed to the primary connection manager 140, forming a connection between the primary connection manager 140 and the client 105. The primary connection manager 140 then forwards the data packets to the primary DBMS 145, forwards a copy of the data packets to the secondary connection manager 150, and monitors statistics related to, for example, the number of connected clients and the status of the secondary connection manager 150. Meanwhile, the secondary connection manager 150 receives the copied data packets, holds them in memory, and monitors statistics related to, for example, the number of connected clients and the status of the primary connection manager 140.
  • The [0032] primary DBMS 145 receives the data packets from the primary connection manager 140, assembles them into operational statements of transactions, and executes the same against the data files 130. The primary DBMS 145 then returns the requested data and/or acknowledgment of the received data packets back to the primary connection manager 140, which in turns forwards a copy to the secondary connection manager 150 and a copy to the respective client 105 through the communication network 115.
  • In an embodiment, the [0033] secondary connection manager 150 can detect a condition of the connection between the primary connection manager 140 and the client 105 from the statistics being monitored. For example, the secondary connection manager 150 can detect a failure of the connection, an unbalanced or overloaded workload on the primary connection, or the like. In such circumstances, the secondary connection manager assumes control of the connection and replays any rolled back transactions against the data files 130 through the secondary DBMS 155 as follows.
  • The [0034] secondary connection manager 150 communicates with the routing device 120 to acknowledge TCP requests from the client 105 to the primary connection manager 140. These acknowledgements advantageously keep the client TCP connection from timing out and failing. Additionally, the secondary connection manager 150 replays any operational statements of transactions rolled back due to, for example, the failure of the primary connection. As is generally known in the art, upon failure of a DBMS, all operational statements of open transactions (for, example, non-committed transactions) executed against the data files 130 are rolled back as if they never occurred. However, because the operational statements of open transactions are stored in the foregoing memory of the secondary connection manager 150, these operational statements from open transactions can be reexecuted against the data files 130 through the secondary DBMS 155. After replaying the foregoing operational statements, the secondary connection manager 150 begins forwarding data packets from the client 105 to the secondary DBMS 155 to be executed against the data files 130.
  • Based on the foregoing disclosure, the [0035] database cluster 110 advantageously moves a connection between the primary DBMS 145 and the client 105 to the secondary DBMS 155 in the cluster 110, when the primary DBMS 145 fails, becomes unbalanced, overloaded, or the like. Additionally, the database cluster 110 advantageously replays any rolled back statements of open transactions during fail-over to the secondary DBMS 155, thereby providing an assumption of the connection that is transparent to the client 105. Accordingly, the cluster 110 avoids failure of non-fault tolerant clients by moving the connection rather than allowing it to fail. Additionally, the cluster 110 advantageously provides communication in the native protocol of the underlying two or more DBMSs, thereby providing fault tolerant connections for the potentially geographically remote and potentially non-fault tolerant client 105.
  • FIG. 2 illustrates a block diagram of embodiments of the primary and secondary connection managers, [0036] 140 and 150, of the cluster 110, according to embodiments of the invention. FIG. 2 shows the primary connection manager 140 including a primary connection 205 communicating with a memory 210 including statistics 215, a monitor process 220 also communicating with the memory 210, and a protocol shadow 225 communicating with the memory 210 and the primary DBMS 145. Moreover, FIG. 2 shows the secondary connection manager 150 including a secondary connection 245 communicating with a memory 250 including statistics 255, a monitor process 260 also communicating with the memory 250, and a protocol shadow 265 communicating with the memory 250 and the secondary DBMS 155. In addition, the secondary connection manager 150 includes an import process 270 communicating with the primary connection 205 and a queue 275. The secondary connection manager 150 also includes a replay process 280 communicating with the queue 275 and the protocol shadow 265. Moreover, while not shown, an additional redo monitor can access one or more log files 285 associated with the primary DBMS 145. The redo monitor also can communicate with the memory 210 and review the statistics 215. FIG. 2 also shows the protocol shadow 265 accessing the one or more log files 285.
  • The following simplified exemplary transactions are disclosed to provide an understanding of the operation of the primary and secondary connection managers, [0037] 140 and 150 respectively, however, they are not intended to limit the scope of the disclosure. Rather, an artisan will recognize from the disclosure herein, alternative arrangements to simplify or expand one or more of the features or aspects disclosed herein.
  • Normal Operation
  • When the [0038] client 105 begins a transaction by issuing an operational statement to be applied against the data files 130, the client 105 distributes the statement across one or more data packets. The data packets are forwarded through the communication network 115 to the routing device 120, where, as disclosed, the routing device 120 forwards the packets to the primary connection 205 and to the secondary connection 245. The primary connection 205 examines statistics in the statistics 215 generated by the redo monitor. These statistics include, for example, the current location of transaction being stored in the log files 285. The primary connection transmits a copy of each data packet along with the current log file location, such as a sequence number, to the import process 270 of the secondary connection manager 150, and places a copy in the memory 210. The import process 270 stores the data packets in the queue 275. The protocol shadow 225 accesses the memory 210 and retrieves the data packets. The protocol shadow 225 forwards the packets to the primary DBMS 145, where the packets are assembled and the operational statement executed against the data files 130. Moreover, as is generally known in the art, the DBMS can also keep a record or log of the executed statement, generally in the log file 285.
  • The [0039] DBMS 145 forwards a result of the statement and/or and acknowledgement of receipt of the same, back to the protocol shadow 225, preferably in one or more acknowledgement data packets. The protocol shadow 225 transfers the data packets back to the memory 210, where they are picked up by the primary connection 205. The primary connection 205 forwards a copy of the data packets to the import process 270 and to the client 105. Thus, the client 105 receives the results and/or acknowledgement of the transmitted statement of an open transaction.
  • The [0040] client 105 may then desire to finalize, or commit the transaction against the data files 130. In such case, the client 105 issues a commit statement, which is forwarded to the primary DBMS 145 and the import process 270, along with the subsequent result and/or acknowledgement, in a manner similar to that disclosed. In one embodiment, the protocol shadow 225 stores sufficient data from the data packets that it can assemble the statements of a given transaction. When the protocol shadow 225 determines the data packets for a commit statement have been sent to the primary DBMS 145, the protocol shadow attaches a marker to the result/acknowledgement data packets associated with the primary DBMS 145 acknowledging execution of the commit statement. According to one embodiment, the marker comprises a location marker, such as, for example, a sequence number from the primary DBMS 145. Then, as disclosed, the result/acknowledgement data packets are transmitted with their marker to the import process 270. According to one embodiment, the import process 270 recognizes the marker placed on the data packets associated with the commit statement, and recognizes that the entire transaction has been executed by the primary DBMS 145 against the data files 130. Therefore, the import process 270 deletes the data packets associated with the now finalized transaction from the queue 275.
  • Based on the foregoing, the [0041] protocol shadow 225 and the import process 270 advantageously work together to ensure that only the data packets associated with open transactions remain in the queue 275.
  • The [0042] primary connection 205 also stores the statistics 215 related to the connection with the client 105 in the memory 210. In one embodiment, the statistics include sufficient information for the monitor process 220 to determine whether the primary connection 205 has failed, is processing an unbalanced or overloaded workload, or the like, and whether the secondary connection 245 has failed, is processing an unbalanced or overloaded workload, or the like. For example, the statistics 215 can include the number of clients seen by the primary connection 205, the number of clients seen by the secondary connection 245, the status of communication with secondary communication manager 150, or the like. The primary connection 205 acquires the statistics 215 corresponding to information from the secondary connection manager 150 through the connection between the primary connection 205 and the secondary connection 245. Moreover, according to one embodiment, the foregoing status of the secondary communication manger 150 can be ascertained through straightforward ping or ping-like commands.
  • Fail-Over
  • FIG. 3 illustrates a flow chart of a fail-over [0043] process 300, according to embodiments of the invention. As shown in FIG. 3, the fail-over process 300 begins with BLOCK 305 where the cluster 110 monitors the statistics of one or more connections with one or more clients. In the foregoing example, the monitoring corresponds to the monitor processes 220 and 260. In BLOCK 310, the cluster 110 detects the need to move the connection from one DBMS to another. For example, the monitor 260 may determine that the primary DBMS 145 has failed, become unbalanced, overloaded, or the like, and determine that the secondary connection manager 150 should assume the connection with the client 105. When the determination that a connection move is desired, the fail-over process 300 proceeds to BLOCK 320, where the cluster 110 moves the connection from one DBMS to another without losing the connection or causing even a non-fault tolerant client to fail. For example, the secondary connection 245 can communicate with the routing device 120 to assume the IP address (or VIP) of the primary DBMS 145. Additionally, the secondary connection manager 150 can replay all statements of open transactions which were rolled back in the data files 130. Accordingly, the move is transparent to the client 105 who does not lose the connection and does not know that a change has been made.
  • According to one embodiment, [0044] BLOCK 320 can include SUBBLOCK 321, where the cluster 110 instructs the routing device 120 to forward communication from the client to another DBMS. For example, as disclosed, the secondary connection 245 can assume the IP address of the primary DBMS 145. BLOCK 320 can also include SUBBLOCK 322, where the cluster 110 can send a keepalive message to one or more clients to ensure against failure of the connection to the same. According to one embodiment, the client 105 resends data packets which are not responded to or otherwise acknowledged by the cluster 110. When the client 105 resends the same data packets a predetermined amount of times, the client 105 may register a failure of the connection, thereby causing non-fault tolerant clients (such as those clients not programmed to recover) to also fail. Thus, during the fail-over process 300, the cluster 110 can respond to the client 105 with a message or acknowledgement that keeps the client 105 from resending the same data packets, therefore keeping the client from determining that the connection has failed. According to one embodiment, the secondary connection 245 sends the foregoing keepalive messages.
  • [0045] BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 324 where the cluster 110 replays any statements from open transactions that were rolled back during the failure of the primary DBMS 145. For example, the replay process 280 can access the queue 275 to retrieve data packets associated with rolled back transactions and to forward them to the protocol shadow 265. For example, as disclosed in the foregoing, the import process 270 removes the statements associated with all finalized or committed transactions, thereby leaving only rolled back transactions in the queue 275.
  • [0046] BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 326 where the cluster 110 removes any leftover committed transactions that may have slipped through. For example, it is possible that Host A 125 can fail after the primary DBMS 145 executes a commit statement for a particular transaction, but before the result/acknowledgement of the same can be transmitted to the import process 270. Thus, the secondary connection manager 150 believes the statements associated with the foregoing transaction were rolled back, e.g., because they were left in the queue 275, and therefore, the replay process 280 will forward the already committed statements to the protocol shadow 265. In one embodiment, the protocol shadow 265 parses the log file 285 of the primary DBMS 145 to ensure a commit statement associated with the open transaction was not received. When the protocol shadow 265 determines that a commit statement was received, the protocol shadow 265 deletes the statements associated therewith before their associated data packets are forwarded to the secondary DBMS 155 to be executed against the data files 130.
  • [0047] BLOCK 320 of the fail-over process 300 can also include SUBBLOCK 328 where the cluster 110 establishes communication between the client and the secondary DBMS. For example, after all rolled back statements are either executed against the data files 130 through the secondary DBMS 155 or deleted from the queue 275 by the protocol shadow 265, the protocol shadow 265 begins accessing new data packets stored in the memory 250 by the secondary connection 245 after it assumed the connection to the client 105 from the primary connection manager 140. Thus, after bringing the secondary DBMS 155 back up to the point of failure of the primary DBMS 145, the secondary connection manager 150 performs operations similar to the normal operations of the primary connection manager 140 as disclosed above.
  • According to one embodiment, the system administrator of the [0048] database cluster 110 can designate whether the secondary connection manager 150 through the monitor process 260 fails-back to the primary connection manager 140 after the cause of failure of the same is repaired, or whether the secondary connection manager 150 simply becomes the primary and vice versa.
  • Although the foregoing invention has been described in terms of certain preferred embodiments, other embodiments will be apparent to those of ordinary skill in the art from the disclosure herein. For example, the data packets captured from the [0049] primary connection manager 140 can be replicated to other DBMSs by replaying the same on the other DBMSs. This replication has several advantages over other replication techniques including a potential reduction in the traffic keeping the database cluster synchronized, thereby advantageously providing economical replication of geographically diverse data files.
  • The captured data packets can also be used to assist a transaction log based replication system. For example, the data packets can be directed to the other databases in the cluster prior to committing the transactions. Accordingly, committed transactions on a particular DBMS are not lost when the DBMS fails, as these transactions may advantageously be replayed on the other DBMSs in the cluster. [0050]
  • The captured data packets can also be used to assist a transaction log based replication system when posting replicated modifications. Some modifications (such as a vertical table update or DDL operation) may be difficult to replicate via a log-based replication. When the original data packets are available, posting the original SQL rather than the data from the transaction log may be more efficient and straightforward. [0051]
  • According to another embodiment, software may be added just below the [0052] client 105, thereby providing a mechanism to replay incomplete transactions. For example, a typical client application does not access the database directly, but instead uses some type of intermediate layer such as ODBC or JDBC, OCI, or the like. The foregoing added software can advantageously replace this intermediate layer.
  • Additionally, other combinations, omissions, substitutions and modifications will be apparent to the skilled artisan in view of the disclosure herein. Accordingly, the present invention is not intended to be limited by the reaction of the preferred embodiments, but is to be defined by reference to the appended claims. [0053]
  • Additionally, all publications, patents, and patent applications mentioned in this specification are herein incorporated by reference to the same extent as if each individual publication, patent, or patent application was specifically and individually indicated to be incorporated by reference. [0054]

Claims (23)

What is claimed is:
1. A database cluster which avoids client failure by connecting to multiple nodes of the cluster, the database cluster comprising:
a first computing system including:
a primary connection manager which forms a client connection with and receives transactions from at least one client, and
a primary database management system (DBMS) which communicates with the primary connection manager to receive the transactions and executes the transactions on data stored in one or more data files; and
a second computing system including:
a secondary connection manager, and
a secondary DBMS which communicates with the secondary connection manager and can access data stored in the one or more data files,
wherein when the second connection manager determines that a predetermined condition is met, the second connection manager receives data from the client connection, replays incomplete portions of open transactions on the data through the secondary DBMS, and begins to receive additional transactions from the at least one client to be executed against the one or more data files.
2. The highly available database cluster of claim 1, wherein the predetermined condition comprises a failure of the first computing system.
3. The highly available database cluster of claim 1, wherein the predetermined condition comprises a failure of the primary DBMS.
4. The highly available database cluster of claim 1, wherein the predetermined condition comprises an unbalanced workload between the first and second computing systems.
5. The highly available database cluster of claim 1, wherein the primary connection manager and the secondary connection manager communicate with one another.
6. The highly available database cluster of claim 5, wherein the primary connection manager transmits copies to the secondary connection manager of data packets which include the transactions and responses or acknowledgements to the transactions.
7. The highly available database cluster of claim 5, wherein the primary connection manager and the secondary connection manager exchange statistics in order to monitor the client connection.
8. The highly available database cluster of claim 7, wherein the statistics include the number of clients connected to the primary connection manager.
9. The highly available database cluster of claim 7, wherein the statistics include the number of clients the secondary connection manager can see connected to the primary connection manager.
10. The highly available database cluster of claim 7, wherein the statistics include whether the secondary connection manager can communicate with the primary connection manager.
11. A primary and at least one secondary connection manager of a database cluster, which manage a connection between at least one client and two or more database management systems (DBMSs), wherein the primary and at least one secondary connection manager can move the connection from the primary connection manager to the at least one secondary connection manager while providing protocols for the connection native to the two or more DBMSs, the primary and secondary connection manager comprising:
a first memory;
a primary connection configured to form a connection with a client and to place statements from transactions from the client into the first memory;
a primary protocol shadow configured to retrieve the statements and forward the statements to a primary DBMS;
a secondary memory;
a secondary connection configured to receive transactions from the connection with the client when one or more predetermined conditions are met and to place new statements from the transactions from the client into the second memory;
at least one process configured to replay any incomplete statements of open transactions; and
a secondary protocol shadow configured to connect to the at least one process until the incomplete statements are forwarded to a secondary DBMS and then to connect to the secondary memory to retrieve the new statements and forward the new statements to the secondary DBMS.
12. The primary and at least one secondary connection manager of claim 11, wherein the protocol native to the two or more DBMSs comprises SQL*Net.
13. The primary and at least one secondary connection manager of claim 11, wherein the at least one process further comprises:
an import process configured to retrieve the statements from the primary connection and store those statements associated with open transactions; and
a replay process configured to access the stored statements and to forward the stored statements to the secondary protocol shadow.
14. The primary and at least one secondary connection manager of claim 11, wherein the secondary protocol shadow is configured to access a log file of the primary DBMS to ensure against replaying of statements of closed transactions.
15. The primary and at least one secondary connection manager of claim 11, wherein the primary and secondary connections communicate with one another.
16. The primary and at least one secondary connection manager of claim 15, wherein the primary connection and the secondary connection exchange statistics in order to monitor the connection.
17. A method of providing native protocol access and transparent fail-over to a client connection thereby avoiding a client failure when a primary host fails, the method comprising:
rerouting a client connection between a first host and a client to a second host;
replaying at least one statement from open transactions, wherein the at least one statement includes a statement received but not committed by the first host when the client connection was moved from the first host; and
establishing communication between the second host and the client over the client connection.
18. The method of claim 17, further comprising sending keepalive messages to the client in order to keep the client from dropping the client connection.
19. The method of claim 17, wherein the replaying at least one statement further comprising removing leftover statements of closed transactions.
20. A method of providing transparent fail-over to a client connection thereby avoiding a client failure when a primary database management system DBMS fails, the method comprising:
monitoring statistics of a client connection between a first DBMS and a client;
determining from the statistics a need to move the client connection to a second DBMS while keeping the client connection alive from a perspective of the client;
rerouting the client connection to the second DBMS;
replaying any statements from open transactions rolled back when the client connection was moved from the first DBMS; and
establishing communication between the second DBMS and the client over the client connection.
21. A data processing system which provides transparent fail-over to a client connection, thereby avoiding a client failure when a primary host fails, the data processing system comprising:
a first host configured to accept a client connection from a client;
a connection manager which reroutes the client connection to a second host without recognition by the client; and
a replay process which forwards to the second host at least one incomplete statement from open transactions when the client connection was moved from the first host, wherein the connection manager establishes communication between the second host and the client over the client connection.
22. The data processing system of claim 21, wherein the client communication comprises a protocol native to the primary host.
23. The data processing system of claim 22, wherein the protocol comprises SQL*Net.
US10/072,317 2001-02-06 2002-02-06 Highly available database clusters that move client connections between hosts Abandoned US20020129146A1 (en)

Priority Applications (5)

Application Number Priority Date Filing Date Title
US10/072,317 US20020129146A1 (en) 2001-02-06 2002-02-06 Highly available database clusters that move client connections between hosts
US10/661,412 US7231391B2 (en) 2001-02-06 2003-09-11 Loosely coupled database clusters with client connection fail-over
US10/662,039 US20040162836A1 (en) 2001-02-06 2003-09-11 System and method for altering database requests and database responses
US11/471,313 US7512682B2 (en) 2001-02-06 2006-06-20 Database cluster systems and methods for maintaining client connections
US11/754,891 US7606839B2 (en) 2001-02-06 2007-05-29 Systems and methods for providing client connection fail-over

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US26690801P 2001-02-06 2001-02-06
US10/072,317 US20020129146A1 (en) 2001-02-06 2002-02-06 Highly available database clusters that move client connections between hosts

Related Child Applications (3)

Application Number Title Priority Date Filing Date
US10/661,412 Continuation-In-Part US7231391B2 (en) 2001-02-06 2003-09-11 Loosely coupled database clusters with client connection fail-over
US10/662,039 Continuation-In-Part US20040162836A1 (en) 2001-02-06 2003-09-11 System and method for altering database requests and database responses
US11/471,313 Continuation US7512682B2 (en) 2001-02-06 2006-06-20 Database cluster systems and methods for maintaining client connections

Publications (1)

Publication Number Publication Date
US20020129146A1 true US20020129146A1 (en) 2002-09-12

Family

ID=26753244

Family Applications (2)

Application Number Title Priority Date Filing Date
US10/072,317 Abandoned US20020129146A1 (en) 2001-02-06 2002-02-06 Highly available database clusters that move client connections between hosts
US11/471,313 Expired - Lifetime US7512682B2 (en) 2001-02-06 2006-06-20 Database cluster systems and methods for maintaining client connections

Family Applications After (1)

Application Number Title Priority Date Filing Date
US11/471,313 Expired - Lifetime US7512682B2 (en) 2001-02-06 2006-06-20 Database cluster systems and methods for maintaining client connections

Country Status (1)

Country Link
US (2) US20020129146A1 (en)

Cited By (26)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20030198228A1 (en) * 2001-06-02 2003-10-23 Naiming Shen Method and apparatus for restart communication between network elements
US20040162836A1 (en) * 2001-02-06 2004-08-19 Eyal Aronoff System and method for altering database requests and database responses
US20050038833A1 (en) * 2003-08-14 2005-02-17 Oracle International Corporation Managing workload by service
US20050038772A1 (en) * 2003-08-14 2005-02-17 Oracle International Corporation Fast application notification in a clustered computing system
US20050125557A1 (en) * 2003-12-08 2005-06-09 Dell Products L.P. Transaction transfer during a failover of a cluster controller
US20050267904A1 (en) * 2004-05-28 2005-12-01 Katsushi Yako Method and system for data processing with high availability
US20060164974A1 (en) * 2005-01-06 2006-07-27 Michael Ramalho Method of moving a transport connection among network hosts
US7231391B2 (en) 2001-02-06 2007-06-12 Quest Software, Inc. Loosely coupled database clusters with client connection fail-over
US20070174655A1 (en) * 2006-01-18 2007-07-26 Brown Kyle G System and method of implementing automatic resource outage handling
US20070294573A1 (en) * 2002-08-08 2007-12-20 Hall Adrian R Method and system for improving the availability of software processes utilizing configurable finite state tables
US7418633B1 (en) * 2004-05-13 2008-08-26 Symantec Operating Corporation Method and apparatus for immunizing applications on a host server from failover processing within a switch
US7430740B1 (en) * 2002-04-12 2008-09-30 724 Solutions Software, Inc Process group resource manager
US20090094477A1 (en) * 2002-12-17 2009-04-09 Kaminsky David L System and program product for detecting an operational risk of a node
US20100020678A1 (en) * 2007-04-25 2010-01-28 Fujitsu Limited Switching program, switching method and full duplex system
US7797565B1 (en) * 2006-04-04 2010-09-14 Symantec Operating Corporation System and method for maintaining communication protocol connections during failover
US7831686B1 (en) 2006-03-31 2010-11-09 Symantec Operating Corporation System and method for rapidly ending communication protocol connections in response to node failure
US7853579B2 (en) 2003-08-14 2010-12-14 Oracle International Corporation Methods, systems and software for identifying and managing database work
US8140888B1 (en) * 2002-05-10 2012-03-20 Cisco Technology, Inc. High availability network processing system
GB2462787B (en) * 2007-06-25 2012-07-25 Equallogic Inc Storage area network with target side recognition and routing table upload
US8713186B2 (en) 2007-03-13 2014-04-29 Oracle International Corporation Server-side connection resource pooling
US20140281672A1 (en) * 2013-03-15 2014-09-18 Aerohive Networks, Inc. Performing network activities in a network
US9282018B2 (en) 2010-07-27 2016-03-08 Aerohive Networks, Inc. Client-independent network supervision application
US20160072823A1 (en) * 2014-09-05 2016-03-10 Qualcomm Incorporated Using multiple credentials for access and traffic differentiation
US9479540B2 (en) 2013-12-13 2016-10-25 Aerohive Networks, Inc. User-based network onboarding
US9948626B2 (en) 2013-03-15 2018-04-17 Aerohive Networks, Inc. Split authentication network systems and methods
US10474653B2 (en) 2016-09-30 2019-11-12 Oracle International Corporation Flexible in-memory column store placement

Families Citing this family (13)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US8126848B2 (en) * 2006-12-07 2012-02-28 Robert Edward Wagner Automated method for identifying and repairing logical data discrepancies between database replicas in a database cluster
US20080140734A1 (en) * 2006-12-07 2008-06-12 Robert Edward Wagner Method for identifying logical data discrepancies between database replicas in a database cluster
US8949421B2 (en) * 2009-07-29 2015-02-03 Teradata Us, Inc. Techniques for discovering database connectivity leaks
US8549354B2 (en) * 2011-05-23 2013-10-01 International Business Machines Corporation Managing rollback in a transactional memory environment
US9087007B2 (en) 2013-03-14 2015-07-21 International Business Machines Corporation Generating fault tolerant connectivity API
CN104679907A (en) * 2015-03-24 2015-06-03 新余兴邦信息产业有限公司 Realization method and system for high-availability and high-performance database cluster
US10255336B2 (en) 2015-05-07 2019-04-09 Datometry, Inc. Method and system for transparent interoperability between applications and data management systems
US10594779B2 (en) * 2015-08-27 2020-03-17 Datometry, Inc. Method and system for workload management for data management systems
US10148607B2 (en) 2015-09-28 2018-12-04 Quest Software Inc. Electronic-messaging system interceptor forwarding client notifications
US11294869B1 (en) 2018-12-19 2022-04-05 Datometry, Inc. Expressing complexity of migration to a database candidate
US11475001B1 (en) 2018-12-19 2022-10-18 Datometry, Inc. Quantifying complexity of a database query
US11615062B1 (en) 2018-12-20 2023-03-28 Datometry, Inc. Emulation of database catalog for migration to a different database
US11941027B2 (en) * 2022-05-18 2024-03-26 Sap Se Scale-out of database systems in cloud computing environments

Citations (64)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4130867A (en) * 1975-06-19 1978-12-19 Honeywell Information Systems Inc. Database instruction apparatus for determining a database record type
US5222235A (en) * 1990-02-01 1993-06-22 Bmc Software, Inc. Databases system for permitting concurrent indexing and reloading of data by early simulating the reload process to determine final locations of the data
US5297279A (en) * 1990-05-30 1994-03-22 Texas Instruments Incorporated System and method for database management supporting object-oriented programming
US5325505A (en) * 1991-09-04 1994-06-28 Storage Technology Corporation Intelligent storage manager for data storage apparatus having simulation capability
US5333314A (en) * 1987-04-20 1994-07-26 Hitachi, Ltd. Distributed data base system of composite subsystem type, and method of fault recovery for the system
US5422979A (en) * 1991-09-11 1995-06-06 Siemens Aktiengesellschaft Fuzzy logic controller with optimized storage organization
US5423037A (en) * 1992-03-17 1995-06-06 Teleserve Transaction Technology As Continuously available database server having multiple groups of nodes, each group maintaining a database copy with fragments stored on multiple nodes
US5440727A (en) * 1991-12-18 1995-08-08 International Business Machines Corporation Asynchronous replica management in shared nothing architectures
US5455945A (en) * 1993-05-19 1995-10-03 Vanderdrift; Richard System and method for dynamically displaying entering, and updating data from a database
US5551020A (en) * 1994-03-28 1996-08-27 Flextech Systems, Inc. System for the compacting and logical linking of data blocks in files to optimize available physical storage
US5553303A (en) * 1990-08-31 1996-09-03 Fujitsu Limited Data processing system for dynamically switching access control process and for performing recovery process
US5596747A (en) * 1991-11-27 1997-01-21 Nec Corporation Method and apparatus for reorganizing an on-line database system in accordance with an access time increase
US5625811A (en) * 1994-10-31 1997-04-29 International Business Machines Corporation Method and system for database load balancing
US5634052A (en) * 1994-10-24 1997-05-27 International Business Machines Corporation System for reducing storage requirements and transmission loads in a backup subsystem in client-server environment by transmitting only delta files from client to server
US5655081A (en) * 1995-03-08 1997-08-05 Bmc Software, Inc. System for monitoring and managing computer resources and applications across a distributed computing environment using an intelligent autonomous agent architecture
US5721915A (en) * 1994-12-30 1998-02-24 International Business Machines Corporation Interaction between application of a log and maintenance of a table that maps record identifiers during online reorganization of a database
US5761687A (en) * 1995-10-04 1998-06-02 Apple Computer, Inc. Character-based correction arrangement with correction propagation
US5774717A (en) * 1995-12-15 1998-06-30 International Business Machines Corporation Method and article of manufacture for resynchronizing client/server file systems and resolving file system conflicts
US5778392A (en) * 1996-04-01 1998-07-07 Symantec Corporation Opportunistic tile-pulling, vacancy-filling method and apparatus for file-structure reorganization
US5778377A (en) * 1994-11-04 1998-07-07 International Business Machines Corporation Table driven graphical user interface
US5796934A (en) * 1996-05-31 1998-08-18 Oracle Corporation Fault tolerant client server system
US5799322A (en) * 1995-01-24 1998-08-25 Tandem Computer, Inc. System and method for stopping updates at a specified timestamp in a remote duplicate database facility
US5822780A (en) * 1996-12-31 1998-10-13 Emc Corporation Method and apparatus for hierarchical storage management for data base management systems
US5848416A (en) * 1994-06-06 1998-12-08 Nokia Telecommunications Oy Method and apparatus for storing and retrieving data and a memory arrangement
US5893924A (en) * 1995-07-28 1999-04-13 International Business Machines Corporation System and method for overflow queue processing
US5933820A (en) * 1996-05-20 1999-08-03 International Business Machines Corporation System, method, and program for using direct and indirect pointers to logically related data and targets of indexes
US5933818A (en) * 1997-06-02 1999-08-03 Electronic Data Systems Corporation Autonomous knowledge discovery system and method
US5940832A (en) * 1994-03-10 1999-08-17 Fujitsu Limited Dynamic database structuring method and apparatus, and database clustering method and apparatus
US5943677A (en) * 1997-10-31 1999-08-24 Oracle Corporation Sparsity management system for multi-dimensional databases
US5948108A (en) * 1997-06-12 1999-09-07 Tandem Computers, Incorporated Method and system for providing fault tolerant access between clients and a server
US5951694A (en) * 1995-06-07 1999-09-14 Microsoft Corporation Method of redirecting a client service session to a second application server without interrupting the session by forwarding service-specific information to the second server
US5951695A (en) * 1997-07-25 1999-09-14 Hewlett-Packard Company Fast database failover
US5956489A (en) * 1995-06-07 1999-09-21 Microsoft Corporation Transaction replication system and method for supporting replicated transaction-based services
US5956504A (en) * 1996-03-04 1999-09-21 Lucent Technologies Inc. Method and system for compressing a data stream in a database log so as to permit recovery of only selected portions of the data stream
US5978594A (en) * 1994-09-30 1999-11-02 Bmc Software, Inc. System for managing computer resources across a distributed computing environment by first reading discovery information about how to determine system resources presence
US5983239A (en) * 1997-10-29 1999-11-09 International Business Machines Corporation Storage management system with file aggregation supporting multiple aggregated file counterparts
US5991761A (en) * 1997-01-10 1999-11-23 Bmc Software, Inc. Method of reorganizing a data entry database
US5995958A (en) * 1997-03-04 1999-11-30 Xu; Kevin Houzhi System and method for storing and managing functions
US6003022A (en) * 1994-06-24 1999-12-14 International Buisness Machines Corporation Database execution cost and system performance estimator
US6016497A (en) * 1997-12-24 2000-01-18 Microsoft Corporation Methods and system for storing and accessing embedded information in object-relational databases
US6029195A (en) * 1994-11-29 2000-02-22 Herz; Frederick S. M. System for customized electronic identification of desirable objects
US6067545A (en) * 1997-08-01 2000-05-23 Hewlett-Packard Company Resource rebalancing in networked computer systems
US6070170A (en) * 1997-10-01 2000-05-30 International Business Machines Corporation Non-blocking drain method and apparatus used to reorganize data in a database
US6119128A (en) * 1998-03-30 2000-09-12 International Business Machines Corporation Recovering different types of objects with one pass of the log
US6122640A (en) * 1998-09-22 2000-09-19 Platinum Technology Ip, Inc. Method and apparatus for reorganizing an active DBMS table
US6151607A (en) * 1997-03-10 2000-11-21 Microsoft Corporation Database computer system with application recovery and dependency handling write cache
US6157932A (en) * 1998-06-04 2000-12-05 Aspect Telecommunications Method of updating a redundant service system while preserving transaction data in a database featuring on-line resynchronization
US6253121B1 (en) * 1998-09-03 2001-06-26 Balboa Instruments, Inc. Control system for bathers with ground continuity and ground fault detection
US6289357B1 (en) * 1998-04-24 2001-09-11 Platinum Technology Ip, Inc. Method of automatically synchronizing mirrored database objects
US6343296B1 (en) * 1999-09-03 2002-01-29 Lucent Technologies Inc. On-line reorganization in object-oriented databases
US6411964B1 (en) * 1998-12-23 2002-06-25 International Business Machines Corporation Methods for in-place online reorganization of a database
US20020091712A1 (en) * 2000-10-28 2002-07-11 Martin Andrew Richard Data-base caching system and method of operation
US6421688B1 (en) * 1999-10-20 2002-07-16 Parallel Computers Technology, Inc. Method and apparatus for database fault tolerance with instant transaction replication using off-the-shelf database servers and low bandwidth networks
US20020111945A1 (en) * 2000-12-15 2002-08-15 Young William J. System for interfacing an application program with diverse databases
US6460048B1 (en) * 1999-05-13 2002-10-01 International Business Machines Corporation Method, system, and program for managing file names during the reorganization of a database object
US6470344B1 (en) * 1999-05-29 2002-10-22 Oracle Corporation Buffering a hierarchical index of multi-dimensional data
US6477535B1 (en) * 1998-11-25 2002-11-05 Computer Associates Think Inc. Method and apparatus for concurrent DBMS table operations
US6523035B1 (en) * 1999-05-20 2003-02-18 Bmc Software, Inc. System and method for integrating a plurality of disparate database utilities into a single graphical user interface
US6584474B1 (en) * 1998-08-31 2003-06-24 Computer Associates Think, Inc. Method and apparatus for fast and comprehensive DBMS analysis
US6691139B2 (en) * 2001-01-31 2004-02-10 Hewlett-Packard Development Co., Ltd. Recreation of archives at a disaster recovery site
US6728780B1 (en) * 2000-06-02 2004-04-27 Sun Microsystems, Inc. High availability networking with warm standby interface failover
US7003531B2 (en) * 2001-08-15 2006-02-21 Gravic, Inc. Synchronization of plural databases in a database replication system
US20060143246A1 (en) * 1999-12-23 2006-06-29 Jeffrey Phillips Method and apparatus for managing information related to storage activities of data storage systems
US7136918B2 (en) * 1997-11-07 2006-11-14 Hitachi, Ltd. Method for monitoring abnormal behavior in a computer system

Family Cites Families (12)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
JP2708386B2 (en) 1994-03-18 1998-02-04 インターナショナル・ビジネス・マシーンズ・コーポレイション Method and apparatus for recovering duplicate database through simultaneous update and copy procedure
US5603026A (en) 1994-12-07 1997-02-11 Xerox Corporation Application-specific conflict resolution for weakly consistent replicated databases
US5761677A (en) 1996-01-03 1998-06-02 Sun Microsystems, Inc. Computer system method and apparatus providing for various versions of a file without requiring data copy or log operations
US6199110B1 (en) * 1997-05-30 2001-03-06 Oracle Corporation Planned session termination for clients accessing a resource through a server
US6148410A (en) * 1997-09-15 2000-11-14 International Business Machines Corporation Fault tolerant recoverable TCP/IP connection router
US6925477B1 (en) 1998-03-31 2005-08-02 Intellisync Corporation Transferring records between two databases
US6253212B1 (en) 1998-06-23 2001-06-26 Oracle Corporation Method and system for maintaining checkpoint values
US6256634B1 (en) 1998-06-30 2001-07-03 Microsoft Corporation Method and system for purging tombstones for deleted data items in a replicated database
US6393485B1 (en) * 1998-10-27 2002-05-21 International Business Machines Corporation Method and apparatus for managing clustered computer systems
US7487232B1 (en) * 2000-09-13 2009-02-03 Fortinet, Inc. Switch management system and method
US20020107966A1 (en) * 2001-02-06 2002-08-08 Jacques Baudot Method and system for maintaining connections in a network
US7213246B1 (en) * 2002-03-28 2007-05-01 Veritas Operating Corporation Failing over a virtual machine

Patent Citations (67)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4130867A (en) * 1975-06-19 1978-12-19 Honeywell Information Systems Inc. Database instruction apparatus for determining a database record type
US5333314A (en) * 1987-04-20 1994-07-26 Hitachi, Ltd. Distributed data base system of composite subsystem type, and method of fault recovery for the system
US5222235A (en) * 1990-02-01 1993-06-22 Bmc Software, Inc. Databases system for permitting concurrent indexing and reloading of data by early simulating the reload process to determine final locations of the data
US5297279A (en) * 1990-05-30 1994-03-22 Texas Instruments Incorporated System and method for database management supporting object-oriented programming
US5553303A (en) * 1990-08-31 1996-09-03 Fujitsu Limited Data processing system for dynamically switching access control process and for performing recovery process
US5325505A (en) * 1991-09-04 1994-06-28 Storage Technology Corporation Intelligent storage manager for data storage apparatus having simulation capability
US5422979A (en) * 1991-09-11 1995-06-06 Siemens Aktiengesellschaft Fuzzy logic controller with optimized storage organization
US5596747A (en) * 1991-11-27 1997-01-21 Nec Corporation Method and apparatus for reorganizing an on-line database system in accordance with an access time increase
US5440727A (en) * 1991-12-18 1995-08-08 International Business Machines Corporation Asynchronous replica management in shared nothing architectures
US5423037A (en) * 1992-03-17 1995-06-06 Teleserve Transaction Technology As Continuously available database server having multiple groups of nodes, each group maintaining a database copy with fragments stored on multiple nodes
US5455945A (en) * 1993-05-19 1995-10-03 Vanderdrift; Richard System and method for dynamically displaying entering, and updating data from a database
US5940832A (en) * 1994-03-10 1999-08-17 Fujitsu Limited Dynamic database structuring method and apparatus, and database clustering method and apparatus
US5551020A (en) * 1994-03-28 1996-08-27 Flextech Systems, Inc. System for the compacting and logical linking of data blocks in files to optimize available physical storage
US5848416A (en) * 1994-06-06 1998-12-08 Nokia Telecommunications Oy Method and apparatus for storing and retrieving data and a memory arrangement
US6003022A (en) * 1994-06-24 1999-12-14 International Buisness Machines Corporation Database execution cost and system performance estimator
US5978594A (en) * 1994-09-30 1999-11-02 Bmc Software, Inc. System for managing computer resources across a distributed computing environment by first reading discovery information about how to determine system resources presence
US5634052A (en) * 1994-10-24 1997-05-27 International Business Machines Corporation System for reducing storage requirements and transmission loads in a backup subsystem in client-server environment by transmitting only delta files from client to server
US5625811A (en) * 1994-10-31 1997-04-29 International Business Machines Corporation Method and system for database load balancing
US5675791A (en) * 1994-10-31 1997-10-07 International Business Machines Corporation Method and system for database load balancing
US5778377A (en) * 1994-11-04 1998-07-07 International Business Machines Corporation Table driven graphical user interface
US6029195A (en) * 1994-11-29 2000-02-22 Herz; Frederick S. M. System for customized electronic identification of desirable objects
US6026412A (en) * 1994-12-30 2000-02-15 International Business Machines Corporation Interaction between application of a log and maintenance of a table that maps record identifiers during online reorganization of a database
US5721915A (en) * 1994-12-30 1998-02-24 International Business Machines Corporation Interaction between application of a log and maintenance of a table that maps record identifiers during online reorganization of a database
US5799322A (en) * 1995-01-24 1998-08-25 Tandem Computer, Inc. System and method for stopping updates at a specified timestamp in a remote duplicate database facility
US5655081A (en) * 1995-03-08 1997-08-05 Bmc Software, Inc. System for monitoring and managing computer resources and applications across a distributed computing environment using an intelligent autonomous agent architecture
US5956489A (en) * 1995-06-07 1999-09-21 Microsoft Corporation Transaction replication system and method for supporting replicated transaction-based services
US5951694A (en) * 1995-06-07 1999-09-14 Microsoft Corporation Method of redirecting a client service session to a second application server without interrupting the session by forwarding service-specific information to the second server
US5893924A (en) * 1995-07-28 1999-04-13 International Business Machines Corporation System and method for overflow queue processing
US5761687A (en) * 1995-10-04 1998-06-02 Apple Computer, Inc. Character-based correction arrangement with correction propagation
US5774717A (en) * 1995-12-15 1998-06-30 International Business Machines Corporation Method and article of manufacture for resynchronizing client/server file systems and resolving file system conflicts
US5956504A (en) * 1996-03-04 1999-09-21 Lucent Technologies Inc. Method and system for compressing a data stream in a database log so as to permit recovery of only selected portions of the data stream
US5778392A (en) * 1996-04-01 1998-07-07 Symantec Corporation Opportunistic tile-pulling, vacancy-filling method and apparatus for file-structure reorganization
US5933820A (en) * 1996-05-20 1999-08-03 International Business Machines Corporation System, method, and program for using direct and indirect pointers to logically related data and targets of indexes
US5796934A (en) * 1996-05-31 1998-08-18 Oracle Corporation Fault tolerant client server system
US5822780A (en) * 1996-12-31 1998-10-13 Emc Corporation Method and apparatus for hierarchical storage management for data base management systems
US5991761A (en) * 1997-01-10 1999-11-23 Bmc Software, Inc. Method of reorganizing a data entry database
US5995958A (en) * 1997-03-04 1999-11-30 Xu; Kevin Houzhi System and method for storing and managing functions
US6151607A (en) * 1997-03-10 2000-11-21 Microsoft Corporation Database computer system with application recovery and dependency handling write cache
US5933818A (en) * 1997-06-02 1999-08-03 Electronic Data Systems Corporation Autonomous knowledge discovery system and method
US5948108A (en) * 1997-06-12 1999-09-07 Tandem Computers, Incorporated Method and system for providing fault tolerant access between clients and a server
US5951695A (en) * 1997-07-25 1999-09-14 Hewlett-Packard Company Fast database failover
US6067545A (en) * 1997-08-01 2000-05-23 Hewlett-Packard Company Resource rebalancing in networked computer systems
US6070170A (en) * 1997-10-01 2000-05-30 International Business Machines Corporation Non-blocking drain method and apparatus used to reorganize data in a database
US6519613B1 (en) * 1997-10-01 2003-02-11 International Business Machines Corporation Non-blocking drain method and apparatus for use in processing requests on a resource
US5983239A (en) * 1997-10-29 1999-11-09 International Business Machines Corporation Storage management system with file aggregation supporting multiple aggregated file counterparts
US5943677A (en) * 1997-10-31 1999-08-24 Oracle Corporation Sparsity management system for multi-dimensional databases
US7136918B2 (en) * 1997-11-07 2006-11-14 Hitachi, Ltd. Method for monitoring abnormal behavior in a computer system
US6016497A (en) * 1997-12-24 2000-01-18 Microsoft Corporation Methods and system for storing and accessing embedded information in object-relational databases
US6119128A (en) * 1998-03-30 2000-09-12 International Business Machines Corporation Recovering different types of objects with one pass of the log
US6289357B1 (en) * 1998-04-24 2001-09-11 Platinum Technology Ip, Inc. Method of automatically synchronizing mirrored database objects
US6157932A (en) * 1998-06-04 2000-12-05 Aspect Telecommunications Method of updating a redundant service system while preserving transaction data in a database featuring on-line resynchronization
US6584474B1 (en) * 1998-08-31 2003-06-24 Computer Associates Think, Inc. Method and apparatus for fast and comprehensive DBMS analysis
US6253121B1 (en) * 1998-09-03 2001-06-26 Balboa Instruments, Inc. Control system for bathers with ground continuity and ground fault detection
US6122640A (en) * 1998-09-22 2000-09-19 Platinum Technology Ip, Inc. Method and apparatus for reorganizing an active DBMS table
US6477535B1 (en) * 1998-11-25 2002-11-05 Computer Associates Think Inc. Method and apparatus for concurrent DBMS table operations
US6411964B1 (en) * 1998-12-23 2002-06-25 International Business Machines Corporation Methods for in-place online reorganization of a database
US6460048B1 (en) * 1999-05-13 2002-10-01 International Business Machines Corporation Method, system, and program for managing file names during the reorganization of a database object
US6523035B1 (en) * 1999-05-20 2003-02-18 Bmc Software, Inc. System and method for integrating a plurality of disparate database utilities into a single graphical user interface
US6470344B1 (en) * 1999-05-29 2002-10-22 Oracle Corporation Buffering a hierarchical index of multi-dimensional data
US6343296B1 (en) * 1999-09-03 2002-01-29 Lucent Technologies Inc. On-line reorganization in object-oriented databases
US6421688B1 (en) * 1999-10-20 2002-07-16 Parallel Computers Technology, Inc. Method and apparatus for database fault tolerance with instant transaction replication using off-the-shelf database servers and low bandwidth networks
US20060143246A1 (en) * 1999-12-23 2006-06-29 Jeffrey Phillips Method and apparatus for managing information related to storage activities of data storage systems
US6728780B1 (en) * 2000-06-02 2004-04-27 Sun Microsystems, Inc. High availability networking with warm standby interface failover
US20020091712A1 (en) * 2000-10-28 2002-07-11 Martin Andrew Richard Data-base caching system and method of operation
US20020111945A1 (en) * 2000-12-15 2002-08-15 Young William J. System for interfacing an application program with diverse databases
US6691139B2 (en) * 2001-01-31 2004-02-10 Hewlett-Packard Development Co., Ltd. Recreation of archives at a disaster recovery site
US7003531B2 (en) * 2001-08-15 2006-02-21 Gravic, Inc. Synchronization of plural databases in a database replication system

Cited By (50)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070226220A1 (en) * 2001-02-06 2007-09-27 Quest Software, Inc. Systems and methods for providing client connection fail-over
US20040162836A1 (en) * 2001-02-06 2004-08-19 Eyal Aronoff System and method for altering database requests and database responses
US7606839B2 (en) 2001-02-06 2009-10-20 Quest Software, Inc. Systems and methods for providing client connection fail-over
US7231391B2 (en) 2001-02-06 2007-06-12 Quest Software, Inc. Loosely coupled database clusters with client connection fail-over
US20030198228A1 (en) * 2001-06-02 2003-10-23 Naiming Shen Method and apparatus for restart communication between network elements
US7430740B1 (en) * 2002-04-12 2008-09-30 724 Solutions Software, Inc Process group resource manager
US8140888B1 (en) * 2002-05-10 2012-03-20 Cisco Technology, Inc. High availability network processing system
US7779129B2 (en) * 2002-08-08 2010-08-17 International Business Machines Corporation Method and system for improving the availability of software processes utilizing configurable finite state tables
US20070294573A1 (en) * 2002-08-08 2007-12-20 Hall Adrian R Method and system for improving the availability of software processes utilizing configurable finite state tables
US20090094477A1 (en) * 2002-12-17 2009-04-09 Kaminsky David L System and program product for detecting an operational risk of a node
US20050038772A1 (en) * 2003-08-14 2005-02-17 Oracle International Corporation Fast application notification in a clustered computing system
US20050038833A1 (en) * 2003-08-14 2005-02-17 Oracle International Corporation Managing workload by service
US7664847B2 (en) 2003-08-14 2010-02-16 Oracle International Corporation Managing workload by service
US7747717B2 (en) * 2003-08-14 2010-06-29 Oracle International Corporation Fast application notification in a clustered computing system
US7853579B2 (en) 2003-08-14 2010-12-14 Oracle International Corporation Methods, systems and software for identifying and managing database work
US20050125557A1 (en) * 2003-12-08 2005-06-09 Dell Products L.P. Transaction transfer during a failover of a cluster controller
US7418633B1 (en) * 2004-05-13 2008-08-26 Symantec Operating Corporation Method and apparatus for immunizing applications on a host server from failover processing within a switch
US8201022B2 (en) 2004-05-28 2012-06-12 Hitachi, Ltd. Method and system for data processing with high availability
US7409588B2 (en) * 2004-05-28 2008-08-05 Hitachi, Ltd. Method and system for data processing with high availability
US20050267904A1 (en) * 2004-05-28 2005-12-01 Katsushi Yako Method and system for data processing with high availability
US20080301161A1 (en) * 2004-05-28 2008-12-04 Katsushi Yako Method and system for data processing with high availability
US7929422B2 (en) * 2005-01-06 2011-04-19 Cisco Technology, Inc. Method of moving a transport connection among network hosts
US20060164974A1 (en) * 2005-01-06 2006-07-27 Michael Ramalho Method of moving a transport connection among network hosts
US20070174655A1 (en) * 2006-01-18 2007-07-26 Brown Kyle G System and method of implementing automatic resource outage handling
US7831686B1 (en) 2006-03-31 2010-11-09 Symantec Operating Corporation System and method for rapidly ending communication protocol connections in response to node failure
US7797565B1 (en) * 2006-04-04 2010-09-14 Symantec Operating Corporation System and method for maintaining communication protocol connections during failover
US8713186B2 (en) 2007-03-13 2014-04-29 Oracle International Corporation Server-side connection resource pooling
JP5201134B2 (en) * 2007-04-25 2013-06-05 富士通株式会社 Redundant system, switching program and switching method
US8174966B2 (en) 2007-04-25 2012-05-08 Fujitsu Limited Switching program, switching method and duplex system
US20100020678A1 (en) * 2007-04-25 2010-01-28 Fujitsu Limited Switching program, switching method and full duplex system
JPWO2008136107A1 (en) * 2007-04-25 2010-07-29 富士通株式会社 Switching processing program, switching processing method, and complete duplex system
GB2462787B (en) * 2007-06-25 2012-07-25 Equallogic Inc Storage area network with target side recognition and routing table upload
US9699055B2 (en) 2010-07-27 2017-07-04 Aerohive Networks, Inc. Client-independent network supervision application
US9282018B2 (en) 2010-07-27 2016-03-08 Aerohive Networks, Inc. Client-independent network supervision application
US10924465B2 (en) 2013-03-15 2021-02-16 Extreme Networks, Inc. Split authentication network systems and methods
US10810095B2 (en) 2013-03-15 2020-10-20 Extreme Networks, Inc. Assigning network device subnets to perform network activities using network device information
US20140281672A1 (en) * 2013-03-15 2014-09-18 Aerohive Networks, Inc. Performing network activities in a network
US9690676B2 (en) * 2013-03-15 2017-06-27 Aerohive Networks, Inc. Assigning network device subnets to perform network activities using network device information
US9965366B2 (en) 2013-03-15 2018-05-08 Aerohive Networks, Inc. Assigning network device subnets to perform network activities using network device information
US10397211B2 (en) 2013-03-15 2019-08-27 Aerohive Networks, Inc. Split authentication network systems and methods
US9948626B2 (en) 2013-03-15 2018-04-17 Aerohive Networks, Inc. Split authentication network systems and methods
US9686319B2 (en) 2013-12-13 2017-06-20 Aerohive Networks, Inc. User-based network onboarding
US10003615B2 (en) 2013-12-13 2018-06-19 Aerohive Networks, Inc. User-based network onboarding
US10320847B2 (en) 2013-12-13 2019-06-11 Aerohive Networks, Inc. User-based network onboarding
US9479540B2 (en) 2013-12-13 2016-10-25 Aerohive Networks, Inc. User-based network onboarding
KR101854095B1 (en) 2014-09-05 2018-05-02 퀄컴 인코포레이티드 Using multiple credentials for access and traffic differentiation
US9794266B2 (en) * 2014-09-05 2017-10-17 Qualcomm Incorporated Using multiple credentials for access and traffic differentiation
US20160072823A1 (en) * 2014-09-05 2016-03-10 Qualcomm Incorporated Using multiple credentials for access and traffic differentiation
US11223628B2 (en) 2014-09-05 2022-01-11 Qualcomm Incorporated Using multiple credentials for access and traffic differentiation
US10474653B2 (en) 2016-09-30 2019-11-12 Oracle International Corporation Flexible in-memory column store placement

Also Published As

Publication number Publication date
US7512682B2 (en) 2009-03-31
US20060242297A1 (en) 2006-10-26

Similar Documents

Publication Publication Date Title
US7512682B2 (en) Database cluster systems and methods for maintaining client connections
US7231391B2 (en) Loosely coupled database clusters with client connection fail-over
US20040162836A1 (en) System and method for altering database requests and database responses
US6421688B1 (en) Method and apparatus for database fault tolerance with instant transaction replication using off-the-shelf database servers and low bandwidth networks
US10990609B2 (en) Data replication framework
EP1963985B1 (en) System and method for enabling site failover in an application server environment
US5513314A (en) Fault tolerant NFS server system and mirroring protocol
US6601101B1 (en) Transparent access to network attached devices
US5828569A (en) Method and apparatus for maintaining network connections across a voluntary process switchover
US7702791B2 (en) Hardware load-balancing apparatus for session replication
US8429450B2 (en) Method and system for coordinated multiple cluster failover
US7428657B2 (en) Method for rolling back from snapshot with log
US20020194015A1 (en) Distributed database clustering using asynchronous transactional replication
US6871296B2 (en) Highly available TCP systems with fail over connections
US20050193245A1 (en) Internet protocol based disaster recovery of a server
US6910150B2 (en) System and method for state preservation in a stretch cluster
US20030014684A1 (en) Connection cache for highly available TCP systems with fail over connections
EP1704480B1 (en) Cluster database with remote data mirroring
JP4295089B2 (en) Method and apparatus for session replication and failover
US20020107966A1 (en) Method and system for maintaining connections in a network
WO2003044697A1 (en) Data replication system and method
WO2003096231A1 (en) Database scattering system
US6061807A (en) Methods systems and computer products for error recovery of endpoint nodes
US20050108187A1 (en) Apparatus and method of heartbeat mechanism using remote mirroring link for multiple storage system
US7694012B1 (en) System and method for routing data

Legal Events

Date Code Title Description
AS Assignment

Owner name: QUEST SOFTWARE, INC., CALIFORNIA

Free format text: ASSIGNMENT OF ASSIGNORS INTEREST;ASSIGNORS:ARONOFF, EYAL;KALDERON, EYAL;ROMINE, BILL;REEL/FRAME:012874/0012

Effective date: 20020507

STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION