US20110047413A1 - Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer - Google Patents

Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer Download PDF

Info

Publication number
US20110047413A1
US20110047413A1 US12/860,343 US86034310A US2011047413A1 US 20110047413 A1 US20110047413 A1 US 20110047413A1 US 86034310 A US86034310 A US 86034310A US 2011047413 A1 US2011047413 A1 US 2011047413A1
Authority
US
United States
Prior art keywords
site
updates
client computer
client
computer
Prior art date
Legal status (The legal status is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the status listed.)
Abandoned
Application number
US12/860,343
Inventor
Robert E. McGill
Igor Tsinman
Current Assignee (The listed assignees may be inaccurate. Google has not performed a legal analysis and makes no representation or warranty as to the accuracy of the list.)
Landmark Technology Partners Inc
Original Assignee
Landmark Technology Partners Inc
Priority date (The priority date is an assumption and is not a legal conclusion. Google has not performed a legal analysis and makes no representation as to the accuracy of the date listed.)
Filing date
Publication date
Application filed by Landmark Technology Partners Inc filed Critical Landmark Technology Partners Inc
Priority to US12/860,343 priority Critical patent/US20110047413A1/en
Assigned to LANDMARK TECHNOLOGY PARTNERS, INC. reassignment LANDMARK TECHNOLOGY PARTNERS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MCGILL, ROBERT E., TSINMAN, IGOR
Assigned to LANDMARK TECHNOLOGY PARTNERS, INC. reassignment LANDMARK TECHNOLOGY PARTNERS, INC. ASSIGNMENT OF ASSIGNORS INTEREST (SEE DOCUMENT FOR DETAILS). Assignors: MCGILL, ROBERT E., TSINMAN, IGOR
Publication of US20110047413A1 publication Critical patent/US20110047413A1/en
Priority to US13/854,581 priority patent/US8949657B2/en
Abandoned legal-status Critical Current

Links

Images

Classifications

    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2002Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where interconnections or communication control functionality are redundant
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/1658Data re-synchronization of a redundant component, or initial sync of replacement, additional or spare unit
    • G06F11/1662Data re-synchronization of a redundant component, or initial sync of replacement, additional or spare unit the resynchronized component or unit being a persistent storage device
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2041Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant with more than one idle spare processing component
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/2097Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements maintaining the standby controller/processing unit updated
    • HELECTRICITY
    • H04ELECTRIC COMMUNICATION TECHNIQUE
    • H04LTRANSMISSION OF DIGITAL INFORMATION, e.g. TELEGRAPHIC COMMUNICATION
    • H04L69/00Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass
    • H04L69/40Network arrangements, protocols or services independent of the application payload and not provided for in the other groups of this subclass for recovering from a failure of a protocol instance or entity, e.g. service redundancy protocols, protocol state redundancy or protocol service redirection
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1415Saving, restoring, recovering or retrying at system level
    • G06F11/1443Transmit or communication errors
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/14Error detection or correction of the data by redundancy in operation
    • G06F11/1402Saving, restoring, recovering or retrying
    • G06F11/1471Saving, restoring, recovering or retrying involving logging of persistent data for recovery
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2038Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant with a single idle spare processing component
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F11/00Error detection; Error correction; Monitoring
    • G06F11/07Responding to the occurrence of a fault, e.g. fault tolerance
    • G06F11/16Error detection or correction of the data by redundancy in hardware
    • G06F11/20Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements
    • G06F11/202Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant
    • G06F11/2048Error detection or correction of the data by redundancy in hardware using active fault-masking, e.g. by switching out faulty elements or by switching in spare elements where processing functionality is redundant where the redundant components share neither address space nor persistent storage
    • GPHYSICS
    • G06COMPUTING; CALCULATING OR COUNTING
    • G06FELECTRIC DIGITAL DATA PROCESSING
    • G06F2201/00Indexing scheme relating to error detection, to error correction, and to monitoring
    • G06F2201/855Details of asynchronous mirroring using a journal to transfer not-yet-mirrored changes

Definitions

  • a distributed service may employ data replication, where a service maintains multiple copies of data to permit local access at multiple locations, or to increase availability when a server process may have crashed.
  • Data caching techniques where a copy of requested data is copied and retained locally, can also be used to reduce the number and frequency of service calls made from a client to a server. For example, web browsers use document caching to speed up access to frequently used documents. Data caching is similar to data replication, but cached data can become stale when it remains on a local computer too long. Other systems employ other techniques to address communication interruptions and service failures.
  • the network system 100 includes client computer 199 , although the network system 100 can include other types of private and public networks that include other devices. Communications, such as requests from client computers 199 and responses from sites 101 , 102 take place over the network system 100 according to standard network protocols, such as the HTTP and TCP/IP protocols, but the principles discussed are not limited to this example and can include other protocols. Further, the network system 100 can include local area networks (LANs), wide area networks (WANs), direct connections and any combination thereof, other types and numbers of network types.
  • LANs local area networks
  • WANs wide area networks
  • direct connections and any combination thereof other types and numbers of network types.
  • the network interface unit 1024 is an FPGA that can include a local memory and be configured with logic to implement one or more aspects of the technology, including by way of example only, automatically detecting service tier failures and rerouting client requests and updates to secondary sites or secondary services on the primary site to maintain application services, although the network interface unit 1024 can include other types of configurable hardware, such as digital signal processors, micro-controllers, ASICs, PLDs, FPLDs, and the like, programmed or configured according to the teachings as described and illustrated herein with respect to FIGS. 1-9 , as well as software executed by the system processor 1020 , combinations thereof, and other components and elements configured in other manners which could implement one or more aspects of the technology.
  • the use of specialized hardware in this example allows the network interface unit 1024 to rapidly process network data packets.
  • each of the servers, sites 101 , 102 , and client computers 199 can include the processor 1020 , memory 1022 , network interface controller 1024 , and I/O interface 1029 coupled together by a bus 1026 , two or more computing systems or devices can be substituted for any one of the devices in the system 100 . Accordingly, principles and advantages of distributed processing, such as redundancy, replication, and the like, also can be implemented as desired to increase the robustness and performance of the devices and systems of the system 100 .
  • the system 100 can also be implemented on a computer system or systems that extend across any network environment using any suitable interface mechanisms and communications technologies including, for example telecommunications in any suitable form (e.g., voice, modem, and the like), Public Switched Telephone Network (PSTNs), Packet Data Networks (PDNs), the Internet, intranets, a combination thereof, and the like.
  • PSTNs Public Switched Telephone Network
  • PDNs Packet Data Networks
  • the Internet e.g., intranets, a combination thereof, and the like.
  • Primary site 101 and secondary site 102 include databases and servers that include one or more computing machines capable of operating one or more web-based or network based applications that can be accessed by network devices in the network system 100 .
  • Client device 199 is an example of a computing machine and can provide other data representing requested resources, such as particular web page(s), image(s) of physical objects, business objects, and any other objects, responsive to the requests, although the servers can perform other tasks and provide other types of resources.
  • servers 103 , 105 , 107 , 109 are shown in the primary site 101 depicted in FIG. 1 , other numbers and types of servers can be coupled to the primary site 101 and network system 100 . It is also contemplated that one or more of the servers 103 , 105 , 107 , 109 can be a cluster of servers managed on the network system 100 .
  • the client 199 can use network traffic characteristics such as connectivity, one-way delay and loss, round-trip delay, delay variation, loss patterns, packet reordering, bulk transport capacity, link bandwidth capacity, packet duplication and network capacity metrics with which to select a secondary site to fail over to. These metrics can be determined and stored using a routing table (not shown separately) in client 199 .
  • the routing table logs and tracks “where” changes made by an intelligent client in a particular session have been sent.
  • a change transaction or a delete transaction received prior to an add transaction can manifest as a group identifier (GID) not found.
  • GID group identifier
  • the system 400 can then direct secondary site 402 to store the change transaction or delete transaction in a pending queue and register interest in the missing GID.

Abstract

Intelligent client computing devices track and record the changes they make to data, applications, and services. Systems, devices, and computer readable media for detecting service tier failures and maintaining application services provide a resilient client architecture that allows a client application on an intelligent client to automatically detect the unavailability of server tiers or sites and re-route requests and updates to secondary sites to maintain application services at the client tier in a manner that is transparent to a user. The resilient client architecture understands the level of currentness of secondary sites in order to select the best secondary site and to automatically and transparently bring this secondary site up to date to ensure no data updates are missing from the secondary site.

Description

    CROSS-REFERENCE TO RELATED APPLICATIONS
  • This application claims priority from U.S. Provisional Patent Application No. 61/235,464, filed on Aug. 20, 2009, the entire contents of which are incorporated herein by reference.
  • TECHNOLOGICAL FIELD
  • This technology generally relates to client-server networks, and more particularly to systems and methods for detecting service failures on a primary site and routing requests to secondary sites to maintain application services at the client tier.
  • BACKGROUND
  • Client-server computing models include a distributed application structure that divides tasks between service providers (servers) and service requesters (clients). Clients and servers communicate over computer networks where clients request content on the servers or service functions provided by the servers. Clients initiate communication sessions with servers, which monitor incoming requests. Servers share their content or functions or computing resources with the requesting clients. Clients and servers form two tiers of a computer architecture, where the first tier, the servers, provides information based upon queries submitted by the second tier, the client application.
  • Servers share functions with clients such as email exchange, web access, and database access. A web browser is a client program running on a client's computer that accesses information stored on a web server on the Internet. Other users access banking services from their computers with a web browser client that sends a request to a web server at a bank. The bank program might perform additional functions on the client request such as forwarding the request to a separate database client program that sends a request to a database server at another bank computer to retrieve account information, such as an account balance. The balance information is then returned to the bank database client, which in turn serves it back to the web browser client displaying the results to the user.
  • The many client requests and server responses provide many points on the network where communication may be interrupted or service of the requests can fail. To reduce the frequency of communication interruptions and service request failures, a distributed service may employ data replication, where a service maintains multiple copies of data to permit local access at multiple locations, or to increase availability when a server process may have crashed. Data caching techniques, where a copy of requested data is copied and retained locally, can also be used to reduce the number and frequency of service calls made from a client to a server. For example, web browsers use document caching to speed up access to frequently used documents. Data caching is similar to data replication, but cached data can become stale when it remains on a local computer too long. Other systems employ other techniques to address communication interruptions and service failures.
  • Resilient computer systems are those computer systems that tolerate faults or communication interruptions or service request failures and provide an acceptable quality of service in spite of the interruptions. Resilient computer systems can withstand service interruptions without permanently failing. Resilient systems recover from service interruptions and make adjustments to continue to provide service.
  • In order to recover quickly or provide service alternatives, a resilient computer system or computer application needs to recognize the temporary loss of communication connection between the client and the server. The client application must be aware that the connection will be temporarily unavailable and must retry access rather than ending or initiating a failover. Similarly, server applications need to be aware that the communication connection is no longer available. Different approaches have been used in the past to provide resilient computer systems that provide persistent service delivery and can justifiably be trusted when servicing user requests.
  • SUMMARY
  • One example of the claimed invention is a resilient client architecture that allows a client application to automatically detect service failures on a server tier and/or a complete site failure at the server tier and re-route requests and updates to secondary sites to maintain application services at the client tier transparently to a user. In addition, the resilient client architecture employed is able to understand the level of currentness of secondary sites in order to select the best secondary site to which requests are to be sent and to automatically and transparently bring this secondary site up to date so that no data updates are missing from the secondary site. The resilient client architecture can be used to recover from a site failure, from a server failure, and/or from a service failure. For example, a site can include multiple servers. The servers, in turn, can each be providing multiple services to a client or clients. The resilient client architecture of the intelligent client keeps track of updates the client makes to data representing requested resources, such as particular web page(s), image(s) of physical objects, business objects, database fields, application processes, and other objects that can be provided by a server in response to client requests. The resilient client architecture tracks and stores changes that the client makes to the data, so that in the event of a service failure, a server failure, or a site failure, the changes to the data may be replayed to at least one of a secondary service, a secondary server, or a secondary site. Regardless of the point of failure, the resilient client architecture provides a process to replay data updates to the particular point of failure regardless of whether the failure occurred at the service level, the server level, or the site level.
  • In this disclosure, many of the examples discuss secondary sites used to provide failover synchronization to a client computer. However, it should be understood that secondary services and secondary servers in accordance with the claimed invention can also provide failover synchronization to a client computer, depending upon where the particular point of failure occurs. Additionally, the secondary services and secondary servers can reside and/or be provided by the same physical computing device as the primary services and primary servers or can reside and/or be provided by a physically separate computing device.
  • In another example implementation, the secondary sites are used to provide application services to clients to offload the primary site and to provide scalability to applications in addition to redundancy. For example, the secondary sites can be used to balance the workload across servers or across clusters of servers.
  • One example in the following disclosure is a resilient client failover synchronization algorithm and architecture. When a communication interruption occurs, and the service of a client request is provided by a secondary server site, the client must assess the “current-ness” of the data on the secondary server site. “Current-ness” is a measure of the amount of time the data has been on the secondary server site. For example, data stored on the secondary server five minutes ago is more current than data stored on the secondary server one hour ago. The currentness of the data is a function of the frequency in which data is offloaded from a primary site to a secondary site. In one example system below, the frequency in which data is offloaded can be established using a heartbeat signal from the primary site to the secondary site, which establishes a regular time frame to store and forward data from the primary site to the secondary site. For example, data can be stored and forwarded from the primary site to the secondary site every five minutes, every five milliseconds, or at other regular intervals.
  • In one example, the resilient client architecture with failover synchronization operates in a networked system that includes at least one intelligent client computer device, a primary site, and at least one secondary site. The client computer can include a client object factory, a transaction log, an application services routing table, and a console. The intelligent client computer device communicates over a computer network and requests content and/or services from the primary site. The primary site can monitor incoming requests and share its content or functions or computing resources with the requesting intelligent client computing device. The primary site can include a web server, an application server, and a database server. The primary site can also include a transaction log, an application services routing table, and a console as well. The servers can be synchronous or asynchronous, or a combination of synchronous and asynchronous. For example, communication events can be synchronous events, and deferred events can be asynchronous and can be handled by an appropriate server.
  • The intelligent client computer can be configured with the resilient client architecture to carry out the failover synchronization methods described below. Alternatively, the primary and/or secondary site(s) can also be configured to operate in conjunction with the intelligent client computer to perform the failover synchronization methods described below. Of course, a combination of the intelligent client computer, the primary site, and/or the secondary site(s) can also be configured to carry out the failover synchronization methods described.
  • In one example, the client server system automatically detects service tier service failures and reroutes client requests and updates to secondary service nodes at the primary site or at secondary sites to maintain application services. The client server system determines the level of currentness of secondary sites and updates the secondary sites accordingly.
  • In another example, the client server system automatically detects service tier service failures, but when the client determines that the secondary service is less current than the primary service, the client server system automatically updates the secondary service with updates from the client's local transaction log to bring the secondary service up to date. The client server system then reroutes client requests to up-to-date secondary service nodes.
  • Similarly, in another example, the client server system automatically detects service tier service failures, but the client employs an application service routing table to determine primary and eligible secondary service locations. The client server system then reroutes client requests and updates to the service location determined by the application service routing table when a primary service tier failure occurs.
  • Additionally, in another example, the client server system automatically detects service tier service failures. Upon detection of the service tier failure, the client retries the primary site a parameter driven number of times. After retrying the primary site the number of times specified by the defined parameter, the client reroutes client requests to the secondary site. The parameter specifying the number of times to retry can be defined by a user or it can be determined by the number and nature of the client request, the network configuration, the transmission parameters, and the like.
  • In one example, the client server system automatically detects service tier failures, and the client notifies a master console whenever a server tier service is unavailable. The master console can be included in the intelligent client, in a primary site, in a secondary site, or as a separate computing device. Additionally, the master console can replicate data used when a server tier service is unavailable to other intelligent clients, primary sites, and/or secondary sites. This master console is accessed when sessions are initiated to determine the sites that are available at the time of session initiation. The available sites can then be used as destinations to receive rerouted client requests in the event of a service tier failure.
  • In each of the examples, the client can be run inside of a web browser or outside of a web browser, such as by a web server in a private network. In all these environments, information and computing resources can be accessed by a software application and retrieved and presented. Information resources, including the primary and secondary sites, can be traversed regardless of whether the software application is a web browser or other application. Additionally, the techniques and processes described in the examples can also be performed when clients are connected to multiple sites that are synchronized automatically.
  • One computer-implemented resilient client failover method to maintain an up-to-date shared application service uses an intelligent client computer to determine a level of currentness of a primary site. The intelligent client computer detects that the primary site is unavailable and assesses a level of currentness of a secondary site. In detecting that the primary site is unavailable, the intelligent client computer can detect at least one of a service tier failure, an application tier failure, or a database failure, for example. The intelligent client computer determines updates made by the intelligent client computer to data for updating the primary site and replays the updates to the data made by the intelligent client computer to the secondary site.
  • Additionally, the method can include establishing a time standard between the primary site and the secondary site to determine the data currentness of the secondary site. The primary site can include at least one of a web server, an application server, or a database server. Similarly, the secondary site can also include at least one of a web server, an application server, or a database server. Further, in performing the method, the updates to data can include at least one of a client request, an application service request, or an update to a data field.
  • In assessing the data currentness of the secondary site, the failover method can include determining the most recent time synchronization of the secondary site. Also, the data currentness can be determined by the date and time of a most recent successful update to a persistent data store on the primary site.
  • In performing methods of the claimed invention, the primary site and the secondary site can reside in the same computing device, or the primary site and the secondary site can reside in separate computing devices.
  • In addition, one method of the claimed invention can also include storing the updates to data made by the intelligent client in a transaction log of the intelligent client. The updates replayed to the secondary site are performed prior to posting transactions or performing read operations on the secondary site. Additionally, replaying the updates to the secondary site can further include tracking the updates in an application service routing table of the intelligent client to determine the primary site and at least one eligible secondary site and rerouting client updates to a secondary site. The secondary site receiving the rerouted client updates can be selected based upon a number of factors, such as data currentness of eligible secondary sites, network traffic characteristics, network capacity, and the like.
  • The method of the claimed invention can also include the intelligent client computer requesting transaction services from a client object factory. The client object factory can serialize updates made by the intelligent computer to the data for updating the primary site as a transaction, and the updates can be stored as the transaction in a transaction log. The transaction can then be replayed to the secondary site when the intelligent client computer determines that the primary site is unavailable.
  • In one example, a method for automatically detecting service tier failures and rerouting client requests and updates to secondary sites maintains application services on a client server network system. The method determines the level of currentness of secondary sites and updates the secondary sites accordingly.
  • In one example, a computer readable medium has instructions stored on it for automatically detecting service tier failures and rerouting client requests and updates to secondary sites to maintain application services on a client server network system. A computer readable medium of the claimed invention includes instructions stored on it to cause at least one processor to perform steps of the failover method described above. For example, one computer readable storage media for maintaining an up-to-date shared application service includes one or more computer-readable instructions configured to cause one or more computer processors to execute operations including using an intelligent client computer to determine a level of currentness of a primary site and detecting that the primary site is unavailable. The computer readable storage media further includes one or more computer-readable instructions configured to cause one or more computer processors to execute operations including using the intelligent client computer to assess a level of currentness of a secondary site and to determine updates made by the intelligent client computer to data for updating the primary site. Additionally, the computer readable storage media includes one or more computer-readable instructions configured to cause one or more computer processors to execute operations including using the intelligent client computer to replay the updates to the data made by the intelligent client computer to the secondary site.
  • The computer readable storage media of the claimed invention can further include computer-readable instructions configured to cause one or more computer processors to execute operations including storing the updates to data made by the intelligent client in a transaction log of the intelligent client. Further, these instructions can be configured to cause one or more computer processors to execute the operations for replaying the updates to the secondary site by tracking the updates in an application service routing table of the intelligent client to determine the primary site and at least one eligible secondary site and rerouting client updates to a secondary site.
  • The computer readable storage media of the claimed invention can further include computer-readable instructions configured to cause one or more computer processors to execute operations including detecting a service tier failure, an application tier failure, and/or a database failure when detecting that the primary site is unavailable.
  • The computer readable storage media of the claimed invention can also include computer-readable instructions configured to cause one or more computer processors to execute operations including requesting transaction services from a client object factory using the intelligent client computer and serializing the updates made by the intelligent client computer to data for updating the primary site as a transaction with the client object factory. Additionally, the stored instructions can be configured to cause a processor to execute operations including storing the updates made by the intelligent client computer as the transaction in a transaction log. The computer readable storage media can also include computer-readable instructions configured to cause one or more computer processors to execute the operation of replaying the transaction to the secondary site when the intelligent client computer determines that the primary site is unavailable.
  • When executed by at least one processor, the instructions on the computer readable medium cause a processor to automatically detect service tier failures and reroute client requests and updates to secondary sites and maintain application services on a client server network system. The instructions on the computer readable medium further cause the processor to perform steps including determining the level of currentness of secondary services on primary sites and secondary sites and updating the secondary sites accordingly.
  • These and other advantages, aspects, and features will become more apparent from the following detailed description when viewed in conjunction with the accompanying drawings. Non-limiting and non-exhaustive embodiments are described with reference to the following drawings. Accordingly, the drawings and descriptions below are to be regarded as illustrative in nature, and not as restrictive.
  • BRIEF DESCRIPTION OF THE DRAWINGS
  • FIG. 1 is an example resilient client architecture system with failover synchronization.
  • FIG. 2 illustrates an example of resilient client architecture system with transaction routing and logging.
  • FIG. 3 shows an example resilient client architecture system with transaction response.
  • FIG. 4 illustrates an example resilient client architecture system with a build failover site.
  • FIG. 5 shows an example resilient client architecture system with a primary site re-build.
  • FIG. 6 shows an example resilient client architecture system with an active secondary site architecture.
  • FIG. 7 illustrates an example resilient client architecture system with server side synchronization.
  • FIG. 8 illustrates an example resilient client architecture system with synchronous event process details.
  • FIG. 9 illustrates an example resilient client architecture system with asynchronous update event process details.
  • FIG. 10 is a block diagram of an example client/server computer device in the system.
  • DETAILED DESCRIPTION
  • When a primary network system component, such as a server, database, or other network device fails or is otherwise shut down, a backup operation can be implemented to automatically switch the other network devices to a secondary database, server or network. This backup operation is often termed a “failover” operation. For example, when a network device fails, the rest of the network switches over to a secondary device. Failover is an important fault tolerance function of networks that rely on constant accessibility. The degree to which failover operations successfully, automatically, and transparently (to the user) redirect requests from the failed or unavailable primary device or system to the secondary device or system and mimic the operations of the primary device or system is a measure of the robustness of the communications network.
  • For example, a web server or application server, or application software component can crash and stop providing services to clients with or without a hardware or device failure. In this case, client requests must be handled by a different service provider running on the same or different hardware.
  • FIG. 1 is an exemplary client-server network 100 that that provides failover synchronization between primary site 101, secondary site 102, and client 199. The primary database connection 111 is between client 199 and primary site 101. In order to best mimic operations of the primary site 101, secondary site 102 must have data that is the same as that on the primary site 101. In the following examples, the primary site and secondary site(s) can include one or more servers and/or one or more shared application services.
  • The client computer 199 in this example can run interface applications such as web browsers and others that can provide an interface to make requests for (and send data to) different web server-based applications via the network system 100. A series of applications and services can run on the primary site 101 and secondary site 102 that allow the transmission of data requested by the client computers 199. The primary site 101 and secondary site 102 can provide data or receive data in response to requests directed toward the respective applications on the respective sites from the client computers 199. In accordance with the transmission control protocol (TCP), packets can be sent to the primary site 101 and secondary site 102 from the requesting client computers 199. The packets from the client computers 199 can include requests for the sites 101, 102 to send data. It is to be understood that the primary site 101 and secondary sites 102 can be hardware or software or can represent a site with a single server or with multiple servers, which can include internal or external networks. Further, additional sites and servers can be coupled to the network system 100, and many different types of applications can be available on sites coupled to the system 100.
  • Generally, client devices such as the client computers 199 can include any computing device capable of connecting to another computing device to send and receive information, including web-based information. The set of such devices can include devices that typically connect using a wired and/or a wireless communications medium, such as personal computers, desktops, laptops, mobile phones and/or smart phones, and the like. In these examples, the client devices can run web browsers that can provide an interface to make requests to different web server-based applications via the network system 100. A series of web-based applications can run on the primary site 101 and secondary site 102 that allow the transmission of data requested by the client computers 199. The client computers 199 can be further configured to engage in a secure communication with other devices and/or the sites 101, 102 using mechanisms such as Secure Sockets Layer (SSL), Internet Protocol Security (IPSec), Tunnel Layer Security (TLS), and the like.
  • In one example, the network system 100 includes client computer 199, although the network system 100 can include other types of private and public networks that include other devices. Communications, such as requests from client computers 199 and responses from sites 101, 102 take place over the network system 100 according to standard network protocols, such as the HTTP and TCP/IP protocols, but the principles discussed are not limited to this example and can include other protocols. Further, the network system 100 can include local area networks (LANs), wide area networks (WANs), direct connections and any combination thereof, other types and numbers of network types. On an interconnected set of LANs or other networks, including those based on different architectures and protocols, routers, switches, hubs, gateways, bridges, and other intermediate network devices can act as links within and between LANs and other networks to enable messages and other data to be sent from and to network devices. Also, communication links within and between LANs and other networks typically include twisted wire pair (e.g., Ethernet), coaxial cable, analog telephone lines, full or fractional dedicated digital lines including T1, T2, T3, and T4, Integrated Services Digital Networks (ISDNs), Digital Subscriber Lines (DSLs), wireless links including satellite links and other communications links known to those skilled in the relevant arts. In essence, the network system 100 includes any communication medium and method by which data can travel between client devices 199 and sites 101, 102 and these example configurations are provided by way of example only.
  • Each of the servers, sites 101, 102, and client computers 199 can include a central processing unit (CPU), controller or processor, a memory, and an interface system which are coupled together by a bus or other link, although other numbers and types of each of the components and other configurations and locations for the components can be used.
  • As shown further in FIG. 10, the client computers and server computers (herein “resilient computing devices”) shown as reference numeral 1010 described in these examples include system processor(s) 1020, system memory 1022, system I/O interface(s) 1029, and network interface controller 1024, which are coupled together by a bus 1026 or other numbers and types of links, although the resilient computing devices 1010 can include other components and elements in other configurations. In this example, the resilient computing device 1010 is implemented as a standalone device, although the resilient computing device 1010 could be implemented as a blade device in a chassis-blade implementation, for example.
  • System processor 1020 includes one or more microprocessors configured to execute computer/machine readable and executable instructions stored in system memory 1022 to implement automatic detection of service tier failures and rerouting client requests and updates to secondary sites to maintain application services on a client server network system, such as network system 100, for example. When executed by at least one processor, the instructions stored in system memory 1022 cause the processor 1020 to automatically detect service tier failures and reroute client requests and updates to secondary sites and/or to secondary services on the primary site, and maintain application services on a client server network system. The instructions on the computer readable medium, including system memory 1022, further cause the processor 1020 to perform steps including determining the level of currentness of secondary sites 102 and/or secondary services on the primary site and updating the secondary sites 102 and services accordingly.
  • The system processor 1020 is configured to execute computer/machine readable and executable instructions stored in system memory 1022 to implement one or more portions of the processes described further below in connection with FIGS. 1-9, for example, although processor 1020 can comprise other types and/or combinations of processors, such as digital signal processors, micro-controllers, switch chips, bus controllers, application specific integrated circuits (“ASICs”), programmable logic devices (“PLDs”), field programmable logic devices (“FPLDs”), field programmable gate arrays (“FPGAs”), and the like, programmed or configured according to the teachings as described and illustrated herein with respect to FIGS. 1-9.
  • System memory 1022 comprises computer readable media, namely computer readable or processor readable storage media, which are examples of machine-readable storage media. Computer readable storage/machine-readable storage media can include volatile, nonvolatile, removable, and non-removable media implemented in any method or technology for storage of information, such as computer readable/machine-executable instructions, data structures, program modules, or other data, which can be obtained and/or executed by one or more processors, such as system processor 1020, to perform actions, including implementing an operating system for controlling the general operation of resilient computing device 1010 to automatically detect service tier failures and reroute client requests and updates in accordance with the processes described further below in connection with FIGS. 1-9, for example.
  • Examples of computer readable storage media include RAM, BIOS, ROM, EEPROM, flash/firmware memory or other memory technology, CD-ROM, digital versatile disks (DVD) or other optical storage, magnetic cassettes, magnetic tape, magnetic disk storage or other magnetic storage devices, or any other medium which can be used to store the desired information, including data and/or computer-executable instructions or other machine-executable instructions, and which can be accessed by a computing or specially programmed device, such as resilient computing device 1010. When the instructions stored in system memory 1022 are run by the system processor 1020, the resilient computing device 1010 implements at least a portion of the processes described further below to detect service tier failures and maintain application services in connection with FIGS. 1-9, in addition to the various client-server management related functions, including redundancy functions, version controls, server load balancing functions, device configuration functions (e.g., defining network security policies), VPN hosting, network traffic management, loss control, and other functions.
  • System I/O interface(s) 1029 includes one or more user input and output device interface mechanisms, such as a computer keyboard, mouse, display device, and the corresponding physical ports and underlying supporting hardware and software to enable the resilient computing device 1010 to communicate with the outside environment for accepting user data input and to provide user output, although other types and numbers of user input and output devices can be used. Alternatively or in addition, as will be described in connection with network interface controller 1024 below, the resilient computing device 1010 can communicate with the outside environment for certain types of operations (e.g., configuration) via a network management port, for example.
  • Network interface controller 1024 provides physical access to a networking medium and provides a low-level addressing system, which enables the resilient computing device 1010 to engage in TCP/IP communications over network system 100 (shown in FIG. 1) and to automatically detect service tier failures and reroute client requests and updates to secondary sites and/or to secondary services on the primary site to maintain application services although the network interface controller 1024 can be constructed for use with other communication protocols and types of networks, and can include other components, and can perform other functions. Network interface controller 1024 is sometimes referred to as a transceiver, transceiving device, or network interface card (NIC), which transmits and receives network data packets to one or more networks, such as network system 100 in this example. When the resilient computing device 1010 includes more than one system processor 1020 (or a processor 1020 has more than one core), each processor 1020 (and/or core) can use the same single network interface controller 1024 or a plurality of network interface controllers 1024. Further, the network interface controller 1024 can include one or more physical ports, such as Ethernet ports, to couple the resilient computing device 1010 with other network devices, such as servers and other sites. Additionally, the network interface controller 1024 can include certain physical ports dedicated to receiving and/or transmitting certain types of network data, such as device management related data for configuring the resilient computing device 1010.
  • In one example, the network interface unit 1024 is an FPGA that can include a local memory and be configured with logic to implement one or more aspects of the technology, including by way of example only, automatically detecting service tier failures and rerouting client requests and updates to secondary sites or secondary services on the primary site to maintain application services, although the network interface unit 1024 can include other types of configurable hardware, such as digital signal processors, micro-controllers, ASICs, PLDs, FPLDs, and the like, programmed or configured according to the teachings as described and illustrated herein with respect to FIGS. 1-9, as well as software executed by the system processor 1020, combinations thereof, and other components and elements configured in other manners which could implement one or more aspects of the technology. The use of specialized hardware in this example allows the network interface unit 1024 to rapidly process network data packets.
  • Bus 1026 includes at least one internal device component communication bus, link, bridge and supporting components, such as bus controllers and/or arbiters. These devices enable the various components of the resilient computing device 1010, such as the system processor 1020, system memory 1022, system I/O interface 1029, and network interface controller 1024, to communicate, although the bus 1026 can enable one or more components of the resilient computing device 1010 to communicate with components in other devices as well. By way of example only, example buses include HyperTransport, PCI, PCI Express, InfiniBand, USB, Firewire, Serial ATA (SATA), SCSI, IDE and AGP buses, although other types and numbers of buses can be used, and the particular types and arrangement of buses will depend on the particular configuration of resilient computing device 1010.
  • While each of the servers, sites 101, 102, and client computers 199 can include the processor 1020, memory 1022, network interface controller 1024, and I/O interface 1029 coupled together by a bus 1026, two or more computing systems or devices can be substituted for any one of the devices in the system 100. Accordingly, principles and advantages of distributed processing, such as redundancy, replication, and the like, also can be implemented as desired to increase the robustness and performance of the devices and systems of the system 100. The system 100 can also be implemented on a computer system or systems that extend across any network environment using any suitable interface mechanisms and communications technologies including, for example telecommunications in any suitable form (e.g., voice, modem, and the like), Public Switched Telephone Network (PSTNs), Packet Data Networks (PDNs), the Internet, intranets, a combination thereof, and the like.
  • For clarity and brevity, in FIG. 1 a single secondary site 102 is shown connected to primary site 101 via heartbeat connection 155, but it should be understood that any number of secondary sites can be employed in the exemplary network 100. Likewise, a single client 199 is shown in FIG. 1, but any number of client devices can also use the exemplary network 100 as well. In the examples discussed, client devices are referred to by reference numeral 199 and secondary sites are referred to by reference numeral 102. Similarly, in the examples that follow, secondary sites 102 are shown, but the methods, devices, and systems shown also apply to secondary services that may be provided by, or located on, the primary site. For brevity, secondary sites 102 are called out, but it should be understood that secondary services on the primary site are also included.
  • Primary site 101 and secondary site 102 include databases and servers that include one or more computing machines capable of operating one or more web-based or network based applications that can be accessed by network devices in the network system 100. Client device 199 is an example of a computing machine and can provide other data representing requested resources, such as particular web page(s), image(s) of physical objects, business objects, and any other objects, responsive to the requests, although the servers can perform other tasks and provide other types of resources. It should be noted that while only four servers 103, 105, 107, 109 are shown in the primary site 101 depicted in FIG. 1, other numbers and types of servers can be coupled to the primary site 101 and network system 100. It is also contemplated that one or more of the servers 103, 105, 107, 109 can be a cluster of servers managed on the network system 100.
  • As detailed below, the primary site 101 and secondary site 102 can receive packets from clients 199 that can include data requested for server applications running on primary site 101 and secondary site 102. The requested packets can be routed from the client computers 199 to the primary site 101 and secondary site 102 via network system 100.
  • In order to provide an effective fail over response or to provide scalability to applications across multiple sites, the data on the secondary site 102 must be current (or nearly current) with respect to primary site 101. An SQL Router 112 can be used to make a group of database replicas spread across multiple secondary sites 102 appear to the application as a single database 110. The SQL router 112 can switch to a new database (not shown) when an existing database 102 is no longer available, can balance the network system 100 load by spreading connections across several copies, and can partition the data by selecting a database server from among multiple secondary sites 102.
  • To establish a manner of synchronizing the primary and secondary sites 101, 102, a time standard is established between the primary and secondary sites 101, 102. A network time protocol (NTP) server (not shown separately) synchronizes the clocks of the computers on the network 100. For example, primary site 101 and/or secondary site 102 can include an NTP server to establish a time synch between the client 199, primary site 101, and secondary site 102 by using an NTP server as one or more of web servers 103, 104, application servers 105, 106, 107, 108, or database servers 109, 110, or as another server connected to the same network system 100.
  • Applications route to primary data stores on the primary site 101 but fail over automatically to secondary data stores when the primary data store is not available. When failing over to the secondary site 102, the client 199 must assess the data currency of the secondary site 102 and “replay” updates that have occurred since the most recent time synchronization of the secondary site 102. The client 199 is an intelligent device in that the client 199 knows which updates have been made in a particular session. Updates are logged and tracked in transaction log 191 on the client 199. The transaction log 191 logs and tracks “what” has been changed in a particular session. The replays provide a resource-efficient series of snapshots of data that have changed. Using replays saves disk storage space and decreases data transmission times. Replays provide the incremental changed data rather than entire files or larger stores of information. The replay of the updates is performed prior to posting any transactions or performing any read operations. Otherwise, these operations may be performed using data that is not current.
  • In FIG. 1, the web server 103 maintains “currentness” time for the site 101 as a whole. The currentness time is determined by the date/time of the last successful update to the persistent data store (DB server 109) on the site 101. The currentness time can be used by the client as one measure in determining the “best” secondary site is to failover to in a multi site scenario. For example, if there are five “secondary” sites 102, each may be assessed as to how often it is updated. For example, one secondary site is updated only nightly through batch file transfer, another secondary site is updated every 30 minutes via store and forward of change logs at the database level, and a third secondary site is updated essentially in real time through the use of a SQL router. The client 199 can “choose” the best secondary site to route transactions to when its primary site fails via this published currentness value. If the client 199 is running on the primary site 101, with primary database 109, the primary site is current. However, if the client 199 is connected to the secondary site 102 with secondary database 110, then the secondary site 102 inherits the time synchronization of the secondary database 110.
  • Of course additional measures can be used to determine the best secondary site to fail over to. The client 199 can use network traffic characteristics such as connectivity, one-way delay and loss, round-trip delay, delay variation, loss patterns, packet reordering, bulk transport capacity, link bandwidth capacity, packet duplication and network capacity metrics with which to select a secondary site to fail over to. These metrics can be determined and stored using a routing table (not shown separately) in client 199. The routing table logs and tracks “where” changes made by an intelligent client in a particular session have been sent.
  • To perform the synchronization, transaction routing and logging is performed by networked system 200 as shown in FIG. 2. FIG. 2 includes reference elements 1-5 which represent the flow of information between the components of networked system 200. For example, in client 299, client application process (Apps 297) requests transaction services in block 1 from Client Object Factory (COF) 295. COF 295 is a software program or hardware device that renders services to client 299 necessary for the resilience features described, such as transaction logging and transaction routing. COF 295 can be used to create objects more simply than obtaining an object of a particular kind. In any event, in block 2, the COF 295 serializes the contents of a transaction cache (not shown separately) and then posts a completed communication event (transaction 293) to the client side transaction log 291. The COF 295 takes just the changes made to data and sends them to the transaction log 291. In this fashion, the client 199 can identify exactly what updates are needed to go to the secondary site 102. The Client Object Factory transaction log 291 keeps track of the changes. The transaction 293 can be posted to the transaction log 291 in memory, or the transaction 293 can be persisted depending on whether a browser is deployed or if the client 299 is using a disconnected version of software which allows persistence of application data between sessions.
  • In block 3, the COF 295 maintains the address of the primary site 201 by checking in internal application services routing table 289, which lists the primary site 201 and available secondary sites (not shown in FIG. 2). In block 4, the COF 295 sends the serialized transaction 293 to the IP address resolved from the app services routing table 289 and waits for a response. App services routing table 289 can include the transaction type, the site being used, such as primary site 201 or secondary site(s) (not shown in FIG. 2), and the on-line or off-line status of the sites as well as the resolved IP address of the site. In this example, the IP address resolved from the app services routing table 289 corresponds to primary site 201. When the serialized transaction 293 is received at primary site 201, the web server 203 component of the primary site 201 posts a copy of the transaction 293 to a server side transaction log 287 in block 4 a. As the transfers proceed, in block 5 the COF 295 issues a regularly scheduled heartbeat message which, among other things, refreshes updates to the app services routing table 289.
  • The process continues in FIG. 3, where the transaction response is serviced by networked system 300. In processing a transaction request, client 399 initiates a transfer with primary site 301. In block 6, the COF 395 either receives a transaction response 385 from primary site 301 within a specified time window or it does not receive a response within the time window. If COF 395 receives a transaction response 385, COF 395 delivers the transaction response 385 to the requesting application. For example, the COF 395 can make requests using an HTTP/S request response model. When a response is received, it is delivered to the proper listener (not shown separately) on the client. The listener application can run on the client and listen for and accept client connections. For example, when a user updates a customer record, a request to do is sent to the primary site 301 and when successful, primary site 301 returns a response indicating a successful update. While the application is waiting an hourglass can be displayed, and the application listens for a response. When a response is received, it is processed by the COF, which notifies the application which then closes the window and removes the hourglass. If no transaction response 385 is received from the requesting application, and the retry counter is less than a configurable maximum number of tries, the web server 303 is probed for diagnostic reasons for the lack of response and to confirm that a service interruption has occurred. For example, in block 7 a, the COF 395 pings the server 303 to determine if it is reachable. Additionally, in block 7 b, the COF 395 sends a web request to assess if the web tier (for example, web server 303) is responding. Further, in block 7 c, the COF 395 sends an app request to assess if the app tier (for example, app servers 305, 307) is responding. Similarly, in block 7 d, the COF 395 sends a dB request to assess if the db tier (for example, dB server 309) is responding. If all four queries in blocks 7 a-d are successful, then the client 399 loops back to block 6, increments the retry counter, and the COF 395 continues to attempt to receive a response from the requesting application.
  • If one or more of the four queries in blocks 7 a-7 d fails, then the app services routing table 389 is updated for that transaction in block 9. In block 9 a, a notification message is issued to a control console 383, and the message is resent to the next IP address on the app services routing table 389. That is, the transaction is resent. The heartbeat messages between client and server tiers (see block 5 in FIG. 2, for example) are running in a different thread, and they will be re-established with the secondary site once the primary site 301 is determined to have failed. For example, when the heartbeat runs every 30 seconds, once a transaction is rerouted to a secondary site and the app services routing table 389 is marked indicating that the primary site 301 is unavailable, the heartbeat thread will run down the application services routing table 389 in the same fashion as the transaction, find a new site to “heartbeat” with and initiate heartbeat communications with that secondary site. For example, if primary site 301 does not respond, the COF 395 will attempt to deliver the transaction (not shown in FIG. 3) to a secondary site (not shown in FIG. 3), which is the next IP address on the app services routing table 389. Additionally, the networked system 300 may be set up to mark all transactions as offline for an IP address that fails to respond to the pings (block 7 a) or web tier probes (block 7 b). If only the app request (block 7 c) fails, then only the application requests, known as communication events (CEs), related to that app request (synch/asynch) are marked as offline or unavailable. The layers of services running on the server tier and failures at different layers can be set to failover in different fashions. For example, when an app server fails, the system can route to a different app server instance running on the same physical server. But if a database server fails and there is no local failover copy, then the failover goes all the way to a secondary site.
  • FIG. 4 shows an example manner in which a failover site is built using the system 400. The COF 495 connects with secondary site 402 in block 10 and determines the currentness of the secondary site 402. As described above, the currentness of the secondary site 402 can be determined based upon the frequency in which data was offloaded from a primary site, such as primary site 401, to the secondary site 402. The COF 495 polls transaction log 491 in block 11 to determine if secondary site 402 is missing any transactions that are contained in the local transaction log 491. If the secondary site 402 is missing transactions, the COF 495 sends the missing transactions 493 in block 12 to web server 404 in chronological sequence.
  • In block 13, the secondary site 402 applies the missing transactions 493 by routing them from the web server 404 and application servers 406, 408 to database server 410. In a system 400 that consists of more than one active client 499, each client can failover to the secondary site at a different time as each client 499 tries to access the primary site 401 and fails at a different time. As each client 499 fails over to the secondary site 402 the secondary site 402 is rebuilt from the distributed transaction logs 491 at the client 499 as they failover to the secondary site. Because clients can failover at different times and in no predetermined order, transaction exceptions can occur when the distributed transaction logs 491 are processed. Transaction exceptions can be handled in a number of ways. For example, a change transaction or a delete transaction received prior to an add transaction can manifest as a group identifier (GID) not found. The system 400 can then direct secondary site 402 to store the change transaction or delete transaction in a pending queue and register interest in the missing GID.
  • If a change transaction is received out of order, the result can appear as an old value on the transaction 493 not equaling an old value on the database 410. The system 400 then directs secondary site 402 to check the time stamps of the change transaction. If the time stamp on the change transaction is earlier than the latest applied transaction for that GID, then the change transaction is discarded. If the time stamp on the change transaction is not earlier than the latest applied transaction for that GID, then the change transaction is applied. The timestamp corresponding to the current value on the database 410 is resolved by checking the local transaction log 488 to assess when the update to the database 410 was made. If the update on the transaction log 488 is not found—either because it was done prior to when the transaction log 488 was cleaned or by an overnight batch importer job—the system 400 assumes that the received change transaction 493 is more current than the database 410 and applies the change.
  • If a delete transaction is received out of order, this can also manifest as an old value on the transaction not equal to the old value on the database. In this case, however, the system will simply apply the transaction.
  • As shown in FIG. 5, once the failover site (secondary site 502) is built, the primary site 501 can be re-built. When the primary site 501 returns to on-line status, the system 500 can initiate a start up process to rebuild the primary site 501. As part of the start up process, primary site 501 queries the app services routing table 523 (on the console server 583) in block 14 to determine where the new master database is located. In one example, the new master database 510 is currently at the secondary site 502. Primary site 501 will check with console server 583 first, but if the console server 583 is not available or is otherwise unable to provide the location of the new master database, the primary site 501 will check other sites defined in its local copy of the app services routing table 579 to determine the location of the new master database.
  • Once the primary site 501 resolves to the new master database 510, the primary site 501 uses web server 503 to request all changes from the new master site (secondary site 502 in this example) from the point at which the primary site 501 went down or otherwise became unavailable. In block 15, web server 503 requests all changes from the transaction log 588 on the secondary site 502. The changes from the point at which the primary site 501 went down are determined from the last entry to the local transaction log 588 to the instant time at which the web server 503 requests the changes.
  • Once the primary site 501 has been restored and is fully current, the secondary site 502 continues to send changes to the primary site 501 to keep it current. Secondary site 502 then initiates a switch of the master site designation back to the primary site 501 in block 16 by updating the app service routing tables 589 and notifying all connected clients 599 to retry the primary site 501. Connected clients 599 then reconnect to the primary site 501 in block 17.
  • FIG. 6 illustrates an example system configuration with active secondary site architecture. In an advanced configuration such as in system 600, instead of having a primary site where all users connect and a secondary site that is used only when a service failure occurs, users are connected to all sites 601, 602 in the network system 600, thereby spreading the user load across more than one site. This type of configuration is useful in a network of more than two sites, because less “standby” capacity is required. That is, fewer “backup” sites or servers are required because all users are connected to all sites, and failovers can occur to any of the connected servers and/or sites. For example, in a network of ten sites, an application could have 20% extra capacity (2 extra sites) and load all users across all ten sites. Then, up to two sites could fail, and the users would be automatically redistributed across the remaining eight sites with no loss of capacity or diminishment of the user experience. To achieve the same result with a primary/secondary site orientation (as discussed above in the examples of FIGS. 1-5) would require sixteen sites—eight primary and eight secondary. An active secondary site architecture approach has added value in that sites can be located in areas close to end users to reduce message latency and to isolate sites from risks such as natural and man-made disasters.
  • To effectively implement an active secondary site architecture, such as the example depicted as system 600 in FIG. 6, the sites 601, 602 and the clients 699, 655 need to be kept very close to being synchronous. When a user updates one client site to which he is connected then the remaining sites (nine of them in this active secondary site architecture example) should be updated via an asynchronous update process so that they stay in near-time synchronization with all other the other sites. This drives inter site message traffic as one application data change needs n messages sent and processed where n=the number of sites. As such, this approach is well-suited for applications that are read oriented with relatively few updates relative to reads.
  • In FIG. 6, client 699 is connected to site 601 and client 655 is connected to site 602, and each client 699, 655 is making changes to its own database. The changes are processed as discussed in FIGS. 1-5, but after the changes are posted to the respective local databases 609, 610, the changes are routed to all other sites through a dedicated asynchronous process 651, 652. For example, when client 699 makes a change to data or applications in its own database, the change is sent to the primary site 601 as a synchronous update. Primary site 601 consults a local copy of the application services routing table 679 to determine other sites (such as secondary site 602) to which an update must be sent. The update to secondary site 602 is sent asynchronously to synchronous CE Synch process 652, which is then used by secondary site 602 to effect the update. Similarly, when updates are made by client 655, secondary site 602 routes the update to primary site 601 via a similar asynchronous process 651. Due to asynchronous processing queues and delays, transactions may be processed out of order and will be resolved in the same fashion as described above with regard to FIG. 4.
  • FIG. 7 shows an example configuration of system 700 showing server side synchronization details where each site 701, 702 services a unique set of clients 799, 755. The updates from each site 701, 702 are propagated to the other sites. Synchronous and asynchronous update transactions shown in blocks 20 and 21 are propagated at the application layer with application servers 705, 706, 707, 708 through a transaction synchronization process. For example, a synchronous transaction 20 is applied by the receiving site 702 as if the transaction 20 had been submitted locally. Likewise, transaction 21 is applied by site 701 as if the transaction 21 had been submitted locally.
  • In one example system 700, asynchronous update transactions lock down portions of the application to be updated. This is of importance when an application needs to perform a large update process, such as a request to update securities position values with end of day closing prices for hundreds of thousands of positions, for example. In this case, prior to processing this large batch update, the impacted parts of the data base 709, 710 are locked on all sites 701, 702 using transaction blocks 20 and 21. That is, asynchronous update transactions use a pessimistic application layer locking model that blocks operations of the transaction if they can cause a violation of integrity rules. Integrity rules can include properties that guarantee database transactions are processed reliably. For example, the properties may include atomicity, consistency, isolation, and durability. Once the asynchronous transactions are complete, database servers 709, 710 propagate the results of the asynchronous updates at the data base layer as shown in blocks 22 and 23 and release the remote application layer locks and return control to the submitting site.
  • FIG. 8 shows an example configuration of system 800 showing synchronous event process details where each site 801, 802 services a unique set of clients 899, 855. System 800 can be employed when an update or change to a first site is made by one (set of) client(s) and the update or change needs to be implemented on a second site. For example, a synch update is made by client 899 in block 81, and the update transaction is passed to site 801. In block 82, the synch update is applied by the application server 807 to the database server 809 at site 801 and the client transaction is completed. In block 83, application server 805 determines synch update propagation instructions from app services routing table 879. In block 84, the synch update is sent by application server 807 to distributed sites, such as site 802.
  • Distributed site 802 receives the distributed synch update 84 and uses application server 806 to apply it to the distributed data base 810 in block 85. If the distributed site data base 810 detects a conflict, the conflict can be resolved in a number of ways depending upon the type of conflict detected as described above with regard to transaction exceptions and FIG. 4.
  • FIG. 9 shows an example configuration of system 900 showing asynchronous event update process details where each site 901, 902 services a unique set of clients 999, 955. For example, an asynchronous update to be incorporated at all sites can include a commissions recalculation for a sales force. In such an example, it would be important to lock down all sites as the global asynchronous update is performed to avoid data conflicts. The lock is effected synchronously, and the application of the resulting data updates is performed asynchronously. For example, in block 91, an asynch update is made by a client in client set 999 and passed to site 901. In block 92, site 901 uses application server 907 to pass the deferred event notification 993 to the distributed sites, such as site 902. In block 93, the distributed site 902 uses application server 906 to lock the application functions that can update data so that the site 902 can be updated by the asynch update transaction 993 (target data). The site 902 then returns a lock complete message to the originating site 901(return lock block not shown).
  • In block 94, the originating site 901 uses application server 907 and runs the asynch update transaction (DE, or “deferred event”) on the database server 909 and logs the data changes to a deferred event data change log 911. After the data change is logged, in block 95 the originating site 901 uses data change log 911 and sends the deferred event 973 (DE) changes to the distributed site 902. In blocks 96 a and 96 b, data change application 940 applies the changes to the database server 910 and unlocks application components in application server 906 to effect the changes to distributed site 902.
  • To prevent a deadly embrace when two asynch update requests are made on two different sites, the originating site will check the local transaction log to see if any competing asynch update transactions are running Conflicts detected at this stage will be resolved by the user. If a distributed site receives an asynch update request that conflicts with one already in process on the local transaction log, it will return a fail message to the lock request. All asynch update requests must get success messages from all distributed sites prior to beginning execution.
  • A further example of the processing described in FIG. 9 is the commission calculation for a set of end user sales persons. Commissions calculation is typically a computationally intense process that is run in batch in off hours to minimize impact to other online users. In this example, the system will recalculate commissions for all users. The users are connected to multiple sites. First, a commissions recalculation request is made (block 91) to a site, which then has to signal all other sites that a commissions recalculation update will be run (block 92). Each remote site will receive this message and lock all data that is potentially updated by the commissions process (block 93). This lock is an application level lock that prevents users from changing data that will be changed by the batch commissions recalculation process. When each remote site responds that the locks are in place (the return lock block is not shown separately), the commissions recalculation will take place on the site on which the original commissions recalculation request was made and the results will be stored as a batch of data base changes (block 94). When the commissions recalculation is complete, the data base updates are sent to the remote sites (block 95) and applied (block 96 b). Once each site has applied the data base changes, the application level locks at that site are released (block 96 a).
  • As outlined above, each of the client computers, sites, and servers can include a central processing unit (CPU), controller or processor, a memory, and an interface system which are coupled together by a bus or other link, although other numbers and types of each of the components and other configurations and locations for the components can be used. The processors in the client computers, the servers, and the sites can execute a program of stored instructions for one or more aspects of the methods and systems as described herein, although the processor could execute other types of programmed instructions. The memory can store these programmed instructions for one or more aspects of the methods and systems as described herein, although some or all of the programmed instructions could be stored and/or executed elsewhere. A variety of different types of memory storage devices, such as a random access memory (RAM) or a read only memory (ROM) in the system or a floppy disk, hard disk, CD ROM, DVD ROM, or other computer readable medium which is read from and/or written to by a magnetic, optical, or other reading and/or writing system that is coupled to the processor, can be used for the memory. The user input device can include a computer keyboard and a computer mouse, although other types and numbers of user input devices can be used. The display can include a computer display screen, such as a CRT or LCD screen by way of example only, although other types and numbers of displays could be used.
  • Although an example of the client computers 199, sites 101, 102, and servers are described and illustrated in connection with FIG. 10, each of the computers and computing devices of the system 100 could be implemented on any suitable computer system or computing device. It is to be understood that the example devices and systems of the system 100 are for exemplary purposes, as many variations of the specific hardware and software used to implement the system 100 are possible, as will be appreciated by those skilled in the relevant art(s).
  • Further, each of the devices of the system 100 can be conveniently implemented using one or more general purpose computer systems, microprocessors, digital signal processors, micro-controllers, application specific integrated circuits (ASIC), programmable logic devices (PLD), field programmable logic devices (FPLD), field programmable gate arrays (FPGA) and the like, programmed according to the teachings as described and illustrated herein, as will be appreciated by those skilled in the computer, software, and networking arts.
  • The operation of example processes to provide a client failover synchronization shown in FIGS. 1-9 can be run on the network system 100. The flow diagrams of FIGS. 1-9 are representative of example machine readable instructions for implementing the failover synchronization and/or the process to determine data transmission failure and activate loss controls. The steps described below are example machine readable instructions for implementing a method in accordance with the examples described in this disclosure. In one example, the machine readable instructions include an algorithm for execution by: (a) a processor, (b) a controller, and/or (c) one or more other suitable processing device(s). The algorithm can be instantiated in software stored on tangible media such as, for example, a flash memory, a CD-ROM, a floppy disk, a hard drive, a digital video (versatile) disk (DVD), or other memory devices, but persons of ordinary skill in the art will readily appreciate that the entire algorithm and/or parts thereof could alternatively be executed by a device other than a processor and/or embodied in firmware or in dedicated hardware in a known manner. For example, the algorithm can be implemented by an application specific integrated circuit (ASIC), a programmable logic device (PLD), a field programmable logic device (FPLD), a field programmable gate array (FPGA), discrete logic, etc.). For example, any or all of the components of the failover synchronization system could be implemented by software, hardware, and/or firmware. Also, some or all of the machine readable instructions described herein can be implemented manually. Further, although an example of the present invention is described and illustrated herein, persons of ordinary skill in the art will readily appreciate that many other methods of implementing the example machine readable instructions can alternatively be used. For example, the order of execution can be changed, and/or some of the steps described can be changed, eliminated, or combined.
  • By performing a failover synchronization method using a system described above, if a primary site fails or is unavailable, the failover to the secondary site(s) is accomplished seamlessly without interruption to the client.
  • Having thus described the basic concept of the invention, it will be rather apparent to those skilled in the art that the foregoing detailed disclosure is intended to be presented by way of example only, and is not limiting. Various alterations, improvements, and modifications will occur and are intended to those skilled in the art, though not expressly stated herein. These alterations, improvements, and modifications are intended to be suggested hereby, and are within the spirit and scope of the invention. Additionally, the recited order of processing elements or sequences, or the use of numbers, letters, or other designations therefore, is not intended to limit the claimed processes to any order except as can be specified in the claims. Accordingly, the invention is limited only by the following claims and equivalents thereto.

Claims (26)

1. A computer-implemented resilient client failover method to maintain an up-to-date shared application service, the method comprising:
determining, with an intelligent client computer, a level of currentness of a primary site;
detecting, with the intelligent client computer, that the primary site is unavailable;
assessing, with the intelligent client computer, a level of currentness of a secondary site;
determining, with the intelligent client computer, updates made by the intelligent client computer to data for updating the primary site; and
replaying, by the intelligent client computer, the updates to the data made by the intelligent client computer to the secondary site.
2. The method of claim 1 further comprising:
establishing a time standard between the primary site and the secondary site to determine the data currentness of the secondary site.
3. The method of claim 1, wherein the primary site includes at least one of a web server, an application server, or a database server.
4. The method of claim 1, wherein the secondary site includes at least one of a web server, an application server, or a database server.
5. The method of claim 1, wherein the updates to data include at least one of a client request, an application service request, or an update to a data field.
6. The method of claim 1, wherein assessing the data currentness of the secondary site includes determining the most recent time synchronization of the secondary site.
7. The method of claim 1, wherein the data currentness is determined by the date and time of a most recent successful update to a persistent data store on the primary site.
8. The method of claim 1 further comprising:
storing the updates to data made by the intelligent client in a transaction log of the intelligent client.
9. The method of claim 8, wherein the updates replayed to the secondary site are performed prior to posting transactions or performing read operations on the secondary site.
10. The method of claim 1, wherein the primary site and the secondary site reside in the same computing device.
11. The method of claim 1, wherein the primary site and the secondary site reside in separate computing devices.
12. The method of claim 1, wherein replaying the updates to the secondary site further comprises:
tracking the updates in an application service routing table of the intelligent client to determine the primary site and at least one eligible secondary site; and
rerouting client updates to a secondary site.
13. The method of claim 12, wherein the secondary site receiving the rerouted client updates is selected based upon at least one of data currentness of eligible secondary sites, network traffic characteristics, and network capacity.
14. The method of claim 1, wherein detecting that the primary site is unavailable includes at least one of detecting a service tier failure, an application tier failure, or a database failure.
15. The method of claim 1 further comprising:
requesting, by the intelligent client computer, transaction services from a client object factory;
serializing as a transaction, by the client object factory, updates made by the intelligent client computer to data for updating the primary site;
storing the updates made by the intelligent client computer as the transaction in a transaction log; and
replaying the transaction to the secondary site when the intelligent client computer determines that the primary site is unavailable.
16. A computer readable storage media for maintaining an up-to-date shared application service, the computer readable storage media comprising one or more computer-readable instructions configured to cause one or more computer processors to execute operations comprising:
determining, with an intelligent client computer, a level of currentness of a primary site;
detecting, with the intelligent client computer, that the primary site is unavailable;
assessing, with the intelligent client computer, a level of currentness of a secondary site;
determining, with the intelligent client computer, updates made by the intelligent client computer to data for updating the primary site; and
replaying, by the intelligent client computer, the updates to the data made by the intelligent client computer to the secondary site.
17. The computer readable storage media of claim 16 further comprising computer-readable instructions configured to cause one or more computer processors to execute the operations comprising:
storing the updates to data made by the intelligent client in a transaction log of the intelligent client.
18. The computer readable storage media of claim 16, wherein the computer-readable instructions configured to cause one or more computer processors to execute the operations for replaying the updates to the secondary site further comprises computer-readable instructions configured to cause one or more computer processors to execute the operations comprising:
tracking the updates in an application service routing table of the intelligent client to determine the primary site and at least one eligible secondary site; and
rerouting client updates to a secondary site.
19. The computer readable storage media of claim 16, wherein detecting that the primary site is unavailable includes at least one of detecting a service tier failure, an application tier failure, or a database failure.
20. The computer readable storage media of claim 16, further comprising computer-readable instructions configured to cause one or more computer processors to execute the operations comprising:
requesting, by the intelligent client computer, transaction services from a client object factory;
serializing as a transaction, by the client object factory, updates made by the intelligent client computer to data for updating the primary site;
storing the updates made by the intelligent client computer as the transaction in a transaction log; and
replaying the transaction to the secondary site when the intelligent client computer determines that the primary site is unavailable.
21. An intelligent client computer device configured to implement a resilient client failover to maintain an up-to-date shared application service, the intelligent client computer device comprising:
a client object factory configured to determine a level of currentness of a primary site; and
a link to a control console storing a latest copy of an application services routing table, the intelligent client computer device configured to detect that the primary site is unavailable by accessing the application services routing table;
wherein the client object factory is further configured to assess a level of currentness of a secondary site, determine updates made by the intelligent client computer to data for updating the primary site, and to replay the updates to the data made by the intelligent client computer to the secondary site.
22. The intelligent client computer device of claim 21, wherein the client object factory is further configured to store the updates to data made by the intelligent client computer device in a transaction log.
23. The intelligent client computer device of claim 21 further comprising:
an application service routing table configured to track the updates to determine the primary site and at least one eligible secondary site and to reroute the updates to a secondary site.
24. The intelligent client computer device of claim 21, wherein the control console is further configured to detect that the primary site is unavailable by detecting at least one of a service tier failure, an application tier failure, or a database failure.
25. The intelligent client computer device of claim 21, wherein the client object factory is further configured to:
serialize updates made by the intelligent client computer to data for updating the primary site as a transaction
store the updates made by the intelligent client computer as the transaction in a transaction log; and
replay the transaction to the secondary site when the control console determines that the primary site is unavailable.
26. A computer-implemented resilient client update method using an active secondary site, the method comprising:
determining, with an intelligent client computer, a level of currentness of a primary site;
determining, with the intelligent client computer, updates made by the intelligent client computer to data for updating the primary site;
synchronously posting, with the intelligent client computer, the updates made by the intelligent client to the primary site;
assessing, with the intelligent client computer, a level of currentness of a secondary site; and
asynchronously routing the updates to the data made by the intelligent client computer from the primary site to the secondary site.
US12/860,343 2009-08-20 2010-08-20 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer Abandoned US20110047413A1 (en)

Priority Applications (2)

Application Number Priority Date Filing Date Title
US12/860,343 US20110047413A1 (en) 2009-08-20 2010-08-20 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US13/854,581 US8949657B2 (en) 2009-08-20 2013-04-01 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Applications Claiming Priority (2)

Application Number Priority Date Filing Date Title
US23546409P 2009-08-20 2009-08-20
US12/860,343 US20110047413A1 (en) 2009-08-20 2010-08-20 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Related Child Applications (1)

Application Number Title Priority Date Filing Date
US13/854,581 Continuation US8949657B2 (en) 2009-08-20 2013-04-01 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Publications (1)

Publication Number Publication Date
US20110047413A1 true US20110047413A1 (en) 2011-02-24

Family

ID=43606250

Family Applications (2)

Application Number Title Priority Date Filing Date
US12/860,343 Abandoned US20110047413A1 (en) 2009-08-20 2010-08-20 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US13/854,581 Active 2030-09-20 US8949657B2 (en) 2009-08-20 2013-04-01 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Family Applications After (1)

Application Number Title Priority Date Filing Date
US13/854,581 Active 2030-09-20 US8949657B2 (en) 2009-08-20 2013-04-01 Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Country Status (1)

Country Link
US (2) US20110047413A1 (en)

Cited By (126)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20110161281A1 (en) * 2009-12-30 2011-06-30 Sybase, Inc. Distributed Transaction Management in a Distributed Shared Disk Cluster Environment
US20110246640A1 (en) * 2010-04-06 2011-10-06 Debashis Saha Method and system for synchronous and asynchronous monitoring
US20120066394A1 (en) * 2010-09-15 2012-03-15 Oracle International Corporation System and method for supporting lazy deserialization of session information in a server cluster
US20130006949A1 (en) * 2011-06-30 2013-01-03 Tarik Essawi Systems and methods for data integrity checking
WO2012166689A3 (en) * 2011-06-01 2013-01-31 Microsoft Corporation Replaying jobs at a secondary location of a service
US20130060932A1 (en) * 2011-09-06 2013-03-07 Shachar Ofek Discovering tiers within an application
US20130091271A1 (en) * 2011-10-05 2013-04-11 Marek Piekarski Connection method
US8446920B2 (en) * 2011-06-14 2013-05-21 Mitel Networks Corporation Providing resilient digital telephony services for wireless device
WO2013074985A1 (en) * 2011-11-16 2013-05-23 Box, Inc. Resource effective incremental updating of a remote client with events which occurred via a cloud-enabled platform
US20130173538A1 (en) * 2012-01-04 2013-07-04 International Business Machines Corporation Managing remote data replication
US8495091B2 (en) * 2011-07-01 2013-07-23 Intel Corporation Dynamically routing data responses directly to requesting processor core
US8515902B2 (en) 2011-10-14 2013-08-20 Box, Inc. Automatic and semi-automatic tagging features of work items in a shared workspace for metadata tracking in a cloud-based content management system with selective or optional user contribution
US8583619B2 (en) 2007-12-05 2013-11-12 Box, Inc. Methods and systems for open source collaboration in an application service provider environment
US20130326266A1 (en) * 2012-05-31 2013-12-05 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US20140059202A1 (en) * 2012-08-23 2014-02-27 Efrat Ben David Determining the type of a network tier
US20140064279A1 (en) * 2012-08-31 2014-03-06 Omx Technology Ab Resilient peer-to-peer application message routing
US8719445B2 (en) 2012-07-03 2014-05-06 Box, Inc. System and method for load balancing multiple file transfer protocol (FTP) servers to service FTP connections for a cloud-based service
US20140129873A1 (en) * 2009-08-20 2014-05-08 Landmark Technology Partners, Inc. Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US8745267B2 (en) 2012-08-19 2014-06-03 Box, Inc. Enhancement of upload and/or download performance based on client and/or server feedback information
US20140188957A1 (en) * 2012-11-30 2014-07-03 Hitachi, Ltd. Hierarchical storage system and file management method
US20140229932A1 (en) * 2010-04-27 2014-08-14 Apple Inc. Dynamic retrieval of installation packages when installing software
US20140279936A1 (en) * 2011-09-02 2014-09-18 Compuverde Ab Method for data retrieval from a distributed data storage system
US8868574B2 (en) 2012-07-30 2014-10-21 Box, Inc. System and method for advanced search and filtering mechanisms for enterprise administrators in a cloud-based environment
US20140324773A1 (en) * 2012-03-09 2014-10-30 Huawei Technologies Co., Ltd. Data synchronization method and system and data access apparatus
US8892679B1 (en) 2013-09-13 2014-11-18 Box, Inc. Mobile device, methods and user interfaces thereof in a mobile device platform featuring multifunctional access and engagement in a collaborative environment provided by a cloud-based platform
US8914900B2 (en) 2012-05-23 2014-12-16 Box, Inc. Methods, architectures and security mechanisms for a third-party application to access content in a cloud-based platform
US20150019901A1 (en) * 2013-07-11 2015-01-15 International Business Machines Corporation Tolerating failures using concurrency in a cluster
US20150016261A1 (en) * 2013-07-12 2015-01-15 Seven Networks, Inc. Transport protocol layer optimization for managing signaling and power consumption
US9015601B2 (en) 2011-06-21 2015-04-21 Box, Inc. Batch uploading of content to a web-based collaboration environment
US9019123B2 (en) 2011-12-22 2015-04-28 Box, Inc. Health check services for web-based collaboration environments
US9027108B2 (en) 2012-05-23 2015-05-05 Box, Inc. Systems and methods for secure file portability between mobile applications on a mobile device
US9054919B2 (en) 2012-04-05 2015-06-09 Box, Inc. Device pinning capability for enterprise cloud service and storage accounts
US9063912B2 (en) 2011-06-22 2015-06-23 Box, Inc. Multimedia content preview rendering in a cloud content management system
US9098474B2 (en) 2011-10-26 2015-08-04 Box, Inc. Preview pre-generation based on heuristics and algorithmic prediction/assessment of predicted user behavior for enhancement of user experience
US20150237002A1 (en) * 2011-10-21 2015-08-20 Aruba Networks, Inc. Centralized Configuration with Dynamic Distributed Address Management
US9117087B2 (en) 2012-09-06 2015-08-25 Box, Inc. System and method for creating a secure channel for inter-application communication based on intents
US9135462B2 (en) 2012-08-29 2015-09-15 Box, Inc. Upload and download streaming encryption to/from a cloud-based platform
GB2498047B (en) * 2011-11-16 2015-09-30 Box Inc Managing updates at clients used by a user to access a cloud-based collaboration service
US9197718B2 (en) 2011-09-23 2015-11-24 Box, Inc. Central management and control of user-contributed content in a web-based collaboration environment and management console thereof
US9195519B2 (en) 2012-09-06 2015-11-24 Box, Inc. Disabling the self-referential appearance of a mobile application in an intent via a background registration
US9195636B2 (en) 2012-03-07 2015-11-24 Box, Inc. Universal file type preview for mobile devices
US20150339113A1 (en) * 2013-05-10 2015-11-26 Box, Inc. Identification and handling of items to be ignored for synchronization with a cloud-based platform by a synchronization client
US9213684B2 (en) 2013-09-13 2015-12-15 Box, Inc. System and method for rendering document in web browser or mobile device regardless of third-party plug-in software
US9237170B2 (en) 2012-07-19 2016-01-12 Box, Inc. Data loss prevention (DLP) methods and architectures by a cloud service
US9292833B2 (en) 2012-09-14 2016-03-22 Box, Inc. Batching notifications of activities that occur in a web-based collaboration environment
US9311071B2 (en) 2012-09-06 2016-04-12 Box, Inc. Force upgrade of a mobile application via a server side configuration file
US9396245B2 (en) 2013-01-02 2016-07-19 Box, Inc. Race condition handling in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US9413587B2 (en) 2012-05-02 2016-08-09 Box, Inc. System and method for a third-party application to access content within a cloud-based platform
US9483473B2 (en) 2013-09-13 2016-11-01 Box, Inc. High availability architecture for a cloud-based concurrent-access collaboration platform
US9495364B2 (en) 2012-10-04 2016-11-15 Box, Inc. Enhanced quick search features, low-barrier commenting/interactive features in a collaboration platform
US9503549B2 (en) 2011-04-20 2016-11-22 International Business Machines Corporation Real-time data analysis for resource provisioning among systems in a networked computing environment
US9507795B2 (en) 2013-01-11 2016-11-29 Box, Inc. Functionalities, features, and user interface of a synchronization client to a cloud-based environment
US9516127B2 (en) 2013-03-25 2016-12-06 Seven Networks, Llc Intelligent alarm manipulator and resource tracker
US9519886B2 (en) 2013-09-13 2016-12-13 Box, Inc. Simultaneous editing/accessing of content by collaborator invitation through a web-based or mobile application to a cloud-based collaboration platform
US9535909B2 (en) 2013-09-13 2017-01-03 Box, Inc. Configurable event-based automation architecture for cloud-based collaboration platforms
US9535924B2 (en) 2013-07-30 2017-01-03 Box, Inc. Scalability improvement in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US9553758B2 (en) 2012-09-18 2017-01-24 Box, Inc. Sandboxing individual applications to specific user folders in a cloud-based service
US9558202B2 (en) 2012-08-27 2017-01-31 Box, Inc. Server side techniques for reducing database workload in implementing selective subfolder synchronization in a cloud-based environment
US9575981B2 (en) 2012-04-11 2017-02-21 Box, Inc. Cloud service enabled to handle a set of files depicted to a user as a single file in a native operating system
US9602514B2 (en) 2014-06-16 2017-03-21 Box, Inc. Enterprise mobility management and verification of a managed application by a content provider
US9603056B2 (en) 2010-07-26 2017-03-21 Seven Networks, Llc Mobile application traffic optimization
WO2017062972A1 (en) * 2015-10-08 2017-04-13 Wake Forest University Methods, systems and computer readable media for providing resilient computing services using systems diversity
US9628268B2 (en) 2012-10-17 2017-04-18 Box, Inc. Remote key management in a cloud-based environment
US9626378B2 (en) 2011-09-02 2017-04-18 Compuverde Ab Method for handling requests in a storage system and a storage node for a storage system
US9633037B2 (en) 2013-06-13 2017-04-25 Box, Inc Systems and methods for synchronization event building and/or collapsing by a synchronization component of a cloud-based platform
US20170134501A1 (en) * 2015-11-09 2017-05-11 International Business Machines Corporation Session management
US9652741B2 (en) 2011-07-08 2017-05-16 Box, Inc. Desktop application for access and interaction with workspaces in a cloud-based content management system and synchronization mechanisms thereof
CN106687929A (en) * 2014-09-04 2017-05-17 精工爱普生株式会社 Data processing system, data processing method, and terminal device
US9665349B2 (en) 2012-10-05 2017-05-30 Box, Inc. System and method for generating embeddable widgets which enable access to a cloud-based collaboration platform
US9691051B2 (en) 2012-05-21 2017-06-27 Box, Inc. Security enhancement through application access control
US9705967B2 (en) 2012-10-04 2017-07-11 Box, Inc. Corporate user discovery and identification of recommended collaborators in a cloud platform
US9712510B2 (en) 2012-07-06 2017-07-18 Box, Inc. Systems and methods for securely submitting comments among users via external messaging applications in a cloud-based platform
US9756022B2 (en) 2014-08-29 2017-09-05 Box, Inc. Enhanced remote key management for an enterprise in a cloud-based environment
US9773051B2 (en) 2011-11-29 2017-09-26 Box, Inc. Mobile platform file and folder selection functionalities for offline access and synchronization
US9794256B2 (en) 2012-07-30 2017-10-17 Box, Inc. System and method for advanced control tools for administrators in a cloud-based service
US9792320B2 (en) 2012-07-06 2017-10-17 Box, Inc. System and method for performing shard migration to support functions of a cloud-based service
US9805050B2 (en) 2013-06-21 2017-10-31 Box, Inc. Maintaining and updating file system shadows on a local device by a synchronization client of a cloud-based platform
US9894119B2 (en) 2014-08-29 2018-02-13 Box, Inc. Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms
US9904435B2 (en) 2012-01-06 2018-02-27 Box, Inc. System and method for actionable event generation for task delegation and management via a discussion forum in a web-based collaboration environment
US9948716B2 (en) 2010-04-23 2018-04-17 Compuverde Ab Distributed data storage
US9953036B2 (en) 2013-01-09 2018-04-24 Box, Inc. File system monitoring in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US9959420B2 (en) 2012-10-02 2018-05-01 Box, Inc. System and method for enhanced security and management mechanisms for enterprise administrators in a cloud-based environment
US9965542B2 (en) 2011-09-02 2018-05-08 Compuverde Ab Method for data maintenance
US9965745B2 (en) 2012-02-24 2018-05-08 Box, Inc. System and method for promoting enterprise adoption of a web-based collaboration environment
US9978040B2 (en) 2011-07-08 2018-05-22 Box, Inc. Collaboration sessions in a workspace on a cloud-based content management system
US10009241B1 (en) 2006-11-15 2018-06-26 Conviva Inc. Monitoring the performance of a content player
US10009242B1 (en) 2009-07-20 2018-06-26 Conviva Inc. Augmenting the functionality of a content player
US10038731B2 (en) 2014-08-29 2018-07-31 Box, Inc. Managing flow-based interactions with cloud-based shared content
US10084845B2 (en) * 2015-09-14 2018-09-25 Uber Technologies, Inc. Data restoration for datacenter failover
US10110656B2 (en) 2013-06-25 2018-10-23 Box, Inc. Systems and methods for providing shell communication in a cloud-based platform
CN108776629A (en) * 2018-05-30 2018-11-09 郑州云海信息技术有限公司 A kind of server chips group carry Micro SD cards switching system and method
US10140109B2 (en) * 2014-02-25 2018-11-27 Ford Global Technologies, Llc Silent in-vehicle software updates
US10148716B1 (en) 2012-04-09 2018-12-04 Conviva Inc. Dynamic generation of video manifest files
US10154074B1 (en) 2006-11-15 2018-12-11 Conviva Inc. Remediation of the impact of detected synchronized data requests in a content delivery network
US10178043B1 (en) 2014-12-08 2019-01-08 Conviva Inc. Dynamic bitrate range selection in the cloud for optimized video streaming
US20190012245A1 (en) * 2014-03-21 2019-01-10 Huawei Technologies Co., Ltd. Failover Method, Apparatus and System
US10182096B1 (en) 2012-09-05 2019-01-15 Conviva Inc. Virtual resource locator
US10200256B2 (en) 2012-09-17 2019-02-05 Box, Inc. System and method of a manipulative handle in an interactive mobile user interface
US10216549B2 (en) 2013-06-17 2019-02-26 Seven Networks, Llc Methods and systems for providing application programming interfaces and application programming interface extensions to third party applications for optimizing and minimizing application traffic
US10229134B2 (en) 2013-06-25 2019-03-12 Box, Inc. Systems and methods for managing upgrades, migration of user data and improving performance of a cloud-based platform
US10235383B2 (en) 2012-12-19 2019-03-19 Box, Inc. Method and apparatus for synchronization of items with read-only permissions in a cloud-based environment
US10305955B1 (en) * 2014-12-08 2019-05-28 Conviva Inc. Streaming decision in the cloud
US10303678B2 (en) 2016-06-29 2019-05-28 International Business Machines Corporation Application resiliency management using a database driver
US10313734B1 (en) 2009-03-23 2019-06-04 Conviva Inc. Switching content
US10452667B2 (en) 2012-07-06 2019-10-22 Box Inc. Identification of people as search results from key-word based searches of content in a cloud-based environment
US10509527B2 (en) 2013-09-13 2019-12-17 Box, Inc. Systems and methods for configuring event-based automation in cloud-based collaboration platforms
US10530854B2 (en) 2014-05-30 2020-01-07 Box, Inc. Synchronization of permissioned content in cloud-based environments
US10554426B2 (en) 2011-01-20 2020-02-04 Box, Inc. Real time notification of activities that occur in a web-based collaboration environment
US10574442B2 (en) 2014-08-29 2020-02-25 Box, Inc. Enhanced remote key management for an enterprise in a cloud-based environment
US20200065176A1 (en) * 2018-08-21 2020-02-27 Red Hat, Inc. Optimizing logging of decision outcomes in distributed transactions
US10585766B2 (en) 2011-06-06 2020-03-10 Microsoft Technology Licensing, Llc Automatic configuration of a recovery service
US10599671B2 (en) 2013-01-17 2020-03-24 Box, Inc. Conflict resolution, retry condition management, and handling of problem files for the synchronization client to a cloud-based platform
US10650022B2 (en) 2008-10-24 2020-05-12 Compuverde Ab Distributed data storage
US10725968B2 (en) 2013-05-10 2020-07-28 Box, Inc. Top down delete or unsynchronization on delete of and depiction of item synchronization with a synchronization client to a cloud-based platform
US10862994B1 (en) 2006-11-15 2020-12-08 Conviva Inc. Facilitating client decisions
US10866931B2 (en) 2013-10-22 2020-12-15 Box, Inc. Desktop application for accessing a cloud collaboration platform
US10873615B1 (en) 2012-09-05 2020-12-22 Conviva Inc. Source assignment based on network partitioning
US10915492B2 (en) 2012-09-19 2021-02-09 Box, Inc. Cloud-based platform enabled with media content indexed for text-based searches and/or metadata extraction
US10990609B2 (en) * 2010-12-28 2021-04-27 Amazon Technologies, Inc. Data replication framework
US11210610B2 (en) 2011-10-26 2021-12-28 Box, Inc. Enhanced multimedia content preview rendering in a cloud content management system
US11232481B2 (en) 2012-01-30 2022-01-25 Box, Inc. Extended applications of multimedia content previews in the cloud-based content management system
CN114189424A (en) * 2021-12-14 2022-03-15 北京有竹居网络技术有限公司 Intelligent device control method, device, electronic device, medium, and program product
CN114338727A (en) * 2022-01-05 2022-04-12 中国建设银行股份有限公司 Service cluster construction method and related device
US11379464B2 (en) * 2019-12-12 2022-07-05 Micro Focus Llc Asymmetric quorum protocol based distributed transaction database consistency control
US11593397B2 (en) * 2020-10-05 2023-02-28 Micro Focus Llc Low latency polling
US11675678B1 (en) * 2022-03-28 2023-06-13 Dell Products L.P. Managing storage domains, service tiers, and failed service tiers

Families Citing this family (6)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US9652336B2 (en) 2015-03-13 2017-05-16 International Business Machines Corporation Resilient programming frameworks for handling failures in parallel programs
US10042722B1 (en) * 2015-06-23 2018-08-07 Juniper Networks, Inc. Service-chain fault tolerance in service virtualized environments
US10277456B2 (en) 2016-08-26 2019-04-30 International Business Machines Corporation Network-enabled devices
US11403319B2 (en) * 2018-06-01 2022-08-02 Hewlett Packard Enterprise Development Lp High-availability network device database synchronization
EP3714372A4 (en) * 2018-07-06 2021-02-24 Snowflake Inc. Data replication and data failover in database systems
US10708122B1 (en) * 2019-10-30 2020-07-07 T-Mobile Usa, Inc. Network fault detection and quality of service improvement systems and methods

Citations (20)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4958270A (en) * 1989-01-23 1990-09-18 Honeywell Inc. Method for control data base updating of a redundant processor in a process control system
US5255388A (en) * 1990-09-26 1993-10-19 Honeywell Inc. Synchronizing slave processors through eavesdrop by one on a write request message directed to another followed by comparison of individual status request replies
US5261092A (en) * 1990-09-26 1993-11-09 Honeywell Inc. Synchronizing slave processors through eavesdrop by one on periodic sync-verify messages directed to another followed by comparison of individual status
US5426774A (en) * 1993-04-06 1995-06-20 Honeywell Inc. Method for maintaining a sequence of events function during failover in a redundant multiple layer system
US6134673A (en) * 1997-05-13 2000-10-17 Micron Electronics, Inc. Method for clustering software applications
US6292905B1 (en) * 1997-05-13 2001-09-18 Micron Technology, Inc. Method for providing a fault tolerant network using distributed server processes to remap clustered network resources to other servers during server failure
US20010056554A1 (en) * 1997-05-13 2001-12-27 Michael Chrabaszcz System for clustering software applications
US20040158766A1 (en) * 2002-09-09 2004-08-12 John Liccione System and method for application monitoring and automatic disaster recovery for high-availability
US20050229024A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US6981174B1 (en) * 2002-07-18 2005-12-27 Extreme Networks, Inc. Method and apparatus for a redundant port
US20060005231A1 (en) * 2002-02-08 2006-01-05 Nir Zuk Intelligent integrated network security device for high-availability applications
US20060074925A1 (en) * 2004-09-21 2006-04-06 Bixby Peter C Pathname caching and protection of the root directory in a nested multilayer directory structure
US7251713B1 (en) * 2002-03-18 2007-07-31 Xiotech Corporation System and method to transport data snapshots
US20070244937A1 (en) * 2006-04-12 2007-10-18 Flynn John T Jr System and method for application fault tolerance and recovery using topologically remotely located computing devices
US7437445B1 (en) * 2004-03-30 2008-10-14 Emc Corporation System and methods for host naming in a managed information environment
US7447855B2 (en) * 2003-09-09 2008-11-04 Hitachi, Ltd. Data processing method providing remote copy in a system having first, second, and third storage systems
US20100299451A1 (en) * 2007-12-01 2010-11-25 Cai Yigang Ims diameter router with load balancing
US20110010560A1 (en) * 2009-07-09 2011-01-13 Craig Stephen Etchegoyen Failover Procedure for Server System
US7925630B1 (en) * 2007-03-30 2011-04-12 Symantec Corporation Method of inserting a validated time-image on the primary CDP subsystem in a continuous data protection and replication (CDP/R) subsystem
US20110106906A1 (en) * 2009-11-04 2011-05-05 Simon Assouad Method and system for offline data access on computer systems

Family Cites Families (3)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US20070234123A1 (en) * 2006-03-31 2007-10-04 Inventec Corporation Method for detecting switching failure
JP2008250695A (en) * 2007-03-30 2008-10-16 Nec Corp Disk array controller and disk array system therewith
US20110047413A1 (en) * 2009-08-20 2011-02-24 Mcgill Robert E Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer

Patent Citations (29)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US4958270A (en) * 1989-01-23 1990-09-18 Honeywell Inc. Method for control data base updating of a redundant processor in a process control system
US5255388A (en) * 1990-09-26 1993-10-19 Honeywell Inc. Synchronizing slave processors through eavesdrop by one on a write request message directed to another followed by comparison of individual status request replies
US5261092A (en) * 1990-09-26 1993-11-09 Honeywell Inc. Synchronizing slave processors through eavesdrop by one on periodic sync-verify messages directed to another followed by comparison of individual status
US5426774A (en) * 1993-04-06 1995-06-20 Honeywell Inc. Method for maintaining a sequence of events function during failover in a redundant multiple layer system
US20050229027A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US6134673A (en) * 1997-05-13 2000-10-17 Micron Electronics, Inc. Method for clustering software applications
US20010056554A1 (en) * 1997-05-13 2001-12-27 Michael Chrabaszcz System for clustering software applications
US6363497B1 (en) * 1997-05-13 2002-03-26 Micron Technology, Inc. System for clustering software applications
US6701453B2 (en) * 1997-05-13 2004-03-02 Micron Technology, Inc. System for clustering software applications
US6292905B1 (en) * 1997-05-13 2001-09-18 Micron Technology, Inc. Method for providing a fault tolerant network using distributed server processes to remap clustered network resources to other servers during server failure
US20050229024A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US20050229026A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US20050229025A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US20050229028A1 (en) * 1997-05-13 2005-10-13 Bruce Findlay System and method for communicating a software-generated pulse waveform between two servers in a network
US20100242093A1 (en) * 2002-02-08 2010-09-23 Juniper Networks, Inc. Intelligent integrated network security device for high-availability applications
US20060005231A1 (en) * 2002-02-08 2006-01-05 Nir Zuk Intelligent integrated network security device for high-availability applications
US7251713B1 (en) * 2002-03-18 2007-07-31 Xiotech Corporation System and method to transport data snapshots
US6981174B1 (en) * 2002-07-18 2005-12-27 Extreme Networks, Inc. Method and apparatus for a redundant port
US20040158766A1 (en) * 2002-09-09 2004-08-12 John Liccione System and method for application monitoring and automatic disaster recovery for high-availability
US20070255977A1 (en) * 2002-09-09 2007-11-01 Messageone, Inc. System and Method for Application Monitoring and Automatic Disaster Recovery for High-Availability
US7447855B2 (en) * 2003-09-09 2008-11-04 Hitachi, Ltd. Data processing method providing remote copy in a system having first, second, and third storage systems
US7437445B1 (en) * 2004-03-30 2008-10-14 Emc Corporation System and methods for host naming in a managed information environment
US20060074925A1 (en) * 2004-09-21 2006-04-06 Bixby Peter C Pathname caching and protection of the root directory in a nested multilayer directory structure
US8166005B2 (en) * 2004-09-21 2012-04-24 Emc Corporation Pathname caching and protection of the root directory in a nested multilayer directory structure
US20070244937A1 (en) * 2006-04-12 2007-10-18 Flynn John T Jr System and method for application fault tolerance and recovery using topologically remotely located computing devices
US7925630B1 (en) * 2007-03-30 2011-04-12 Symantec Corporation Method of inserting a validated time-image on the primary CDP subsystem in a continuous data protection and replication (CDP/R) subsystem
US20100299451A1 (en) * 2007-12-01 2010-11-25 Cai Yigang Ims diameter router with load balancing
US20110010560A1 (en) * 2009-07-09 2011-01-13 Craig Stephen Etchegoyen Failover Procedure for Server System
US20110106906A1 (en) * 2009-11-04 2011-05-05 Simon Assouad Method and system for offline data access on computer systems

Cited By (218)

* Cited by examiner, † Cited by third party
Publication number Priority date Publication date Assignee Title
US10154074B1 (en) 2006-11-15 2018-12-11 Conviva Inc. Remediation of the impact of detected synchronized data requests in a content delivery network
US10009241B1 (en) 2006-11-15 2018-06-26 Conviva Inc. Monitoring the performance of a content player
US10862994B1 (en) 2006-11-15 2020-12-08 Conviva Inc. Facilitating client decisions
US9519526B2 (en) 2007-12-05 2016-12-13 Box, Inc. File management system and collaboration service and integration capabilities with third party applications
US8583619B2 (en) 2007-12-05 2013-11-12 Box, Inc. Methods and systems for open source collaboration in an application service provider environment
US11468088B2 (en) 2008-10-24 2022-10-11 Pure Storage, Inc. Selection of storage nodes for storage of data
US11907256B2 (en) 2008-10-24 2024-02-20 Pure Storage, Inc. Query-based selection of storage nodes
US10650022B2 (en) 2008-10-24 2020-05-12 Compuverde Ab Distributed data storage
US10313734B1 (en) 2009-03-23 2019-06-04 Conviva Inc. Switching content
US10313035B1 (en) 2009-03-23 2019-06-04 Conviva Inc. Switching content
US10009242B1 (en) 2009-07-20 2018-06-26 Conviva Inc. Augmenting the functionality of a content player
US8949657B2 (en) * 2009-08-20 2015-02-03 Landmark Technology Partners, Inc. Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US20140129873A1 (en) * 2009-08-20 2014-05-08 Landmark Technology Partners, Inc. Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US20110161281A1 (en) * 2009-12-30 2011-06-30 Sybase, Inc. Distributed Transaction Management in a Distributed Shared Disk Cluster Environment
US9268664B2 (en) * 2010-04-06 2016-02-23 Paypal, Inc. Method and system for synchronous and asynchronous monitoring
US10050852B2 (en) 2010-04-06 2018-08-14 Paypal, Inc. Method and system for synchronous and asynchronous monitoring
US10785131B2 (en) 2010-04-06 2020-09-22 Paypal, Inc. Method and system for synchronous and asynchronous monitoring
US20110246640A1 (en) * 2010-04-06 2011-10-06 Debashis Saha Method and system for synchronous and asynchronous monitoring
US9948716B2 (en) 2010-04-23 2018-04-17 Compuverde Ab Distributed data storage
US10379831B2 (en) 2010-04-27 2019-08-13 Apple Inc. Dynamic retrieval of installation packages when installing software
US9465600B2 (en) * 2010-04-27 2016-10-11 Apple Inc. Dynamic retrieval of installation packages when installing software
US20140229932A1 (en) * 2010-04-27 2014-08-14 Apple Inc. Dynamic retrieval of installation packages when installing software
US9603056B2 (en) 2010-07-26 2017-03-21 Seven Networks, Llc Mobile application traffic optimization
US9838905B2 (en) 2010-07-26 2017-12-05 Seven Networks, Llc Mobile application traffic optimization
US9495392B2 (en) 2010-09-15 2016-11-15 Oracle International Corporation System and method for parallel multiplexing between servers in a cluster
US20120066394A1 (en) * 2010-09-15 2012-03-15 Oracle International Corporation System and method for supporting lazy deserialization of session information in a server cluster
US9864759B2 (en) 2010-09-15 2018-01-09 Oracle International Corporation System and method for providing scatter/gather data processing in a middleware environment
US9811541B2 (en) * 2010-09-15 2017-11-07 Oracle International Corporation System and method for supporting lazy deserialization of session information in a server cluster
US10990609B2 (en) * 2010-12-28 2021-04-27 Amazon Technologies, Inc. Data replication framework
US10554426B2 (en) 2011-01-20 2020-02-04 Box, Inc. Real time notification of activities that occur in a web-based collaboration environment
US9503549B2 (en) 2011-04-20 2016-11-22 International Business Machines Corporation Real-time data analysis for resource provisioning among systems in a networked computing environment
US9311199B2 (en) 2011-06-01 2016-04-12 Microsoft Technology Licensing, Llc Replaying jobs at a secondary location of a service
US8850261B2 (en) 2011-06-01 2014-09-30 Microsoft Corporation Replaying jobs at a secondary location of a service
KR101970839B1 (en) 2011-06-01 2019-04-19 마이크로소프트 테크놀로지 라이센싱, 엘엘씨 Replaying jobs at a secondary location of a service
CN103562904A (en) * 2011-06-01 2014-02-05 微软公司 Replaying jobs at a secondary location of a service
US9442813B2 (en) 2011-06-01 2016-09-13 Microsoft Technology Licensing, Llc Replaying jobs at a secondary location of a service
KR20140025503A (en) * 2011-06-01 2014-03-04 마이크로소프트 코포레이션 Replaying jobs at a secondary location of a service
WO2012166689A3 (en) * 2011-06-01 2013-01-31 Microsoft Corporation Replaying jobs at a secondary location of a service
US10585766B2 (en) 2011-06-06 2020-03-10 Microsoft Technology Licensing, Llc Automatic configuration of a recovery service
US8446920B2 (en) * 2011-06-14 2013-05-21 Mitel Networks Corporation Providing resilient digital telephony services for wireless device
US9015601B2 (en) 2011-06-21 2015-04-21 Box, Inc. Batch uploading of content to a web-based collaboration environment
US9063912B2 (en) 2011-06-22 2015-06-23 Box, Inc. Multimedia content preview rendering in a cloud content management system
US8719232B2 (en) * 2011-06-30 2014-05-06 Verisign, Inc. Systems and methods for data integrity checking
US20130006949A1 (en) * 2011-06-30 2013-01-03 Tarik Essawi Systems and methods for data integrity checking
US8495091B2 (en) * 2011-07-01 2013-07-23 Intel Corporation Dynamically routing data responses directly to requesting processor core
US9978040B2 (en) 2011-07-08 2018-05-22 Box, Inc. Collaboration sessions in a workspace on a cloud-based content management system
US9652741B2 (en) 2011-07-08 2017-05-16 Box, Inc. Desktop application for access and interaction with workspaces in a cloud-based content management system and synchronization mechanisms thereof
US10579615B2 (en) * 2011-09-02 2020-03-03 Compuverde Ab Method for data retrieval from a distributed data storage system
US10909110B1 (en) 2011-09-02 2021-02-02 Pure Storage, Inc. Data retrieval from a distributed data storage system
US11372897B1 (en) 2011-09-02 2022-06-28 Pure Storage, Inc. Writing of data to a storage system that implements a virtual file structure on an unstructured storage layer
US9626378B2 (en) 2011-09-02 2017-04-18 Compuverde Ab Method for handling requests in a storage system and a storage node for a storage system
US10769177B1 (en) 2011-09-02 2020-09-08 Pure Storage, Inc. Virtual file structure for data storage system
US10430443B2 (en) 2011-09-02 2019-10-01 Compuverde Ab Method for data maintenance
US9965542B2 (en) 2011-09-02 2018-05-08 Compuverde Ab Method for data maintenance
US20140279936A1 (en) * 2011-09-02 2014-09-18 Compuverde Ab Method for data retrieval from a distributed data storage system
US20130060932A1 (en) * 2011-09-06 2013-03-07 Shachar Ofek Discovering tiers within an application
US9197718B2 (en) 2011-09-23 2015-11-24 Box, Inc. Central management and control of user-contributed content in a web-based collaboration environment and management console thereof
CN103858105A (en) * 2011-10-05 2014-06-11 美光科技公司 Connection method
US9798601B2 (en) * 2011-10-05 2017-10-24 Micron Technology, Inc. Connection method
US20130091271A1 (en) * 2011-10-05 2013-04-11 Marek Piekarski Connection method
US8990151B2 (en) 2011-10-14 2015-03-24 Box, Inc. Automatic and semi-automatic tagging features of work items in a shared workspace for metadata tracking in a cloud-based content management system with selective or optional user contribution
US8515902B2 (en) 2011-10-14 2013-08-20 Box, Inc. Automatic and semi-automatic tagging features of work items in a shared workspace for metadata tracking in a cloud-based content management system with selective or optional user contribution
US20150237002A1 (en) * 2011-10-21 2015-08-20 Aruba Networks, Inc. Centralized Configuration with Dynamic Distributed Address Management
US9787632B2 (en) * 2011-10-21 2017-10-10 Aruba Networks, Inc. Centralized configuration with dynamic distributed address management
US9098474B2 (en) 2011-10-26 2015-08-04 Box, Inc. Preview pre-generation based on heuristics and algorithmic prediction/assessment of predicted user behavior for enhancement of user experience
US11210610B2 (en) 2011-10-26 2021-12-28 Box, Inc. Enhanced multimedia content preview rendering in a cloud content management system
GB2498047B (en) * 2011-11-16 2015-09-30 Box Inc Managing updates at clients used by a user to access a cloud-based collaboration service
US9015248B2 (en) 2011-11-16 2015-04-21 Box, Inc. Managing updates at clients used by a user to access a cloud-based collaboration service
US8990307B2 (en) 2011-11-16 2015-03-24 Box, Inc. Resource effective incremental updating of a remote client with events which occurred via a cloud-enabled platform
WO2013074985A1 (en) * 2011-11-16 2013-05-23 Box, Inc. Resource effective incremental updating of a remote client with events which occurred via a cloud-enabled platform
US11853320B2 (en) 2011-11-29 2023-12-26 Box, Inc. Mobile platform file and folder selection functionalities for offline access and synchronization
US10909141B2 (en) 2011-11-29 2021-02-02 Box, Inc. Mobile platform file and folder selection functionalities for offline access and synchronization
US9773051B2 (en) 2011-11-29 2017-09-26 Box, Inc. Mobile platform file and folder selection functionalities for offline access and synchronization
US11537630B2 (en) 2011-11-29 2022-12-27 Box, Inc. Mobile platform file and folder selection functionalities for offline access and synchronization
US9019123B2 (en) 2011-12-22 2015-04-28 Box, Inc. Health check services for web-based collaboration environments
US10037250B2 (en) 2012-01-04 2018-07-31 International Business Machines Corporation Managing remote data replication
US9372910B2 (en) * 2012-01-04 2016-06-21 International Business Machines Corporation Managing remote data replication
US20130173537A1 (en) * 2012-01-04 2013-07-04 International Business Machines Corporation Managing remote data replication
US11243848B2 (en) 2012-01-04 2022-02-08 International Business Machines Corporation Managing remote data replication
US10394665B2 (en) 2012-01-04 2019-08-27 International Business Machines Corporation Managing remote data replication
US10394666B2 (en) 2012-01-04 2019-08-27 International Business Machines Corporation Managing remote data replication
US10037249B2 (en) 2012-01-04 2018-07-31 International Business Machines Corporation Managing remote data replication
US20130173538A1 (en) * 2012-01-04 2013-07-04 International Business Machines Corporation Managing remote data replication
US11243847B2 (en) 2012-01-04 2022-02-08 International Business Machines Corporation Managing remote data replication
US9384255B2 (en) * 2012-01-04 2016-07-05 International Business Machines Corporation Managing remote data replication
US9904435B2 (en) 2012-01-06 2018-02-27 Box, Inc. System and method for actionable event generation for task delegation and management via a discussion forum in a web-based collaboration environment
US11232481B2 (en) 2012-01-30 2022-01-25 Box, Inc. Extended applications of multimedia content previews in the cloud-based content management system
US9965745B2 (en) 2012-02-24 2018-05-08 Box, Inc. System and method for promoting enterprise adoption of a web-based collaboration environment
US10713624B2 (en) 2012-02-24 2020-07-14 Box, Inc. System and method for promoting enterprise adoption of a web-based collaboration environment
US9195636B2 (en) 2012-03-07 2015-11-24 Box, Inc. Universal file type preview for mobile devices
US20140324773A1 (en) * 2012-03-09 2014-10-30 Huawei Technologies Co., Ltd. Data synchronization method and system and data access apparatus
US9054919B2 (en) 2012-04-05 2015-06-09 Box, Inc. Device pinning capability for enterprise cloud service and storage accounts
US10148716B1 (en) 2012-04-09 2018-12-04 Conviva Inc. Dynamic generation of video manifest files
US9575981B2 (en) 2012-04-11 2017-02-21 Box, Inc. Cloud service enabled to handle a set of files depicted to a user as a single file in a native operating system
US9413587B2 (en) 2012-05-02 2016-08-09 Box, Inc. System and method for a third-party application to access content within a cloud-based platform
US9691051B2 (en) 2012-05-21 2017-06-27 Box, Inc. Security enhancement through application access control
US8914900B2 (en) 2012-05-23 2014-12-16 Box, Inc. Methods, architectures and security mechanisms for a third-party application to access content in a cloud-based platform
US9552444B2 (en) 2012-05-23 2017-01-24 Box, Inc. Identification verification mechanisms for a third-party application to access content in a cloud-based platform
US9027108B2 (en) 2012-05-23 2015-05-05 Box, Inc. Systems and methods for secure file portability between mobile applications on a mobile device
US9280613B2 (en) 2012-05-23 2016-03-08 Box, Inc. Metadata enabled third-party application access of content at a cloud-based platform via a native client to the cloud-based platform
US20150082079A1 (en) * 2012-05-31 2015-03-19 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US20160110247A1 (en) * 2012-05-31 2016-04-21 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US20190034265A1 (en) * 2012-05-31 2019-01-31 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US20130326270A1 (en) * 2012-05-31 2013-12-05 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US9244787B2 (en) * 2012-05-31 2016-01-26 International Business Machines Corporation Maximizing use of storage in a data replication environment
US10896086B2 (en) * 2012-05-31 2021-01-19 International Business Machines Corporation Maximizing use of storage in a data replication environment
US10083074B2 (en) * 2012-05-31 2018-09-25 International Business Machines Corporation Maximizing use of storage in a data replication environment
US8918672B2 (en) * 2012-05-31 2014-12-23 International Business Machines Corporation Maximizing use of storage in a data replication environment
US20130326266A1 (en) * 2012-05-31 2013-12-05 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US8930744B2 (en) * 2012-05-31 2015-01-06 International Business Machines Corporation Maximizing use of storage in a data replication environment
US9244788B2 (en) * 2012-05-31 2016-01-26 International Business Machines Corporation Maximizing use of storage in a data replication environment
US20150089277A1 (en) * 2012-05-31 2015-03-26 International Business Machines Corporation Maximizing Use of Storage in a Data Replication Environment
US9021099B2 (en) 2012-07-03 2015-04-28 Box, Inc. Load balancing secure FTP connections among multiple FTP servers
US8719445B2 (en) 2012-07-03 2014-05-06 Box, Inc. System and method for load balancing multiple file transfer protocol (FTP) servers to service FTP connections for a cloud-based service
US9792320B2 (en) 2012-07-06 2017-10-17 Box, Inc. System and method for performing shard migration to support functions of a cloud-based service
US10452667B2 (en) 2012-07-06 2019-10-22 Box Inc. Identification of people as search results from key-word based searches of content in a cloud-based environment
US9712510B2 (en) 2012-07-06 2017-07-18 Box, Inc. Systems and methods for securely submitting comments among users via external messaging applications in a cloud-based platform
US9473532B2 (en) 2012-07-19 2016-10-18 Box, Inc. Data loss prevention (DLP) methods by a cloud service including third party integration architectures
US9237170B2 (en) 2012-07-19 2016-01-12 Box, Inc. Data loss prevention (DLP) methods and architectures by a cloud service
US9794256B2 (en) 2012-07-30 2017-10-17 Box, Inc. System and method for advanced control tools for administrators in a cloud-based service
US8868574B2 (en) 2012-07-30 2014-10-21 Box, Inc. System and method for advanced search and filtering mechanisms for enterprise administrators in a cloud-based environment
US8745267B2 (en) 2012-08-19 2014-06-03 Box, Inc. Enhancement of upload and/or download performance based on client and/or server feedback information
US20140059202A1 (en) * 2012-08-23 2014-02-27 Efrat Ben David Determining the type of a network tier
US9178771B2 (en) * 2012-08-23 2015-11-03 Hewlett-Packard Development Company, L.P. Determining the type of a network tier
US9558202B2 (en) 2012-08-27 2017-01-31 Box, Inc. Server side techniques for reducing database workload in implementing selective subfolder synchronization in a cloud-based environment
US9135462B2 (en) 2012-08-29 2015-09-15 Box, Inc. Upload and download streaming encryption to/from a cloud-based platform
US9450926B2 (en) 2012-08-29 2016-09-20 Box, Inc. Upload and download streaming encryption to/from a cloud-based platform
US9774527B2 (en) * 2012-08-31 2017-09-26 Nasdaq Technology Ab Resilient peer-to-peer application message routing
US20140064279A1 (en) * 2012-08-31 2014-03-06 Omx Technology Ab Resilient peer-to-peer application message routing
US10027585B2 (en) 2012-08-31 2018-07-17 Nasdaq Technology Ab Resilient peer-to-peer application message routing
US10873615B1 (en) 2012-09-05 2020-12-22 Conviva Inc. Source assignment based on network partitioning
US10848540B1 (en) 2012-09-05 2020-11-24 Conviva Inc. Virtual resource locator
US10182096B1 (en) 2012-09-05 2019-01-15 Conviva Inc. Virtual resource locator
US9195519B2 (en) 2012-09-06 2015-11-24 Box, Inc. Disabling the self-referential appearance of a mobile application in an intent via a background registration
US9311071B2 (en) 2012-09-06 2016-04-12 Box, Inc. Force upgrade of a mobile application via a server side configuration file
US9117087B2 (en) 2012-09-06 2015-08-25 Box, Inc. System and method for creating a secure channel for inter-application communication based on intents
US9292833B2 (en) 2012-09-14 2016-03-22 Box, Inc. Batching notifications of activities that occur in a web-based collaboration environment
US10200256B2 (en) 2012-09-17 2019-02-05 Box, Inc. System and method of a manipulative handle in an interactive mobile user interface
US9553758B2 (en) 2012-09-18 2017-01-24 Box, Inc. Sandboxing individual applications to specific user folders in a cloud-based service
US10915492B2 (en) 2012-09-19 2021-02-09 Box, Inc. Cloud-based platform enabled with media content indexed for text-based searches and/or metadata extraction
US9959420B2 (en) 2012-10-02 2018-05-01 Box, Inc. System and method for enhanced security and management mechanisms for enterprise administrators in a cloud-based environment
US9495364B2 (en) 2012-10-04 2016-11-15 Box, Inc. Enhanced quick search features, low-barrier commenting/interactive features in a collaboration platform
US9705967B2 (en) 2012-10-04 2017-07-11 Box, Inc. Corporate user discovery and identification of recommended collaborators in a cloud platform
US9665349B2 (en) 2012-10-05 2017-05-30 Box, Inc. System and method for generating embeddable widgets which enable access to a cloud-based collaboration platform
US9628268B2 (en) 2012-10-17 2017-04-18 Box, Inc. Remote key management in a cloud-based environment
US20140188957A1 (en) * 2012-11-30 2014-07-03 Hitachi, Ltd. Hierarchical storage system and file management method
US10235383B2 (en) 2012-12-19 2019-03-19 Box, Inc. Method and apparatus for synchronization of items with read-only permissions in a cloud-based environment
US9396245B2 (en) 2013-01-02 2016-07-19 Box, Inc. Race condition handling in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US9953036B2 (en) 2013-01-09 2018-04-24 Box, Inc. File system monitoring in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US9507795B2 (en) 2013-01-11 2016-11-29 Box, Inc. Functionalities, features, and user interface of a synchronization client to a cloud-based environment
US10599671B2 (en) 2013-01-17 2020-03-24 Box, Inc. Conflict resolution, retry condition management, and handling of problem files for the synchronization client to a cloud-based platform
US10178199B1 (en) 2013-03-25 2019-01-08 Seven Networks, Llc Intelligent alarm manipulator and resource tracker
US9516127B2 (en) 2013-03-25 2016-12-06 Seven Networks, Llc Intelligent alarm manipulator and resource tracker
US10725968B2 (en) 2013-05-10 2020-07-28 Box, Inc. Top down delete or unsynchronization on delete of and depiction of item synchronization with a synchronization client to a cloud-based platform
US20150339113A1 (en) * 2013-05-10 2015-11-26 Box, Inc. Identification and handling of items to be ignored for synchronization with a cloud-based platform by a synchronization client
US10846074B2 (en) * 2013-05-10 2020-11-24 Box, Inc. Identification and handling of items to be ignored for synchronization with a cloud-based platform by a synchronization client
US10877937B2 (en) 2013-06-13 2020-12-29 Box, Inc. Systems and methods for synchronization event building and/or collapsing by a synchronization component of a cloud-based platform
US9633037B2 (en) 2013-06-13 2017-04-25 Box, Inc Systems and methods for synchronization event building and/or collapsing by a synchronization component of a cloud-based platform
US10216549B2 (en) 2013-06-17 2019-02-26 Seven Networks, Llc Methods and systems for providing application programming interfaces and application programming interface extensions to third party applications for optimizing and minimizing application traffic
US9805050B2 (en) 2013-06-21 2017-10-31 Box, Inc. Maintaining and updating file system shadows on a local device by a synchronization client of a cloud-based platform
US11531648B2 (en) 2013-06-21 2022-12-20 Box, Inc. Maintaining and updating file system shadows on a local device by a synchronization client of a cloud-based platform
US10110656B2 (en) 2013-06-25 2018-10-23 Box, Inc. Systems and methods for providing shell communication in a cloud-based platform
US10229134B2 (en) 2013-06-25 2019-03-12 Box, Inc. Systems and methods for managing upgrades, migration of user data and improving performance of a cloud-based platform
US20150019900A1 (en) * 2013-07-11 2015-01-15 International Business Machines Corporation Tolerating failures using concurrency in a cluster
US9176833B2 (en) * 2013-07-11 2015-11-03 Globalfoundries U.S. 2 Llc Tolerating failures using concurrency in a cluster
US9176834B2 (en) * 2013-07-11 2015-11-03 Globalfoundries U.S. 2 Llc Tolerating failures using concurrency in a cluster
US20150019901A1 (en) * 2013-07-11 2015-01-15 International Business Machines Corporation Tolerating failures using concurrency in a cluster
US9973965B2 (en) * 2013-07-12 2018-05-15 Seven Networks, Llc Transport protocol layer optimization for managing signaling and power consumption
US20150016261A1 (en) * 2013-07-12 2015-01-15 Seven Networks, Inc. Transport protocol layer optimization for managing signaling and power consumption
US9535924B2 (en) 2013-07-30 2017-01-03 Box, Inc. Scalability improvement in a system which incrementally updates clients with events that occurred in a cloud-based collaboration platform
US10509527B2 (en) 2013-09-13 2019-12-17 Box, Inc. Systems and methods for configuring event-based automation in cloud-based collaboration platforms
US11435865B2 (en) 2013-09-13 2022-09-06 Box, Inc. System and methods for configuring event-based automation in cloud-based collaboration platforms
US9519886B2 (en) 2013-09-13 2016-12-13 Box, Inc. Simultaneous editing/accessing of content by collaborator invitation through a web-based or mobile application to a cloud-based collaboration platform
US8892679B1 (en) 2013-09-13 2014-11-18 Box, Inc. Mobile device, methods and user interfaces thereof in a mobile device platform featuring multifunctional access and engagement in a collaborative environment provided by a cloud-based platform
US9704137B2 (en) 2013-09-13 2017-07-11 Box, Inc. Simultaneous editing/accessing of content by collaborator invitation through a web-based or mobile application to a cloud-based collaboration platform
US9535909B2 (en) 2013-09-13 2017-01-03 Box, Inc. Configurable event-based automation architecture for cloud-based collaboration platforms
US9483473B2 (en) 2013-09-13 2016-11-01 Box, Inc. High availability architecture for a cloud-based concurrent-access collaboration platform
US10044773B2 (en) 2013-09-13 2018-08-07 Box, Inc. System and method of a multi-functional managing user interface for accessing a cloud-based platform via mobile devices
US9213684B2 (en) 2013-09-13 2015-12-15 Box, Inc. System and method for rendering document in web browser or mobile device regardless of third-party plug-in software
US11822759B2 (en) 2013-09-13 2023-11-21 Box, Inc. System and methods for configuring event-based automation in cloud-based collaboration platforms
US10866931B2 (en) 2013-10-22 2020-12-15 Box, Inc. Desktop application for accessing a cloud collaboration platform
US10140109B2 (en) * 2014-02-25 2018-11-27 Ford Global Technologies, Llc Silent in-vehicle software updates
US20190012245A1 (en) * 2014-03-21 2019-01-10 Huawei Technologies Co., Ltd. Failover Method, Apparatus and System
US10795785B2 (en) * 2014-03-21 2020-10-06 Huawei Technologies Co., Ltd. Failover method, apparatus and system
US10530854B2 (en) 2014-05-30 2020-01-07 Box, Inc. Synchronization of permissioned content in cloud-based environments
US9602514B2 (en) 2014-06-16 2017-03-21 Box, Inc. Enterprise mobility management and verification of a managed application by a content provider
US11876845B2 (en) 2014-08-29 2024-01-16 Box, Inc. Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms
US9894119B2 (en) 2014-08-29 2018-02-13 Box, Inc. Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms
US10038731B2 (en) 2014-08-29 2018-07-31 Box, Inc. Managing flow-based interactions with cloud-based shared content
US10574442B2 (en) 2014-08-29 2020-02-25 Box, Inc. Enhanced remote key management for an enterprise in a cloud-based environment
US10708323B2 (en) 2014-08-29 2020-07-07 Box, Inc. Managing flow-based interactions with cloud-based shared content
US10708321B2 (en) 2014-08-29 2020-07-07 Box, Inc. Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms
US11146600B2 (en) 2014-08-29 2021-10-12 Box, Inc. Configurable metadata-based automation and content classification architecture for cloud-based collaboration platforms
US9756022B2 (en) 2014-08-29 2017-09-05 Box, Inc. Enhanced remote key management for an enterprise in a cloud-based environment
US20170214769A1 (en) * 2014-09-04 2017-07-27 Seiko Epson Corporation Data processing system, data processing method, and terminal device
CN106687929A (en) * 2014-09-04 2017-05-17 精工爱普生株式会社 Data processing system, data processing method, and terminal device
US10757224B2 (en) * 2014-09-04 2020-08-25 Seiko Epson Corporation Data processing system, data processing method, and printer
US10305955B1 (en) * 2014-12-08 2019-05-28 Conviva Inc. Streaming decision in the cloud
US10848436B1 (en) 2014-12-08 2020-11-24 Conviva Inc. Dynamic bitrate range selection in the cloud for optimized video streaming
US10887363B1 (en) 2014-12-08 2021-01-05 Conviva Inc. Streaming decision in the cloud
US10178043B1 (en) 2014-12-08 2019-01-08 Conviva Inc. Dynamic bitrate range selection in the cloud for optimized video streaming
US10666705B2 (en) 2015-09-14 2020-05-26 Uber Technologies, Inc. Data restoration for datacenter failover
US10084845B2 (en) * 2015-09-14 2018-09-25 Uber Technologies, Inc. Data restoration for datacenter failover
US11277464B2 (en) 2015-09-14 2022-03-15 Uber Technologies, Inc. Data restoration for datacenter failover
WO2017062972A1 (en) * 2015-10-08 2017-04-13 Wake Forest University Methods, systems and computer readable media for providing resilient computing services using systems diversity
US11100231B2 (en) 2015-10-08 2021-08-24 Errin Wesley Fulp Methods, systems and computer readable media for providing resilient computing services using systems diversity
US20170134501A1 (en) * 2015-11-09 2017-05-11 International Business Machines Corporation Session management
US10284621B2 (en) * 2015-11-09 2019-05-07 International Business Machines Corporation Session management
US10303678B2 (en) 2016-06-29 2019-05-28 International Business Machines Corporation Application resiliency management using a database driver
CN108776629A (en) * 2018-05-30 2018-11-09 郑州云海信息技术有限公司 A kind of server chips group carry Micro SD cards switching system and method
US20200065176A1 (en) * 2018-08-21 2020-02-27 Red Hat, Inc. Optimizing logging of decision outcomes in distributed transactions
US11720429B2 (en) 2018-08-21 2023-08-08 Red Hat, Inc. Optimizing logging of decision outcomes in distributed transactions
US11163625B2 (en) * 2018-08-21 2021-11-02 Red Hat, Inc. Optimizing logging of decision outcomes in distributed transactions
US11379464B2 (en) * 2019-12-12 2022-07-05 Micro Focus Llc Asymmetric quorum protocol based distributed transaction database consistency control
US11593397B2 (en) * 2020-10-05 2023-02-28 Micro Focus Llc Low latency polling
CN114189424A (en) * 2021-12-14 2022-03-15 北京有竹居网络技术有限公司 Intelligent device control method, device, electronic device, medium, and program product
CN114338727A (en) * 2022-01-05 2022-04-12 中国建设银行股份有限公司 Service cluster construction method and related device
US11675678B1 (en) * 2022-03-28 2023-06-13 Dell Products L.P. Managing storage domains, service tiers, and failed service tiers

Also Published As

Publication number Publication date
US8949657B2 (en) 2015-02-03
US20140129873A1 (en) 2014-05-08

Similar Documents

Publication Publication Date Title
US8949657B2 (en) Methods and devices for detecting service failures and maintaining computing services using a resilient intelligent client computer
US11899684B2 (en) System and method for maintaining a master replica for reads and writes in a data store
US20200358848A1 (en) Methods, systems, and media for providing distributed database access during a network split
US8904231B2 (en) Synchronous local and cross-site failover in clustered storage systems
US7702947B2 (en) System and method for enabling site failover in an application server environment
EP3635558B1 (en) Service state preservation across nodes
JP6084624B2 (en) Split brain tolerant failover in high availability clusters
US20180324033A1 (en) System and method for data replication using a single master failover protocol
US9361348B1 (en) Database replication
US11893264B1 (en) Methods and systems to interface between a multi-site distributed storage system and an external mediator to efficiently process events related to continuity
US7653668B1 (en) Fault tolerant multi-stage data replication with relaxed coherency guarantees
US8719225B1 (en) System and method for log conflict detection and resolution in a data store
US9002939B2 (en) Adaptive and dynamic replication management in cloud computing
US11709743B2 (en) Methods and systems for a non-disruptive automatic unplanned failover from a primary copy of data at a primary storage system to a mirror copy of the data at a cross-site secondary storage system
US9319267B1 (en) Replication in assured messaging system
US11740811B2 (en) Reseeding a mediator of a cross-site storage solution
US11537314B1 (en) Resynchronization of individual volumes of a consistency group (CG) within a cross-site storage solution while maintaining synchronization of other volumes of the CG
US11892982B2 (en) Facilitating immediate performance of volume resynchronization with the use of passive cache entries
US20180074925A1 (en) Systems and methods for resynchronizing mirroring partners in a storage system
WO2015196692A1 (en) Cloud computing system and processing method and apparatus for cloud computing system
US11934670B2 (en) Performing various operations at the granularity of a consistency group within a cross-site storage solution
US20240036997A1 (en) Methods and systems to improve input/output (i/o) resumption time during a non-disruptive automatic unplanned failover from a primary copy of data at a primary storage system to a mirror copy of the data at a cross-site secondary storage system
US20220317897A1 (en) Performing various operations at the granularity of a consistency group within a cross-site storage solution
US20240036732A1 (en) Methods and systems to improve resumption time of input/output (i/o) operations based on prefetching of configuration data and early abort of conflicting workflows during a non-disruptive automatic unplanned failover from a primary copy of data at a primary storage system to a mirror copy of the data at a cross-site secondary storage system
Overby Enabling Continuous Availability with IBM Multi-site Workload Lifeline

Legal Events

Date Code Title Description
STCB Information on status: application discontinuation

Free format text: ABANDONED -- FAILURE TO RESPOND TO AN OFFICE ACTION