Patents
Literature
Hiro is an intelligent assistant for R&D personnel, combined with Patent DNA, to facilitate innovative research.
Hiro

154results about How to "Avoid contention" patented technology

Reentrant read-write lock algorithm

InactiveUS9471400B1Avoiding further contentionSignificant bottleneckResource allocationProgram synchronisationReal-time computingCurrent mode
Access to a shareable resource between threads is controlled by a lock having shared, optimistic and exclusive modes and maintaining a list of threads requesting ownership of said lock. A shared optimistic mode is provided. A lock state descriptor is provided for each desired change of mode comprising a current mode in which a thread has already acquired the lock. When a thread acquires the lock in shared optimistic mode, other threads are allowed to acquire the lock in shared or optimistic mode. When a thread which acquired the lock in shared optimistic mode wants to acquire the lock in exclusive mode, other threads which have acquired the lock in shared or optimistic mode are prevented from acquiring the lock in exclusive mode until the thread which acquired the lock in shared optimistic mode and requested to acquire the lock in exclusive mode releases the lock.
Owner:IBM CORP

Database system, server, query posing method, and data updating method

An object of the present invention is to avoid contention for resources in a parallel / decentralized database system so as to improve the performance of the system. A query posing method is implemented in a database system comprising a plurality of database servers each of which includes a database from which the same content can be retrieved and searches the database in response to a query request a front-end server that poses a query according to a predetermined rule and a management server that manages rules to be used by the front-end server. The management server acquires a processed query log relevant to a database server, and produces a rule according to the compatibility value of a query calculated using the acquired processed query log. Moreover, the front-end server poses the query according to the rule produced by the management server.
Owner:HITACHI LTD

Contention free pipelined broadcasting within a constant bisection bandwidth network topology

InactiveUS20110228789A1Avoiding contention latencyEasy to useData switching by path configurationBroadcastingComputer science
In an interconnection network, multiple nodes are connected to one of a first layer of switches. The first layer of switches is connected to one another through a second layer of switches, wherein each of the nodes is connected through one of multiple shared links connecting the first layer switches and the second layer of switches. A pipelined broadcast manager schedules a hierarchical pipelined broadcast through at least one switch of the first layer switches comprising a multiple non-root nodes by selecting two nodes of multiple non-root nodes connected to the at least one switch and scheduling each of multiple broadcast steps for the pipelined broadcast with at least one of an inter-switch broadcast phase and an intra-switch broadcast phase. The inter-switch broadcast phase is scheduled with a first node of the two nodes receiving a first data packet from another switch from among the first layer of switches and a second node of the two nodes sending a second data packet previously received from another switch to one other switch from among the first layer switches. The intra-switch broadcast phase is scheduled with the first node acting as a source for sending a previously received data packet to at least one other non-root node and the second node acting as a sink for receiving the previously received data packet, wherein each of the non-root nodes sends and receives the previously received data packet once throughout the broadcast steps and the sink receives the previously received data packet last. In scheduling the hierarchical pipelined broadcast, the first node and the second node alternate roles each broadcast step.
Owner:IBM CORP

Data migration with reduced contention and increased speed

Methods and apparatus are provided for managing data in a hierarchal storage subsystem. A plurality of volumes is designated as a storage group for Level 0 storage; a high threshold is established for the storage group; space is allocated for a data set to a volume of the storage group, storing the data set to the volume; the high threshold is compared with a total amount of space consumed by all data sets stored to volumes in the storage group; and data sets are migrated from the storage group to a Level 1 storage if the high threshold is less than or equal to the total amount of space used by all of the data sets stored to volumes in the storage group. Optionally, high threshold are assigned to each storage group and, when the space used in a storage group reaches or exceeds the high threshold, migration of data will begin from volumes in the storage group, beginning with the volume having the least free space. Thus, contention between migration and space allocation is reduced. Also optionally, when a volume is selected for migration, a flag is set which prevents space in the volume from being allocated to new data sets. Upon completion of the migration, the flag is cleared and allocation is allowed. Thus, contention between migration and space allocation is avoided.
Owner:IBM CORP

System and methods for task management

In a storage area network (SAN), different tasks may expect different types of commands for commencing execution, such as interactive or offline, and may have different formats for reporting status and completion, such as log files or message based. A framework for defining the business logic enveloped in a particular task, and providing a common manner of deploying, or enabling invocation, of the task provides consistent operator control for scheduling, monitoring, ensuring completion, and tracking errors and other events. Business logic modules are identified as commands corresponding to a task. Transactions including a set of the commands define an ordered sequence for completing the task. The operator requests a particular set of tasks, using a selection tree, and the task manager builds a corresponding execution tree to identify and map the transactions and commands of the task to the execution tree to optimize execution and mitigate redundancies.
Owner:EMC IP HLDG CO LLC
Who we serve
  • R&D Engineer
  • R&D Manager
  • IP Professional
Why Eureka
  • Industry Leading Data Capabilities
  • Powerful AI technology
  • Patent DNA Extraction
Social media
Try Eureka
PatSnap group products