In concurrent programming, a deadlock is a situation in which two or more
competing actions are each waiting for the other to finish, and thus neither
ever does.
In a transactional database, a deadlock
happens when two processes each within its own transaction updates two rows of
information but in the opposite order. For example, process A updates row 1
then row 2 in the exact timeframe that process B updates row 2 then row 1.
Process A can't finish updating row 2 until process B is finished, but process
B cannot finish updating row 1 until process A is finished. No matter how much
time is allowed to pass, this situation will never resolve itself and because
of this database management
systems will typically kill the
transaction of the process that has done the least amount of work.
In an operating system, a deadlock is a
situation which occurs when a process or thread enters a waiting state because a resource requested is being held by
another waiting process, which in turn is waiting for another resource held by
another waiting process. If a process is unable to change its state
indefinitely because the resources requested by it are being used by another
waiting process, then the system is said to be in a deadlock.
Deadlock
is a common problem in multiprocessing systems, parallel computing and distributed
systems, where software and hardware locks are
used to handle shared resources and implement process
synchronization.
In telecommunication systems, deadlocks
occur mainly due to lost or corrupt signals instead of resource contention.
Examples
A
simple computer-based example is as follows. Suppose a computer has three CD
drives and three processes. Each of the three processes holds one of the
drives. If each process now requests another drive, the three processes will be
in a deadlock. Each process will be waiting for the "CD drive
released" event, which can be only caused by one of the other waiting
processes. Thus, it results in a circular
chain.
Moving
onto the source code level, a deadlock can occur even in the case of a single
thread and one resource (protected by a mutex).
Assume there is a function func1() which does some work on the resource,
locking the mutex at the beginning and releasing it after it's done. Next,
somebody creates a different function func2() following that pattern on the same
resource (lock, do work, release) but decides to include a call to func1() to delegate a part of the job. What
will happen is the mutex will be locked once when enteringfunc2() and then again at the call to func1(), resulting in a
deadlock if the mutex is not reentrant (i.e. the plain "fast mutex"
variety).
Necessary Conditions
A dead
lockers situation can arise if all of the following conditions hold
simultaneously in a system:
1.
Mutual Exclusion: At least one resource must be held in a
non-shareable mode. Only one process can use the resource at any given instant
of time.
2.
Hold and Wait or Resource
Holding: A process is
currently holding at least one resource and requesting additional resources
which are being held by other processes.
3.
No Preemption: a resource can be released only voluntarily
by the process holding it.
4.
Circular Wait: A process must be waiting for a resource
which is being held by another process, which in turn is waiting for the first
process to release the resource. In general, there is a set of
waiting processes, P = {P1, P2, ...,
PN}, such that P1 is waiting for a resource held by P2, P2 is waiting for a resource held by P3 and so on until PN is
waiting for a resource held by P1
These
four conditions are known as the Coffman
conditions from their first
description in a 1971 article by Edward
G. Coffman, Jr. Unfulfillment of
any of these conditions is enough to preclude a deadlock from occurring.
Deadlock Handling
Most
current operating systems cannot prevent a deadlock from occurring. When a deadlock occurs, different
operating systems respond to them in different non-standard manners. Most
approaches work by preventing one of the four Coffman conditions from
occurring, especially the fourth one. Major approaches are as follows.
Ignoring
deadlock
In this
approach, it is assumed that a deadlock will never occur. This is also an
application of the Ostrich
algorithm. This approach was initially used by MINIX and UNIX.
This is used when the time intervals between occurrences of deadlocks are large
and the data loss incurred each time is tolerable.
Detection
Under
deadlock detection, deadlocks are allowed to occur. Then the state of the
system is examined to detect that a deadlock has occurred and subsequently it
is corrected. An algorithm is employed that tracks resource allocation and
process states, it rolls back and restarts one or more of the processes in
order to remove the detected deadlock. Detecting a deadlock that has already
occurred is easily possible since the resources that each process has locked
and/or currently requested are known to the resource scheduler of the operating
system.
Deadlock
detection techniques include, but are not limited to, model checking. This approach
constructs a finite state-model
on which it performs a progress analysis and finds all possible terminal sets
in the model. These then each represent a deadlock.
After a
deadlock is detected, it can be corrected by using one of the following
methods:
1.
Process Termination: One or more processes involved in the
deadlock may be aborted. We can choose to abort all processes involved in the
deadlock. This ensures that deadlock is resolved with certainty and speed. But
the expense is high as partial computations will be lost. Or, we can choose to
abort one process at a time until the deadlock is resolved. This approach has
high overheads because after each abort an algorithm must determine whether the
system is still in deadlock. Several factors must be considered while choosing
a candidate for termination, such as priority and age of the process.
2.
Resource Preemption: Resources allocated to various processes
may be successively preempted and allocated to other processes until the
deadlock is broken.
Prevention
Deadlock
prevention works by preventing one of the four Coffman conditions from
occurring.
·
Removing the mutual
exclusion condition means
that no process will have exclusive access to a resource. This proves
impossible for resources that cannot be spooled.
But even with spooled resources, deadlock could still occur. Algorithms that
avoid mutual exclusion are called non-blocking
synchronization algorithms.
·
The hold and
wait or resource holding conditions may be removed by requiring
processes to request all the resources they will need before starting up (or
before embarking upon a particular set of operations). This advance knowledge
is frequently difficult to satisfy and, in any case, is an inefficient use of
resources. Another way is to require processes to request resources only when
it has none. Thus, first they must release all their currently held resources
before requesting all the resources they will need from scratch. This too is
often impractical. It is so because resources may be allocated and remain
unused for long periods. Also, a process requiring a popular resource may have
to wait indefinitely; as such a resource may always be allocated to some
process, resulting in resource
starvation. (These algorithms,
such as serializing tokens, are
known as the all-or-none
algorithms.)
·
The no preemption condition may also be difficult or
impossible to avoid as a process has to be able to have a resource for a
certain amount of time, or the processing outcome may be inconsistent or thrashing may occur. However, inability to
enforce preemption may interfere with a priority algorithm. Preemption of a
"locked out" resource generally implies a rollback, and is to be avoided, since
it is very costly in overhead. Algorithms that allow preemption include lock-free and wait-free algorithms and optimistic
concurrency control. This condition may be removed as follows : If a
process holding some resources and requests for some another resource(s) which
cannot be immediately allocated to it, then by releasing all the currently
being held resources of that process.
·
The final condition is the circular
wait condition. Approaches
that avoid circular waits include disabling interrupts during critical sections
and using a hierarchy to determine a partial ordering of resources. If no obvious hierarchy
exists, even the memory address of resources has been used to determine
ordering and resources are requested in the increasing order of the enumeration, Dijkstra's solution can also be used.
In a database, a deadlock is a situation in which two or more transactions are waiting for one another to give up locks. For example, Transaction A might hold a lock on some rows in the Accounts table and needs to update some rows in the Orders table to finish.
Comments
Post a Comment