Blocking (computing)
In computing, a process is an instance of a computer program that is being executed. A process always exists in exactly one process state. A process that is blocked is one that is waiting for some event, such as a resource becoming available or the completion of an I/O operation.[1]
In a multitasking computer system, individual tasks, or threads of execution, must share the resources of the system. Shared resources include: the CPU, network and network interfaces, memory and disk.
When one task is using a resource, it is generally not possible, or desirable, for another task to access it. The techniques of mutual exclusion are used to prevent this concurrent use. When the other task is blocked, it is unable to execute until the first task has finished using the shared resource.
Programming languages and scheduling algorithms are designed to minimize the over-all effect blocking. A process that blocks may prevent local work-tasks from progressing. In this case "blocking" often is seen as not wanted.[2] However, such work-tasks may instead have been assigned to independent processes, where halting one has no or little effect on the others, since scheduling will continue. An example is "blocking on a channel" where passively waiting for the other part (no polling or spin loop) is part of the semantics of channels.[3] Correctly engineered any of these may be used to implement reactive systems.
Deadlock means that processes pathologically wait for each other in a circle. As such it is not directly associated with blocking.
Once the event occurs for which the process is waiting ("is blocked on"), the process is advanced from blocked state to an imminent one, such as runnable.
See also
References
- Stallings, William (2004). Operating Systems: Internals and Design Principles (5th ed.). Prentice Hall.
- C++ and Beyond 2012: Herb Sutter - C++ Concurrency
- Go concurrency patterns by Rob Pike