language-icon Old Web
English
Sign In

Black box

In science, computing, and engineering, a black box is a device, system or object which can be viewed in terms of its inputs and outputs (or transfer characteristics), without any knowledge of its internal workings. Its implementation is 'opaque' (black). Almost anything might be referred to as a black box: a transistor, an algorithm, or the human brain.The constitution and structure of the box are altogether irrelevant to the approach under consideration, which is purely external or phenomenological. In other words, only the behavior of the system will be accounted for.The child who tries to open a door has to manipulate the handle (the input) so as to produce the desired movement at the latch (the output); and he has to learn how to control the one by the other without being able to see the internal mechanism that links them. In our daily lives we are confronted at every turn with systems whose internal mechanisms are not fully open to inspection, and which must be treated by the methods appropriate to the Black Box.(...) This simple rule proved very effective and is an illustration of how the Black Box principle in cybernetics can be used to control situations that, if gone into deeply, may seem very complex. A further example of the Black Box principle is the treatment of mental patients. The human brain is certainly a Black Box, and while a great deal of neurological research is going on to understand the mechanism of the brain, progress in treatment is also being made by observing patients' responses to stimuli. In science, computing, and engineering, a black box is a device, system or object which can be viewed in terms of its inputs and outputs (or transfer characteristics), without any knowledge of its internal workings. Its implementation is 'opaque' (black). Almost anything might be referred to as a black box: a transistor, an algorithm, or the human brain. To analyse something modeled as an open system, with a typical 'black box approach', only the behavior of the stimulus/response will be accounted for, to infer the (unknown) box. The usual representation of this black box system is a data flow diagram centered in the box. The opposite of a black box is a system where the inner components or logic are available for inspection, which is most commonly referred to as a white box (sometimes also known as a 'clear box' or a 'glass box'). The modern meaning of the term 'black box' seems to have entered the English language around 1945. In electronic circuit theory the process of network synthesis from transfer functions, which led to electronic circuits being regarded as 'black boxes' characterized by their response to signals applied to their ports, can be traced to Wilhelm Cauer who published his ideas in their most developed form in 1941. Although Cauer did not himself use the term, others who followed him certainly did describe the method as black-box analysis. Vitold Belevitch puts the concept of black-boxes even earlier, attributing the explicit use of two-port networks as black boxes to Franz Breisig in 1921 and argues that 2-terminal components were implicitly treated as black-boxes before that. In cybernetics, a full treatment was given by Ross Ashby in 1956. A black box was described by Norbert Wiener in 1961 as an unknown system that was to be identified using the techniques of system identification. He saw the first step in self-organization as being to be able to copy the output behavior of a black box. Many other engineers, scientists and epistemologists, such as Mario Bunge, used and perfected the black box theory in the 1960s. In systems theory, the black box is an abstraction representing a class of concrete open system which can be viewed solely in terms of its stimuli inputs and output reactions: The understanding of a black box is based on the 'explanatory principle', the hypothesis of a causal relation between the input and the output. This principle states that input and output are distinct, that the system has observable (and relatable) inputs and outputs and that the system is black to the observer (non-openable). An observer makes observations over time. All observations of inputs and outputs of a black box can be written in a table, in which, at each of a sequence of times, the states of the box’s various parts, input and output, are recorded. Thus, using an example from Ashby, examining a box that has fallen from a flying saucer might lead to this protocol: Thus every system, fundamentally, is investigated by the collection of a long protocol, drawn out in time, showing the sequence of input and output states. From this there follows the fundamental deduction that all knowledge obtainable from a Black Box (of given input and output) is such as can be obtained by re-coding the protocol (the observation table); all that, and nothing more.

[ "Control theory", "Artificial intelligence" ]
Parent Topic
Child Topic
    No Parent Topic