D1 and D2 Constraints:

In today’s post, I am following up from my last post and looking further at the idea of constraints as proposed by Dr. Lila Gatlin. Gatlin was an American biophysicist, who used the idea of information theory to propose an information-processing aspect of life. In information theory, the ‘constraints’ are the ‘redundancies’ utilized for the transmission of the message. Gatlin’s use of this idea from an evolutionary standpoint is quite remarkable. I will explain the idea of redundancies in language using an example I have used before here. This is the famous idea that if a monkey had infinite time on its hands and a typewriter, it will at some point, type out the entire works of Shakespeare, just by randomly clicking on the typewriter keys. It is obviously highly unlikely that a monkey can actually do this. In fact, this was investigated further by William R. Bennett, Jr., a Yale professor of Engineering. As Jeremy Campbell, in his wonderful book, Grammatical Man, notes:

Bennett… using computers, has calculated that if a trillion monkeys were to type ten keys a second at random, it would take more thana trillion times as long as the universe has been in existence merely to produce the sentence “To be, or not to be: that is the question.”

This is mainly because the keyboard of a typewriter does not truly reflect the alphabet as they are used in English. The typewriter keyboard has only one key for each letter. This means that every letter has the same chance of being struck. From an information theory standpoint, this represents a maximum entropy scenario. Any letter can come next since they all have the same probability of being struck. In English, however, the distribution of letters is not the same. Some letters such as “E” are more likely to occur than say “Q”. This is a form of “redundancy” in language. Here redundancy refers to regularities, something that occurs on a regular basis. Gatlin referred to this redundancy as “D1”, which she described as divergence from equiprobability. Bennett used this redundancy next in his experiment. This will be like saying that some letters now had lot more keys on the typewriter so that they are more likely to be clicked. Campbell continues:

Bennett has shown that by applying certain quite simple rules of probability, so that the typewriter keys were not struck completely at random, imaginary monkeys could, in a matter of minutes, turn out passages which contain striking resemblances to lines from Shakespeare’s plays. He supplied his computers with the twenty-six letters of the alphabet, a space and an apostrophe. Then, using Act Three of Hamlet as his statistical model, Bennett wrote a program arranging for certain letters to appear more frequently than others, on the average, just as they do in the play, where the four most common letters are e, o, t, and a, and the four least common letters are j, n, q, and z. Given these instructions, the computer monkeys still wrote gibberish, but no it had a slight hint of structure.

The next type of redundancy in English is the divergence from independence. In English, we know that certain letters are more likely to come together. For example, “ing” or “qu” or “ion”. If we see an “i” and “o”, then there is high chance that the next letter is going to be an “n”. If we see a “q”, we can be fairly sure that the next letter is going to be a “u”. The occurrence of one letter makes the occurrence of another letter highly likely. In other words, this type of redundancy makes the letter interdependent rather than independent. Gatlin referred to this as “D2”. Bennett utilized this redundancy for his experiment:

Next, Bennett programmed in some statistical rules about which letters are likely to appear at the beginning and end of words, and which pairs of letters, such as th, he, qu, and ex, are used most often. This improved the monkey’s copy somewhat, although it still fell short of the Bard’s standards. At this second stage of programming, a large number of indelicate words and expletives appeared, leading Bennett to suspect that one-syllable obscenities are among the most probable sequences of letters used in normal language. Swearing has a low information content! When Bennett then programmed the computer to take into account triplets of letters, in which the probability of one letter is affected by the two letters which come before it, half the words were correct English ones and the proportion of obscenities increased. At a fourth level of programming, where groups of four letters were considered, only 10 percent of the words produced were gibberish and one sentence, the fruit of an all-night computer run, bore a certain ghostly resemblance to Hamlet’s soliloquy:

TO DEA NOW NAT TO BE WILL AND THEM BE DOES

DOESORNS CALAWROUTOULD

We can see that as Bennett’s experiment started using more and more redundancies found in English, a certain structure seems to emerge. With the use of redundancies, even though it might appear that the monkeys were free to choose any key, the program made it such that certain events were more likely to happen than others. This is the basic premise of constraints. Constraints make certain things more likely to happen than others. This is different than a cause-and-effect phenomenon like a billiard ball hitting another billiard ball. Gatlin’s brilliance was to use this analogy with evolution. She pondered why some species were able to evolve to be more complex than others. She concluded that this has to do with the two types of redundancies, D1 and D2. She considered the transmission of genetic material to be similar to how a message is transmitted from the source to the receiver. She determined that some species were able to evolve differently because they were able to use the two redundancies in an optimal fashion.

If we come back to the analogy with the language, and if we were to only use D1 redundancy, then we would have a very high success rate of repeating certain letters again and again. Eventually, the strings we would generate would become monotonous, without any variety. It would be something like EEEAAEEEAAAEEEO. Novelty is introduced when we utilize the second type of redundancy, D2. Using D2 introduces a more likelihood of emergence since there are more connections present. As Campbell explains the two redundancies further:

Both kinds lower the entropy, but not in the same way, and the distinction is a critical one. The first kind of redundancy, which she calls D1, is the statistical rule that some letters likely to appear more often than the others, on the average, in a passage of text. D1 which is context-free, measures the extent to which a sequence of symbols generated by a message source departs from the completely random state where each symbol is just as likely to appear as any other symbol. The second kind of redundancy, D2, which is context-sensitive, measures the extent to which the individual symbols have departed from a state of perfect independence from one another, departed from a state in which context does not exist. These two types of redundancy apply as much to a sequence of chemical bases strung out along a molecule of DNA as to the letters and words of a language.

Campbell suggests that D2 is a richer version of redundancy because it permits greater variety, while at the same time controlling errors. Campbell also notes that Bennett had to utilize the D1 constraint as a constant, whereas he had to keep on increasing the D2 constraints to the limit of his equipment until he saw something roughly similar to sensible English. Using this analogy to evolution, Gatlin notes:

Let us assume that the first DNA molecules assembled in the primordial soup were random sequences, that is, D2 was zero, and possibly also D1. One of the primary requisites of a living system is that it reproduces itself accurately. If this reproduction is highly inaccurate, the system has not survived. Therefore, any device for increasing the fidelity of information processing would be extremely valuable in the emergence of living forms, particularly higher forms… Lower organisms first attempted to increase the fidelity of the genetic message by increasing redundancy primarily by increasing D1, the divergence from equiprobability of the symbols. This is a very unsuccessful and naive technique because as D1 increases, the potential message variety, the number of different words that can be formed per unit message length, declines. Gatlin determined that this was the reason why invertebrates remained “lower organisms”.

A much more sophisticated technique for increasing the accuracy of the genetic message without paying such a high price for it was first achieved by vertebrates. First, they fixed D1. This is a fundamental prerequisite to the formulation of any language, particularly more complex languages… The vertebrates were the first living organisms to achieve the stabilization of D1, thus laying the foundation for the formulation of a genetic language. Then they increased D2 at relatively constant D1. Hence, they increased the reliability of the genetic message-without loss of potential message variety. They achieved a reduction in error probability without paying too great a price for it… It is possible’ within limits to increase the fidelity of the genetic message without loss of potential message variety provided that the entropy variables change in just the right way, namely, by increasing D2 at relatively constant D1. This is what the vertebrates have done. This is why we are “higher” organisms.

Final Words:

I have always wondered about the exponential advancement of technology and how we as a species were able to achieve it. Gatlin’s ideas made me wonder if they are applicable to our species’ tremendous technological advancement. We started off with stone tools and now we are on the brink of visiting Mars. It is quite likely that we first came across a sharp stone and cut ourselves on it and then thought of using it for cutting things. From there, we realized that we could sharpen certain stones to get the same result. Gatlin puts forth that during the initial stages, it is extremely important that errors are kept to a minimum. We had to first get better at the stone tools before we could proceed to higher and more complex tools. The complexification happened when we were able to make connections – by increasing D2 redundancy. As Gatlin states – D2 endows the structure, The more tools and ideas we could connect, the faster and better we could invent new technologies. The exponentiality only came by when we were able to connect more things to each other.

I was introduced to Gatlin’s ideas through Campbell and Alicia Juarrero. As far as I could tell, Gatlin did not use the terms “context-free” or “context-sensitive”. They seem to have been used by Campbell. Juarrero refers to “context-free constraints” as “governing constraints” and “context-sensitive constraints” as “enabling constraints”. I will be writing about these in a future post. I will finish with a neat observation about the ever-present redundancies in English language from Claude Shannon, the father of Information Theory.:

The redundancy of ordinary English, not considering statistical structure over greater distances than about eight letters, is roughly 50%. This means that when we write English half of what we write is determined by the structure of the language and half is chosen freely.

In other words, if you follow basic rules of English language, you could make sense at least 50% of what you have written, as long as you use short words!

Please maintain social distance, wear masks and take vaccination, if able. Stay safe and always keep on learning… In case you missed it, my last post was More Notes on Constraints in Cybernetics:

More Notes on Constraints in Cybernetics:

In today’s post, I am looking further at constraints. Please see here for my previous post on this. Ross Ashby is one of the main pioneers of Cybernetics, and his book “Introduction to Cybernetics” still remains an essential read for a cybernetician. Alicia Juarrero is a Professor Emerita of Philosophy at Prince George’s Community College (MD), and is well known for her book, “Dynamics in Action: Intentional Behavior as a Complex System”.

I will start off with the basic idea of a system and then proceed to complexity from a Cybernetics standpoint. A system is essentially a collection of variables that an observer has chosen to make sense of something. Thus, a system is a mental construct and not something that is an objective reality. A system from this standpoint is entirely contingent upon the observer. Ashby’s view on complexity was regarding variety. Variety is the number of possible states of a system. A good example of this is a light switch. It has two states – ON or OFF. Thus, we can state that a light switch has a variety of 2. Complexity is expressed in terms of variety. The higher variety a system has, the more possibilities it possesses. A light switch and a person combined has indefinite variety. The person is able to communicate via messages simply by turning the light switch ON and OFF in a certain logical sequence such as Morse code.

Now let’s look at constraints. A constraint can be said to exist when the variety of a system is said to have diminished or decreased. Ashby gives the example of a boys only school. The variety for sex in humans is 2. If a school has a policy that only boys are allowed in that school, the variety has now decreased to 1 from 2. We can say that a constraint exists at the school.

Ashby indicated that we should be looking at all possibilities when we are trying to manage a situation. Our main job is to influence the outcomes so that certain outcomes are more likely than others. We do this through constraints. Ashby noted:

The fundamental questions in regulation and control can be answered only when we are able to consider the broader set of what it (system) might do, when ‘might’ is given some exact specification.

We can describe what we have been talking about so far with a simple schematic. We can try to imagine the possible outcomes of the system when we interact with it and utilize constraints so that certain outcomes, P2 and P4 are more likely to occur. There may be other outcomes that we do not know of or can imagine. Ashby advises that cybernetics is not about trying to understand what a system is, but what a system does. We have to imagine a set of all possible outcomes, so that we can guide or influence the system by managing variety. The external variety is always more than the internal variety. Therefore, to manage a situation, we have to at least match the variety of the system. We do this by attenuating the unwanted variety and by amplifying our internal variety so that we can match the variety thrown at us by the system. This is also represented as Ashby’s Law of Requisite Variety – only variety can absorb variety. Ashby stated:

Cybernetics looks at the totality, in all its possible richness, and then asks why the actualities should be restricted to some portion of the total possibilities.

Ashby talked about several versions of constraints. He talked about slight and severe constraints. He gave an example of a squad of soldiers. If the soldiers are asked to line up without any instructions, they have maximum freedom or minimum constraints to do so. If the order was given that no man may stand next to a man whose birthday falls on the same day, the constraint would be slight, for of all the possible arrangements few would be excluded. If, however, the order was given that no man was to stand at the left of a man who was taller than himself, the constraint would be severe; for it would, in fact, allow only one order of standing (unless two men were of exactly the same height). The intensity of the constraint is thus shown by the reduction it causes in the number of possible arrangements.

Another way that Ashby talked about constraints was by identifying constraint in vectors. Here, multiple factors are combined in a vector such that the resultant constraint is considered. The example that Ashby gave was that of an automobile. He gave the example of the vector shown below:

(Age of car, Horse-power, Color)

He noted that each component has a variety that may or may not be dependent on the other components. If the components are dependent on each other the final constraint will be less than the sum of individual component constraints. If the components are all independent, then the resultant constraints would be the sum of individual constraints. This is an interesting point to further look at. Imagine that we are looking at a team here of say Person A, B and C. Each person here is able to come up with indefinite possibilities, the resultant variety of the team would be also indefinite. If we allow for the indefinite possibilities to emerge, as in innovation or invention of new ideas or products, the constraints could play a role. When we introduce thinking agents to the mix, the number of possibilities goes up.

Complexity is about managing variety – about allowing room for possibilities to tackle complexity. Ashby famously noted that a world without constraints is totally chaotic. His point is that if a constraint exists, it can be used to tackle complexity. Allowing parts to depend upon each other introduces constraints that could cut down on unwanted variety and at the same time allow for innovative possibilities to emerge. The controller’s goal is to manage variety and allow for certain possible outcomes to be more likely than others. For this, the first step to imagine the total set of possible outcomes to best of their abilities. This means that the controller also has to have a good imagination and creative mind. This points to the role of the observer when it comes to seeing and identifying the possibilities. Ashby referred to the set of possibilities as “product space.” Ashby noted that its chief peculiarity is that it contains more than actually exists in the real physical world, for it is the latter that gives us the actual, constrained subset.

The real world gives the subset of what is; the product space represents the uncertainty of the observer. The product space may therefore change if the observer changes; and two observers may legitimately use different product spaces within which to record the same subset of actual events in some actual thing. The “constraint” is thus a relation between observer and thing; the properties of any particular constraint will depend on both the real thing and on the observer. It follows that a substantial part of the theory of organization will be concerned with properties that are not intrinsic to the thing but are relational between the observer and thing.

A keen reader might be wondering how the ideas of constraints stack up against Alicia Juarrero’s versions of constraints. More on this in a future post.  I will finish with a wonderful tribute to Ross Ashby from John Casti:

The striking fact is that Ashby’s idea of the variety of a system is amazingly close to many of the ideas that masquerade today under the rubric “complexity.”

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning… In case you missed it, my last post was Towards or Away – Which Way to Go?

The Reality of Informationally Closed Entities:

In today’s post, I am looking at the idea of “informationally closed”. The idea of informational closure was first proposed by Ross Ashby. Ashby defined Cybernetics as a study of systems that are informationally tight. Ashby wanted cyberneticians to look at all the possibilities that a system can be in. Here the system refers to a selection of variables that the observer has chosen. Ashby noted that we should not look at what individual act a system produces ‘here and now’, but at all the possible behaviors it can produce. For example, he asked why does the ovum grows into a rabbit, and not a dog or a fish? Ashby noted that this is strictly related to information, and not energy:

Growth of some form there will be; cybernetics asks “why should the changes be to the rabbit-form, and not to a dog-form, a fish-form or even to a teratoma-form?” Cybernetics envisages a set of possibilities much wider than the actual, and then asks why the particular case should conform to its usual particular restriction. In this discussion, questions of energy play almost no part – the energy is simply taken for granted. Even whether the system is closed to energy or open is often irrelevant; what is important is the extent to which the system is subject to determining and controlling factors. So, no information or signal or determining factor may pass from part to part without its being recorded as a significant event. Cybernetics might, in fact, be defined as the study of systems that are open to energy by closed to information and control – systems that are information-tight.

Ashby’s main point regarding this is that the machine or the system under observation selects its actions from a set of possible actions, and this will remain the same until there is a significant event that causes it to alter the set of possible actions. The action of the system is entirely based on its structure, and not because an external agent is choosing that action for the system. The external agent is only triggering or perturbing the system, and the system in turn reacts. This idea of informational closure was further taken up by Humberto Maturana and Francisco Varela. The idea of “informationally closed” is a strong premise for constructivism – the idea that all knowledge is constructed rather than perceived through senses. They noted that as cognizant beings, we are informationally closed. We do not have information enter us externally. We are instead perturbed by the environment, and we react in ways that we are accustomed to. Jonathan D. Raskin expands on this further:

People are informationally closed systems only in touch with their own processes. What an organism knows is personal and private. In adhering to such a view, constructivism does not conceptualize knowledge in the traditional manner, as something moving from “outside” to “inside” a person. Instead, what is outside sets off, triggers, or disrupts a person’s internal processes, which then generate experiences that the person treats as reflective of what is outside. Sensory data and what we make of it are indirect reflections of a presumed outside world. This is why different organisms experience things quite differently. How Jack’s backyard smells to his dog is different from how it smells to him because he and his dog have qualitatively different olfactory systems. Of course, how Jack’s backyard smells to him may also differ from how it smells to Sara because not only is each of them biologically distinct but each has a unique history that informs the things to which they attend and attribute meaning. The world does not dictate what it “smells” like; it merely triggers biological and psychological processes within organisms, which then react to these triggers in their own ways. The kinds of experiences an organism has depend on its structure and history. Therefore, what is known is always a private and personal product of one’s own processes.

Raskin gives an example of a toaster or a washing machine to provide more clarity on the informational closure.

Maturana asserts that from the point of view of a biologist living systems are informationally closed–that is, things don’t get in and they don’t get out. From the outside, you can trigger a change, but you cannot directly instruct. Think of it as having a toaster and a washing machine. And, the toaster is going to toast no matter what you do. And, the washing machine is going to wash no matter what you do. And they both can be triggered by electricity. But the electricity doesn’t tell the toaster what to do. The toaster’s structure tells the toaster what to do. So similarly, we trigger organisms, but what they do has to do with their internal structure–including their nervous system–and the way it responds to various perturbations.

The idea of informational closure forces us to bring a new perspective to how we view the world. How are we able to know about reality? From a constructivism standpoint, we do not have a direct access to the external reality. What we can truly say is how we experience the world, not how the world really is. We do not construct a representation of the external world. This is not possible, if we are informationally closed. What we do is actually construct how we experience the world. As Raskin points out, the world is not a construction; only our experience of it is. Distinguishing experiential reality from external reality (even a hypothetical, impossible-to-prove-for-sure external reality) is important in maintaining a coherent constructivist stance.

All knowledge from this standpoint is personal, and cannot be passed on as a commodity. In constructivism, there is an idea called as the myth of instructive interaction. This means that we cannot be directly instructed. A teacher cannot teach a student with a direct and exact impact. All the teacher can do is to perturb the student so that the student can construct their personal knowledge based on their internal structure. Raskin notes – once people’s internal systems are triggered, they organize their experiential responses into something meaningful and coherent. That is to say, they actively construe. Events alone do not dictate what people know; constructive processes play a central role as people impose meaning and order on sensory data. 

The more interactions we have with a phenomenon, the better we can experience the phenomenon, and it aids in our construction of the stable experiential reality of that phenomenon. Repetition is an important ingredient for this. Ernst von Glasersfeld notes:

Without repetition there would be no reason to claim that a given experiential item has some kind of permanence. Only if we consider an experience to be the second instance of the self-same item we have experienced before, does the notion of permanence arise.

From this point, I will try to look at some questions that might help to further our understanding of constructivism.

What is the point of constructivism if it means that we cannot have an accurate representation of the real world? The ultimate point about constructivism is not about an ontological stance, it is about viability. It is about being able to continue to survive. All organisms are informationally closed, and they continue to stay viable. The goal is to fit into the real world. Raskin explains – the purpose of this knowledge is not to replicate a presumed outside world but to help the organism survive. In Cybernetics, we say that we need to have a model of what we are trying to manage or control. This “model” does not have to be an exact representation of the “system” we are trying to control. We can treat it as a black box where we have no idea about the inner workings of the system. As long as we are able to come up with a set of possibilities and possible triggers for possible outcomes, we can manage the system. A true representation is not needed.

How would one account for a social realm if we are informationally closed? If each of us are informationally closed, and our knowledge are personal, how we do account for the social realm, where we all acknowledge a version of stable social reality. Raskin provides some clarity on this. He notes:

Von Glasersfeld held that people create a subjective internal environment that they populate with “repeatable objects.” These repeatable objects are experienced as “external and independent” by the person constructing internal representations of them. Certain repeatable objects–those we identify as sentient, primarily other people–are treated as if they have the same active meaning-making abilities that we attribute to ourselves. Consequently, we are able to experience an intersubjective reality whenever other people respond to us in ways that we interpret as indicating they experience things the same way we do. Once again, this alleviates concerns about constructivism being solipsistic because people do relationally coordinate with one another in confirming and maintaining their constructions. 

For von Glasersfeld, it means that people construe one another as active meaning makers and consequently treat their personal understandings as communally shared when others’ behavior is interpreted as affirming those understandings. As I stated elsewhere, “when experiencing sociality or an intersubjective reality, we come to experience our constructions as socially shared to the extent that they appear to be (and, for all functional purposes, can be treated as if they are) also held by others”.

Each one of us construct an experiential reality of the external world. This external world includes other people in it. Our ongoing interaction with these people enhances and updates our own experiential world. We come to see the external world as a social construction. Our personal construction gets triggered in a social setting resulting in a social version of that construction. The more frequent and diverse interactions we get, the more viable this construction becomes. The other people are part of this experiential reality and thereby become cocreators of the social reality. In many regards, what we construct are not representations of the external world, but more a domain of constraints and possibilities. Making sense of the external world is a question about viability. If it does not affect viability, one may very well believe in a God or think that the world is flat. The moment, the viability is impacted, the constructions of the reality will have to adjusted/modified.

The image I have chosen for the post is an artwork by the Japanese Zen master, Nakahara Nantenbō (1839 – 1925). The artwork is a depiction of ensō (circle). The caption reads:

Born within the ensō (circle) of the world, the human heart must also become an ensō (circle).

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning…

In case you missed it, my last post was The Ghost in the System:

This post is also available as a podcast – https://anchor.fm/harish-jose/episodes/The-Reality-of-Informationally-Closed-Entities-e16ke0d

References:

  1. An Introduction to Cybernetics, Ross Ashby (1956)
  2. An introductory perturbation: what is constructivism and is there a future in it?, Raskin, Jonathan D. (2015)

The Ghost in the System:

In today’s post, I am looking at the idea of ‘category mistake’ by the eminent British philosopher Gilbert Ryle. Ryle was an ardent opponent of Rene Descartes’ view of mind-body dualism. Ryle also came up with the phrase ‘the ghost in the machine’ to mock the idea of dualism. Cartesian dualism is the idea that mind and body are two separate entities. Descartes was perhaps influenced by his religious beliefs. Our bodies are physical entities that will wither away when we die. But our minds, Descartes concluded are immaterial and can “live on” after we die. Descartes noted:

There is a great difference between mind and body, inasmuch as body is by nature always divisible, and the mind is entirely indivisible… the mind or soul of man is entirely different from the body.

Ryle called this idea the official doctrine:

The official doctrine, which hails chiefly from Descartes, is something like this. With the doubtful exceptions of idiots and infants in arms every human being has both a body and a mind. Some would prefer to say that every human being is both a body and a mind. His body and his mind are ordinarily harnessed together, but after the death of the body his mind may continue to exist and function.

Ryle referred to the idea of Cartesian dualism as the dogma of the ghost in the machine – the physical body being the machine, and the mind being the ghost. Ryle pointed out that Descartes was engaging in a category mistake by saying that mind and body are separate things. A category mistake happens when we operate with an idea as if it belongs to a particular category. Loosely put, it is like comparing apples to oranges, or even better, comparing apples to hammers. The two items do not belong to the same category and hence, a comparison between the two is a futile and incorrect attempt. The mind is not separate from the body. In fact, the two are interconnected and influence each other in a profound manner. Ryle talked about the idea of dualism as the absurdity of the official doctrine:

I shall often speak of it, with deliberate abusiveness, as ‘the dogma of the Ghost in the Machine’. I hope to prove that it is entirely false, and false not in detail but in principle. It is not merely an assemblage of particular mistakes. It is one big mistake and a mistake of a special kind. It is, namely, a category-mistake. It represents the facts of mental life as if they belonged to one logical type or category (or range of types or categories), when they actually belong to another. The dogma is therefore a philosopher’s myth.

Ryle explained the category mistake with some examples. One of the examples was that of a foreigner visiting Oxford or Cambridge:

A foreigner visiting Oxford or Cambridge for the first time is shown a number of colleges, libraries, playing fields, museums, scientific departments and administrative offices. He then asks ‘But where is the University? I have seen where the members of the Colleges live, where the Registrar works, where the scientists experiment and the rest. But I have not yet seen the University in which reside and work the members of your ‘University’. It has then to be explained to him that the University is not another collateral institution, some ulterior counterpart to the colleges, laboratories and offices which he has seen. The University is just the way in which all that he has already seen is organized. When they are seen and when their co-ordination is understood, the University has been seen. His mistake lay in his innocent assumption that it was correct to speak of Christ Church, the Bodleian Library, the Ashmolean Museum and the University, to speak, that is, as if ‘the University’ stood for an extra member of the class of which these other units are members. He was mistakenly allocating the University to the same category as that to which the other institutions belong.

The foreigner committed the category mistake by assuming that the university is a material entity just like different buildings he saw. He could not understand that the university is a collective whole made up of the different buildings, the students, the staff etc. I will discuss one more example that Ryle gave:

The same mistake would be made by a child witnessing the march-past of a division, who, having had pointed out to him such and such battalions, batteries, squadrons, etc., asked when the division was going to appear. He would be supposing that a division was a counterpart to the units already seen, partly similar to them and partly unlike them. He would be shown his mistake by being told that in watching the battalions, batteries and squadrons marching past he had been watching the division marching past. The march-past was not a parade of battalions, batteries, squadrons and a division; it was a parade of the battalions, batteries and squadrons of a division.

Similar to the foreigner, the child was looking for a separate entity called “the division”. He could not understand that the division is what he is seeing. It was not a parade of battalions, batteries, squadrons and a division; it was a parade of the battalions, batteries and squadrons of a division.

Ryle also gave another example of a visitor who was getting an explanation of the game of Cricket. He saw and understood the different players in the field such as the batsman, the bowler, the fielder etc. After he looked at each one of the players, he asked who is in charge of the team spirit. “But there is no one left on the field to contribute the famous element of team-spirit. I see who does the bowling, the batting and the wicket-keeping; but I do not see whose role it is to exercise esprit de corps.” Ryles explained:

Once more, it would have to be explained that he was looking for the wrong type of thing. Team-spirit is not another cricketing-operation supplementary to all of the other special tasks. It is, roughly, the keenness with which each of the special tasks is performed, and performing a task keenly is not performing two tasks. Certainly exhibiting team-spirit is not the same thing as bowling or catching, but nor is it a third thing such that we can say that the bowler first bowls and then exhibits team-spirit or that a fielder is at a given moment either catching or displaying esprit de corps.

The reader would have noticed that I titled the post – The Ghost in the System. I am alluding to the category mistakes we make in systems thinking. Most often we commit the category mistake of assuming that the system is a standalone objective entity. This is an ontological error. We talk of a hospital system or a transportation system as if it is a physical entity that is visible for everyone to see and understand. We talk about optimizing the system or changing the system as if it is a machine that we can repair by changing out a faulty part with another. In actuality, the system we refer to is a mental construct of how we imagine the different chosen components interact with each other producing specific outcomes we are interested. When we talk of the issues haunting the hospital system, we might be meaning the long waits we have to endure, or the expensive tests that we had to go through. Each one of us construct a version of a “system” and yet we use the same term “system” to talk about different aspects. It is a category mistake to assume that we know what the others are saying. Coming back to the example of the hospital system, when we speak of a hospital system, we point to the hospital buildings, the equipment in the hospitals, the waiting rooms, the doctors, the staff, or the patients. But that is not a hospital system, not really because a system is mental construct that is entirely dependent on who is doing the observing. The observer has a specific thing in mind when they use that word. It is a category mistake to assume that you know what was said. The artifacts are not the system. 

Ryle viewed category mistakes occurring due to problems in vocabulary. He wrote:

These illustrations of category-mistakes have a common feature which must be noticed. The mistakes were made by people who did not know how to wield the concepts University, division and team-spirit. Their puzzles arose from inability to use certain items in the English vocabulary.

Wittgenstein famously wrote – The limits of language are the limits of my world. Our use of language limits what we can know or tell about the world. To go further with this idea, I am looking at the idea of systems from West Churchman’s viewpoint. Churchman advised us that a systems approach begins when first you see the world through the eyes of another. We live in a social realm and by social realm, I mean that we live in a world where “reality” is co-constructed with the other inhabitants of the realm. We define and redefine reality on an ongoing basis through continual interactions with the other cocreators. We should have a model or an image of what we are trying to manage. But if social realm is cocreated, we need to be aware of others in the realm and treat it as a cocreation rather than an objective reality that we have access to. Systems do not have an objective existence. Each one of us view and construct systems from our viewpoint. Thus, how we define a system is entirely dependent on us, the observers. What we have to do is to seek understanding before we rush in to change or optimize a system. The first step is to be aware of the others in the realm. The next step is to seek understanding and see how each one of them views the world. We have to better our vocabulary so that we can speak their language.

There is no ghost in the machine. There is only the machine.

I will finish with a wonderful reflexive nugget from Ryle:

In searching for the self, one cannot be the hunter and the hunted.

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning…

In case you missed it, my last post was The Cybernetics of Complexity:

This post is also available as a podcast – https://anchor.fm/harish-jose/episodes/The-Ghost-in-the-System-e169men

Observations on Observing, The Case Continues:

Art by Audrey Jose

In today’s post, I am continuing from the last post, mainly using the ideas of Dirk Baecker. We noted that every observation is an operation of distinction, where an observer crosses a line, entering a marked state. This is shown in the schematic below. Here “a” refers to the marked state that the observer is interested in. The solid corner of a square is the distinction that was used by the observer, and “n” refers to the unmarked state. The entire schematic with the two sides and the three values (“a”, “n” and the distinction) are notated as a “form”. The first order observer is observing only the marked state “a”, and is not aware of or paying attention to the distinction(s) utilized. They are also not aware of the unmarked state “n”. When a second order observer enters the picture, they are able to see the entire form including the distinction employed by the first order observer.  

However, it is important to note that the observation made by the second order observer is also a first order observation. This means that they also have a distinction and an unmarked state, another “n” that they are not aware of. Baecker explains this:

We have to bring in second-order observers in order to introduce consciousness or self-observation. Yet to be able to operate at all, these second-order observers must also be first-order observers… Second-order observers intervene as first-order observers, thereby presenting their own distinction to further second-order observation.

We also discussed the idea of “reentry” in our last post. Reentry is a means to provide closure so that the first order and second order observations taken together leads to a stable meaning.

So, to recap, the first order observer is interested in “a”.

The second order observer observes the first order observer, and understands that the first order observer made a distinction. They see where the first order observer is coming from, and the context of their observation. Let’s call the context as “b”. This will be the unmarked state for the first observer.

The second order observer engages with the first order observer in an ongoing back and forth discussion. The second order observer is able to combine both their “dealing with the world” approaches and come together to a nuanced understanding. This understanding is an effect of distinguishing “a” from “b”, and also combining “a” and “b” – an action of implication and negation taken together. This is an operation of sensemaking in the medium of meaning. This is depicted as the reentry in the schematic below.

Baecker explains reentry further:

Any operation that is able to look at both sides of the distinction – that is, at its form – is defined by Spencer Brown as an operation of reentry. It consists of reentering the distinction into the distinction, thereby splitting the same distinction into one being crossed and the same one being marked by another distinction that is deferred. The general idea of the reentry is to note and use the fact that distinctions occur in two versions: the distinction actually used, and the distinction looked at or reflected on.

Let’s look further at the form by using a famous syllogism from philosophy to further enhance our understanding:

All Men are Mortals;

Socrates is a man;

Therefore, Socrates is a mortal.

 This can be depicted as a form as shown below:

By distinguishing Socrates from Men, and Men from Mortals, and by putting it all together, we get to “Socrates is Mortal”. In this case, we did not have to do a lot of work to come to the final conclusion. However, as the complexity increases, we will need to perform reentry on an ongoing basis to bring forth a stable meaning. Reentry introduces temporality to the sensemaking operation. No matter how many distinctions we employ, we can only get to a second order observation. All observations are in all actuality first order observations. And what is being distinguished is also dependent entirely on the observer.

I will also look at another example. A manager is required to maintain the operations of a plant while at the same time they need to make modifications to the operations to ensure that the plant can stay viable in an everchanging environment. In other words, the operations are maintained as consistent as possible until it needs to be changed. This can be depicted as shown below:

Another way to look at this is to view a plant as needing centralized structure as well as decentralized structure or top-down and bottom-up structure. This can be depicted as shown below. Here the two states are not shown as nested, but adjacent to each other.

Dirk Baecker saw a firm as follows:

Baecker notes that the product is the first distinction that we have to make. Our first distinction is the distinction of the product. Whatever else the firm may be doing, it has to recursively draw the distinction of which product it is to produce. This may be a material or immaterial, a tangible or intangible, an easy or difficult to define product, but it has to be a product that tells employees, managers and clients alike just what the firm is about. He continues- The technology is part of the form of the first distinction. Indeed, it is the outside or the first context of the first distinction, as observed by a second-order observer who may be the first-order observer observing him/herself. This means that a firm distinguishes only those products for which it has, or hopes to acquire, the necessary technology. Technology here means all kinds of ways of making sure that we can do what we want to do. This includes material access to resources, knowledge of procedures, technologies, availability of people to do the job and ways to convince society that you are doing what you are doing in the proper way.

Baecker explains “work” as follows:

We add the assumption of communication between first-order observers who at the same time act as second-order observers. The firm observes itself. By working, it relates products to technology and technology back to products.

Additional information can be found on Dirk Baecker’s The Form of the Firm.

In all that we have seen so far, we have not yet talked about the unmarked state. The unmarked state “n” is always present in the form and is not accessible to the observer. The observation can have as many distinctions as needed, dependent on the observer. The “n” represents everything that can be further added to the distinctions to improve our “meaning” as needed. The more distinctions there are, the more complex the observations. The observers deal with the complexity of the phenomena to be understood by applying as many or as few distinctions as needed.

We are able to better help with someone else’s problems because we can engage in second order observations. As second order observers, we can see the distinctions they made which are not accessible to them in the first order observation. The second order observer is able to understand the distinctions that the first order observer was able to make. The distinctions lay in the blind spots for the first order observer. The second order observation can be completed by the first order observer themselves as an operation of self-reflection. As cognitive beings, we must reproduce existing patterns by continually engaging with the external world, our local environment. We have to keep evaluating and adjusting these patterns on an ongoing self-correcting basis.

The basic structure of what we have discussed so far can be depicted as the following form:

We need to be mindful that there is always “n” that is not part of our observation. We may gain a better understanding of our distinctions if we engage in second order observation, but we will still not be able to access the unmarked state. We will not be able to access the unmarked state unless we create a new distinction in the unmarked state cutting “n” to a marked state and an unmarked state, yielding a new “n”. Second-order observation, noting one’s own distinctions, can lay the groundwork for epistemic humility.

This brings into question – how many distinctions are really needed? We will answer this with going to the first distinction we made. The first cross that we started with leading to the first distinction is the most important thing that we care about. Every other distinction is based on this first one. To answer – how many distinctions are really needed? – we need as many distinctions as needed until we are fully satisfied with our understanding. This includes understanding our blind spots and the distinctions we have made.

I will finish with a Peter Drucker story from Baecker. Peter Drucker was working with a hospital to improve their Emergency Room. Baecker noted that it took the hospital staff two days to come up with the first distinction, their “a”. Their “a” was to bring immediate relief to the afflicted. The afflicted needing relief may not always be the patient. In Drucker’s words:

Many years ago, I sat down with the administrators of a major hospital to think through the mission statement of the emergency room. It took us a long time to come up with the very simple, and (most people thought) too obvious statement that the emergency room was there to give assurance to the afflicted.

To do that well, you have to know what really goes on. And, much to the surprise of the physicians and nurses, it turned out that in a good emergency room, the function is to tell eight out of ten people there is nothing wrong that a good night’s sleep won’t take care of. You’ve been shaken up. Or the baby has the flu. All right, it’s got convulsions, but there is nothing seriously wrong with the child. The doctors and nurses give assurance.

We worked it out, but it sounded awfully obvious. Yet translating that mission statement into action meant that everybody who comes in is now seen by a qualified person in less than a minute. That is the mission; that is the goal. The rest is implementation.

Some people are immediately rushed to intensive care, others get a lot of tests, and yet others are told: “Go back home, go to sleep, take an aspirin, and don’t worry. If these things persist, see a physician tomorrow.” But the first objective is to see everybody, almost immediately — because that is the only way to give assurance.

This post is also available as a podcast – https://anchor.fm/harish-jose/episodes/Observations-on-Observing–The-Case-Continues-e15kpc1

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning…

In case you missed it, my last post was The Case of the Distinguished Observer:

The Case of the Distinguished Observer:

In today’s post, I am looking at observation. This will be a general overview and I will follow up with more posts in the future. I am inspired by the ideas of George Spencer-Brown (GSB), Niklas Luhman, Dirk Baecker and Heinz von Foerster. In Cybernetics, observation does not mean just to utilize your eyes and look at something. It has a deeper “sensemaking” type meaning. Observation in Cybernetics does not follow the rigid subject-object relationship. Toth Benedek explains this:

Heinz von Foerster tried to develop a point of view that replaces the linear and rigid structure of the object-subject (observer-observed) distinction. According to von Foerster, the observer is really constructed by the observed and vice versa: ‘observation’ is nothing else but the circular relation between them. Observation as a relation defines the observer and the observed, so the observer refers not only to the observed, but also to himself by the act of observation.

Observation is an operation of distinction. The role of an observer is to generate information. If no information is being generated, then no observation has been made. An observation is an act of cognition. GSB in his seminal work, Laws of Form noted:

A universe comes into being when a space is severed or taken apart. The skin of a living organism cuts off an outside from an inside. So does the circumference of a circle in the plane. By tracing the way we represent such a severance, we can begin to reconstruct, with an accuracy and coverage that appear almost uncanny, the basic forms underlying linguistic, mathematical, physical, and biological science, and can begin to see how the familiar laws of our own experience follow inexorably from the original act of severance.

GSB advises us to draw a distinction. He proposed a notation called as “mark” to do this. A basic explanation of a mark is shown below. It separates a space into two sections; one part that is observed and the other that is not observed. We can look at a space, and identify a difference, a distinction that allows us to identify a part of the space as something and the remaining of that space as NOT that something. For example, we can distinguish a part of a house as kitchen and everything else is “not kitchen”. At that point in time, we are looking only at the kitchen, and ignoring or not paying attention to anything else. What is being observed is in relation to what is not being observed. A kitchen is identified as “kitchen” only in the context of the remaining of the house.

Dirk Baecker explains this:

Spencer-Brown’s first propositions about his calculus is the distinction being drawn itself, considered to be “perfect continence”, that is to contain everything. A distinction can only contain everything when one assumes that it indeed contains (a) its two sides, that is the marked state and the unmarked state, (b) the operation of the distinction, that is the separation of the two sides by marking one of them, and (c) the space in which all this occurs and which is brought forth by this occurrence.

From the context of GSB, we can view a distinction as a first order observation. We can only see what is inside the box, and not what is outside the box. What is outside the box is our “blind spot.”

Hans-Georg Moeller explains this very well:

A first-order observation can simply observe something and, on the basis of this, establish that thing’s factuality: I see that this book is black—thus the book is black. Second-order observation observes how the eye of an observer constructs the color of this book as black. Thus, the simple “is” of the expression “the book is black” becomes more complex—it is not black in itself but as seen by the eyes of its observer. The ontological simplicity is lost and the notion of “being” becomes more complex. What is lost is the certainty about the “essential” color of this book.

The first order observer is confident about the observation he makes. He may view his observation as necessary and not contingent. However, a second order observer is able to also see what the first order observer is not. The second order observer is able to understand to an extent how the first order observer is making his distinctions. The second order observer thus comes to the conclusion that the distinction made by the first order observer is in fact contingent and somewhat arbitrary.

The most important point about the first order observation is that the first order observer cannot see that he does not see what he does not see. In other words, the first order observer is unaware that he has a blind spot. A second order observer observing a first order observer is able to see what the first order observer is not able to see, and he is also able to see that the first order observer has a blind spot. This is depicted in the schematic below:

As the schematic depicts, the second order observer is also making a distinction. In other words, what he is doing is also a first order observation! This means that the second order observer also has a blind spot, and he not aware that he has a blind spot! As Benedek further notes:

the first order of observation (our eye’s direct observation) is unable to get a coherent and complete image about the world out there. What we can see is something we learnt to see: the image we “see” is a result of computing processes.

The second order observation can also be carried out as a self-observation, where the observer doing the first order observation is also the observer doing the second order observation. This may appear paradoxical. GSB talked about an idea called “reentry” in Laws of Form. Reentry is the idea of reentering the form again. In other words, we are re-introducing the distinction we used onto the form again. The reentry is depicted in the schematic below:

Dirk Baecker explains:

Spencer-Brown’s calculus of form consists in developing arithmetic with distinctions from a first instruction—”draw a distinction”—to the re-introduction (“re-entry”) of the distinction into the form of distinction, in order to be able to show in this way that the apparently simple, but actually already complex beginning involved in making a distinction can only take place in a space in which the distinction is for its part introduced again. The observer who makes this distinction through it becomes aware of the distinction, to which he is himself indebted.

Self-observation requires a reentry. In order to become aware that we have cognitive blind spots, we have to perform reentry. The re-entry includes what was not part of the original distinction. This allows us to understand (to a point) how we make and utilize distinctions. To paraphrase Heinz von Foerster, we come to see when we realize that we cannot see.

The reentry is a continuous operation that is self-correcting in nature. There is no end point to this per-se and it oscillates between the inside and the outside. This leads to an emergent stability as an eigenform. As noted before, the second order observation is still a form of first order observation even with reentry. There are still cognitive blind spots and we are still subject our biases and limitations of our interpretative framework. We are affected by what we observe and we can only observe what our interpretative framework can afford. As noted at the start of the post, the role of the observer is to generate information. If the observer is not able to make a distinction, then no information can be generated. This has the same effect as us being in a closed system where the entropy keeps on increasing. Borrowing a phrase from Stafford Beer, this means that observers are negentropic pumps. We engage in making dynamic distinctions which allows us to gather the requisite information/knowledge to remain viable in an everchanging environment.

The discussion about first order and second order observations may bring up the question – is it possible to have a third order observation? Heinz von Foerster pointed out that there is no need for a third order observation. He noted that a reflection of a reflection is still a reflection. Hans-Georg Moeller explains this further:

While second-order observation arrives at more complex notions of reality or being, it still only observes—it is a second-order observation, because it observes as a first-order observation another first-order observation. It is, so to speak, the result of two simultaneous first-order observations. A third-order observation cannot transcend this pattern—for it is still the first-order observation of a first-order observation of a first-order observation… No higher-order observation—not even a third-order observation—can observe more “essentially” than a lower-order observation. A third-order observation is still an observation of an observation and thus nothing more than a second-order observation. There is no Platonic climb towards higher and higher realities—no observation brings us closer to the single light of truth.

I will finish with some wise words from Dirk Baecker:

Draw a distinction.

Watch its form.

Work its unrest.

Know your ignorance.

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning… In case you missed it, my last post was The Cybernetics of Magic and the Magic of Cybernetics:

The Cybernetics of Magic and the Magic of Cybernetics:

In today’s post, I am looking at magic and cybernetics. From a young age, I have been a fan of magic. I have talked about magic here before. I see magic as the art of paradoxes.  The word paradox stems from the Greek words – “para” and “dokein”, and taken together it means contrary to expectation.

Take for example a simple magic trick where the magician shows you an open empty hand. The magician closes the hand, and does a gentle wiggle and then opens his hand to reveal a coin. He again closes his hands, and does another gentle wiggle and then opens the hand to show that his hand is empty. The magic happens from a self-referential operation. The spectator (or the observer) sees an empty hand and describes it to themselves as an empty hand. Later, when the magician shows their hand again, the hand now contains a coin. The spectator has to reference back to the previous state of empty hand, and face the moment of paradox. The hand that was empty now has a coin. The moment of magic comes only when the spectator can reference back to the empty hand. If we denote the empty hand as A, the value of the hand now is !A or in other words, not an empty hand. If the spectator cannot reference back to their original observation, they will not see the magic. From the magician’s standpoint, he should take care to make sure that this experience is as strong as possible. For example, he should take care to maintain the image of the hand with and without the coin, the same. This means that the position of the fingers, the gap between them, the gesture etc. are all maintained the same for the two states – one where the hand has no coin, and the second where the hand has a coin. This reinforces the “magic” for the spectator.

The idea of self-reference is of great importance in cybernetics. In logic, the idea of self-reference is shunned because it normally leads to paradoxes. A great example for a paradox is the liar paradox. One of the oldest forms of liar paradox is the statement that Epimenides, the Cretan made. He said that, “all Cretans are liars.” Since he himself was a Cretan, that would mean that he is also a liar, but that would mean that what he is saying is true, which means that he must be a liar… and so on. This goes into a paradox from the self-reference. There have been many solutions suggested for this conundrum. One of the ways to resolve any apparent paradox is to introduce temporality into this sentence. We can do this by making the statement slightly ambiguous and add the word “sometimes”. So, the sentence becomes, “all Cretans are liars sometimes.” The temporality suggests that the value for the statement and the person uttering the statement changes with time and this dissolves the paradox.

Paradoxes don’t exist in the “real world.” The reasonable conclusion is that they have something to do with our stubborn and rigid thinking. When we are unwilling to add temporality or ambiguity, we get stuck with our thinking. Another way to look at this is from a programmer’s standpoint. The statement a = a + 1, is valid from a computer program standpoint. Here the variable, “a” does not stand for a constant value. It is a placeholder for a value at a given point in time. Thus, although the equation a = a +1 is self-referential, it does not crash the computer because we introduce temporality to it, and we do not see “a” having one unique value at all times.

In Cybernetics, self-reference is accepted as a normal operation. Cyberneticians talk about second order concepts such as “understanding understanding” and “observing observing”. One of my favorite description of Cybernetics comes from Larry Richards. He describes cybernetics as a way of thinking about ways of thinking (of which it – cybernetics – is one). This is form of self-reference.

In Cybernetics, self-reference does not lead to paradox. Instead, it leads to a stable outcome. As cognizing agents, we build a stable reality based on self-reference. We can do activities such as thinking about thinking or learning about learning from this approach. Louis Kauffman talks about this:

Heinz von Foerster in his essays has suggested the enticing notion that “objects are tokens for eigen behaviors.” … The short form of this meaning is that there is a behavior between the perceiver and the object perceived and a stability or repetition that “arises between them.” It is this stability that constitutes the object (and the perceiver). In this view, one does not really have any separate objects, objects are always “objects perceived,” and the perceiver and the perceived arise together in the condition of observation.

We identify the world in terms of how we shape it. We shape the world in response to how it changes us. We change the world and the world changes us. Objects arise as tokens of behavior that leads to seemingly unchanging forms. Forms are seen to be unchanging through their invariance under our attempts to change, to shape them.

My post was inspired by the ideas of Spencer-Brown, Francisco Varela and Heinz von Foerster. I will finish with another gem from Heinz von Foerster:

I am the observed relation between myself and observing myself.

This post is also available as a podcast here – https://anchor.fm/harish-jose/episodes/The-Cybernetics-of-Magic-and-the-Magic-of-Cybernetics-e14a257

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning… In case you missed it, my last post was TPS’s Operation Paradox:

Note: The point of a = a+ 1, was made also by Elena Esposito (Kalkul der Form).