Cybernetics of Kindness:

In today’s post, I am looking at the Socrates of Cybernetics, Heinz von Foerster’s ethical imperative:

“Always act so as to increase the number of choices.”

I see this as the recursive humanist commandment. This is very much applicable to ethics, and how we should treat each other. Von forester said the following about ethics:

Whenever we speak about something that has to do with ethics, the other is involved. If I live alone in the jungle or in the desert, the problem of ethics does not exist. It only comes to exist through our being together. Only our togetherness, our being together, gives rise to the question, How do I behave toward the other so that we can really always be one?

Von Foerster’s views align with that of constructivism, the idea that we construct our knowledge about our reality. We construct our knowledge to “re-cognize” a reality through the intercorrelation of the activities of the various sense organs. It is through these computed correlations that we recognize a reality. No findings exist independently of observers. Observing systems can only correlate their sense experiences with themselves and each other.

Paul Pangaro reminded me that von Foerster did not mean “options” or “possibilities”. Von Foerster specifically chose the word “choices”. By choices, he meant those selections among options that you might “actually take” depending on who “you are” right now. Here choices narrow down to the few that apply most to what you are now in this moment and in this context, down to a decision that makes you who you are. As von Foerster said, “Don’t make the decision, let the decision make you.” You and your choice you take are indistinguishable.

Since we are the ones doing the construction, we are also ultimately responsible for what we construct. No one should take this away from us. Ernst von Glasersfeld, father of radical constructivism explained this well:

The moment you begin to think that you are the author of your knowledge, you have to consider that you are responsible for it. You are responsible for what you are thinking, because it’s you who’s doing the thinking and you are responsible for what you have put together because it’s you who’s putting it all together. It’s a disagreeable idea and it has serious consequences, because it makes you truly responsible for everything you do. You can no longer say “well, that’s how the world is”, or “sono così”; you know, that’s not good enough.

Cybernetics is about communication and control in the animal and machine, as Norbert Wiener viewed it. When we view control in terms of von Foerster’s ethical imperative, interesting thoughts come about. Control is about reducing the number of choices so that only certain pre-selected activities are available for the one being controlled. For example, a steersman has to control their ship such that it maintains a specific course, and here the ship’s “available options” to move are drastically reduced. When we use this view of control and apply it to human beings, we should do so in light of von Foerster’s ethical imperative.

Von Foerster also said – A is better off when B is better off. This also provides further clarity on the recursiveness. If I am to make sure that I act so as to increase the number of choices for B, then B also in turn does the same. How I act impacts how others (re)act, which in turn impacts how I act back… on and on. This might remind the reader of the golden rule – Treat others as you would like others to treat you. However, this is missing the point about constructivism and the ongoing interaction that leads to the construction of a social reality. I see this as part of a social contract. As Jean-Jacques Rousseau noted, Man is born free, but everywhere he is in chains. The social contract comes about from the ongoing interactions and the contexts we are in with our fellow human beings as part of being in a society or social groups. This also means that this is dynamic and contingent in nature. What was “good” before may not be “good” today. This requires an ongoing framing and reframing though interactions.

John Boyd, father of OODA loop, shed more light on this:

Studies of human behavior reveal that the actions we undertake as individuals are closely related to survival, more importantly, survival on our own terms. Naturally, such a notion implies that we should be able to act relatively free or independent of any debilitating external influences — otherwise that very survival might be in jeopardy. In viewing the instinct for survival in this manner we imply that a basic aim or goal, as individuals, is to improve our capacity for independent action. The degree to which we cooperate, or compete, with others is driven by the need to satisfy this basic goal. If we believe that it is not possible to satisfy it alone, without help from others, history shows us that we will agree to constraints upon our independent action — in order to collectively pool skills and talents in the form of nations, corporations, labor unions, mafias, etc — so that obstacles standing in the way of the basic goal can either be removed or overcome. On the other hand, if the group cannot or does not attempt to overcome obstacles deemed important to many (or possibly any) of its individual members, the group must risk losing these alienated members. Under these circumstances, the alienated members may dissolve their relationship and remain independent, form a group of their own, or join another collective body in order to improve their capacity for independent action.

In a similar fashion, Dirk Baecker also noted the following:

Control means to establish causality ensured by communication. Control consists in reducing degrees of freedom in the self-selection of events. This is why the notion of “conditionality” is certainly one of the most important notions in the field of systems theory. Conditionality exists as soon as we introduce a distinction which separates subsets of possibilities and an observer who is forced to choose, yet who can only choose depending on the “product space” he is able to see. If we assume observers on both sides of the control relationship, we end up with subsets of possibilities selecting each other and thereby experiencing, and solving, the problem of “double contingency” so much cherished by sociologists. In other words, communication is needed to entice observers into a self-selection and into the reduction of degrees of freedom that goes with it. This means there must be a certain gain in the reduction of degrees of freedom, which for instance may be a greater certainty in the expectation of specific things happening or not happening.

Ultimately, this is all about what we value for ourselves and for the society we are part of. Our personal freedom makes sense only in light of other’s personal freedoms. That is the context – in relation to another human being, one who may be less fortunate than us. Making the world easier for those less fortunate than us makes the world better for everyone of us. I will finish with a great quote from one of my favorite science fiction character, Doctor Who:

“Human progress isn’t measured by industry. It’s measured by the value you place on a life. An unimportant life. A life without privilege. The boy who died on the river, that boy’s value is your value. That’s what defines an age. That’s what defines a species.”

Please maintain social distance, wear masks and take vaccination, if able. Stay safe and always keep on learning…

In case you missed it, my last post was The Constraint of Custom:

The Cybernetics of “Here & Now” and “There & Then”:

This is available as part of a book offering that is free for community members of Cyb3rSynLabs. Please check here (https://www.cyb3rsynlabs.com/c/books/) for Second Order Cybernetics Essays for Silicon Valley. The e-book version is available here (https://www.cyb3rsyn.com/products/soc-book)

In case you missed it, my last post was The Cybernetics of Bayesian Epistemology:

Direct and Indirect Constraints:

In today’s post, I am following on the theme of Lila Gatlin’s work on constraints and tying it up with cybernetics. Please refer to my previous posts here and here for additional background. As I discussed in the last post, Lila Gatlin used the analogy of language to explain the emergence of complexity in evolution. She postulated that lower complex organisms such as invertebrates focused on D1 constraints to ensure that the genetic material is passed on accurately over generations, while vertebrates maintained a constant level of D1 constraints and utilized D2 constraints to introduce novelty leading to complexification of the species. Gatlin noted that this is similar to Shannon’s second theorem which points out that if a message is encoded properly, then it can be sent over a noisy medium in a reliable manner. As Jeremy Campbell notes:

In Shannon’s theory, the essence of successful communication is that the message must be properly encoded before it is sent, so that it arrives at its destination just as it left the transmitter, intact and free from errors caused by the randomizing effects of noise. This means that a certain amount of redundancy must be built into the message at the source… In Gatlin’s new kind of natural selection, “second-theorem selection,” fitness is defined in terms very different and abstract than in classical theory of evolution. Fitness here is not a matter of strong bodies and prolific reproduction, but of genetic information coded according to Shannon’s principles.

The codes that made possible the so-called higher organisms, Gatlin suggests, were redundant enough to ensure transmission along the channel from DNA to protein without error, yet at the same time they possessed an entropy, in Shannon’s sense of “amount of potential information,” high enough to generate a large variety of possible messages.

Gatlin viewed that complexity arose from the ability to introduce more variety while at the same time maintaining accuracy in an optimal mix, similar to human language where there is always constant emergence of new and new ideas while the main grammar, syntax etc. are maintained. As Campbell continues:

In the course of evolution, certain living organisms acquired DNA messages which were coded in this optimum way, giving them a highly successful balance between variety and accuracy, a property also displayed by human languages. These winning creatures were the vertebrates, immensely innovative and versatile forms of life, whose arrival led to a speeding-up of evolution.

As Campbell puts it, vertebrates were agents of novelty. They were able to revolutionize their anatomy and body chemistry. They were able to evolve more rapidly and adapt to their surroundings. The first known vertebrate is a bottom-dwelling fish that lived over 350 million years ago. They had a heavy external skeleton that anchored them to the floor of the water-body. They evolved such that some of the spiny parts of the skeleton grew into fins. They also evolved such that they developed skull with openings for sense organs such as eyes, nose, ears etc. Later on, some of them developed limbs from the bony supports of fins, leading to the rise of amphibians.

What kind of error-correcting redundancy did he DNA of these evolutionary prize winners, the vertebrates, possess? It had to give them the freedom to be creative, to become something markedly different, for their emergence was made possible not merely by changes in the shape of a common skeleton, but rather by developing whole new parts and organs of the body. Yet this redundancy also had to provide them with the constraints needed to keep their genetic messages undistorted.

Gatlin defined the first type of redundancy, one that allows deviation from equiprobability as ‘D1 constraint’. This is also referred to as ‘governing constraint’. The second type of redundancy, one that allows deviation from independence was termed by Gatlin as ‘D2 constraint’, and this is also referred to as ‘enabling constraint’. Gatlin’s speculation was that vertebrates were able to use both D1 and D2 constraints to increase their complexification, ultimately leading to a high cognitive being such as our species, homo sapiens.

One of the pioneers in Cybernetics, Ross Ashby, also looked at a similar question. He was looking at the biological learning mechanisms of “advanced” organisms. Ashby identified that for lower complex organisms, the main source of regulation is their gene pattern. For Ashby, regulation is linked to their viability or survival. He noted that the lower complex organisms can rely just on their gene pattern to continue to survive in their environment. Ashby noted that they are adapted because their conditions have been constant over many generations. In other words, a low complex organism such as a hunting wasp can hunt and survive simply based on their genetic information. They do not need to learn to adapt, they can adapt with what they have. Ashby referred to this as direct regulation. With direct regulation, there is a limit to the adaptation. If the regularities of the environment change, the hunting wasp will not be able to survive. It relies on the regularities of the environment for its survival. Ashby contrasted this with indirect regulation. With indirect regulation, one is able to amplify adaptation. Indirect regulation is the learning mechanism that allows the organism to adapt. A great example for this is a kitten. As Ashby notes:

This (indirect regulation) is the learning mechanism. Its peculiarity is that the gene-pattern delegates part of its control over the organism to the environment. Thus, it does not specify in detail how a kitten shall catch a mouse, but provides a learning mechanism and a tendency to play, so that it is the mouse which teaches the kitten the finer points of how to catch mice.

The learning mechanism in its gene pattern does not directly teach the kitten to hunt for the mice. However, chasing the mice and interacting with it, trains the kitten how to catch the mice. As Ashby notes, the gene pattern is supplemented by the information supplied by the environment. Part of the regulation is delegated to the environment.

In the same way the gene-pattern, when it determines the growth of a learning animal, expends part of its resources in forming a brain that is adapted not only by details in the gene-pattern but also by details in the environment. The environment acts as the dictionary, while the hunting wasp, as it attacks its prey, is guided in detail by its genetic inheritance, the kitten is taught how to catch mice by the mice themselves. Thus, in the learning organism the information that comes to it by the gene-pattern is much supplemented by information supplied by the environment; so, the total adaptation possible, after learning, can exceed the quantity transmitted directly through the gene-pattern.

Ashby further notes:

As a channel of communication, it has a definite, finite capacity, Q say. If this capacity is used directly, then, by the law of requisite variety, the amount of regulation that the organism can use as defense against the environment cannot exceed Q.  To this limit, the non-learning organisms must conform. If, however, the regulation is done indirectly, then the quantity Q, used appropriately, may enable the organism to achieve, against its environment, an amount of regulation much greater than Q. Thus, the learning organisms are no longer restricted by the limit.

In the same way the gene-pattern, when it determines the growth of a learning animal, expends part of its resources in forming a brain that is adapted not only by details in the gene-pattern but also by details in the environment. The environment acts as the dictionary, while the hunting wasp, as it attacks its prey, is guided in detail by its genetic inheritance, the kitten is taught how to catch mice by the mice themselves. Thus, in the learning organism the information that comes to it by the gene-pattern is much supplemented by information supplied by the environment; so the total adaptation possible, after learning, can exceed the quantity transmitted directly through the gene-pattern.

As I look at Ashby’s ideas, I cannot help but see similarities between the D1/D2 constraints and Direct/Indirect regulation respectively. Indirect regulation, similar to enabling constraints, helps the organism adapt to its environment by connecting things together. Indirect regulation has a second order nature to it such as learning how to learn. It works on being open to possibilities when interacting with the environment. It brings novelty into the situation. Similar to governing constraints, direct regulation focuses only on the accuracy of the ‘message’. Nothing additional or any form of amplification is not possible. Direct regulation is hardwired, whereas indirect regulation is enabling. Direct regulation is context-free, whereas indirect regulation is context-sensitive. What the hunting wasp does is entirely reliant on its gene pattern, no matter the situation, whereas, what a kitten does is entirely dependent on the context of the situation.

Final Words:

Cybernetics can be looked at as the study of possibilities, especially why out of all the possibilities only certain outcomes occur. There are strong undercurrents to information theory in Cybernetics. For example, in information theory entropy is a measure of how many messages might have been sent, but were not. In other words, if there are a lot of possible messages available, and only one message is selected, then it eliminates a lot of uncertainty. Therefore, this represents a high information scenario. Indirect regulation allows us to look at the different possibilities and adapt as needed. Additionally, indirect regulation allows retaining the successes and failures and the lessons learned from them.

I will finish with a great lesson from Ashby to explain the idea of the indirect regulation:

If a child wanted to discover the meanings of English words, and his father had only ten minutes available for instruction, the father would have two possible modes of action. One is to use the ten minutes in telling the child the meanings of as many words as can be described in that time. Clearly there is a limit to the number of words that can be so explained. This is the direct method. The indirect method is for the father to spend the ten minutes showing the child how to use a dictionary. At the end of the ten minutes the child is, in one sense, no better off; for not a single word has been added to his vocabulary. Nevertheless, the second method has a fundamental advantage; for in the future the number of words that the child can understand is no longer bounded by the limit imposed by the ten minutes. The reason is that if the information about meanings has to come through the father directly, it is limited to ten-minutes’ worth; in the indirect method the information comes partly through the father and partly through another channel (the dictionary) that the father’s ten-minute act has made available.

Please maintain social distance, wear masks and take vaccination, if able. Stay safe and always keep on learning…

In case you missed it, my last post was D1 and D2 Constraints:

D1 and D2 Constraints:

In today’s post, I am following up from my last post and looking further at the idea of constraints as proposed by Dr. Lila Gatlin. Gatlin was an American biophysicist, who used the idea of information theory to propose an information-processing aspect of life. In information theory, the ‘constraints’ are the ‘redundancies’ utilized for the transmission of the message. Gatlin’s use of this idea from an evolutionary standpoint is quite remarkable. I will explain the idea of redundancies in language using an example I have used before here. This is the famous idea that if a monkey had infinite time on its hands and a typewriter, it will at some point, type out the entire works of Shakespeare, just by randomly clicking on the typewriter keys. It is obviously highly unlikely that a monkey can actually do this. In fact, this was investigated further by William R. Bennett, Jr., a Yale professor of Engineering. As Jeremy Campbell, in his wonderful book, Grammatical Man, notes:

Bennett… using computers, has calculated that if a trillion monkeys were to type ten keys a second at random, it would take more thana trillion times as long as the universe has been in existence merely to produce the sentence “To be, or not to be: that is the question.”

This is mainly because the keyboard of a typewriter does not truly reflect the alphabet as they are used in English. The typewriter keyboard has only one key for each letter. This means that every letter has the same chance of being struck. From an information theory standpoint, this represents a maximum entropy scenario. Any letter can come next since they all have the same probability of being struck. In English, however, the distribution of letters is not the same. Some letters such as “E” are more likely to occur than say “Q”. This is a form of “redundancy” in language. Here redundancy refers to regularities, something that occurs on a regular basis. Gatlin referred to this redundancy as “D1”, which she described as divergence from equiprobability. Bennett used this redundancy next in his experiment. This will be like saying that some letters now had lot more keys on the typewriter so that they are more likely to be clicked. Campbell continues:

Bennett has shown that by applying certain quite simple rules of probability, so that the typewriter keys were not struck completely at random, imaginary monkeys could, in a matter of minutes, turn out passages which contain striking resemblances to lines from Shakespeare’s plays. He supplied his computers with the twenty-six letters of the alphabet, a space and an apostrophe. Then, using Act Three of Hamlet as his statistical model, Bennett wrote a program arranging for certain letters to appear more frequently than others, on the average, just as they do in the play, where the four most common letters are e, o, t, and a, and the four least common letters are j, n, q, and z. Given these instructions, the computer monkeys still wrote gibberish, but no it had a slight hint of structure.

The next type of redundancy in English is the divergence from independence. In English, we know that certain letters are more likely to come together. For example, “ing” or “qu” or “ion”. If we see an “i” and “o”, then there is high chance that the next letter is going to be an “n”. If we see a “q”, we can be fairly sure that the next letter is going to be a “u”. The occurrence of one letter makes the occurrence of another letter highly likely. In other words, this type of redundancy makes the letter interdependent rather than independent. Gatlin referred to this as “D2”. Bennett utilized this redundancy for his experiment:

Next, Bennett programmed in some statistical rules about which letters are likely to appear at the beginning and end of words, and which pairs of letters, such as th, he, qu, and ex, are used most often. This improved the monkey’s copy somewhat, although it still fell short of the Bard’s standards. At this second stage of programming, a large number of indelicate words and expletives appeared, leading Bennett to suspect that one-syllable obscenities are among the most probable sequences of letters used in normal language. Swearing has a low information content! When Bennett then programmed the computer to take into account triplets of letters, in which the probability of one letter is affected by the two letters which come before it, half the words were correct English ones and the proportion of obscenities increased. At a fourth level of programming, where groups of four letters were considered, only 10 percent of the words produced were gibberish and one sentence, the fruit of an all-night computer run, bore a certain ghostly resemblance to Hamlet’s soliloquy:

TO DEA NOW NAT TO BE WILL AND THEM BE DOES

DOESORNS CALAWROUTOULD

We can see that as Bennett’s experiment started using more and more redundancies found in English, a certain structure seems to emerge. With the use of redundancies, even though it might appear that the monkeys were free to choose any key, the program made it such that certain events were more likely to happen than others. This is the basic premise of constraints. Constraints make certain things more likely to happen than others. This is different than a cause-and-effect phenomenon like a billiard ball hitting another billiard ball. Gatlin’s brilliance was to use this analogy with evolution. She pondered why some species were able to evolve to be more complex than others. She concluded that this has to do with the two types of redundancies, D1 and D2. She considered the transmission of genetic material to be similar to how a message is transmitted from the source to the receiver. She determined that some species were able to evolve differently because they were able to use the two redundancies in an optimal fashion.

If we come back to the analogy with the language, and if we were to only use D1 redundancy, then we would have a very high success rate of repeating certain letters again and again. Eventually, the strings we would generate would become monotonous, without any variety. It would be something like EEEAAEEEAAAEEEO. Novelty is introduced when we utilize the second type of redundancy, D2. Using D2 introduces a more likelihood of emergence since there are more connections present. As Campbell explains the two redundancies further:

Both kinds lower the entropy, but not in the same way, and the distinction is a critical one. The first kind of redundancy, which she calls D1, is the statistical rule that some letters likely to appear more often than the others, on the average, in a passage of text. D1 which is context-free, measures the extent to which a sequence of symbols generated by a message source departs from the completely random state where each symbol is just as likely to appear as any other symbol. The second kind of redundancy, D2, which is context-sensitive, measures the extent to which the individual symbols have departed from a state of perfect independence from one another, departed from a state in which context does not exist. These two types of redundancy apply as much to a sequence of chemical bases strung out along a molecule of DNA as to the letters and words of a language.

Campbell suggests that D2 is a richer version of redundancy because it permits greater variety, while at the same time controlling errors. Campbell also notes that Bennett had to utilize the D1 constraint as a constant, whereas he had to keep on increasing the D2 constraints to the limit of his equipment until he saw something roughly similar to sensible English. Using this analogy to evolution, Gatlin notes:

Let us assume that the first DNA molecules assembled in the primordial soup were random sequences, that is, D2 was zero, and possibly also D1. One of the primary requisites of a living system is that it reproduces itself accurately. If this reproduction is highly inaccurate, the system has not survived. Therefore, any device for increasing the fidelity of information processing would be extremely valuable in the emergence of living forms, particularly higher forms… Lower organisms first attempted to increase the fidelity of the genetic message by increasing redundancy primarily by increasing D1, the divergence from equiprobability of the symbols. This is a very unsuccessful and naive technique because as D1 increases, the potential message variety, the number of different words that can be formed per unit message length, declines. Gatlin determined that this was the reason why invertebrates remained “lower organisms”.

A much more sophisticated technique for increasing the accuracy of the genetic message without paying such a high price for it was first achieved by vertebrates. First, they fixed D1. This is a fundamental prerequisite to the formulation of any language, particularly more complex languages… The vertebrates were the first living organisms to achieve the stabilization of D1, thus laying the foundation for the formulation of a genetic language. Then they increased D2 at relatively constant D1. Hence, they increased the reliability of the genetic message-without loss of potential message variety. They achieved a reduction in error probability without paying too great a price for it… It is possible’ within limits to increase the fidelity of the genetic message without loss of potential message variety provided that the entropy variables change in just the right way, namely, by increasing D2 at relatively constant D1. This is what the vertebrates have done. This is why we are “higher” organisms.

Final Words:

I have always wondered about the exponential advancement of technology and how we as a species were able to achieve it. Gatlin’s ideas made me wonder if they are applicable to our species’ tremendous technological advancement. We started off with stone tools and now we are on the brink of visiting Mars. It is quite likely that we first came across a sharp stone and cut ourselves on it and then thought of using it for cutting things. From there, we realized that we could sharpen certain stones to get the same result. Gatlin puts forth that during the initial stages, it is extremely important that errors are kept to a minimum. We had to first get better at the stone tools before we could proceed to higher and more complex tools. The complexification happened when we were able to make connections – by increasing D2 redundancy. As Gatlin states – D2 endows the structure, The more tools and ideas we could connect, the faster and better we could invent new technologies. The exponentiality only came by when we were able to connect more things to each other.

I was introduced to Gatlin’s ideas through Campbell and Alicia Juarrero. As far as I could tell, Gatlin did not use the terms “context-free” or “context-sensitive”. They seem to have been used by Campbell. Juarrero refers to “context-free constraints” as “governing constraints” and “context-sensitive constraints” as “enabling constraints”. I will be writing about these in a future post. I will finish with a neat observation about the ever-present redundancies in English language from Claude Shannon, the father of Information Theory.:

The redundancy of ordinary English, not considering statistical structure over greater distances than about eight letters, is roughly 50%. This means that when we write English half of what we write is determined by the structure of the language and half is chosen freely.

In other words, if you follow basic rules of English language, you could make sense at least 50% of what you have written, as long as you use short words!

Please maintain social distance, wear masks and take vaccination, if able. Stay safe and always keep on learning… In case you missed it, my last post was More Notes on Constraints in Cybernetics:

More Notes on Constraints in Cybernetics:

In today’s post, I am looking further at constraints. Please see here for my previous post on this. Ross Ashby is one of the main pioneers of Cybernetics, and his book “Introduction to Cybernetics” still remains an essential read for a cybernetician. Alicia Juarrero is a Professor Emerita of Philosophy at Prince George’s Community College (MD), and is well known for her book, “Dynamics in Action: Intentional Behavior as a Complex System”.

I will start off with the basic idea of a system and then proceed to complexity from a Cybernetics standpoint. A system is essentially a collection of variables that an observer has chosen to make sense of something. Thus, a system is a mental construct and not something that is an objective reality. A system from this standpoint is entirely contingent upon the observer. Ashby’s view on complexity was regarding variety. Variety is the number of possible states of a system. A good example of this is a light switch. It has two states – ON or OFF. Thus, we can state that a light switch has a variety of 2. Complexity is expressed in terms of variety. The higher variety a system has, the more possibilities it possesses. A light switch and a person combined has indefinite variety. The person is able to communicate via messages simply by turning the light switch ON and OFF in a certain logical sequence such as Morse code.

Now let’s look at constraints. A constraint can be said to exist when the variety of a system is said to have diminished or decreased. Ashby gives the example of a boys only school. The variety for sex in humans is 2. If a school has a policy that only boys are allowed in that school, the variety has now decreased to 1 from 2. We can say that a constraint exists at the school.

Ashby indicated that we should be looking at all possibilities when we are trying to manage a situation. Our main job is to influence the outcomes so that certain outcomes are more likely than others. We do this through constraints. Ashby noted:

The fundamental questions in regulation and control can be answered only when we are able to consider the broader set of what it (system) might do, when ‘might’ is given some exact specification.

We can describe what we have been talking about so far with a simple schematic. We can try to imagine the possible outcomes of the system when we interact with it and utilize constraints so that certain outcomes, P2 and P4 are more likely to occur. There may be other outcomes that we do not know of or can imagine. Ashby advises that cybernetics is not about trying to understand what a system is, but what a system does. We have to imagine a set of all possible outcomes, so that we can guide or influence the system by managing variety. The external variety is always more than the internal variety. Therefore, to manage a situation, we have to at least match the variety of the system. We do this by attenuating the unwanted variety and by amplifying our internal variety so that we can match the variety thrown at us by the system. This is also represented as Ashby’s Law of Requisite Variety – only variety can absorb variety. Ashby stated:

Cybernetics looks at the totality, in all its possible richness, and then asks why the actualities should be restricted to some portion of the total possibilities.

Ashby talked about several versions of constraints. He talked about slight and severe constraints. He gave an example of a squad of soldiers. If the soldiers are asked to line up without any instructions, they have maximum freedom or minimum constraints to do so. If the order was given that no man may stand next to a man whose birthday falls on the same day, the constraint would be slight, for of all the possible arrangements few would be excluded. If, however, the order was given that no man was to stand at the left of a man who was taller than himself, the constraint would be severe; for it would, in fact, allow only one order of standing (unless two men were of exactly the same height). The intensity of the constraint is thus shown by the reduction it causes in the number of possible arrangements.

Another way that Ashby talked about constraints was by identifying constraint in vectors. Here, multiple factors are combined in a vector such that the resultant constraint is considered. The example that Ashby gave was that of an automobile. He gave the example of the vector shown below:

(Age of car, Horse-power, Color)

He noted that each component has a variety that may or may not be dependent on the other components. If the components are dependent on each other the final constraint will be less than the sum of individual component constraints. If the components are all independent, then the resultant constraints would be the sum of individual constraints. This is an interesting point to further look at. Imagine that we are looking at a team here of say Person A, B and C. Each person here is able to come up with indefinite possibilities, the resultant variety of the team would be also indefinite. If we allow for the indefinite possibilities to emerge, as in innovation or invention of new ideas or products, the constraints could play a role. When we introduce thinking agents to the mix, the number of possibilities goes up.

Complexity is about managing variety – about allowing room for possibilities to tackle complexity. Ashby famously noted that a world without constraints is totally chaotic. His point is that if a constraint exists, it can be used to tackle complexity. Allowing parts to depend upon each other introduces constraints that could cut down on unwanted variety and at the same time allow for innovative possibilities to emerge. The controller’s goal is to manage variety and allow for certain possible outcomes to be more likely than others. For this, the first step to imagine the total set of possible outcomes to best of their abilities. This means that the controller also has to have a good imagination and creative mind. This points to the role of the observer when it comes to seeing and identifying the possibilities. Ashby referred to the set of possibilities as “product space.” Ashby noted that its chief peculiarity is that it contains more than actually exists in the real physical world, for it is the latter that gives us the actual, constrained subset.

The real world gives the subset of what is; the product space represents the uncertainty of the observer. The product space may therefore change if the observer changes; and two observers may legitimately use different product spaces within which to record the same subset of actual events in some actual thing. The “constraint” is thus a relation between observer and thing; the properties of any particular constraint will depend on both the real thing and on the observer. It follows that a substantial part of the theory of organization will be concerned with properties that are not intrinsic to the thing but are relational between the observer and thing.

A keen reader might be wondering how the ideas of constraints stack up against Alicia Juarrero’s versions of constraints. More on this in a future post.  I will finish with a wonderful tribute to Ross Ashby from John Casti:

The striking fact is that Ashby’s idea of the variety of a system is amazingly close to many of the ideas that masquerade today under the rubric “complexity.”

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning… In case you missed it, my last post was Towards or Away – Which Way to Go?

Observations on Observing, The Case Continues:

Art by Audrey Jose

In today’s post, I am continuing from the last post, mainly using the ideas of Dirk Baecker. We noted that every observation is an operation of distinction, where an observer crosses a line, entering a marked state. This is shown in the schematic below. Here “a” refers to the marked state that the observer is interested in. The solid corner of a square is the distinction that was used by the observer, and “n” refers to the unmarked state. The entire schematic with the two sides and the three values (“a”, “n” and the distinction) are notated as a “form”. The first order observer is observing only the marked state “a”, and is not aware of or paying attention to the distinction(s) utilized. They are also not aware of the unmarked state “n”. When a second order observer enters the picture, they are able to see the entire form including the distinction employed by the first order observer.  

However, it is important to note that the observation made by the second order observer is also a first order observation. This means that they also have a distinction and an unmarked state, another “n” that they are not aware of. Baecker explains this:

We have to bring in second-order observers in order to introduce consciousness or self-observation. Yet to be able to operate at all, these second-order observers must also be first-order observers… Second-order observers intervene as first-order observers, thereby presenting their own distinction to further second-order observation.

We also discussed the idea of “reentry” in our last post. Reentry is a means to provide closure so that the first order and second order observations taken together leads to a stable meaning.

So, to recap, the first order observer is interested in “a”.

The second order observer observes the first order observer, and understands that the first order observer made a distinction. They see where the first order observer is coming from, and the context of their observation. Let’s call the context as “b”. This will be the unmarked state for the first observer.

The second order observer engages with the first order observer in an ongoing back and forth discussion. The second order observer is able to combine both their “dealing with the world” approaches and come together to a nuanced understanding. This understanding is an effect of distinguishing “a” from “b”, and also combining “a” and “b” – an action of implication and negation taken together. This is an operation of sensemaking in the medium of meaning. This is depicted as the reentry in the schematic below.

Baecker explains reentry further:

Any operation that is able to look at both sides of the distinction – that is, at its form – is defined by Spencer Brown as an operation of reentry. It consists of reentering the distinction into the distinction, thereby splitting the same distinction into one being crossed and the same one being marked by another distinction that is deferred. The general idea of the reentry is to note and use the fact that distinctions occur in two versions: the distinction actually used, and the distinction looked at or reflected on.

Let’s look further at the form by using a famous syllogism from philosophy to further enhance our understanding:

All Men are Mortals;

Socrates is a man;

Therefore, Socrates is a mortal.

 This can be depicted as a form as shown below:

By distinguishing Socrates from Men, and Men from Mortals, and by putting it all together, we get to “Socrates is Mortal”. In this case, we did not have to do a lot of work to come to the final conclusion. However, as the complexity increases, we will need to perform reentry on an ongoing basis to bring forth a stable meaning. Reentry introduces temporality to the sensemaking operation. No matter how many distinctions we employ, we can only get to a second order observation. All observations are in all actuality first order observations. And what is being distinguished is also dependent entirely on the observer.

I will also look at another example. A manager is required to maintain the operations of a plant while at the same time they need to make modifications to the operations to ensure that the plant can stay viable in an everchanging environment. In other words, the operations are maintained as consistent as possible until it needs to be changed. This can be depicted as shown below:

Another way to look at this is to view a plant as needing centralized structure as well as decentralized structure or top-down and bottom-up structure. This can be depicted as shown below. Here the two states are not shown as nested, but adjacent to each other.

Dirk Baecker saw a firm as follows:

Baecker notes that the product is the first distinction that we have to make. Our first distinction is the distinction of the product. Whatever else the firm may be doing, it has to recursively draw the distinction of which product it is to produce. This may be a material or immaterial, a tangible or intangible, an easy or difficult to define product, but it has to be a product that tells employees, managers and clients alike just what the firm is about. He continues- The technology is part of the form of the first distinction. Indeed, it is the outside or the first context of the first distinction, as observed by a second-order observer who may be the first-order observer observing him/herself. This means that a firm distinguishes only those products for which it has, or hopes to acquire, the necessary technology. Technology here means all kinds of ways of making sure that we can do what we want to do. This includes material access to resources, knowledge of procedures, technologies, availability of people to do the job and ways to convince society that you are doing what you are doing in the proper way.

Baecker explains “work” as follows:

We add the assumption of communication between first-order observers who at the same time act as second-order observers. The firm observes itself. By working, it relates products to technology and technology back to products.

Additional information can be found on Dirk Baecker’s The Form of the Firm.

In all that we have seen so far, we have not yet talked about the unmarked state. The unmarked state “n” is always present in the form and is not accessible to the observer. The observation can have as many distinctions as needed, dependent on the observer. The “n” represents everything that can be further added to the distinctions to improve our “meaning” as needed. The more distinctions there are, the more complex the observations. The observers deal with the complexity of the phenomena to be understood by applying as many or as few distinctions as needed.

We are able to better help with someone else’s problems because we can engage in second order observations. As second order observers, we can see the distinctions they made which are not accessible to them in the first order observation. The second order observer is able to understand the distinctions that the first order observer was able to make. The distinctions lay in the blind spots for the first order observer. The second order observation can be completed by the first order observer themselves as an operation of self-reflection. As cognitive beings, we must reproduce existing patterns by continually engaging with the external world, our local environment. We have to keep evaluating and adjusting these patterns on an ongoing self-correcting basis.

The basic structure of what we have discussed so far can be depicted as the following form:

We need to be mindful that there is always “n” that is not part of our observation. We may gain a better understanding of our distinctions if we engage in second order observation, but we will still not be able to access the unmarked state. We will not be able to access the unmarked state unless we create a new distinction in the unmarked state cutting “n” to a marked state and an unmarked state, yielding a new “n”. Second-order observation, noting one’s own distinctions, can lay the groundwork for epistemic humility.

This brings into question – how many distinctions are really needed? We will answer this with going to the first distinction we made. The first cross that we started with leading to the first distinction is the most important thing that we care about. Every other distinction is based on this first one. To answer – how many distinctions are really needed? – we need as many distinctions as needed until we are fully satisfied with our understanding. This includes understanding our blind spots and the distinctions we have made.

I will finish with a Peter Drucker story from Baecker. Peter Drucker was working with a hospital to improve their Emergency Room. Baecker noted that it took the hospital staff two days to come up with the first distinction, their “a”. Their “a” was to bring immediate relief to the afflicted. The afflicted needing relief may not always be the patient. In Drucker’s words:

Many years ago, I sat down with the administrators of a major hospital to think through the mission statement of the emergency room. It took us a long time to come up with the very simple, and (most people thought) too obvious statement that the emergency room was there to give assurance to the afflicted.

To do that well, you have to know what really goes on. And, much to the surprise of the physicians and nurses, it turned out that in a good emergency room, the function is to tell eight out of ten people there is nothing wrong that a good night’s sleep won’t take care of. You’ve been shaken up. Or the baby has the flu. All right, it’s got convulsions, but there is nothing seriously wrong with the child. The doctors and nurses give assurance.

We worked it out, but it sounded awfully obvious. Yet translating that mission statement into action meant that everybody who comes in is now seen by a qualified person in less than a minute. That is the mission; that is the goal. The rest is implementation.

Some people are immediately rushed to intensive care, others get a lot of tests, and yet others are told: “Go back home, go to sleep, take an aspirin, and don’t worry. If these things persist, see a physician tomorrow.” But the first objective is to see everybody, almost immediately — because that is the only way to give assurance.

This post is also available as a podcast – https://anchor.fm/harish-jose/episodes/Observations-on-Observing–The-Case-Continues-e15kpc1

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning…

In case you missed it, my last post was The Case of the Distinguished Observer:

Complexity is in the Middle:

In today’s post, I am inspired by the idea of a rhizome by Félix Guattari and Gilles Deleuze. They spoke about it in their fascinating book, A Thousand Plateaus. A rhizome is defined in Oxford dictionary as a continuously growing horizontal underground stem which puts out lateral shoots and adventitious roots at intervals. Common examples of rhizomes include crab grass and ginger. Guattari and Delueze or G&D as often notated, used the idea of a rhizome as a metaphor. They put the idea of a rhizome against what they called as “arborescent” or tree-thinking. A tree has a very definite structure; one that is hierarchic with the branches, main stalk and the root system. G&D viewed tree-thinking as being focused on a central idea and building a world view upon that. They noted:

The tree is already the image of the world, or the root the image of the world-tree.

Tree-thinking believes in having a true image of the world. As G&D noted, the tree-thinkers’ law is the law of reflection. They believe that they can simply copy the rules and apply them to any situation. Any situation has a clear structure that is hierarchical and centralized. This can be understood by all if they just follow the logic presented. With this thinking, things can be separated out to distinct categories that do not overlap. Most times this leads to a dichotomy – either this or that, with no middle ground. As G&D noted – binary logic is the spiritual reality of the root-tree. Additionally, the arborescent thinking is also linear thinking, where things follow a linear pattern and rarely lead to paradoxes or confusion.

In a contrast to this, G&D presented rhizome. A rhizome does not have a central structure. It does not have a beginning or an end. Wherever you are, you can start from there. A rhizomic plant can grow from any point in the horizontal structure. If you cut a rhizome in half, each half can grow separately.

A pack of organisms can act as a rhizome. Structures such as a burrow or a city can be a rhizome. There is a collective identification that can be started at any point in the structure. You can start from any point in a city and walk around the city to absorb its culture. It is not specific to one point that we can pinpoint as the start or the end. Just like in a map, we can start anywhere and move around in a map. There is not start or an end. A torn map still remains a map. A rhizome includes the best and the worst.

G&D also calls a collection of elements that are connected together in an intricate relationship as a rhizome. One of the examples they give is that of a certain type of wasp and an orchid. The orchid flower resembles the female wasp, and this leads to a relationship where the wasp becomes part of the reproductive cycle of the orchid. There is a lot more going on in this relationship. This is explained in a very poetic language by G&D:

The orchid deterritorializes by forming an image, a tracing of a wasp; but the wasp reterritorializes on that image. The wasp is nevertheless deterritorialized, becoming a piece in the orchid’s reproductive apparatus. But it reterritorializes the orchid by transporting its pollen. Wasp and orchid, as heterogeneous elements, form a rhizome. It could be said that the orchid imitates the wasp, reproducing its image in a signifying fashion (mimesis, mimicry, lure, etc.). But this is true only on the level of the strata-a parallelism between two strata such that a plant organization on one imitates an animal organization on the other. At the same time, something else entirely is going on: not imitation at all but a capture of code, surplus value of code, an increase in valence, a veritable becoming, a becoming-wasp of the orchid and a becoming-orchid of the wasp. Each of these becomings brings about the deterritorialization of one term and the reterritorialization of the other; the two becomings interlink and form relays in a circulation of intensities pushing the deterritorialization ever further. There is neither imitation nor resemblance, only an exploding of two heterogeneous series on the line of flight composed by a common rhizome that can no longer be attributed to or subjugated by anything signifying.

A rhizome has a circular relationship amongst the elements of its assemblage. A book’s relationship with the world is one such example. A book is never a copy of the world. Its meaning changes with the world. The book changes how we view the world, and this in turn changes how we view the book. G&D noted:

contrary to a deeply rooted belief, the book is not an image of the world. It forms a rhizome with the world, there is an aparallel evolution of the book and the world; the book assures the deterritorialization of the world, but the world effects a reterritorialization of the book, which in turn deterritorializes itself in the world (if it is capable, if it can).

G&D noted that a rhizome is characterized by connections and heterogeneity – any point of a rhizome can be connected to anything other, and must be. Heterogeneity simply means the different or non-identical components in the rhizome. Coming back to the example of the pack of organisms, I am reminded of the idea of complexity. Often, complexity is denoted by the numerous connections within a collective that lead to unforeseen and nonlinear results. Things somewhat make sense when we look backwards. A very good example of a complex phenomenon is child rearing. No matter how many kids you raise, every experience is unique. There is nothing that you can do that will ensure a fixed outcome. There are however several heuristics that might help you along the way. Giving a loving and caring home is a great heuristic for example.

Understanding the idea of a rhizome helps me also understand complexity better. To me, complexity is about possibilities. It is about the numerous connections that are made. Every point is able to connect to any other point. There is no fixed outcome expected. There are mostly nonlinear relationships in a rhizome. The start and the end are boring parts; the excitement is always in the middle. Complexity is in the middle. G&D noted each chapter as a plateau in their book. From this standpoint, a rhizome is also a plateau – just the middle. G&D were French, and they used the term “milieu” to denote the middle. They used the term also because it stood for context. Complexity is all about context. There is no one way for a rhizome. A rhizome is what a rhizome does. You cannot copy what worked in one situation and expect the same outcome from a different situation. A rhizome changes with time. Complexity changes with time. This implies that along with asking what is complexity, we should also ask WHEN is complexity?

Stafford Beer, the eminent Management Cybernetician, viewed variety as the unit for complexity. In Cybernetics, variety is the number of possible states of a collective. For example, a light switch has two states, ON and OFF. The more connections an assemblage has, the more variety it possesses. The more variety something has, the more complex it becomes. A human being has more variety than a switch. A switch is somewhat predictable, while a human being is not. A collection of human beings is even more complex. A human is a rhizome. A collection of human beings is a rhizome. A collection of human beings in their environment is also a rhizome. As I noted before, I see complexity in terms of possibilities. A light switch does not have a lot of possibilities. A light switch, some wires, circuit boards, electronic components and a very curious child have a lot of possibilities. Wherever there are connections, there is a rhizomatic possibility. Wherever elements come together as an assemblage and interact, there is a rhizomatic possibility. The possibility comes from a decentralized space. Every word and every thought are part of a rhizome. This post is also a rhizome with you, the reader.

A rhizome has to remain only a metaphor for complexity or else it fails what G&D intended. It cannot be an exact image of complexity. It cannot be the only way to explain complexity.

G&D were inspired by the great cybernetician and anthropologist Gregory Bateson. They got the idea of a plateau from Bateson. I will finish with a great quote from Bateson:

What is the pattern that connects the crab to the lobster and the orchid to the primrose, and all four of them to me? And me to you?

This post is also available as a podcast here – https://anchor.fm/harish-jose/episodes/Complexity-is-in-the-Middle-e134o61

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning… In case you missed it, my last post was View from the Left Eye – Modes of Observing:

View from the Left Eye – Modes of Observing:

I was introduced to the drawing above through Douglas Harding who wrote the Zen book, “The Headless Way.” The drawing was drawn by Ernst Mach, the 19th Century Austrian physicist. He called the drawing, “the view from the left eye.” What is beautiful about the drawing is that it is sort of a self-portrait. This is the view we all see when we look around (without using a mirror or other reflective surfaces). If we could draw what we see of ourselves, this would be the most accurate picture. This brings me to the point about the different modes of observing.

Right now, you are most likely reading this on a screen of some sort or perhaps you are listening to this as a podcast. You were not paying attention to the phone or computer screen – until I pointed it out to you. You were not paying attention to how your shoes or socks or clothes feel on your body – until I pointed them out to you. This is mostly how we are in the world. We are just being in the world most of the time. Everything that we interact with is invisible to us. They just flow along the affordances we can afford. The keyboard clacks away when we hit on the keys, the door knobs turn when we turn them, etc. We do not see them until we have to see them. The 20th century German philosopher, Martin Heidegger called this ready-to-handedness. Everything is connected to everything else. We interact with the objects in order to achieve something. We open the door to go inside a building to do something else. We get in the car to get to a place. We use a hammer to hammer a nail in order to build something. Heidegger called these things equipment, and he called the interconnectedness, the totality of the equipment. The items are in the background to us. We do not pay attention to them. This is how we generally see the world by simply being in the world.

Now let’s say that the general flow of things breaks down for some reason. We picked up the hammer, and it is heavier than we thought and we pay attention to the hammer. We look at the hammer as a subject looking at an object. We start seeing that it has a red handle and a steel head. The hammer is not ready-to-hand anymore. The hammer has become an object and in the foreground. Heidegger called this as present-at-hand. When we really look at something, we realize that we, the subjects, are looking at something, the object. We no longer have the affordances to interact with it in a nonchalant manner. We have to pay attention in order to engage with the object, if needed.

With this background, I turn to observing again. In my view(no pun intended), there are three modes of observing:

  1. No self – similar to ready-to-hand, you just “are” in the world, enacting in the world. You just see things without any thought to self. There is no distinction of self in what you observe. Perhaps, we can refer to this as the zero person or zero order view.
  2. Seeing self – you make a distinction with this. You draw a line between you the subject, and the world out there. The world is out there and you are separate from the world. This is similar to present-at-hand. The world is out there. This is also the first order in First Order Cybernetics.
  3. Seeing self through self/others – Here you are able to see yourself through self or others. You are able to observe yourself observing. This is the second order in Second Order Cybernetics. In this case, the world is in here, within you, as a constructed stable reality.

In the first mode, you are being in the world. Heidegger would call this as “dasein.” In the second mode, you see the world as being outside. And in the third mode, you see the world as being inside. There are no hierarchies here. Each mode is simply just a mode of observing. In the second and third modes, you become aware of others who are like you in the world. In the third mode, you will also start to see how the others view the world since you are looking through others’ eyes. You realize that just as you construct a world, they too construct a world. Just like you have a perspective, they too have a perspective. The different modes of observing lead to a stable reality for us based on our interpretative framework. We cognize a reality by constructing it based on the stable correlations we infer from our being in the world. Sharing this with others lead to a stable societal realm through our communication with others. A community is formed when we share and something common emerges. It is no accident that the word “community” stems from the root word “common.”

When we observe a system, we also automatically stipulate a purpose for it. Systems are not real-world entities, but a means for the observer to make sense of something. We may call a collection of automobiles on the road as the transportation system just so that we can explain the congestion in the traffic. The same transportation system might be entirely different for the construction worker working on the pavement.

We have to go through the different modes of observation to help further our understanding. Seeing through the eyes of others is a practice for empathy. And this is something that we have to continuously practice to get better at. Empathy requires continuous practice.

I will finish with Ernst Mach’s explanation for his drawing:

Thus, I lie upon my sofa. If I close my right eye, the picture represented in the accompanying cut is presented to my left eye. In a frame formed by the ridge of my eyebrow, by my nose, and by my moustache, appears a part of my body, so far as visible, with its environment. My body differs from other human bodies beyond the fact that every intense motor idea is immediately expressed by a movement of it, and that, if it is touched, more striking changes are determined than if other bodies are touched by the circumstance, that it is only seen piecemeal, and, especially, is seen without a head

It was about 1870 that the idea of this drawing was suggested to me by an amusing chance. A certain Mr L., now long dead, whose many eccentricities were redeemed by his truly amiable character, compelled me to read one of C. F. Krause’s writings, in which the following occurs:

“Problem : To carry out the self-inspection of the Ego.

Solution : It is carried out immediately.”

In order to illustrate in a humorous manner this philosophical “much ado about nothing,” and at the same time to shew how the self-inspection of the Ego could be really “carried out,” I embarked on the above drawing. Mr L.’s society was most instructive and stimulating to me, owing to the naivety with which he gave utterance to philosophical notions that are apt to be carefully passed over in silence or involved in obscurity.

This post is also available as a podcast episode – https://anchor.fm/harish-jose/episodes/View-from-the-Left-Eye–Modes-of-Observing-e1297um

Please maintain social distance and wear masks. Please take vaccination, if able. Stay safe and Always keep on learning…

In case you missed it, my last post was The Stories We Live By: