Often the ideas in our heads start off as unfocused images. As we continue to study ours thoughts, things come into view. The ideas start getting crystallized out from the melt. The nebulous starts becoming solid. This blog contains those thoughts as they develop and become planets in our minds.
At first glance, the two statements might be thought to be the same. If you’re a winner, you’re not a loser. However the two statements actually ask something different. ‘Not being a loser’ are actions that aim to mitigate all negative aspects in your life. ‘Being a winner’ are actions that add value to your life.
Don’t Be A Loser
How many times have you gotten into an argument with someone, only for both parties to only ultimately be angry, feel resentment, yet have achieved no further value from the argument? Imagine an angry couple asking for a divorce, each side spending all of their money on lawyers in hopes of “being victorious” over the other side. These wars are only destructive. At the end of them, you’ll only ever have less than you started with.
Perhaps something more relatable – how about those times we’ve mindlessly flicked through Facebook or Instagram for a few hours. What have we gained? Nothing. What have we lost? Time – arguably our most valuable asset. These mindless time-stealers are destructive in the most insidious way: we downplay our time’s value. “There will be more time tomorrow”. Yet – unlike money, which we can always acquire more of – we will only have a finite amount of time, no matter how hard we try to reverse it.
These different ways of destruction and loss are paths we choose (or fail to see that we choose) that go towards ‘being a loser’. In learning how to avoid these negative actions in our lives – in learning how not to be a loser – by cultivating good habits and being mindful of our actions, we can mitigate these negative actions. However, after mitigation of the negative, we have only stopped ourselves from being in a deficit. We are only still neutral. We may have kept our time and our money, but we need to know how to use it in a way that adds value.
Be A Winner
‘Being a winner’ contains all the obvious actions that we associate with doing well in life. It’s getting the girl. It’s scoring the winning goal in the last minute. It’s becoming a genius millionaire. It’s putting thousands of hours into a discipline to become an expert in that field (or hours into learning social skills to get the girl, or hours of honing our sporting ability every tiring evening to score the goal). These are the things that people focus on, a lot of the time. These are the things that take us from neutral to ‘winning’.
The Monty Hall Problem
The idea of “being a winner” and “not being a loser” can even be applied to Mathematics and Probability. Perhaps it would be more accurate to say that, rather, it can be applied to our perceptions of probability so that we don’t fall into any logical traps; so that we can see the problem from one more perspective: allowing us to see outcomes more clearly.
“Let’s Make A Deal” was a game-show that was popular in the 1960’s in America. Contestants can go on this game-show and go home with the car of their dreams. It was hosted by a famous guy called Monty Hall. Contestants would come onto the show and be presented, by Monty, three doors. Behind one of the doors lay your dream car. Behind the other two doors lay ‘zonks’: objects you didn’t really want.
So, Monty gave you a choice of the three doors. Let’s say you chose door number 1. You have a 1/3 chance of having picked the correct door. Monty knows which doors contains the ‘zonks’. Now, Monty goes up to one of the other two doors and opens one of them, to reveal a zonk. He turns to you and says:
“Do you like door number one? Or do you want to switch? There’s only one other door to choose from now.”
What would you do? You had a 1/3 chance to pick the correct door. You picked. A door was effectively removed. Now you have a 1/2 chance between the two remaining doors, right?
Do you like that door you picked? Is that door a winner?
For years, contestants agonized over this question. There was speculation about what was the correct strategy to use, but there was never a definitive, well known answer during the running of the show.
Well, how about we see the door from the other perspective. Do you dislike that door?
To answer the question from this new perspective, let’s start over. You’ve got a 2/3 chance of picking the incorrect door. That means that you have a 2/3 chance of picking a door with a zonk. Monty knows which doors contain the zonks. So if you pick a zonk, he is forced to open the other door which contains the remaining zonk. The remaining door that hasn’t been picked by you nor opened by Monty, therefore, has a 2/3 chance of containing the car. How much do you dislike your door now?
Clearly, the best option can now be seen to switch. This statistics problem was popularised during the 90’s when a reader’s letter quoted in Marilyn vos Savant’s “Ask Marilyn” column in Parade magazine. Amazingly, many readers of vos Savant’s column refused to believe switching is beneficial despite her explanation. After the problem appeared in Parade, approximately 10,000 readers, including nearly 1,000 with PhDs, wrote to the magazine, most of them claiming vos Savant was wrong. Even when given explanations, simulations, and formal mathematical proofs, many people still do not accept that switching is the best strategy. Paul Erdős, one of the most prolific mathematicians in history – and the Mathemetician who holds the record for the largest amount of published mathematical papers (counting in at 1500) – remained unconvinced until he was shown a computer simulation demonstrating the predicted result.
I was reminded yesterday of a friend who, a few years ago, who didn’t pass her grades. She was distraught.
“I’m a failure!” she cried. I’ve always had an adversity to calling someone a failure. It’s so absolute. Just as when some annoys you, you shouldn’t say, “you’re annoying,” you might say, “you’re annoying me right now”. The first is absolute, The second is temporary. You don’t get annoyed by their very presence (well, I hope no-one has that effect on you). The annoyance comes and goes.
“You’re not a failure, you’ve just failed this time around,” I replied. I was told that I was an optimist (dubious). I didn’t say anything back, but I wish that I had replied that, rather, I was just a rationalist.
We are born without any knowledge bestowed to us – other than the unconscious neural patterns and cognitive biases that allow us to survive as a baby (breathing, shitting, forming attachments with our mothers): which could arguably be said to processes rather than knowledge. As we get older, we accumulate more and more knowledge. After a time, we start testing our knowledge. Sometimes we pass, sometimes we fail. But should we blame ourselves for the knowledge we haven’t yet come across yet and acquired?
Yes, sometimes all of the knowledge is laid out in front of us, and the reason for why we fail is simply that the rate of knowledge acquisition wasn’t good enough. But even then, the skills of acquiring knowledge need to be taught. We need to learn how to learn. And some of us may learn in different preferential ways. Can we be blamed for not coming across a method that suits our style of learning?
I believe that all we can do is to try. Focus our attention at different levels: work at learning how to learn, then on learning. This is a balancing act; there’s no point planning our revision strategies if, by the time we’ve finished planning, we’ve given ourselves no time left to actually revise. Yet still, all we can do is try. When we fail, we highlight our absence of knowledge in certain areas. Failure is a useful learning experience that will help us succeed in the future. With each failure, our chances of being a winner in the future increase, as long as we learn from the failure.
Ultimately, if you don’t buy a ticket, you can’t win. The only time we are ever an absolute failure is when we give up. Because every individual instance of when we fail is an opportunity to glean one more hint into how to succeed next time.
The follow post is my argument on why we can’t know how long it will take to develop AGI. It is based on various assumptions so I’d be interested to read any feedback you have regarding it!
Explanation of Terms: AGI & ANI
There are three broad forms of artificial intelligence: artificial narrow intelligence (ANI), artificial general intelligence (AGI) and artificial super intelligence (ASI). The subject of this post is relating to AGI, however ANI is also needed to explain where we are, so I will briefly explain both… by copying and pasting the explanation used on WaitButWhy:
AI Caliber 1) Artificial Narrow Intelligence (ANI): Sometimes referred to as Weak AI, Artificial Narrow Intelligence is AI that specializes in one area. There’s AI that can beat the world chess champion in chess, but that’s the only thing it does. Ask it to figure out a better way to store data on a hard drive, and it’ll look at you blankly.
AI Caliber 2) Artificial General Intelligence (AGI): Sometimes referred to as Strong AI, or Human-Level AI, Artificial General Intelligence refers to a computer that is as smart as a human across the board—a machine that can perform any intellectual task that a human being can. Creating AGI is a much harder task than creating ANI, and we’re yet to do it. Professor Linda Gottfredson describes intelligence as “a very general mental capability that, among other things, involves the ability to reason, plan, solve problems, think abstractly, comprehend complex ideas, learn quickly, and learn from experience.” AGI would be able to do all of those things as easily as you can.
The Explosive Growth
Knowledge – and the access to that knowledge – is a key component for any AGI, and the internet is used for precisely that reason. The internet has completely changed the landscape for how humans interact with their daily environment compared to twenty years ago. No-one imagined culture would change so much twenty years ago due to our newly found connectivity. No doubt, it will be fundamental in how AGIs will interact with their environment in the future.
We’ve also made an ANI that can drive cars, that can beat humans at Go, that can exhibit creativity in a specific field by creating music or art, an ANI that can learn – through neural networks – how to play a game of Mario:
There is even a computer that has passed the Turing test (well, passed it 33% of the time): Eugene Goostman. And tech companies like Google are getting into serious bidding wars for experts in the field of AI. That must mean we’re close, right?
The Naive Optimism
In the 1950’s, thanks to scientific advances from Turing and von Neuman et al, the scientific community started looking more and more into the prospect of AI, and the field of AI was born.
There was huge optimism with regards to the rate of growth within AI development. In 1958, H. A. Simon and Allen Newell stated, in : “within ten years a digital computer will be the world’s chess champion” and “within ten years a digital computer will discover and prove an important new mathematical theorem.” We know, with hindsight, that the former prediction took four times longer than expected – with Deep Blue beating Kasparov in 1997 rather than 68 – and the latter prediction is still yet to be realised.
Then, in even more optimism, the following claims that have yet to succeed were made:
In 1965, H. A. Simon stated: “machines will be capable, within twenty years, of doing any work a man can do.”
In 1967, Marvin Minsky said: “Within a generation … the problem of creating ‘artificial intelligence’ will substantially be solved.”
And in 1970, Marvin Minsky (in Life Magazine), claimed: “In from three to eight years we will have a machine with the general intelligence of an average human being.”
Why were these great scientists so off their mark? Can we predict just how long it will take us to achieve AGI?
“When eating an elephant take one bite at a time.” – Creighton Abrams
A standard technique when faced with large tasks is to break it down into smaller chunks. Many British readers will be reminded of this strategy when BBC Bitesize helped us use it to study for our GCSE. This technique has become so ubiquitous that it shapes the way we view tasks. When we see a large task, many of us unconsciously break it up into smaller, more manageable chunks. Applying this lens to knowledge is constantly being reinforced, as examples of its application are everywhere: the categorisation of subjects, science, the hierarchical nature of companies are even founded on this idea. Large tasks are split into smaller chunks, where the small chunks of work are categorised in nature and then given to experts in those related field, such as legal tasks to lawyers, tasks relating to the understanding and creation of complex objects to engineers, the tasks of selling objects/services to the sales team, the task of keeping people & work loads on track, prioritized and organised is given to managers.
So how does this all affect our perception of how close we are to AGI? Well, naturally, scientists will break the task down into easier chunks. In their minds, these chunks might look something like this:
1. Create an artificial neuron (= the transistor). Check (But not really. We’ll re-explore this later)
2. Connect millions of neurons together so that it forms something similar to a neural network (= a modern day computer, which contains more transistors than a fruit fly contains neurons). Check.
3. Write the artificial equivalent of the “software of nature”. This field is growing today and is called Machine Learning. No check yet.
So we’re pretty close, right? We’ve got 2 out of 3 tasks complete. Well, no, not really. At this stage, many people could be – and probably generally are – assuming that each task is roughly the same size with regards to complexity, and thus the time needed to complete the task is the same. But we just don’t know. The problem is even bigger than that: we don’t know what we don’t know to figure out the rest of the puzzle. Let me explain…
“Reports that say that something hasn’t happened are always interesting to me, because as we know, there are known knowns; there are things we know we know. We also know there are known unknowns; that is to say we know there are some things we do not know. But there are also unknown unknowns – the ones we don’t know we don’t know. And if one looks throughout the history of our country and other free countries, it is the latter category that tend to be the difficult ones.” – Donald Rumsfeld
What on earth is Rumsfeld talking about? Well, actually, he’s talking about something very valid. Psychologists would call it metacognition: our ability to perceive and make judgments on our knowledge and skills. There are many models that have been created from metacognition, such as the 4 Stages Of Competency and the Dunning-Kruger Effect. If we take the phrase ‘unknown unknowns’, the first ‘unknown’ would relate to our perception of our knowledge, and the second ‘unknown’ would relate to whether we hold that specific packet of knowledge.
There are a few attributes that are associated with each part, however the main one of interest is that unknown unknowns, due to their very nature, cannot be measured in terms of the size of the knowledge. I think we’re starting to see the problem.
Two Paths To Innovation
There are two main paths to take when creating something new.
1. The joining of two already existing sub-components – which had never been joined before – to create something new
The wheel has been around for a while. The suitcase has also been around for a while. But a wheely suitcase? It took us humans until the 1970’s – dragging heavy suitcases around for every holiday – until one clever inventor finally came up with the novel idea of putting the wheel and the suitcase together.
In this situation, all we need to do is get the expert of each component part in to talk about how the relationship between the new sub-components will work. If it’s a simple amalgamation, maybe not even that. For example, if we’d like to build a wheely suitcase, we might need a designer – to make it looks aesthetically pleasing – and an engineer – to make it structurally and physically sound (and to add the wheels). Maybe also a separate production engineer to find the correct tools & develop a production line for the creation of the product. All of the knowledge required to do this already exists, which makes it easier to predict how long it might take to develop this product. We just need to find people with the knowledge already in their brains and add it together in the novel way.
2. Research and development: applying new science to create new technology
This is the trickier route to new products. But where the barrier to entry is greater, the incentive may also be larger: in monetary terms, but also in prestige and personal sense of achievement. This is also the form of invention that AGI takes when it tries to replicate the brain.
Replicating the brain to create artificial AGI requires a two-pronged strategy: the first is in further research into the mechanisms of the brain, so we can better understand how it works. This will create a blueprint of how to create the artificial brain. The second part of the strategy is to further the development of artificial brain parts (computer chips) so we have better and better devices to emulate a brain with. The two paths start at opposite ends of the room, but as each side develops, they get closer and closer together… until finally at some time in the future, they’ll meet: the point at which we have the knowledge and skills to create a AGI by replicating an artificial brain. However, the distance left to cover will remain unknown until it has been covered. Like seeing light at the end of the tunnel: we can see the end, but no matter how far we travel, the light seems just as far away as it was previously. Then, before we realise it, we can make out shapes beyond the end of the tunnel, we keep walking and we suddenly find ourselves in the outside world with the wind on our faces.
The Hindsight Bias
Just a quick one before moving on, it might be worthwhile to touch upon the Hindsight Bias. This is usually applied to History: where in the past, it seems very obvious which factors contributed to the events that unfolded in History. It seems clear, in hindsight, that certain financial instruments and sub-prime mortgages contributed massively to the financial crash of 2008. Was an imminent crash obvious at the time? No, of course not. To look back in History is to remove all the noise, and to see just pure cause and effect.
In the same way that the hindsight bias applies to history, ‘development bias’ (just made this one up) applies to the production of a new product. It might be clear how much complexity is involved in steps 1 and 2 of the AGI task list, in hindsight, but it is still very unclear regarding the path ahead for step 3. It is easy to see, in hindsight, how to create an object when the path has been set out. During product development, however, there are a multitude of attractive looking paths for creation, yet we never know which one can be successful until we walked that path.
Create an artificial neuron (= the transistor). Uncheck.
We humans work by analogy. An electric circuit is a water pump: the water represents electrons being pumps around the circuit. The mind’s neurons are a logic gate: they can either be on or off.
All analogies and models have thresholds at which they break down, however. Analogies of the brain are arguably getting stronger: from clay and water during Greek ages, to being more like a mechanical machines during the industrial revolution, to being like a computer chip in our current era. However we still underestimate the complexity of the brain by over-simplifying to an analogy we can understand, and this has tricked us into believing we can seamlessly just replace a neuron with a transistor. A neuron can be turned ‘on’ from being given an electrical pulse greater than a certain voltage, however this ‘on’ state is temporary. A neuron has no way of storing information in an on-off state as the transistor does: as a result, both components use completely different mechanisms to operate. A computer, as we know them at the moment, will never be able to operate like a brain. This means that we only have 1 of the 3 boxes checked, and we’ll have to traverse the lands of unknown unknowns in the search of our artificial neurons.
The Application of Unknown Unknowns to the AGI Task Hierarchy
“Fail fast, fail forward” – Silicon Valley Mantra
Tasks that have never been undertaken before are unknown unknowns. We known roughly how long it would take to undertake tasks that others have achieved before, because we can gauge how long others took to do them. However with the unknown unknowns, all we can do is to fail forward: to work in a way that moves us from not perceiving what we don’t know in order to achieve the task, to perceiving what we don’t know. To be cartographers of our own map of knowledge: sketching out in ever more detail the areas we’ve acquired and the areas we haven’t acquired. The areas yet to be discovered are dark and potentially limitless in size.
This brings us to task no. 3 for creating AGI. We just don’t know how long it will take us to replicate a human brain, due to what we have yet to achieve. We still only understand a tiny fraction of the complexity of the human brain. As well as this, machine learning is still in it’s relative infancy. We might argue that based on our current understanding of quantum mechanics, and our ability to manipulate at the nanoscale, we’re fairly close to understanding the brain, but that’s assuming that we already have the broad strokes of what is involved filled in. There might be a whole scientific discipline that we have yet to master before we understand the inner mechanics of the human brain, and we don’t even realise we don’t know it yet.
It seemed like a small step, back in 1960, from transistors to AGI. With hindsight, we can clearly see that there was still so much to learn that the scientists of the time hadn’t even perceived yet. Looking back, we can see the path needed to walk for the development of AGI has been far longer than previously anticipated. It might be tempting to think again that the step from where we are to AGI must be small: our success is almost palpable. But in reality we just don’t know how far we still have yet to go to make an AGI perform in the same way as the human brain.
Yet just because our brain performs as an AGI, this doesn’t mean all AGIs need to behave like a brain. To induce – from seeing a swan that is white – that all swans are white, is a fallacy. Are there other ways to create an AGI?
More Than One Way To Skin A Cat
We can use task break-down again, but this time to break down the steps taken by an AGI to perform tasks. If we break each task that an AGI (a human) performs down enough, we will start to see that even the most complex tasks are just the combination of many simple processes. We can break down the task into many different sub-tasks that an individual ANI can perform.
This bottom-up ANI amalgamation is the strategy Google employs in hopes of achieving an AGI. A single ANI is fairly limited by itself in scope, but it can perform a specific task. Maybe it can defeat the best chess players, or maybe it can just tell you what time it is. Hook this ANI up to the internet, though, and it can be used when needed. If a second layer is placed over the top of all these ANIs – a layer which is able to assess the task, then access and apply the appropriate ANIs to complete this task – we have just made our first AGI.
With this form of AGI, we may be able to class invention as the joining of many objects that already exist – ANIs – to form something new: the AGI. The top layer part of the AGI which judges the task and accesses each ANI would still need to be developed, but even this would be more like an adaption of already existing technology. A prediction may not be able to be made with completely accuracy: there are still areas outside the view of the analyst. Yet if the amount of ANIs necessary to form a functioning AGI can be counted, and the speed at which the ANIs which don’t exist can be created can be estimated, a prediction could be made. Perhaps the next question is: is it possible to calculate how many ANIs are needed until an AGI is created?
To Be Continued…?
So maybe the first AGI won’t pop into existence. It will crawl slowly into view, assimilating more and more ANIs until it is able to do any task that a human can do. No doubt, in time, it will continue to grow different ANI limbs in all directions, until it has completely surpassed humanity.
I’m all up for “following your heart” sometimes. But sometimes… we just shouldn’t. There have been a few instances recently were I’ve relied on my intuition – my feelings – to make decisions and after weighing up these decisions in hindsight, I’ve realised they weren’t the best option to choose. I’d like to share these instances with you to highlight how easy it is to make flawed decisions.
1. I went to a conference recently to learn about Vacuum gauges, vacuum pumps, how to create a vacuum etc. The conference was given by one of the experts in the industry (who was high in the hierarchy of a vacuum company) so I thought it would be worth-while listening to the conference. I brushed up on my knowledge with a pdf I found on the internet by the company this guy worked for before the talk. It was pretty comprehensive – 134 pages (I didn’t read all of it), but it perfectly explained everything I needed to know. So I read this pdf for a few hours and then set off to the conference. When I sat down and the expert started talking, I realised he was using slides almost directly taken out of the pdf. However, this information now was coming from an expert: he would still fill in a few interesting bits here and there, right? And I didn’t know it inside out. At the end of the day, I felt like I had got a lot done: I had gone into Glasgow Uni, made it to a vacuum conference to listen to an expert talk. Productive day!
In hindsight, though, it took about an hour to get from my room to into the conference hall (so 2 hours travelling there and back). Commuting somewhere might feel like getting things done – “I’m moving forward!” – but actually it’s sunk cost that you accept because the value of the thing you’re travelling to is still greater than the cost. Parking cost £4.40. And I listened to this lecture for about 1.5 hours. So that’s 3 hours of time for 1.5 hours of learning. Compare this to the amount of time I could have spent learning the information at home – 3 hours for 3 hours, it starts looking like I should have stayed at home. The weird thing is, I don’t feel like I get a lot of stuff done if I’ve just sat at my computer all day. It doesn’t have the gravitas of telling yourself, “I went to a conference to listen to an expert.” The problem is that I’ve attached some inherent value to “listening to an expert,” where really there is none. The value is in the content that the expert can get into our heads. I somehow feel that because this guy is an expert, he can teach me more content than just a pdf. And the conference makes it seem important. I’m an important person with all these other important businessmen, gleaning some knowledge from an expert. You can’t feel that important sitting at your computer at home, reading a pdf.
2. My route home can take me either two ways: a longer route where I can move more quickly, or a shorter route that goes through a road with huge speed-bumps that I need to take at around 5 mph. I always go the fast route: it just feels like I’m getting home quicker: I’m going faster, after all!
I recently found out, though, that the speed-bump route is quicker. Even with this new information, I still sometimes don’t believe it and go the route where I can travel quicker – it just feels quicker.
I’ve even heard of some people, waiting for a late bus, who actually walk to the next bus station so that they can get the bus further down the line. So that that they can catch the same bus just from the next station… only the walk to the next bus station makes them feel like they’ll get to their destination quicker.
Our feelings/intuitions can be sneaky. And maybe the most sneaky thing about them is that because they’re decisions made on feelings, we don’t consciously realise we’ve made them. We don’t rationally study the decision making, because it’s a case that’s been opened and closed within the intuitive realm. It’s only when we either catch ourselves making them, or look back in hindsight, that we can see.