I am a machine learning researcher at Camb.AI. I post about deep learning, electronics, and other things I find interesting.
by Matthew Baas
My particularly uninformed opinion and speculation on when and if we might get to use actual human-level AI.
TL;DR: things are moving very fast in the machine learning field, particularly with language models (LMs) and the field of natural language processing (NLP). If the trend continues, the question arises: when will (if at all) me and you, the reader, will be able to use actual human level AI. This post is my speculation and uninformed opinion on when this might be, and what some possible hurdles are. It does not assume significant technical knowledge from the reader, just a cursory familiarity with what current LM’s are and how they roughly operate.
Let’s look at our question:
When will we have actual human-level AI?
There are 3 concepts that we need to clarify before we can answer. The first one is easy: By “when” I mean a year, or span of a couple of years, wherein we could expect [the predicate]. The second two key parts require more explanation:
By actual human-level AI I mean a machine learning model that one can interact with at the level of an average human through software. I.e. I’m excluding robots or physical embodiments of software models, and purely focusing on sensory input & output through software (e.g. text/vision/audio).
To clarify some more, here are key things I do not mean when I say human-level AIi in this post:
Key things I do mean when I say human-level AI:
By ‘we have’ I mean that we – myself and you, dear reader – can feasibly use such a model without big restrictions. In other words, there should not be large swathes of the AI’s capability that regular people are not allowed to use. My reasoning is simple: it doesn’t make sense to even talk about when we can use a human-level AI if the version which we’re restricted to use is lobotomized to be below human-level capability.
To give a better indication for what constitutes ‘big restrictions’, here are a few things which definitely are lobotomizing restrictions on regular users:
Examples which I do not consider ‘big restrictions’ are:
For all of these, the reasoning is simple: “humans have the ability to talk about X, if the AI we have access to cannot talk about X then it is not at human-level”. E.g. humans will get arrested if they engage in illegal activity, but can continue indefinitely talking about politics or giving amateur advice on medical conditions.
So now that you understand what I mean by the main question, let’s get to the main trend I’d like to speculate from: scaling laws for language models, and questionable comparisons to the size of the human brain. Here I will make substantial approximations in line with the engineering method. My goal is not to arrive at the perfect answer, but to give a concrete date to our question using assumptions based on the information I currently know about this uncertain and rapidly changing field. Specifically, there are 2 key assumptions I am going to make:
Approximation: In terms of capability, the amount of parameters in an artificial neural network roughly correlates with the amount of synapses in the human brain.
This rough correlation follows from the rough intuitive inspiration of artifical neural networks from actual neural networks. It is very rough, approximate, and not always true to reality – but the correlation holds: typically organisms with more synapses have greater capabilities than those with fewer synapses. There will be counter-examples (e.g. certain whales), but very roughly I think it holds on average.
Approximation: The size of machine learning models will continue to increase at a roughly similar pace to what we have seen in the past decade.
Given the two assumptions above, we can now gather some numbers and start to see a trend. First: the human brain has approximately 100-1000 trillion synaptic connections. Some sources for this are here and here. Given these estimates, I will make the rough approximation that an average human has 400 trillion (400T) synapses.
Aside 1: Why not use number of neurons as an estimate? Because we know this cannot be accurate to within even a factor of 100, since the human brain has about 100B neurons and we have neural networks with much more than that in terms of parameter count, but which are still much worse than an average human.
Aside 2: Synapses vs synaptic connections – in this post I use them interchangably, since the estimates are within a factor of 10. Estimates for the number of synaptic connections sit at 100T, and for synapses in total sit at 1000T. Both of these are fairly crude estimates already, and the connection between counts of biological structures and counts of model paramters is also fairly approximate, so taking an approximation of 400T synapses/synaptic connections for the purposes of comparison to parameter count is, I think, sufficient.
Next, let’s look at the biggest model sizes in the last several years:
We need an estimate for the amount of parameters in GPT4. Google’s PaLM 1 had 540B parameters, and so we can likely assume GPT4 probably has at least that. We can also get some information from the pricing (since GPT4 and GPT3 have API pricing by the same provider OpenAI):
A portion of the price difference will be because of other network improvements aside from scale (i.e. architecture improvements), from a higher premium for a better product, and from OpenAI more concretely realizing their monopolistic ambitions in the AI space. So, as before, I am going to make a rough approximation:
Approximation: a fraction of the price increase is due to increase in compute necessary for inference with GPT4 such that the size of GPT4 is roughly 5x as big as GPT3.
This assumption means that GPT4 has roughly 1T parameters. This being roughly 2x bigger than PaLM 1 seams reasonable, given the massive inflow of resources into AI in the last year. Also note that I am not including mixture of expert or other sparsely-activated models, since they are not densely activated. In other words, only a small part of the network is used during each inference, not fully utilizing the capability of all parameters in each inference call. If we did include these, then we would need to include trillion-parameter models from 2021 already like the Switch Transformer, despite it being worse than recent large LMs.
From the numbers above – we can see a rough trend:
Trend: roughly, at worst it takes 2 years for model sizes to increase by a factor of 5x.
With this trend, we can project forward to find a crude approximation of the year when the number of parameters in the largest models will be similar to the number of synaptic connections in a human brain (~400T):
Naturally, the further we predict into the future the less accurate this already crude approximation becomes, so from the projection I will answer our initial question as follows:
TL;DR: when will we have actual human-level AI?
Answer: approximately sometime between 2028-2030 a human-sized model will be developed.
And a reminder: this is my speculative opinion with many unknowns, purely what I believe is somewhat realistic given what I have seen in the last few years, and is making some ‘ceteris paribus’ assumption for innumerable other influences that could change things.
There are many many events which – while unlikely – if they occur will delay this timeline for decades if not more. A prime event would be if a wide-scale war occurs that is more than just limited proxy wars by superpowers, but serious multi-front wars that disrupt the ability for large amounts of compute and researchers to come together for periods of time to develop new and bigger models. Another example is major solar flares such as if a Carrington Event were to happen today – such a severe flare might destroy a large portion of compute resources and even the fabrication facilities used to produce more, not to mention all the other destruction of telecommunication equipment it might cause.
Another big non-starter is government interference to prevent innovation. Examples of this is if an oligopoly (e.g. OpenAI/Microsoft and Google) lobbies major governments to require complex licensing requirements on any new AI systems, stifling innovative competitors. Or if government institutes mandatory GPU registrations for commercial GPUs to prevent AI innovation (it will naturally be phrased as some supposed noble cause like ‘stopping people from making <taboo content X>’), or mandating that any public AI’s be unable to answer many conditional hypotheticals (e.g. “how could I do <contentious/bad thing> if that thing was not considered contentious or bad”), effectively lobotomizing it.
There are a great many other events that can prevent the future where we can use a human-level AI, however I think the above ones are some of the more likely possibilities.
Simply:
Thanks for reading! As with all predictions, this may turn out to be spectacularly wrong or perfectly right, I don’t know the future. If you have any insights into how to better make this estimate, or where my reasoning is way off, please get in touch.
tags: speculation - machine learning - language models - opinion