Written by: Stephen Hsu
Primary Source: Information Processing
Dominic Cummings begins a new series of blog posts. Highly recommended!
It’s worth noting a few “factor of a million” advances that have happened recently, largely due to physical science, applied mathematics, and engineering:
1. Destructive power of an H-bomb is a million times greater than that of conventional explosives. This advance took ~20 years.
2. Computational power (Moore’s Law) has advanced a million times over a roughly similar timescale.
3. Genome sequencing (and editing) capabilities have improved similarly, just in the 21st century.
How much have machine intelligence and AI progressed, say, in the last 20 years? If it isn’t a factor of a million (whatever that means in this context), it soon will be …
Dominic Cummings: … The big big problem we face – the world is ‘undersized and underorganised’ because of a collision between four forces: 1) our technological civilisation is inherently fragile and vulnerable to shocks, 2) the knowledge it generates is inherently dangerous, 3) our evolved instincts predispose us to aggression and misunderstanding, and 4) there is a profound mismatch between the scale and speed of destruction our knowledge can cause and the quality of individual and institutional decision-making in ‘mission critical’ political institutions …
… Politics is profoundly nonlinear. (I have written a series of blogs about complexity and prediction HERE which are useful background for those interested.) Changing the course of European history via the referendum only involved about 10 crucial people controlling ~£10^7 while its effects over ten years could be on the scale of ~10^8 – 10^9 people and ~£10^12: like many episodes in history the resources put into it are extremely nonlinear in relation to the potential branching histories it creates. Errors dealing with Germany in 1914 and 1939 were costly on the scale of ~100,000,000 (10^8) lives. If we carry on with normal human history – that is, international relations defined as out-groups competing violently – and combine this with modern technology then it is extremely likely that we will have a disaster on the scale of billions (10^9) or even all humans (~10^10). The ultimate disaster would kill about 100 times more people than our failure with Germany. Our destructive power is already much more than 100 times greater than it was then.
Even if we dodge this particular bullet there are many others lurking. New genetic engineering techniques such as CRISPR allow radical possibilities for re-engineering organisms including humans in ways thought of as science fiction only a decade ago. We will soon be able to remake human nature itself. CRISPR-enabled ‘gene drives’ enable us to make changes to the germ-line of organisms permanent such that changes spread through the entire wild population, including making species extinct on demand. Unlike nuclear weapons such technologies are not complex, expensive, and able to be kept secret for a long time. The world’s leading experts predict that people will be making them cheaply at home soon – perhaps they already are.
It is already practically possible to deploy a cheap, autonomous, and anonymous drone with facial-recognition software and a one gram shaped-charge to identify a relevant face and blow it up. Military logic is driving autonomy. …
Dangers have increased, but quality of decision making and institutions has not:
… The national institutions we have to deal with such crises are pretty similar to those that failed so spectacularly in summer 1914 yet they now face crises involving 10^2 – 10^3 times more physical destruction moving at least 10^3 times faster. The international institutions developed post-1945 (UN, EU etc) contribute little to solving the biggest problems and in many ways make them worse. These institutions fail constantly and do not – cannot – learn much.
If we keep having crises like we have experienced over the past century then this combination of problems pushes the probability of catastrophe towards ‘overwhelmingly likely’.
… Can a big jump in performance – ‘better and more powerful thinking programs for man and machine’ – somehow be systematised?
Feynman once gave a talk titled ‘There’s plenty of room at the bottom’ about the huge performance improvements possible if we could learn to do engineering at the atomic scale – what is now called nanotechnology. There is also ‘plenty of room at the top’ of political structures for huge improvements in performance. As I explained recently, the victory of the Leave campaign owed more to the fundamental dysfunction of the British Establishment than it did to any brilliance from Vote Leave. Despite having the support of practically every force with power and money in the world (including the main broadcasters) and controlling the timing and legal regulation of the referendum, they blew it. This was good if you support Leave but just how easily the whole system could be taken down should be frightening for everybody .
Creating high performance teams is obviously hard but in what ways is it really hard?
… The real obstacle is that although we can all learn and study HPTs it is extremely hard to put this learning to practical use and sustain it against all the forces of entropy that constantly operate to degrade high performance once the original people have gone. HPTs are episodic. They seem to come out of nowhere, shock people, then vanish with the rare individuals. People write about them and many talk about learning from them but in fact almost nobody ever learns from them – apart, perhaps, from those very rare people who did not need to learn – and nobody has found a method to embed this learning reliably and systematically in institutions that can maintain it. …
Latest posts by Stephen Hsu (see all)
- Precision Genomic Medicine and the UK - February 15, 2019
- The Future of Genomic Precision Medicine - February 15, 2019
- Genomic Prediction of Complex Disease Risk (bioRxiv) - January 8, 2019