Three Rules for Banking Bots
- Chris Skinner , Chairman at Financial Services Club
- 13.04.2017 12:15 pm Banking Bots , Chris Skinner is known as an independent commentator on the financial markets and fintech through his blog, the Finanser.com, as author of the bestselling bookDigital Bank and its new sequel ValueWeb. In his day job, he is Chair of the European networking forum: the Financial Services Club. He is on the Advisory Boards of many companies including Innovate Finance, Moven and Meniga, and has been voted one of the most influential people in banking by The Financial Brand (as well as one of the best blogs), a FinTech Titan (Next Bank), one of the Fintech Leaders you need to follow (City AM, Deluxe and Jax Finance), one of the Top 5 most influential people on BankInfoSecurity’s list of information security leaders, as well as one of the Top 40 most influential people in financial technology by theWall Street Journal’s Financial News.
Writing more and more about chatbots, robots and AI, I can see a day coming in the not too distant future where we won’t be able to tell the difference between the human and the machine. That’s not scary science fiction, as it’s now almost science fact. However, there is still a long way to go as demonstrated by IBM’s Watson Avatar.
Their AI human agent – usually a woman, as all the folks creating chatbots are sexist – looks humanish, but no avatar or chatbot today feels properly human. In IBM Watson’s case, it’s the mouth movements that give away she’s a machine, even though she’s in HD. Fifteen years ago, AT&T were working on exactly the same ideas, but with scripts for avatars.
As you can see, the idea – and sexism – hasn’t changed much in that time. What is changing however, is the technology behind this idea and, as with all great technological innovation, if something looks worthwhile to develop – biometrics, communicators, health technologies, life sciences, artificial intelligence, robots and so on – then eventually these technologies will develop sufficiently to become mainstream and acceptable.
That will be somewhere over the next 10-25 years to achieve. It feels nearer, but when a group of experts on Artificial General Intelligence were asked this question in 2012, their view was that it would not be achieved until 2040. The question here being: When will achieve the Singularity? From Wikipedia:
The technological singularity (also, simply, the singularity) is the hypothesis that the invention of artificial superintelligence will abruptly trigger runaway technological growth, resulting in unfathomable changes to human civilization. According to this hypothesis, an upgradable intelligent agent (such as a computer running software-based artificial general intelligence) would enter a ‘runaway reaction’ of self-improvement cycles, with each new and more intelligent generation appearing more and more rapidly, causing an intelligence explosion and resulting in a powerful superintelligence that would, qualitatively, far surpass all human intelligence. John von Neumann first uses the term “singularity” (c. 1950s), in the context of technological progress causing accelerating change: “The accelerating progress of technology and changes in the mode of human life, give the appearance of approaching some essential singularity in the history of the race beyond which human affairs, as we know them, can not continue”. Subsequent authors have echoed this viewpoint. I. J. Good‘s “intelligence explosion”, predicted that a future superintelligence would trigger a singularity. Science fiction author Vernor Vinge said in his 1993 essay The Coming Technological Singularity that this would signal the end of the human era, as the new superintelligence would continue to upgrade itself and would advance technologically at an incomprehensible rate.
As I thought about these things, it made me think along the lines of other science fiction authors and visionaries, such as Isaac Asimov. An amazing thinker about the future, he wrote a fantastic series about robots, and came up with the three laws of robotics:
- A robot may not injure a human being or, through inaction, allow a human being to come to harm.
- A robot must obey orders given it by human beings except where such orders would conflict with the First Law.
- A robot must protect its own existence as long as such protection does not conflict with the First or Second Law.
This makes so much sense, and has been used in many other places, notably in the film Robocop who has three prime directives:
- Serve the public trust
- Protect the innocent
- Uphold the law
This means that Robocop can kill humans, but only if it conforms to the above.
It made me wonder what the three rules for banking robots should be? We need to have some, otherwise they may run away with our money. After all, robots won’t adhere to national laws and rules, only the way they program and reprogram themselves. So, here’s my speculative take on three rules for banking robots:
Do not rip customers off
Ensure you are secure
Freeze any activity that breaks the law
I think they speak for themselves, but happy to take comments or questions.
Finally, if you missed it, this was my favourite recent article about robots: