Open your eyes – your AI is biased

Computations have no ethical subroutine. And understanding bias in AI is an important eye opener. Building an AI-facilitated future without properly understanding the algorithms behind the conclusions and actions is leading us to into unexpected pitfalls.

We are all very excited about machine learning and AIs. We see them as the ultimate way of automating daily life from driverless cars to personal health and medical diagnostics. But garbage in = garbage out. And to eliminate the garbage we need to be able to identify it. Long after our little helper has started working.

The main reason we need to watch out is that AI algorithms are not necessarily retraceable and retrackable. Not even the programmer understands it fully once the machine starts accumulating and filtering data. Despite its ability to learn it can only conclude based on the original assumptions built into the underlying algorithms.

vlcsnap-2011-01-17-16h21m27s224-1024x422

Looking for the ultimate answer

Whenever we rely on algorithms to make decisions – or at least recommendations – it is because we seek a simple answer to a complex question.

If the collection of big data for data driven decision making is used to create simple answers to complex questions, the complexity is solved through algorithms that in effect filter and collate based on what the human programmer considered applicable. And it concerns us more than you would expect. I recently read that the AI concept is being used within the US judicial system: Judges rely on the AI’s suggestion on whether an inmate should be granted parole based on assumptions of future behavior of set individual after release. In isolation this would seem like a statistically viable method, as there will be vast amount of available data to substantiate the conclusion.

But if the original algorithms input by a human were in fact influenced by bias such as race, name, gender, age etc., are the conclusions any better than the answer 42?

When Douglas Adams in his science-fiction Hitchiker’s Guide to the Galaxy series introduced Deep Thought, the biggest  computer ever built in the history of men and mice, the builders asked for the answer – and added that they wanted it to be nice and simple. So after millions of years Deep Thought concluded that the answer to life the universe and everything was 42. But by now, this insight was useless because nobody really understood the question.

If we see AI and machine learning as the ultimate answer to complex scenarios, then we must be able to go back to the original question in order to be able to process the answer. Not just to understand but to analyse and apply what the computer is missing – the ethical subroutine.

What will the AI choose in a no-win scenario?

One of the hot topics in the current discussion around self driving cars is whether the AI would make proper ethical decisions in a no-win scenario. Should it risk the life of the passenger by veering off the street and over a cliff to avoid running over another individual in in the street? The decision would be entirely based on the original algorithms which overtime have become inscrutable even for the engineers themselves.

Of course, this is a simplified example. An AI, as opposed to a human behind the wheel, would be able to process more details regarding the potential outcome of either option. What would the statistical probability of successfully avoiding hitting the person on the street be when taking into account elements such as speed, space available without going over the cliff, the chance of the person acknowledging the danger and moving out-of-the-way in time before the collision etc.

self-driving-car-drive-ix-625-u

(Image from Nvidia Marketing Material)

But the self preservation instinct of a human being behind the wheel would most likely lead to the obvious conclusion: Hitting the person is preferable to dying by plunging over the cliff! Would the original programmer not have input exactly this type of bias?

What I believe Douglas Adams was getting at with the magic number 42 was that there is no simple answer to complex questions. If as indicated above the AI is victim to its own programming when making complex decisions or recommendations, then as a tool we must make it as transparent and thereby manageable as any tool developed by humans since the invention of the wheel.

MIT Technology Review addressed this in detail in the article published by Will Knight in April 2017  The Dark Secret at the Heart of AI 

No one really knows how the most advanced algorithms do what they do. That could be a problem.”

willKnightHe goes on to explain that while mathematical models are being used to make life changing decisions such as who gets parole, who gets a loan in the bank, or who gets hired for a job, it remains possible to understand the reasoning. But when it comes to what Knight calls Deep Learning or machine learning, the complexity increases and the continuosly evolving program eventually becomes impossible to backtrack even for the engineer who built it.

Despite the inscrutable nature of the mechanisms that lead to the decisions made by the AI, we are all too happy to plunge in with our eyes closed.

Later the same year another MIT Technology Review article explores the results of a study of the algorithms behind COMPAS (Inspecting Algorithms for Bias ) COMPAS is a risk assessment software which is being used to forecast which criminals are most likely to reoffend.

Without going into detail – I highly recommend you read the article – the conclusion was that there was a clear bias towards blacks. The conclusions later turned out to be incorrect assumptions: Blacks were expected to more frequently reoffend, but in reality did not. And vice versa for the white released prisoners.

The author of the article, German journalist Matthias Spielkamp, is one of the founders of the non-profit AlgorithmWatch which has taken up the mission to watch and explain the effects of algorithmic decision making processes on human behaviour and to point out ethical conflicts.

Spielkamp

Mattias Spielkamp, Founder of AlgorithmWatch

The proverbial top of the iceberg

Even strong advocates of applying artifical intelligence/cognitive intelligence and machine-learning (deep learning) to everyday life applications, such as IBM with its Watson project, are aware of this threat and use strong words such as mitigation to explain how this potential outcome of widespread use of the technology can be handled better.

In a very recent article published February 2018 entitled  Mitigating Bias in AI models , Ruchir Purri, Chief Architect and IBM Fellow, IBM Watson and Cloud Platform stresses that “AI systems are only as effective as the data they are trained on. Bad training data can lead to higher error rates and biased decision making, even when the underlying model is sound… Continually striving to identify and mitigate bias is absolutely essential to building trust and ensuring that these transformative technologies will have a net positive impact on society.”

IBM is undertaking a long range of measures to minimize bias but this is only addressing the top of the iceberg. The real challenge is that we are increasingly dehumanizing complex decisions by relying on algorithms that are too clever for their own good.

Actually – all of this isn’t exactly news.

More than 20 years ago, human bias was already identifed as an important aspect of computer programming

“As early as 1996, Batya Friedman and Helen Nissenbaum developed a typology of bias in computer systems that described the various ways human bias can be built into machine processes: “Bias can enter a [computer] system either through the explicit and conscious efforts of individuals or institutions, or implicitly and unconsciously, even in spite of the best of intentions”.  (Source:  Ethics and Algorithmic Processes for decision making and decision support )

Beating the Kobayashi Maru – or the human vs machine experiment with Watson

“I don’t believe in no-win scenarios.” (James T. Kirk, Starship Captain)

When you are a strong believer in datadriven decision making, building strategies on real insights, and always sticking to facts rather than fiction – it’s a hard blow when one of the world’s leading artificial intelligence systems tells you, that you are not a nice person. It’s based on data – so it’s a fact.

Many industry leaders have evangelists who are excellent presenters and subject matter experts. It’s always a privilege when you get a chance to interview an evangelist. I met IBM’s Rashik Parmar, Watson evangelist, at IPExpo Nordic a few weeks ago.

WatsonDescription.png

Source: IBM

There is so much potential for big data analysis and the learnings and insights we gain, from combining the many available sources of accssible data to draw new conclusions and find answers. That’s basically what Watson does. And then makes the logical connections. Simply put.

IBM developed a small demo engine that would analyse your Twitter personality and generate those awesome charts we all love; and few of us know how to interpret. It was reassuring to see what a nice guy President Obama is on Twitter. And my friend, Rashik, had a similar profile – so all good.

Potus.png

Source: IBM’s demo app TweetMeWatson

Lucky for me, we couldn’t make it work for my profile until I got home. When I ran it, I found out I was

“Inconsiderate and a bit shrewd”

I will spare you the rest. Either I am very delusional about how I express myself, or there was something fishy going on here. But it’s based on data! It has to be true!

Before digging a hole in the garden to hide from the world – or the equivalent: deleting my Twitter account – I decided to think it through. What was Watson picking up on, what were the actual parameters used?

The Big Five (FFM) Personality Traits

Watson is grouping our personalities according to the Five Factor Model (FFM) Wikipedia explains:

The Big Five personality traits, also known as the five factor model (FFM), is a model based on common language descriptors of personality (lexical hypothesis). These descriptors are grouped together using a statistical technique called factor analysis (i.e. this model is not based on experiments).

This widely examined theory suggests five broad dimensions used by some psychologists to describe the human personality and psyche.[1][2] The five factors have been defined as openness to experienceconscientiousnessextraversionagreeableness, and neuroticism, often listed under the acronyms OCEAN or CANOE. Beneath each proposed global factor, a number of correlated and more specific primary factors are claimed. For example, extraversion is said to include such related qualities as gregariousness, assertiveness, excitement seeking, warmth, activity, and positive emotions.

220px-francis_galton_1850s It all sounds very reassuring, the term “Lexical Hypothesis” makes sense –  it was analysing words. This is a principle which was developed by British and German psychologists to identify a personality characteristic. It was used to determine risk of mental illness or criminal behaviour. Invented in 1884, by the way, by Sir Fancis Galton – a stern looking fellow.

But something as elusive and intangible as the human mind is so very hard to classify and illustrate in data points and charts. By creating a lexicon of words and adjectives that at the time were considered to be indicators for certain behaviours, they provided a tool to build profiles – and categorise people based by their choice of words.

Note that the method has also received a lot of criticism – many of them quite reassuring when you are on the receiving end of this exercise. Read more here. 

Phew – that means I can still be a nice person, just not when I tweet. Or speak.

It seemed safe to climb back out of the hole in the garden and meet the world face on. But knowing now what triggered my unpleasant profile, I decided to challenge Watson to a duel.

A duelling experiment

kobayashi-maru-02

@echrexperiment is the experimental Twitter profile where tweets were worded more carefully, where people and followers were thanked and nothing bad was happening in the world. No politics, no injustice, no gender inequality, no discrimination. And lots of cats.

exhrexperiment.png

 

After three weeks, I was a much nicer person. The traits that I seem to be exploiting negatively in my original profile are now contributing to a positive image.

Suddenly, uncompromising was a good thing.

Spock.png

“In academic vernacular, you cheated”

Like Captain James T. Kirk in Star Trek challenging Mr. Spock’s designed program, I cheated to win.

Most of my tweets were carefully drafted using positive semantics but remaining true to my usual topics of interest. I was not trying to be someone else, just focusing on being nice. Here’s a list of the parameters I introduced to make Watson love me more:

  • Following back – anyone who followed me, unless an obvious business account or egghead, was followed back as soon as I spotted them.
  • #FF – sometimes I used the FollowFriday hashtag to thank select people. It generated some nice interactions even between those mentioned, so I grouped them into categories – e.g. Danes, analysts, etc.
  • I thanked, and loved, and “awesome’d” and “great’ed” a lot.
  • Sharing – giving credit, not taking it. I always mentioned the source or the account where I had picked up a link.
  • Sharing the love – retweets were focused on positive news, positive sentiments and uplifting current events. I also checked the wording of the original tweet before RT’ing to avoid contamination of my positivity.
  • Getting personal – my personality and emotions were conveyed more by sharing private interests such as books, cats, travel and science fiction.
  • Language Disclaimer – all of the above choices were based on my non-native perception of the English language, and may have been different from Webster’s Dictionary which is the basic semantic interpreter used in lexical hypothesis.

What I didn’t do

Humour doesn’t travel well, so any jokes, irony, satire and cartoons were not part of echrexperiment. I may have gotten carried away occasionally, but consciously tried to avoid it.

Politics are a powerful emotional trigger, so I avoided RT’ing or engaging in conversations with political statements. That wasn’t the mission.

Automation is a powerful tool to increase the quantity of your social media posts, but with automation things like timing and engagement suffer. Sometimes, due to other news, automation may even lead to displaying insensitivity.

Automatic response is a convenient way to further promote your services and invite people to connect. But it just isn’t personal. Despite all these lovely people addressing me by name. I did not send messages to thank people for the follow, but I checked their profile and retweeted where I could to show my appreciation.

What Watson had to say about @echrexperiment

The app itself produces a lot of detail as you can see from above. Below I grouped the result into more familiar charts to share some highlights. To make sure I picked a really nice person as control, I chose President Obama’s Twitter @potus. But please remember – it’s probably mostly his staff tweeting. And they seem to have done an excellent job.

watson1

Obama – it appears – is very agreeable on Twitter, and my experimental lovely/nicey/catsy account matches this impression very nicely. We are both very open, although I am lagging on conscientousness, but hey –  I am not the President.

Digging deeper into selected parameters, revealed some interesting characteristics related to being a President or just trying to be a nice person.

We can all agree that values should be an important parameter if you are President of the United States. Strangely enough Obama wasn’t all that keen on change, and more inclined to be conservative. For self enhancement … we have identified the villain – the one parameter that makes my original Twitter account so repugnant. I leave the graph to stand on its own.

Meanwhile, President Obama scored a resounding Zero on self-enhancement – but he made it to the top already.

watson2

President Obama’s most distinguishing need is the need for structure. Love – it seems – he gets a plenty.

On the other hand, my original self seems to have enough structure in her life.

Watson3.png

But here’s the greatest insight from this entire exercise – other than confirming that it is possible to change who you are, or rather how you are perceived:

When it comes to curiosity, all you need to do is be a positive tweeter and include lots of cats.

 

IMG_0042.JPG

 

What really happened at IPExpo Nordic Sept 2016

Cloud or not Cloud is no longer a discussion we need to have at IT trade shows – for most companies today, it’s by default. The discussion is not about the technology platform, it revolves around the transformation which cloud has enabled.

IP Expo is a trade show  on Cloud and Infrastructure, Cyber Security, Datacentre, Data Analytics and Developers. Vendors use it as a platform to show off new technology, and delegates to educate themselves with training sessions and workshops – and some stellar keynotes. It’s held held annually in the UK, but this year for the first time in Stockholm as a pan-Nordic event.

Overall, there was a gap between the vendors’ view of the IT shop floor and how we work, and the visions and experiences of the  Nordic IT professionals listening to thshowflooreir pitches.

 

Among the 52 exhibitors, there very few were outside the box, so to speak. Mostly, they were addressing singular IT challenges such as cyber security, infrastructure or offering access to the “Cloud” ( eh, what!?).  I am not saying it’s not important, just saying that we need to maintain a broader perspective.

 

You can check out some of the exhibitors and their messages and posts on the IP Expo Nordic Facebook page

Transform – don’t wait for digital, it’s already here. And bring your own lunch

Of course, as the conference program is very congested – no scheduled lunch or coffee break (a NO-NO in Sweden, organizers!) – you cannot possibly get the complete picture if you are not a machine and can sustain yourself on battery energy alone. So naturally, I did not see all the aspects of the messaging and content delivered at the sessions, but delegates I spoke with agreed that from their perspective there was really very little news. As a learning experience, it did not quite deliver.

Out of 81 speakers, only 9 were women (!)

The non-profit organisation Womengineer with their programs to engage large corporations in empowering more women in engineering both in education, trainee programs and in careers was given a small corner on the trade show. Luckily, they also had their own session which was extremely well attended, despite the more glamourous main programme in the auditoriums. Similar to the worldwide efforts to inspire girls to code, Womengineer holds Introduce a Girl to Engineering Days – mark your calendars for the next event on March 17, 2017.

Aroshine Munasinghe, Head of Business Relations at Womengineer and Jenny Stenström, blogger had checked out the gender balance  in the conference program – and including themselves, there were 9 female presenters out of 81. Quite unusual for a high profile conference in Sweden.

And I encountered many professional women in the various sessions.

 

Barbará and Carla from Lisbon, Portugal and were extremely satisfied with the content of the sessions and the high technical standard of many of the presenters. Among their personal highlights was Susanne Fuglsang in the Digital Transformation Panel who literally took center stage in challenging a lot of preconceptions on what digital susannefuglsangtransformation is about.

 We should stop speaking about digital transformation – it has already happened. We should focus on making the transformation a strategic objective in top management and more importantly middle management where resistence to change is prevalent.

(Susanne Fuglsang, Executive Producer, Another Tomorrow)

 

Len Padilla of NTT echoed this very well in his Digital Transformation session tugged away in the basement and coinciding with Trend Guru Alexander Bard’s glamourous keynote. Happy to say, we were at least 25-30 people who were brave enough to go against the celebrity flow and took some practical advice in the dungeon, as Len called it.

Not only do you need to involve more levels of your organisation in the process, you need to enable those managers who’s job it is to “keep the lights on”. They are the least willing to take risks, so to truly transform you must learn to encourage and reward risk taking and the consequent potential failure. (Len Padilla, NTT)

Smarter citizens – not devices

What IPExpo Nordic showed, was the obvious focus these large players and other software vendors have on the Nordics as a market, and the effort they are making to gain momentum with their key messages.

The keynotes from Microsoft, IBM and Amazon Web Services were all delivered by some of the company’s top speakers and all on what PR professionals call “on message”.  Gleefully interrupted by a worldclass presentation on how the CIO of the City of Stockholm, Ann Hellenius, and her team will make Stockholm the world’s smartest city.

How? Not by adding infrastructure: We have that already with Fiber Optics Bands reaching 30 times around the globe covering the greater Stockholm area.

No, by defining a smart city as the merger of human, financial and technological interactions to achieve the highest quality of life and the best environment for business. The main focus here is the interaction between citizens and services, between human and machine.

stockholmvision

The best digitalisation is when you do not notice that it’s there (Ann Hellenius, Stockholm City CIO)

Making sense of the data – the Big data

IBM’s visionary Rashik Parmar  addresses similar thoughts on the concept of smarter cities– but at IP Expo Nordic, illustrating that the focus of the presentations was not on the vision but on the product, he showcased IBM’s cognitive intelligence Watson instead.

(The program also had an encounter with Furhat the Robot Head, but alas, Furhat and I are yet to meet in person, so to speak).

The potential of this technology of course is enormous, and the Nordic audience, albeit not new to the concept, can relate to the potential of using cognitive intelligence to make the world a better place.

Some of us covering IT innovation, gadgets and business opportunities combined, are always looking for the killer app. For Watson and alike there are many, and Rashik Parmar introduced just one: using Watson analytics to “listen” for cracked wheels on long distance trains.

This is a really important job that used to be a manual one, with a railroad worker hitting with a hammer on each and every wheel at each stop to listen for cracks. According to Rashik Parmar, the ability to filter out everything but the sound defining a crack is unique to the Watson technology – the sheer amount of data involved and the processing required to make this a seamless process and not create delays is unique (at this stage) to IBM’s technology. And this saves lives.

No more -aaS abbreviations, let’s all code

Microsoft’s James Staten, Chief Strategist Cloud and Enterprise Division (The Era of IaaS is Coming to an End) warmly enthused about the hybrid cloud being the future and where Microsoft will continue to invest heavily: The combination of on premise computing and public cloud as a strategy.

jamesandpatrick

What Microsoft has probably realised is that they need a broader audience for their products and concepts in order to meet today’s real customers – the so-called citizen developers  as defined by Gartner.

Basically, it means that end users are taking over the world of software applications because everyone today can learn to and uses code. And Microsoft wants to help IT departments stay in control via Azure Security Center, among other concepts. Read more on the visions and strategies for Cloud in my exclusive interview with James Staten in my next blog post due in a few days.

Listening to Amazon Web Services Technical Evangelist Danilo Poccia was perhaps not as inspiring for leaps of thought, but all the same very useful and constructive. The Amazon Journey to the Cloud was just that – his step by step introduction to and presentation of the various services in the Amazon Web Services portfolio addressing today’s needs among developers and IT departments.

img_0956

The journey as a strategy is what captured the interest of one of the delegates I asked right after the lights were turned on. Per Nordahl, IT Strategy Manager at Telia, found that Amazon could focus more on the journey when walking through the services as a case – many larger enterprises can relate to this story of digital transformation.

 

What is innovation – can we capture this elusive pimpernell

Amazon defines innovation as (f = mechanism + culture).

Unfortunately I could not reach Danilo Poccia to ask how to attach some real metrics to this. But I think an equation like that could help many businesses quantify and qualify their rate of innovation ultimately to get more senior management buy-in for those crazy ideas we all know we need to stay ahead of the game in a digital world.