Pitfalls in AI - smalsresearch.be

42
Pitfalls in AI Joachim Ganseman 20/03/2019 1

Transcript of Pitfalls in AI - smalsresearch.be

Smals, ICT for society20/03/2019
1
S m a l s R e s e a r c h
w w w . s m a l s r e s e a r c h . b e
Robotic Process Automation
Analytics
Databases
I n n o v a t i o n w i t h n e w t e c h n o l o g i e s
C o n s u l t a n c y & e x p e r t i s e
I n t e r n a l & e x t e r n a l k n o w l e d g e t r a n s f e r
Advanced Cryptography
Blockchain
S u p p o r t f o r g o i n g l i v e
Today’s menu
• From data to decision – Data collection issues (bias vs. fairness)
– Data processing issues (confounding variables)
– Goal (mis)formulation
– Adversarial examples
– (personalized) disinformation
• Defense against the Dark Arts – Transparency & explainability
– Digital Skepticism
Someone’s AI screws with you
About data
Garbage in, garbage out
• Training data is ideally
• In reality, this is rarely the case!
4
• Need lots of data quickly
Crowdsourcing?
• Q: What offer to make to a new F engineer?
• AI:
6
Hidden correlations
• We’ll fix it by not taking gender into account, right?
• … well…
– In CVs, men/women mention different things (hobbies…)
Gender as prominent confounding factor in Amazon’s model
7
• Semantic gap between
– What the AI actually derives from the training data
AI does not necessarily learn what you think it’s learning!
• Mitigations:
– Thorough (statistical) data analysis
10
– https://en.wikipedia.org/wiki/List_of_cognitive_biases
– Prostate cancer data is biased towards men
– Cervix cancer data is biased towards women
• Unfair bias can have serious consequences
– Security decisions (airport controls / inspections)
– Legal decisions (bail, parole)
– Economic decisions (insurance, mortgage)
– AI exploits bugs
– AI gets stuck in endless loops
14
• Youtube: wants you to keep watching (ads)
promotes content that “pushes buttons”
– Conspiracy theories
16
Takeaways
AI is only as reliable as the data it’s trained on.
AI maximizes its objective, nothing more.
17
• From data to decision – Data collection issues (bias vs. fairness)
– Data processing issues (confounding variables)
– Goal (mis)formulation
– Adversarial examples
– (personalized) disinformation
• Defense against the Dark Arts – Transparency & explainability
– Digital Skepticism
– Intentionally mislabeled data
20
regardless of data format
21
Takeaways
22
• From data to decision – Data collection issues (bias vs. fairness)
– Data processing issues (confounding variables)
– Goal (mis)formulation
– Adversarial examples
– (personalized) disinformation
• Defense against the Dark Arts – Transparency & explainability
– Digital Skepticism
directed at a specific individual/company
• AI can personalize the message to the target
(as in advertising)
– Verifiably false or misleading information
– Disseminated for economic gain or to intentionally deceive
– May cause public harm
– Unions, lobbying, advocacy, campaigning, …
– Selective presentation of information
– Grammar
28
Gimli was a tall and powerful man, and he had a beard and a moustache. He was also a dwarf, and he had a strong build, and he was covered in tattoos. He was not a man who looked like a hobbit.
Amplification through recommendation
– Videos about vegetarianism lead to veganism
– Videos about jogging lead to ultramarathons
• Similar on many other (free) content platforms
30
obtains a higher ranking in search results
31
obtains a higher ranking in search results
32
Takeaways
It’s still “spam vs. spamfilter”, but on a higher level.
If it sounds too good to be true,
it’s probably too good to be true.
34
• From data to decision – Data collection issues (bias vs. fairness)
– Data processing issues (confounding variables)
– Goal (mis)formulation
– Adversarial examples
– (personalized) disinformation
• Defense against the Dark Arts – Transparency & explainability
– Digital Skepticism
– Why this outcome / decision?
• Still in its infancy
38
• Awareness
– Anything you post can be used against you
• Rely on authoritative, transparent sources
– Peer-reviewed science
– Quality journalism
Requires some Competences / Literacy
• GDPR (ratified in Belgium: law of 30 July 2018)
40
– Reach and impact of social media
– Information warfare
– Culture of permanent learning
41
• 03/2015: Stratcom Task Force (euvsdisinfo.eu)
• 03/2018: recommendations of EU HLEG
• 10/2018: EU code of practice on disinformation
– Signed by Google, Facebook, Twitter, Mozilla etc.
– (Initial) choice for industry self-regulation
• 12/2018: EU action plan on disinformation
42
Takeaways
46
• For a Meaningful Artificial Intelligence (“Villani rapport”, 03/2018)
• Information Manipulation, A Challenge for Our Democracies (CAPS & IRSEM, France, 08/2018)
• EU Action Plan against Disinformation (05/12/2018)
47
Fonsnylaan 20 / Avenue Fonsny 20
1060 Brussel / 1060 Bruxelles
Fonsnylaan 20 / Avenue Fonsny 20
1060 Brussel / 1060 Bruxelles