The limit case is obvious – some people are fully illiterate so lots of AI gets more from text than they do. I'm illiterate in Chinese, for example. The OECD at their 2017 AI meeting talked about a study that showed that 70% of EU citizens can't do analyses top-end AI can 1/2
-
-
Replying to @j2bryson @GaryMarcus and
Actually, the report of the 2017
@OECD meeting I participated in (AI: Intelligent Machines, Smart Policies :-) https://www.oecd-ilibrary.org/docserver/f1a650d9-en.pdf?expires=1573939644&id=id&accname=guest&checksum=A76A1177FCC38BEE4F8E51BF870C358C … says on 11% of EU adults are above the level of AI (see picture), and cites this paper https://www.oecd-ilibrary.org/education/computers-and-the-future-of-skill-demand_9789264284395-en …pic.twitter.com/SdwZWybDGx
3 replies 0 retweets 2 likes -
Replying to @j2bryson @GaryMarcus and
What my tweet means is the fact that AI can in some sense read more than humans (certainly books remember longer & better...) doesn't actually make humans redundant to humans. 3/2
#aiethics2 replies 0 retweets 3 likes -
reading is ultimately about extracting a cognitive model of something an author writes about. books can’t do that; current AI can’t do that either. for now, only (literate) people can do that.
3 replies 15 retweets 34 likes -
Replying to @GaryMarcus @prem_k and
I partially agree. Reading (or language understanding more generally) is about doing something sensible with linguistic input. AI can do that better than people in many circumstances. I agree books don't really do anything, and that humans are special to humans, but…
1 reply 0 retweets 0 likes -
Replying to @j2bryson @GaryMarcus and
…but I don't believe that a "cognitive model" is a sensible delineation. I don't think there's one shape of model all humans & no machines use. This dichotomy is already harder than we generally acknowledge, but that this in no way undermines humans' special ethical status…
2 replies 0 retweets 4 likes -
strawperson alert. i didn’t say one model i said “models”; to some extent we each build our own, but there is also some degree of convergence.
2 replies 0 retweets 1 like -
Replying to @GaryMarcus
you said "a cognitive model" but anyway I wasn't assuming a single model, you're the one who just made the straw person. Please don't be defensive, I do respect you & honestly want to communicate.
1 reply 0 retweets 1 like -
Replying to @j2bryson
“I don't think there's one shape of model all humans & no machines use” is what i was responding to. everyone who reads The Hobbit comes away with some idea of who Bilbo is & who Gandalf is & what is unusual about the ring. Can any AI do that?
7 replies 0 retweets 2 likes -
Replying to @GaryMarcus
Not everyone can make sense of those books, but of course you are right that humans are more likely to make similar models given similar hardware phenomenology inputs contexts capacities etc. My point was that we value humans for being humans, this is an aspect of that identity.
1 reply 0 retweets 0 likes
not about similarity, it’s about not devaluing the human ability to derive cognitive models from natural language input - at all. study you mentioned conflates literacy w ability to interpret discourse; no current machine can do so. comparing illiterate humans to AI distorts.
-
-
Replying to @GaryMarcus
I don’t want to devalue humans, which is why I don’t support tying that value too tightly to abstracted computational capacities when it derives more from shared experience, which derives partly from highly specified modularised computational (and other physical) capacities.
0 replies 0 retweets 0 likesThanks. Twitter will use this to make your timeline better. UndoUndo
-
Loading seems to be taking a while.
Twitter may be over capacity or experiencing a momentary hiccup. Try again or visit Twitter Status for more information.