Try   HackMD

AI Overview Session Plan

Start with Audience

I usually start with asking the audience to volunteer definitions, thoughts on AI. I ask them to tell me What AI is. I wait a few minutes then usually a few participants will start saying ideas. It's important to remind them that there is no good answer, and they can say something moderately silly. If there's no volunteers or very few, I sometimes prompt the audience (anyone thought of robots?)

I write each thought on my iPad (or a whiteboard) and then I spend time discussing each. Saying why they are good thoughts, and perhaps highlighting ways in which they don't fully capture AI.

Humans vs Machines figure

I then say 'Here is how I usually think about AI' and draw a Venn diagram. On the left, I draw the set of tasks humans can accomplish (explaining that this includes everything from filing your taxes, to solving maths problems or doing the dishes) and on the right, overlapping, I draw the set of tasks machines/computation/computers can accomplish.

I then talk about how there are examples everywhere. If I have time I ask participants to volunteer ideas of things only humans can do, or things only machines can do, etc.

I usually focus on examples within chess:

  • playing chess, in terms of finding the next move in a game, is in the intersection. I mention Kasparov vs DeepBlue, etc
  • actually picking up a chess piece and competently manipulating physical objects is something that is still a human ability. I say that you can program a robot to do it, but if you use a different kind of chessboard, or change the weight of the chess pieces, usually robots would still fail. This may change in a few years though
  • going through a very large dataset of past chess games and couting the number of times a certain pattern occurred is an example of a task computers are typically more competent at. Humans can of course do it, but it usually takes ages.
  • as an example of a task outside both sets, I mention that we don't currently have a mathematical proof whether an optimal strategy exists in chess (one that would guarantee that white is never beaten by black if white follows the strategy). Such proof exists for trivial games like TicTacToe, other games, like Otello, which is a recent example, but for chess it's beyond us.

One set is growing fast

Then I say that while human competence largely stagnated for a long time, the right-hand set of things that computers can autonomously achieve has expanded very rapidly over the past few generations/decades. I draw radial arrows pointing out from the right-hand set in the Venn diagram illustrating that it is getting inflated.

Then I say that AI is always about what happens at this expanding frontier, especially when it comes to computers gaining abilities that were previously reserved for humans. Thanks to the constant expansion, this is a moving target. Playing chess was once considered a challenging task for AI, but it is no longer surprising that chess can be played by computers. After chess, Go, was considered a difficult benchmark. This is significantly more difficult for computers (as well as for humans) to play. Yet in 2017 or so, this has fallen to computers as well. Most recently, it is the generation and understanding of human language which has been 'claimed' by computers. Subtle languange manipulation such as understaning jokes, etc, were thought to be way beyong ML capabilities just 5 years ago.

Examples of very rapid growth

I then usually try to illustrate the extremely rapid growth of AI capabilities over the last decade. A particularly visual way of doing this is through image generation. For this I have to explain what image generation, as a task, is about, which takes a bit of time. But then a comparison like this one I think is striking:

Progress-in-synthetic-face-generation-due-to-advances-in-self-supervised-generative-AI

The greyscale thing on the left labelled 2014 is the original GAN. When this paper and result came out, everyone was very surprised (now of course I can authentically say this since I was there, but this was really very exciting). Very quickly, year by year, the quality of generations has improved. And then I point out that it this progress is absolutely incredible if you consider the expansion in resolution. The thing on the left is 32 by 32 greyscale pixels (That's described by about a 1000 numbers, each of which has to be set to the right value so the whole thing shows a face). The one on the right is 1000x1000 pixels, and in colour.

I sometimes then show current examples of image and video generation. I start with thispersondoesnotexist.com which is the 2018 tech, where every time you randomly refresh, you get a new generated face. Explain that this person does not exist.

Then I show something like DALL-e, a description-based image generator.

Then I show video generation, something like SORA.

Then I show the demo of Channel1.ai of completely AI generated news anchors, go to the time in the video when it demonstrates talking to completely different languages smoothly. I explain how human speech generation is another breakthrough, and now NotebookLM can generate a 20 minute podcast episode that sounds incredibly real about any topic you ask. ChatGPT advanced speech mode can fluently speak in multiple languages, and it can even do various accents in English, like a fake French accent.

General AI

Depending on where the conversation is going, I might talk about AGI and ASI.

When talking about the expanding capabilities of computers, you can simply pose the question whether and when will computers be able to accomplish any task that a human can. Say that something like this is called AGI. And if computers become much smarter than the smartest human, that is called ASI. I usually don't go into that.

Another way to illustrate AGI is pointing out that up until recently, AI worked very differently from human intelligence. Human intelligence usually consists of a broad range of core competences, that all or most humans are able to do (learn languages, walk, do the dishes), while each human may specialise in some tasks that they will be good at. You can draw an individual human's competences as a set consisting of this core, and some specialisations. This will look a bit like an amoeba, with a fat core, and some arms going into specialities. For example I say that beyond core competences, I know more than most people about coffee, machine learning research, and a little bit of basketball. Then I draw a different person, who overlaps with me on the core, is much better at basketball, knows nothing about coffee or machine learning research, but is also good at something else I don't know anything about. This is a good opportunity to put in a joke.

Up until recently, AI worked very differently, Each "AI" was a narrow AI with a very limited scope. One AI could be developed to detect pedestrians, a different one to fold proteins, another one to translate from English to Hungarian, another one to play go, etc. You draw this as lots of small bubbles, each representing a different program or model. You can explain that no matter how smart AlphaGo became, and how much better it was at Go than the best humans, it wasn't able to do anything else. Say this is narrow AI.

Current AI tools, especially large language models have changed this picture. These monolithic models, also called foundation models, show remarkable generality. You train a single model, a LLM, and it can than carry out lots of tasks, such as writinig your boring homework, doing your taxes, recommending you movies to watch, explaining you a topic you're interested in, write computer code, whatever.

If you draw language models, they are no longer isolated bubbles, they are a little bit more like human intelligence, in that they, too, share a set of core competences, and this overlaps quite significantly with some of those core competences that we thought were uniquely human. This is why some of these new technologies might feel a little scarier, this is why we are starting to anthropomorphise them.