Episode 24: What is Artificial Intelligence?

  • Links to this episode: Spotify / Apple Podcasts
  • This transcript was generated with AI using PodcastTranscriptor.
  • Unofficial AI-generated transcripts. These may contain mistakes. Please check against the actual podcast.
  • Speakers are denoted as color names.

Transcript

[00:00:00]  Blue: The Theory of Anything Podcast could use your help. We have a small but loyal audience, and we’d like to get the word out about the podcast to others so others can enjoy it as well. To the best of our knowledge, we’re the only podcast that covers all four strands of David Deutsch’s philosophy as well as other interesting subjects. If you’re enjoying this podcast, please give us a five -star rating on Apple Podcasts. This can usually be done right inside your podcast player, or you can Google The Theory of Anything Podcast Apple or something like that. Some players have their own rating system and giving us a five -star rating on any rating system would be helpful. If you enjoy a particular episode, please consider tweeting about us or linking to us on Facebook or other social media to help get the word out. If you are interested in financially supporting the podcast, we have two ways to do that. The first is via our podcast host site, Anchor. Just go to anchor.fm -4 -strands. There’s a support button available that allows you to do reoccurring donations. If you want to make a one -time donation, go to our blog, which is 4strands.org. There is a donation button there that uses PayPal. Thank you. Welcome back to The Theory of Anything Podcast. Hey, cameo, how’s it going?

[00:01:29]  Red: Great. Bruce, how about you?

[00:01:31]  Blue: I’m doing fairly well. I’m excited about today’s episode. We’re going to talk about artificial intelligence, not artificial general intelligence, but specifically artificial intelligence, which is a subject that it’s what I’ve been studying for the last four years in school. I find it to be a really interesting subject. I know a lot of people are really only interested in artificial general intelligence. Once they realize artificial intelligence isn’t the same thing, and it’s not artificial. It’s not even a path to artificial general intelligence. They kind of lose interest in it. In this episode, I’m going to make the case that it is, in and of itself, a really interesting area of study that ties very deeply into David Deutsch’s four strands, two of the four strands, really all four of the four strands. I’m going to make the case. Therefore, it is a particularly interesting subject in and of itself, even though it’s not a path to artificial general intelligence. This presentation, you’ve actually seen once before, cameo. This is the one that you hired me to give at your company. Oh, yes. I’ve changed it some, and I made it so

[00:02:44]  Red: it’s a lot like, I mean, you were presenting it to engineers, so it was already pretty nerdy.

[00:02:58]  Blue: It was always pretty nerdy. Yeah. But then I didn’t have to worry about philosophers. So now I’m presenting to both engineers and philosophers. I see. That is literally as nerdy as it gets. Right. Yes.

[00:03:10]  Red: I think you’re right.

[00:03:12]  Blue: All right. So I do have visuals for this, but they really aren’t necessary. I’m going to explain everything so the audio version can of the podcast shouldn’t really be missing out much at all. So let’s start. Okay. Let’s talk about what is artificial intelligence. Okay. So Peter Norvig. So Peter Norvig is a huge name in artificial intelligence. That’s why I’m quoting him. The most popular textbook on artificial intelligence is called artificial intelligence, a modern approach, and it’s by Peter Norvig and Stuart Russell. And when I started studying Diverdeutsche’s Forstrands, one of the things that I became interested in was artificial intelligence. And I found that textbook. I bought it. It’s a thousand pages. I read the textbook. And by read I mean skimmed. But I went through the whole thing and I’ve got like notes all throughout the textbook. And that was what led to my interest in going back to school and studying artificial intelligence. So he is a very famous individual because of that textbook and because of his other work related to that. So he says artificial intelligence is all about what to do when you don’t know what to do. Okay. So what does that mean? By the way, that quote comes from Peter Norvig visiting my artificial intelligence class. It has video lectures. He was one of the professors in the video lectures. He’s not the main professor, but he was like visiting Dad Starner, who is my professor. Dad Starner went on to say, AI is the study of finding clever hacks to exponential problems. Okay. Now, that may not be very meaningful. It will be to you probably because you’ve gone through our podcasts included computational theory.

[00:05:07]  Blue: So you have some background in this now and you can understand why these statements have some significance and how they tie to computational theory, which is one of the four strands. So what they’re saying is, well, I’ll give a graph here in a second that explains it better. What they’re saying is, is that AI has a specific purpose and it is how do you handle a situation that has an exponential problem? Okay. So that the problem is essentially in theory, solvable, but in practice, you can’t solve it. So to address that artificial intelligence came up with a concept of intelligence as being a quote, quote, intelligence as being something they call quote, a rational agent. Now, why am I putting that in scare quotes? Well, the reason why is because their definition of intelligence and rational doesn’t match the paparian definition. I see. And is therefore in a certain sense wrong. Okay. However, as we’ll see what they came up with isn’t, it captures something important still, it captures something that’s meaningful and proved to be very fruitful because of that. Okay. So what is intelligence and for what, for that matter, what is rationality? Okay. So from a, from the standpoint of Carl Popper’s theory of knowledge, rationality is essentially from our first four podcasts. It’s going with better theories and discarding worse theories. And there’s no, and as we talked about in, in back in those episodes, there’s never any certainty. You don’t ever prove anything. You simply have the ability to measure between two theories and to say this one is better. It solves more problems. It has fewer problems, some combination of that, various other criteria that we talked about in those episodes.

[00:06:58]  Blue: And you’re able to pick which theory is better and it’s rational to go with the theories that have survived criticism rather than the ones that have criticism has caused problems for.

[00:07:10]  Unknown: Okay.

[00:07:10]  Blue: And so that’s what rationality and even intelligence actually is, but nobody really knows that. Like the artificial intelligence community unfortunately doesn’t know much about Popper. You can find references to him in artificial intelligence, modern approach, but just a few and they haven’t really caught on to the importance of the connection between Popper’s theory and what we call intelligence is in terms of human intelligence. What they did instead is they defined rationality as taking the best possible action in a situation. Okay. And that’s straight from artificial intelligence and modern approach. Okay. Page 30. Now, if you really think about that, that’s not a terrible definition of rationality. The word rationality, if I, if someone says it, they probably don’t necessarily mean Popper’s theory. They probably don’t even know Popper’s theory, right? So they probably use the word in ways in other ways that are related but not quite the same. And this is the way words are. We talked about this in our last episode that words don’t have definitive definitions behind them. That’s a kind of just not true. Instead they have uses. And if I’m using the word rational in say an economic sense, that’s very close to the artificial intelligence sense. Okay. One could criticize that when someone talks about economics as rational agents as the agent makes choices for themselves that match their values. Okay. It’s kind of a circular definition. And therefore you could, you could argue it’s not a good explanation of rationality. In many ways it doesn’t matter that it’s not a good explanation of rationality. We understand what they’re talking about in economics and thinking of it in that way allows us to work out economic theory that is true. So we’re talking about rationality in that sense.

[00:09:05]  Blue: So if, if I’m trying to solve a problem, there is in principle an optimal solution to that problem, perhaps. Okay. But maybe I can’t compute that option, optimal solution because it’s intractable. It’s an exponential problem.

[00:09:21]  Red: Yeah.

[00:09:22]  Blue: So, but we know it exists. We know there is an optimal solution. And so what we’re really looking for is a method of getting something, of getting to the optimal solution tractably or getting to something near to the optimal solution that’s good enough for our purposes. Okay. This is the definition of rationality as artificial, the artificial intelligence community defines it. Okay. And it’s important to understand in some ways, you might say the fact they’ve quote defined it wrong is why they’re on the wrong path to artificial general intelligence. And actually I think that’s true. On the other hand, they’re on the path of something really important and something really interesting that is different than the study of artificial general intelligence. It’s about how to solve problems. And that’s of use, right? Coming up with automated ways to solve problems, that’s, that’s absolutely what we want to do. Therefore, this is a science that deserves recognition all on its own. So this, this version of rationality, it’s computable. And that’s another thing that becomes really important. Since they’re trying to design algorithms, they need a definition of rationality that allows them to compute and poppers isn’t really computable at this point. We don’t understand it well enough to come up with definitive, a definitive way to put it, if we, if we could, presumably that would mean we could build an artificial general intelligence. So the fact that they’ve come up with something that they know how to start implementing, that’s important to starting to make progress. But it has little to do with human intelligence. And it’s probably more a matter of defining it in terms of computation, computationality or computability. So just to quickly review, there’s a set of problems called the NP problems.

[00:11:04]  Blue: These are the decision and search problems. This is stuff that came from our computational theory episodes, right? There’s P problems, the polynomial problems, the ones that can be solved in polynomial time and thus are considered tractable, although even polynomial algorithms could be far too slow in some cases. Then you have anything outside of that set is exponential. Then you have a set called NP complete, which is the hardest exponential problems, the ones that are universal. Okay, so there’s, there’s not a lot of hope that we will ever find a P version of those. We don’t have a proof that there doesn’t, we don’t have a proof that shows that P and NP don’t equate, but because those problems are universal and solving one would mean you could solve every problem in NP if you could come up with a polynomial time algorithm for it. Nobody really believes that those two are equated sets. They believe they’re separate sets. We’ve basically theorized they are separate sets and there is not a single counterpiece of evidence otherwise at this point. Therefore that theory has to be refuted and it’s not. It’s, it so far has survived all attempts to refute it.

[00:12:22]  Red: All right.

[00:12:22]  Blue: So that makes it a good scientific theory. So what we’re talking about for artificial intelligence then is that, that exponential set, the stuff that we know for sure is exponential. How do you solve those problems as best as possible? Okay, now this leaves a little bit of error here because machine learning isn’t really, is part of artificial intelligence, but it’s not really about exponential problems. So I’m going to say that artificial intelligence, the non machine learning kind is primarily about solving exponential problems rationally, but machine learning is really a bit right different. There’s a relationship, but they’re not quite the same. Okay. And then just for our definitional purposes, artificial intelligence of the umbrella term underneath it, there are two subsets. One is machine learning. And the other one is other kinds of artificial intelligence that aren’t machine learning. And then under machine learning, there’s supervised learning, unsupervised learning and reinforcement learning, or at least there’s probably more than that. And there’s like that my supervised learning. I mean, there’s, there’s a lot of other things, but there’s those are three main categories we’re going to talk about that encompass enough. Okay.

[00:13:36]  Red: Will you go back just one step? Because this is, I think, for a lay person, pretty much everyone just thinks when they, when they’re hearing words like artificial intelligence, they don’t know about all this stuff that’s underneath, you know, the, and, and even a little bit in technology, you’ll hear people talking about ML and they also don’t really know very much about all, you know, those subgroups underneath supervised learning, unsupervised learning and reinforcement learning.

[00:14:08]  Blue: Yes.

[00:14:09]  Red: And I do appreciate for anybody who’s just listening, the reinforcement learning has a cute little dog icon. So I love it. I’ve loved dog. Yeah.

[00:14:20]  Blue: We will do a future episode where I will go through my presentation on reinforcement learning, which you’ve also seen before. Yeah. And that is a really fascinating field for, for various reasons. And it’s the one that I’ve found the most interesting. It’s not the most productive, though. And that’s kind of in some ways, that’s what gets me excited about it is the fact that they haven’t really figured out how to make it super useful yet. There’s, there’s clearly more work to be done and big breakthroughs to come in that field. It’s also the most general of the algorithms, right? It’s, it’s, it can solve, it has more general ability to solve problems than any of the other algorithms that we’re going to, that exist in under any of the other umbrellas. As we’ll see, though, it’s not that general.

[00:15:08]  Red: Okay.

[00:15:09]  Blue: Okay. Well, you can, you can see it’s very general, but it’s far less general than it first appears.

[00:15:15]  Red: Sure. Sure. I’m like, yeah.

[00:15:17]  Blue: Okay. So let’s talk about some subcategories of artificial intelligence. Now we’re talking about the non machine learning varieties. Okay. And one of

[00:15:27]  Unknown: the

[00:15:27]  Blue: things that you’ll notice right off the bat is that they don’t seem particularly intelligent in the human sense of the word. So I kind of want to emphasize that that we’re not talking about human intelligence. We’re talking about quote, economic rationality, how to get a near optimal answer or an optimal answer. Okay. So the first one would be what’s called a search problem. So let’s imagine that you have, you’re in a city and you have a map that shows how the cities are connected and how many miles are between the different cities. And you want to get to another city and you want to take the optimal path that has the shortest number of miles.

[00:16:06]  Red: Sure.

[00:16:07]  Blue: Okay. We’ll define optimal here as the shortest number of miles. You could also define optimal in a different way. You could define it as the shortest amount of hours because not many people drive different speeds on different roads. And either would be fine. But our purposes, we’re going to say smallest number of miles. So they got the shortest path. So now one obvious way to do that is you try every path. Okay. You start at your start location. You see what happens if you go this path. You see what happens if you go that path. You keep track of it. You keep, when you finally find what your destination, you keep track of how many miles it took to get there by that path. And then you simply pick the best path. Right. And then, so now that by itself would be an AI algorithm, not a very good one, but that would be an AI algorithm. Artificial intelligence would be about the study of how to make that algorithm more optimal. Okay. So for instance, there are things you could do where you could say, well, you know, the need to search every single possible path might involve, you know, going to Tahiti, you know, you can’t really drive to Tahiti, but and be back. Right. I mean, like there’s an potentially infinite number of paths, well, not infinite, but very large number of paths possible. It would be really nice if you could use some sort of heuristic that said, you know, at this point we’ve gone so far. I’m, I’m certain that this isn’t the right path anymore. Right. I need to understand in the search. Right. So AI would be the study of how to come up with those heuristics. And

[00:17:43]  Blue: one of them, one of the most useful ones is that you could say, well, I know what the distance is between me and my destination as the crow flies. And I know what my desk, what the distance is. If I, you know, take this path and that location, what the distance is from as the crow flies, you could use, you know that the path can’t be any shorter than that. Right.

[00:18:08]  Red: Certainly. Yes.

[00:18:10]  Blue: Okay. Just the fact that lots of physics demand that you can now take that, you know, a real graph might have a negative on it, but in a physical world, you can’t have a negative distance. So you could then use that to make your search faster. Okay. You use that heuristic and you could probably just, I won’t go into the specifics, but you could imagine how you might use that heuristic to clip off searches that aren’t relevant. Okay. And this is a huge part of what we call artificial intelligence. And in fact, you’ll see that it’s kind of the basis for a lot of the other types of artificial intelligence. I’m about to explain. Okay. So another one might be non -optimal search algorithms. So they have the kind of the ones you would hear about the most would be hill climbing, simulated annealing and genetic algorithms. Like if you’ve seen the visual version, I’ve got some pictures of what those look like. Hill climbing is very literal. Sometimes it’s really hill descending, but we usually still call it hill climbing. It’s very literally like imagine that you’re on a hill. And so you decide, okay, from where I am on this hill, which direction is up? And so that’s the direction you go.

[00:19:24]  Red: Okay.

[00:19:24]  Blue: And it takes you up to a peak. Now it may not be the global maximum peak, but it will find you some sort of local peak. And then once you’re there, you can say, well, is this good enough? Okay. Okay. And then simulated annealing does something similar. It adds onto there this idea of temperature where it causes you to jump out of the local peak at higher temperatures so that you’re more likely to find the global peak. In fact, simulated annealing offers certain guarantees of finding the global maximum if you turn the temperature up slowly enough. I think the problem is that you never know how much to turn up the temperature. And so you can never really guaranteed in real life. But the reason why it’s called simulated annealing is because that’s how in physics, annealing actually works. Right. And how it finds a near optimal energy state is by doing something very similar to this. It’s been inspired by physics. Genetic algorithm is one that’s been inspired by biology. You do the search by having a population. Now genetic algorithms are different than genetic programming. So I should make a distinction there. You’re typically just dealing with a string of numbers or something that you’re trying to find the right optimal string. And you have a population of candidates and then they would breed and they would share information between them and then you might do mutations on a few. And then over time, the better candidates will survive better because they’re the ones that get bred into the next generation. And it’s very cool. I’ve actually written programs that work like this. It’s just almost strange how it just ends up working.

[00:21:10]  Blue: It just happens to through an evolutionary process, it happens to find an ideal set of strings or something close to it. And genetic algorithms are better than like hill climbing or simulated annealing because it actually learns something about the fitness landscape that it’s trying to explore. It gets kind of modeled in the population just by the way the algorithm works. Interesting. Okay. So that would be nonoptimal search algorithms. That’s a branch of artificial intelligence. Now adversarial search is another branch. Now, so far, all of these are kind of types of search. Okay. Let me want to come back to. So adversarial search is what you would be doing if you were playing a game chest with a computer or tic -tac -toe. Basically, what you do is you take the board position as it currently is, whatever it is, and you try every single possible move and you see what board position that creates. Then you say, okay, I’m going to assume my opponent is trying to maximize their score. So it then tries to find the optimal move of the opponent. And then given the optimal move of the opponent, it then says, then what’s my optimal move against that move? And it just continues to search out until it finds, if it could, it would try to find a successful strategy that wins the game. But that would be a search that no computer could do. It’d be intractable as per computational theory. So instead, it tries to go out and search as far as it can. Let’s say maybe seven moves ahead. It very quickly becomes intractable to go much beyond seven moves. What they typically do is they try to figure out which moves are more likely to be interesting.

[00:22:56]  Blue: And then they try to search just those. And then maybe it goes out seven moves. And then it has some way of evaluating the board. And it says, okay, of all my different possible moves seven generations out, this is the optimal board position I could hope for out of everything I’ve searched so far. And so I’m going to make that move that leads to that optimal board position. And that’s how a game of chess works. Now, how do you evaluate a board? A really simple not great algorithm might be something like, you know, you know how to give points for pieces on the chess board or Pondsworth one point and Queensworth 10 points or something like that. You could maybe evaluate the board based on that. That would be a very simple thing that would probably work well enough. A grand master would beat you easy if you were using that as your board evaluation algorithm. So they have to come up with better ones than that. But that is how adversarial search works. This is the min max algorithm. And this is how you do adversarial games and do artificial intelligence for them. Okay. One of the things that’s interesting is like deep blue, this belongs in the maybe the next discussion with the reinforcement learning, but deep blue it had a handmade board evaluation algorithm. So it was coded by a human. Whereas AlphaGo and AlphaChess, they actually don’t have a board evaluation coded by a human. They use machine learning to code the board evaluation algorithm, use of reinforcement learning in fact. And that’s why those are better than what a human can do.

[00:24:34]  Blue: Because using machine learning, you can come up with a better, more effective board evaluation algorithm than what a human knows how to program by hand. This is one of the main distinctions between artificial intelligence that isn’t machine learning and artificial intelligence that is machine learning. Okay. Now a constraint satisfaction problem would now would also be a kind of artificial intelligence. The easiest example of this would be solving a Sudoku problem.

[00:25:03]  Red: Okay, cool.

[00:25:04]  Blue: So a Sudoku problem, you’ve got certain numbers they think they’re supposed to add up to, you know, certain values. I don’t even remember how to play my wife loves.

[00:25:14]  Red: They have to have a one through nine in each column, each square. That’s what it is. Yeah.

[00:25:22]  Blue: And so they give you certain numbers and then you have to find numbers that will fit those constraints. So a constraint satisfaction algorithm would solve a Sudoku problem by intelligently trying things out until it found a set of numbers that fits the constraints. Sometimes Sudoku problems have one solution. Sometimes they have more than one solution.

[00:25:44]  Red: And in fact,

[00:25:45]  Blue: that sometimes affects how hard the problem actually is for a human to solve.

[00:25:50]  Red: Okay. I’m still with you.

[00:25:53]  Blue: Okay. And then logic and planning problems would be another kind of artificial intelligence. The kind of canonical example would be like the Wompus world, which you either know what that is or you don’t. It doesn’t matter. You know what? I used to have a game, a handheld electronic game back in the day. It was called Dungeons and Dragons based towards the pen and paper game. And it was actually Wompus world. I didn’t know that at the time. You were a little archer and you were going through a maze and there were pits and you could tell if you were next to a pit or not. So you had to be cautious not to fall

[00:26:26]  Red: into the pit by moving into another cell. I see. And now that I remember playing that game. Yeah, that’s Wompus world.

[00:26:32]  Blue: Okay. Okay. And then you’re trying to find the arrow and then you’re trying to find the dragon and then you need to shoot the dragon with the arrow. And you have to use logic to figure out is it safe to move in this direction? Is this the right direction which to shoot the arrow to kill the dragon? Okay.

[00:26:49]  Red: So that’s the Wompus world.

[00:26:51]  Blue: And so you have to use logic to figure out have I eliminated the bad moves and am I down to the good moves? And do I have enough evidence that I know that this is the right move or what is the probability maybe if you can’t actually determine for certain that this is the right move? Along those same lines you’ve got the block worlds. This is another one that’s very famous in artificial intelligence where you’re trying to start with block ABC and you want to stack them as ABC. What are the moves you need to make to stack them? And they’ll try to write agents that know how to do this. It’s really cool when they do it like with robotics where the robotic hand has to reach out and it has to read the block and it has to move the block and things like that but they’ll also do it virtually to make it a little easier. And then artificial intelligence has all sorts of things with just propositional logic or first order logic and solving problems with that. One of my favorite programs that I wrote was an HPLL algorithm just because I was curious about it and it would like try to figure out like you would give it a set of logical statements that you’re saying are true and then you would say is given those statements is X true and then it would figure out if X is entailed from the statements that have been given so far.

[00:28:12]  Red: I see.

[00:28:13]  Blue: And so very cool stuff. So all of these are examples of artificial intelligence. Now this next one’s a little different and so I want to dwell on it just for a second. It’s probabilistic reasoning. Okay. So I’m sure you’re familiar with probability theory. Trying to work with probability theory is a huge part of artificial intelligence. Okay. And in fact there’s this sub branch of probability of statistics and probability theory which is called Bayesian statistics. Okay. And many, many, many people like lots and lots of people think it’s a path to artificial intelligence and it’s not because it captures sort of certain intuitive sounding ideas about how humans feel like they reason about things. And so we’ll have to do like a separate episode on Bayesian reasoning and I’ll explain why it isn’t a good epistemology. It is good artificial intelligence. You can do really useful things with Bayesian reasoning. It’s not good epistemology. It’s seen as a competitor to popper but it’s actually a false competitor. Okay. So but let’s talk about just a second. So there’s a thing called a Bayes net. So Judea Pearl, I don’t know if he actually invented them or if he simply popularized them but like when you look up anything with Bayes nets, he’s quoted extensively, right? Like he just dominates. And then he’s the one who went on to do causal inference which again is probably a podcast worth doing. Okay. And causal inference actually takes things like Bayes nets and moves them to a whole new level. It’s very cool stuff that you can do with them.

[00:30:01]  Blue: So but Bayes nets, basically the idea of a Bayes net is it’s a graph, it’s a graph, a diagram, a graph that it’s a computational device that allows you to solve certain kinds of probability problems. So the example that I’ve got on the screen, which I’ll describe for the audio listeners, and this is the kind of a canonical example that comes from a textbook, is you have an alarm and your alarm goes off when there’s a burglary, but there’s also a chance it gets set off by accident if there’s an earthquake. And then you don’t get reports directly from the alarm because you’re at work. Instead you have a deal with your neighbors, John and Mary to call you if the alarm goes off. Okay. Now, John is more likely to call if the alarm actually goes off. There’s a 90 % chance that 95 % chance that John will call and a 5 % chance he won’t, but he gives false alarms, you know, 5 % of that 95 % is, is false alarms. Mary is far less likely to get false alarm. She only does it 1 % of the time, but she only calls 71 % of the time. Totally.

[00:31:09]  Red: So

[00:31:09]  Blue: what I want to know is using, you know, this is something you’ll recognize from from our podcast on statistics, is we want to know the probability and event given evidence. So in this case, what’s the probability that there actually is a burglary if John calls and Mary doesn’t call. Okay. So the Bayes net is this computational device that allows me to tractably and very quickly solve that problem and come up with what the probability is. Okay. So that I can say, okay, given the fact that John called and Mary didn’t, the chances that there is a burglary are X % and then I kind of know based on that, whether this is, I need to worry or not. Okay. And very cool what you can do with this. This is another thing that I had to study in my artificial intelligence class in school. Now, what is Bayes rule and how does it relate to this? So, and again, this deserves its own podcast. Bayes rule is basically exactly what I put up on the screen here. The probability of Y given X is equal to some alpha times the probability of X given Y times the probability of Y. What does that mean? Don’t need to worry about it. Okay. The key thing to realize is that the X and Y’s flip on the two sides of the equation. Can you see that?

[00:32:30]  Red: Okay.

[00:32:31]  Blue: That’s important. Because if you want to know the probability of a cause giving an effect, you probably don’t have any way to collect that data using just straight statistics, great statistical regular methods. But you probably do have a way to collect the data on the probability and effect given a cause to put this kind of more straightforwardly. You want to know the probability of a disease given a symptom. There’s no way to collect the data on that. Okay. Think about how you would try to go about collecting the data on that. There’s no tractable way to do it right now. Okay. Really stop and think about this for a second. You’ll see there just isn’t a good way to do it. Okay. But if you can collect the probability of a symptom given the disease, so someone who you know has the disease, what are the odds that they have that symptom. And if you know what the probability of that disease is, so how many people in the population have that disease, how rare is the disease, you can calculate the probability of the disease given the symptom. Right. So you never actually have to go collect that data directly to get it. You can derive it using Bayes rule, using things that are collectible and data that is available to you. So Bayes rule is an amazing thing. And a lot of my Paparian friends, I worry that because Bayesian epistemology is seen as a competitor to Paparian epistemology, that they end up denigrating all of Bayesian reasoning. And that’s not correct. Right. Bayesian reasoning is super important and true and correct in the domain in which it is meant to be used.

[00:34:20]  Blue: The problem with the Bayesian epistemology is that they’re trying to they’re trying to apply it to domains where it doesn’t make sense. Probabilistic reasoning isn’t limited to Bayes stuff though. Probabilistic reasoning includes like the original version of the Google car. Now today, the Google automated car would use machine learning, but back in the day, it did not use any machine learning. It used the theory of probabilistic robots, which I took a class in this. It’s a class that was in part taught by the guy who did the Google car. And it was entirely just non machine learning that he used. He used particle filters. I don’t need to describe what those are. But basically techniques that model probability distributions of where the car is. And by comparing it to what it expects to see using Google maps, say, like those cars actually had a connection through satellite to Google maps. And then using its detectors, it could say, okay, this is what I’m detecting. What are the different places I might be? And then over time, as it continues detecting, it can narrow down to this is my location. And it’s actually more accurate than GPS. GPS is seriously not are only accurate to within, you know, at least civil GPS.

[00:35:48]  Unknown: I heard military ones are better.

[00:35:50]  Blue: They don’t get you within a few feet. And if you’re going to drive a car, you can’t be off by 10 feet.

[00:35:54]  Red: Sure.

[00:35:55]  Unknown: Sure.

[00:35:56]  Red: I’m not.

[00:35:58]  Blue: Okay, so this is an overview, then, of artificial intelligence examples, kind of main examples of what artificial intelligence is. Now let’s talk about machine learning, which is the other branch. Okay. So credit to my instructor, Dr. Kira on this for the graphics I’m borrowing from his class. So machine learning’s main use isn’t actually to solve exponential problems like the rest of artificial intelligence. It’s actually to create functions that human beings don’t know how to directly program. Okay. So an example of this would be, I want to write a function that recognizes a face, your face. I want a function that recognizes cameo in a picture.

[00:36:44]  Unknown: Okay.

[00:36:44]  Blue: Okay. Now, back in the day, there is this branch of artificial intelligence. Actually, I don’t think this branch was considered part of artificial intelligence on its own, but it had associations and links into artificial intelligence. But just the study of computer vision. Okay. And there’s all sorts of techniques that exist in computer vision. I never took that class, but I did go through all the lectures because I was interested, because I intended to take it. But they have ways to like detect where a line is or detect where how to segment things. And they try to use those sorts of techniques to recognize things, maybe. Okay. Okay. You can get so far with those techniques. They did have face recognition algorithms before they introduced machine learning into it, but they didn’t work great. And humans would come, there would be humans that would spend their whole career trying to come up with what are the best ways to create features out of an image that’s going to allow me to recognize a face. And they got fairly good at it, but it was kind of mediocre results. And when machine learning, specifically when deep learning started to become popular, I can’t remember who this was, but I was to a conference and told the people who were doing things like facial recognition, he says, your jobs are all gone in five years. So it’s time to switch fields. Machine learning is going to dominate this field. And he was completely right. Nobody does the by hand approach anymore. We use deep learning to figure out how to do a face recognition algorithm. And it will give results way better than anything any human had come up with. Okay.

[00:38:36]  Blue: And that’s because we humans can recognize a face, but we don’t really understand how we do it. It’s something our brain is able to do that is beyond our current understanding.

[00:38:46]  Red: Right, right. It’s, it happens. So on autopilot, we have no comprehension of how we do it.

[00:38:54]  Blue: Yeah. And in fact, there seems to be like a module in the brain meant to recognize faces that can get disabled. And yes, certainly, there are people who can’t recognize faces because that module has become damaged and they, they, they can’t relearn it. Right. They, they, it’s once it’s gone, it seems to just be gone. At least that’s what the theory currently suggests for based on everything we’ve been able to observe so far. And so we really don’t understand how it works inside our brains. So we don’t know how to program it because we don’t fully understand it. So using machine learning is kind of this shortcut where we don’t have to understand it. We just have to understand how to create the algorithm that creates the algorithm.

[00:39:43]  Red: Sure. Okay.

[00:39:45]  Blue: Now, the images that I’m putting up here that for those who can see it, they’re kind of cool to see. And the idea is, is that what we, what we see with the image that’s on the right here is we see what inside the neural network looks like. Okay. So the neural network that is learning to recognize in this case, a yellow car, it is creating parameters as it generates the final neural network that’s going to be the one that is the, the recognition algorithm that we’re going to use. Okay. And what you can see is that what the neural network is doing is it’s learning features. It’s learning what features it needs to know. Automatically it’s discovering the features that matter in terms of recognizing what a car is. And you can, you can just, just by looking at it, you can see that the lowest level of the, of the neural network, the features are, there’s these very low level features and there’s kind of mid level features and there’s these really high level features. When you get to the high level features, a human can start to recognize what it is that the neural network is discovering. Like a wheel. Right. It’s, it’s looking for a wheel, you can tell. And it’s very cool that a deep learning algorithm searches through the space of features and finds the right features to solve the problem we want to solve that humans just don’t know how to come up with on their own. Okay. And then you can do cool things with this. Okay. You can kind of do things in a different order instead of recognizing a face. You can create something that generates faces.

[00:41:22]  Blue: So no one had ever tried to make a program. No one would have even known where to begin to make a program that generates a realistic looking photo realistic looking face that’s not a real person machine learning can do that very easily. This has led though, of course, to things like deep fakes, right? Where now we can we can actually generate an image of Barack Obama saying something that he never actually said in his voice. Right.

[00:41:52]  Red: Right. Terrifying.

[00:41:53]  Blue: We can’t trust photographs anymore. We probably shouldn’t have ever trusted them to begin with. But now we really can’t trust them. We can’t trust video. It’s possible to generate these fakes that make it look like somebody said something they never did. So everything is now suspect, which is probably good because everything probably should have been suspect from the outside. So this is what machine learning is about. Machine learning is really about generating creating functions that human beings do not know how to directly program. Okay. Okay. And this is why machine learning is such an exciting field is because it allows you to come up with programs using a machine learning approach, which my other image here kind of shows the difference between those programming you input in you put inputs into an algorithm and output comes out with machine learning. You have a bunch of data that goes into an algorithm. You have a bunch of labels. This is for supervised learning that goes into an algorithm outcomes a model. And then you use the inputs that go into the model and an output comes out. So the input might be the image of the picture. It goes into the model. The output is, Oh, this is a car. And it recognizes that it’s a car recognizes it recognizes that it cameos face.

[00:43:05]  Red: Interesting. Okay.

[00:43:06]  Blue: All right. So what are the types of machine learning there? We said there’s three types. So supervised learning, unsupervised learning, and reinforcement learning. Okay. Now, each of these, it’s interesting that the relationship between the three sub umbrellas is how much how much help it gets from a human. And this is actually a very interesting point for a number of reasons. David Deutsch, in the beginning of infinity, he talks about how we have these things called genetic algorithms. And we use them. And he kind of, he seems to be very much in doubt that they create knowledge because he feels like the human is inputting so much knowledge. He doubts that there’s any knowledge created. We’re going to see that he’s going a little too far there. But he’s kind of right, too, in that we often give machine learning the credit for everything when really there’s just this huge input, this huge injection of knowledge that comes from the human working with the algorithm. And the true algorithm that does the knowledge creation usually runs across both the human brain and the computer algorithm.

[00:44:19]  Red: Okay. Okay.

[00:44:21]  Blue: However, this distinction matters. So for supervised learning, you’re intentionally injecting in what the correct labels are that you want. So a human has to go through by hand and figure it out. So there’s an obvious injection of knowledge from the human. With unsupervised learning, there’s less obvious of an injection of knowledge from the human. With reinforcement learning, you’re trying to generate the knowledge out of some sort of signal in the environment. For the sake of describing what a signal is, imagine you’re trying to make a machine learning program that figures out when to buy and sell in the market. The signal would be how much money you’ve lost or made. I see. Okay. So you’re picking some signal that exists in the real world. And then you’re using that to inform the algorithm how to evolve itself. Okay. So supervised learning, canonical example. Imagine that we have a number of square feet and a number of bedrooms. We’ve got data on houses with a number of square feet and a number of bedrooms and what the sales prices were. Okay. So the features would be the square feet and the number of bedrooms. And the correct values, what the value we want to predict is the sales price. So I feed that information into a machine learning algorithm. Let’s say linear regression. And it comes up with a line that best matches that set of data. In this case, I’m cheating and I’m making it look like way better than it really ever would be in real life. Sure.

[00:45:54]  Blue: And then based on that, you use that line and you say, okay, given some number of square feet and bedrooms that is in a house that has yet to sell, I predict that the sales price will be this amount. Now, it’s not too hard to guess that this algorithm is not going to be a great algorithm because number of square feet and number of bedrooms probably tells you very little unless you also know your, what state you’re in or what your location is within the state, things like that. So part of machine learning is figuring out what are the right features to actually make a good prediction, right? So that you can actually make a fairly accurate prediction. And you know, typically a human figures that out. This is a case where we have the injection of knowledge coming from the human that’s subtle, that’s more subtle than the actual labels that we’re placing on things. Okay.

[00:46:43]  Red: Okay.

[00:46:44]  Blue: Now unsupervised learning would be like the image I’ve got here. Imagine just an image full of lots of dots. And I say, I want you to take all these dots, you know, they’re based on this data, whatever it is. And I want you to pull it off and make it into three clusters. And so the algorithm will say, okay, I found these three clusters that seem to be together. So they may be similar that this could then let’s say this data happened to be sales based on sales, you would suddenly find, oh, I’ve got three kinds of customers that have different needs. Or maybe you say, oh, this collection of customers buy similar things. So I can use this as a recommendation engine.

[00:47:29]  Red: Interesting. Okay. Much how Amazon. Yes, you’ve got data. Okay.

[00:47:34]  Blue: Okay. And then finally reinforcement learning. The characteristics of reinforcement learning are exploration, delayed rewards and continuous learning. So how does it compare to it’s probably easiest to try to compare it to supervise unsupervised learning. It’s similar to supervised learning in that there are rewards for good behavior. So there’s got kind of a loss function. So loss function would be, you know, how, what, you know, basically the way machine learning works is, is that you’ve got some sort of loss function that tells you how far off you are from the ideal results. And then it tries to minimize that loss function. It tries to search out which parameters are the best ones, weights in the neural network, let’s say, are the best ones that minimize the loss function. So you’ve got something similar going on with reinforcement learning via rewards. You’ve also got something similar going on with unsupervised learning in that they have no correct results to work with. There’s no labels that come from a human. Okay. Okay. But it’s not, so it’s tempting to say something like reinforcement learning is a kind of semi supervised learning, but you shouldn’t say that because semi supervised learning is an actual thing. Okay. And it’s altogether different. So semi supervised learning would be that you label some of the data, but not all of it by a human. And then you use the partial labels to resolve the problem using the data that’s not labeled also. So reinforcement learning is not that it’s really quite different than from that actually. So reinforcement learning is the algorithm that allowed us to create AlphaGo.

[00:49:06]  Blue: And so when we, when we do a future presentation on reinforcement learning, I’ll talk about how reinforcement learning was used to beat the Go Master.

[00:49:14]  Red: Okay.

[00:49:15]  Blue: It’s important to know though that Lee Sedol, who was the Go Master, the world champion of Go that AlphaGo beat, he was in, there’s a documentary on this, which is an awesome documentary, absolutely recommended. Easy. It’s free. You just go Google for it. Just look, you know, Google AlphaGo on YouTube or something.

[00:49:34]  Red: Okay.

[00:49:35]  Blue: And Lee Sedol going into that match, he was completely confident that there was not a chance a computer could beat him. And by the time he was done, he won one match out of five. And he was ecstatic that he managed to win one match because AlphaGo was so good. And you can see like when they were training AlphaGo, they got a Go Master, kind of not a Go Master, but someone really good at Go who was kind of middle ranked or whatever to come in and help. And he almost quit because he was so mad that AlphaGo could beat him because everybody knew that it was impossible to use AI to make a good Go program. Everybody knew that it was common knowledge that Go is so hard to solve for using the Minmax algorithm like chess that we talked about a moment ago, because AlphaGo has got so many possible moves that drastically change whether you’re in where you are in the fitness landscape. It was just known no computer can program can be written. And in fact, they used to say that’s going to require real intelligence, which obviously didn’t turn out to be true. But they were so convinced it was impossible to write a good Go program. So whenever these Go players got beat by AlphaGo, it was this initial embarrassment that really shocked them. And then they would start to realize, oh my gosh, this is better than a human. At least the fact that he beat it once, this is like bragging rights now. So anyhow, just kind of an interesting side point there. Okay, so now up to this point, it should be obvious that we are overlapping with one of the four strands.

[00:51:22]  Blue: And that’s computational theory. I’ve kind of really tried to emphasize that overlap. Okay.

[00:51:28]  Red: Yeah,

[00:51:29]  Blue: absolutely. But there’s actually a second more important overlap in my mind. And that is through paparion epistemology, specifically through something known as universal Darwinism. Universal Darwinism, this is a teaser for next time. So universal Darwinism is the idea that Darwinian evolution principles apply to many things outside of biology. Okay, now if you’re if you understand paparion epistemology, that’s an application of Darwinian theory outside of biology. So that was the first step of realizing that there’s this universal use for Darwinism that’s got nothing to do with biology. And now we’ve got two areas it gets applied in science in the case of paparion epistemology and in biology. So Wikipedia defines universal Darwinism as this is a quote from Wikipedia starting in the 1950s, Donald T. Campbell was the one of the first and most influential authors to revive the tradition of universal Darwinism and to formulate a generalized Darwinian algorithm directly applicable to phenomena outside of biology. Okay, okay. Now, Campbell who was just is given credit for Campbell himself never used the term universal Darwinism. He’s credit and the term exists before Campbell, but he’s given credit for starting the theory modernly. Does that make sense?

[00:52:58]  Red: Yes, yes, like like resetting it and getting people focused back on it again. Yes.

[00:53:04]  Blue: And the way it is formulated today is based on somehow rooted in his formulation, not the ones that came before him. Okay. So Campbell got his ideas primed almost entirely from Carl Popper. Okay.

[00:53:20]  Red: Why

[00:53:21]  Blue: there’s this direct tie to preparing epistemology. Okay. So Campbell, this is a quote from Campbell in evolutionary epistemology. He says it is primarily through the works of Carl Popper that a natural selection epistemology is available today. In fact, if you read that article evolutionary epistemology that I’m quoting from, like significant portions of it is, is Campbell saying, here’s what Popper said on the subject.

[00:53:46]  Red: Right, right,

[00:53:48]  Blue: right. He principally quotes Popper to start off this concept of universal Darwinism. Okay. Furthermore, Popper strongly endorsed Campbell’s work. So Popper in an article called Campbell on evolutionary theory of knowledge said Professor Campbell’s remarkable contribution shows the greatest agreement with my epistemology and what he cannot know an astonishing anticipation of some of which I have not yet published when he wrote his paper.

[00:54:18]  Red: Oh, well, that’s that’s a nice compliment.

[00:54:22]  Blue: So what I’m going to argue in the next episode is that artificial intelligence has deep ties to universal Darwinism and therefore has deep ties to preparing epistemology, even though it’s got nothing to do with scientific explanatory knowledge for the most part. This is in some way a general, so universal Darwinism could be thought of as a generalization of Popper’s epistemology. Popper’s epistemology was meant originally to be about science specifically. Universal Darwinism is about how we come up with solutions to problems. We typically would call this knowledge. There’s some debate over whether this is the right definition of knowledge or not. But solutions to problems, whether it’s through scientific explanatory knowledge or not, that would be universal Darwinism. Okay. And so artificial intelligence is the algorithmic study of universal Darwinism. I am going to make that claim.

[00:55:25]  Red: Okay. And that’s going to be our starting point for next time. Yes. Okay. I like it.

[00:55:32]  Blue: Okay. And then we will also go on to discuss a lot more about machine learning. I’ve kind of just barely touched the surface. I want to talk about how machine learning gets used in real life and kind of some of the cool things that people have done with it.

[00:55:46]  Red: Okay.

[00:55:47]  Blue: I think that sounds exciting. All right. Well, thank you for joining us, everybody. We’ll see you next time.

[00:55:55]  Red: Bye.


Links to this episode: Spotify / Apple Podcasts

Generated with AI using PodcastTranscriptor. Unofficial AI-generated transcripts. These may contain mistakes; please verify against the actual podcast.