BI 201 Rajesh Rao: From Predictive Coding to Brain Co-Processors

December 18, 2024 01:37:22
BI 201 Rajesh Rao: From Predictive Coding to Brain Co-Processors
Brain Inspired
BI 201 Rajesh Rao: From Predictive Coding to Brain Co-Processors

Dec 18 2024 | 01:37:22

/

Show Notes

Support the show to get full episodes, full archive, and join the Discord community.

Today I'm in conversation with Rajesh Rao, a distinguished professor of computer science and engineering at the University of Washington, where he also co-directs the Center for Neurotechnology. Back in 1999, Raj and Dana Ballard published what became quite a famous paper, which proposed how predictive coding might be implemented in brains. What is predictive coding, you may be wondering? It's roughly the idea that your brain is constantly predicting incoming sensory signals, and it generates that prediction as a top-down signal that meets the bottom-up sensory signals. Then the brain computes a difference between the prediction and the actual sensory input, and that difference is sent back up to the "top" where the brain then updates its internal model to make better future predictions.

So that was 25 years ago, and it was focused on how the brain handles sensory information. But Raj just recently published an update to the predictive coding framework, one that incorporates actions and perception, suggests how it might be implemented in the cortex - specifically which cortical layers do what - something he calls "Active predictive coding." So we discuss that new proposal, we also talk about his engineering work on brain-computer interface technologies, like BrainNet, which basically connects two brains together, and like neural co-processors, which use an artificial neural network as a prosthetic that can do things like enhance memories, optimize learning, and help restore brain function after strokes, for example. Finally, we discuss Raj's interest and work on deciphering an ancient Indian text, the mysterious Indus script.

Read the transcript.

0:00 - Intro 7:40 - Predictive coding origins 16:14 - Early appreciation of recurrence 17:08 - Prediction as a general theory of the brain 18:38 - Rao and Ballard 1999 26:32 - Prediction as a general theory of the brain 33:24 - Perception vs action 33:28 - Active predictive coding 45:04 - Evolving to augment our brains 53:03 - BrainNet 57:12 - Neural co-processors 1:11:19 - Decoding the Indus Script 1:20:18 - Transformer models relation to active predictive coding

View Full Transcript

Episode Transcript

[00:00:03] Speaker A: The goal of the brain, at least one goal of the brain is to learn a generative model of the world, an internal model of the world. And it's constantly making these predictions or doing essentially hypotheses testing, right? So it's essentially generating hypotheses and it's trying to then match those hypotheses is what's coming in through the senses. And when there are deviations or mismatches, those are called prediction errors and those errors are what are used to then update the hypothes. If you want to have a very rich way of modeling the world, then what you want to do is have the higher level modulate the dynamics of the lower level. So you need to change the function being computed at the lower level depending on the task. And so we said, okay, why don't we use these primitive modes of interacting with the brain, so EEG and tms and let's build a brain to brain, direct brain to brain communication system with them. [00:01:06] Speaker B: This is brain inspired, powered by the transmitter. Hello, I am Paul. Today I am in conversation with Rajesh Rao. Rajesh is a distinguished professor of computer science and engineering at the University of Washington where he also co directs the center for neurotechnology. Back in 1999, Raj and Dana Ballard published what became quite a famous paper which proposed how predictive coding might be implemented in brains. What is predictive coding? You might be wondering. It is roughly the idea that your brain is constantly predicting, or at least back then it was roughly the idea that your brain is constantly predicting incoming sensory signals and it generates that prediction as a top down signal that meets the bottom up sensory signals. And when those two signals meet the top down prediction with the bottom up sensory input, the brain computes a difference between the prediction and the actual sensory input. And that difference, that error is sent back up to the top where the brain then updates its own internal model to make better future predictions. So that was 25 years ago and it was focused on how the brain handles sensory information. But Raj just recently published and proposed an update to the predictive coding framework. One that incorporates perception and action, suggesting how it might be implemented in the cortex. Specifically which cortical layers do what, something that he calls active predictive coding. So we discussed that new proposal. We also talk about his engineering work on brain computer interface technologies like Brainnet, which basically connects two brains together and like neural co processors which use an artificial neural network as a prosthetic essentially that can do things like enhance memories, optimize learning and help restore brain function. After Strokes, for example. Finally, we discuss Raj's interest and work on deciphering an ancient Indian text, the mysterious Indus script. All right, I link to all of the work that we discuss in the show [email protected] podcast201. 201. I'm in the process of setting up our next live discussion for Patreon supporters. So if you're a Patreon, check your Patreon messages soon or check in on the Brain inspired discord. Thank you for listening. Here's Rajesh. Raj, by the way, this is the first long motorcycle ride I've taken since I got a motorcycle. And I'm in a hotel, so I may not be on the top of my game. Not that there is a top of my game, but anyway, you look great. [00:04:12] Speaker A: And yeah, I think, yeah, I think it should be, you know, interesting to see what kind of interesting questions come out of done that long motorcycle ride. [00:04:21] Speaker B: Well, I actually wanted to start with, well, maybe you won't tell me this. What year were you born? [00:04:29] Speaker A: Oh, I mean, it's on the Internet, right? It's on my wikipedia page. It's 1970. [00:04:33] Speaker B: 70. Okay, so when it was the 60s and you're already good at academia, perhaps before you were born. I understand that you excelled in early years in your schooling, but here's what I want to get to. I would like you to tell me the origins of the predictive coding story. So that's what I. Yeah, when it was in the 60s and you were. Before you were born and you were thinking about predictive coding, how did, how did you start thinking about that? [00:05:09] Speaker A: That's right. Yeah. It was in my previous, you know, life. Right. As a. Been reincarnated now to bring up predictive coding. No, no. So basically it was, you know, the. The story is essentially that I grew up in India, in Hyderabad, and did my high school there. And I got a chance to come to the U.S. as a, you know, 11th grader because I got a rank in the Indian central science exam that they have, and as part of that rank, they were able to send me with another group of people to do some research. Basically, I got hooked onto research, right, in 11th grade, went to University of Maryland, did some research. I would call that. Put that in codes research and superconductivity. It was the time of high temperature, super connectivity at the University of Maryland, and I got a taste of research then, and then, you know, took the SAT exam. Never applied to any universities. I was just waiting for universities to contact me, thinking if my SAT score is high enough, somebody will contact me. And obviously that never happened because that's the way it works in the IIT exams in India. Right? You write an exam, the rank is high enough, you get to go to an iit. [00:06:21] Speaker B: And so they will contact you. They'll contact you. [00:06:23] Speaker A: Yeah, they'll let you know. At least there'll be like a ranking. There'll be announcement and things like that of who. Who got in, who didn't. Right. So I was under that impression. And then, luckily for me, there were a couple of universities that did write to me saying, hey, look, we have scholarships for international students. And one of them was in Texas, a tiny university called Angeles State University. Another one was in Alaska, University of Alaska Fairbanks. And given my preference for weather, you can imagine where I went. So I had some family in Texas as well. So I went to Texas for my undergrad. [00:06:52] Speaker B: All right. I grew up in Texas. So I was curious about that aspect of your development and how you found Texas in general. But that was in the, what, early 80s? [00:07:02] Speaker A: Yes. No, not early 80s. I'm not that old. I'm kind of old, but I'm not that. It's. It was like late 80s. [00:07:08] Speaker B: Late 80s. [00:07:09] Speaker A: So basically late 80s. And it was the time before the Internet. And, you know, it was. It was interesting. It was an interesting culture shock for somebody coming from, you know, a bustling place in India to like a small town in Texas with cowboys for roommates. Right. You can imagine all the interesting things that happened. And I could talk about that over beer sometime. [00:07:29] Speaker B: Yeah, yeah, we'll do that over a beer. Okay. Well, all right. So maybe fast forward, I suppose, because I do want to buy you a beer now about. And talk about Texas, but predictive coding, right? Yeah. Like what? I don't understand the origins of it. [00:07:46] Speaker A: Right, right. [00:07:46] Speaker B: I mean, I understand Helmholtz, so, I mean, we don't have to go that far back. So, Herman Vaughn Helmholtz with inference and predicting the best. Yes, Predicting. Predicting the best outcome. Predicting, sorry, Predicting what you will perceive. And I know that origin story, but in terms of neurons, do you like, what is that origin story? [00:08:10] Speaker A: Yeah, I think for me and my advisor at the time, Dana Ballard. So I went to University of Rochester for my PhD, and I was going to do a PhD in theoretical computer science, but then I happened to meet Dana Ballard in the copy room, and he said, hey, I have the summer ra, so can you work with me for a summer? I said, okay, I'll give that a try. Sure enough, that got me hooked to at that time, computer vision. And we were looking at this problem of how do you reconstruct objects that are behind occlusions? So the idea was, let's take the representation from the visual cortex. You have these Gabor filters, these oriented filters. Now, if you can recognize objects using this representation based on the responses of these filters, can you now also reconstruct what's behind these occlusions that the objects may be behind? And then it turned out that those filters, you cannot just combine them linearly with the responses of the filters and then reconstruct an image like you would do for PCA filters, Principal component analysis, or the eigenvectors, because they're not just. [00:09:14] Speaker B: Linear, which is linear, I'll just linear. [00:09:16] Speaker A: But even if you do a linear combination of these Gabor filters, they are non orthogonal, orthogonal to each other. So that's when we got into. So the idea was, can we then do gradient descent on a reconstruction error cost function? Right? And so you have this idea that, okay, we can reconstruct images that are based on non orthogonal filters, right? The responses. So then it leads to this idea of optimization of the responses of neurons that are trying to reconstruct an image that's been presented based on these Gabor filters. And then if you flip that and say, okay, what if you can not just estimate the responses but also learn those filters? Then you get this really interesting idea of at a fast timescale, you do inference, which is similar to what Helmholtz was talking about, which is inference of what object you're perceiving right now. And you do that using this dynamical system, which is based on gradient descent. The interesting thing there, which was not very appreciated at that time, and even I didn't appreciate as much, was the equation that you get from gradient descent. So you have this cost function of, you know, minimizing the prediction error, the reconstruction error. When you do the gradient descent equations, it basically falls out of the equation that you need an architecture that has feedforward and feedback connections. And the feedforward connections in that architecture would be carrying the prediction errors and the feedback connections would be carrying the predictions or the reconstruction of the image. [00:10:42] Speaker B: For fear of losing the audience. I'm just going to back up and we'll go Lego bricks. [00:10:46] Speaker A: Sure. [00:10:47] Speaker B: Yeah, right. So what is the idea of predictive coding? The idea. Do you want me to say it and then you can correct me? Oh, yeah, go ahead. Yeah, yeah. So predictive coding is basically that you have some idea forward in the brain somewhere forward in the brain Some idea of what you're expecting and you're sending that backward through the brain and those incoming sensory signals that are coming up toward forward of the world of shapes, if we're talking about vision, then they get met with this prediction. And then there's a difference between what you're expecting and what those signals are propagating forward. And it's the difference in the predict between the prediction and the signals that gets propagated forward. How does that sound? [00:11:46] Speaker A: That's great. That's great. Yeah. So I guess we were good at conveying the message in those papers. So I'm glad to hear that that's so. Yeah. So basically the idea is the traditional model going back to Hubel and Wiesel was always that you had this feed forward pass. So you flash an image, you get a feed forward pass all the way from V1, V2, V4 in the visual system up to it, and then you get recognition and then you get cognition of that. You can make a decision. Maybe you press a button, right? So that's action. [00:12:11] Speaker B: But was it back in those days, was it always like, voila, there's the cognition of it, like if it just builds up to an abstract enough edges to a table leg to a table, right? Oh, there's a table. And it magically happened. [00:12:25] Speaker A: I mean, that's the way it was thought even in the AI field. Right. The whole field of AI was partitioned into all these different, you know, there's people doing vision, there's people doing motor control, robotics, there's people doing, you know, higher level AI based on logic. It was very similar to how even in neuroscience and cognitive science we have these, you know, people focusing on particular modalities and so on. And then the. For me, as somebody coming into that area of neuroscience and looking at the anatomy, it seemed really surprising that people were completely ignoring the feedback connection. So if you look at every cortical area, it's sending not just the feed forward connections, but also receiving feedback connections from a higher order area. And then if you talk to people at that time who are really famous visual neuroscientists and ask them, okay, what about those feedback connections? And they're like, no, no, no, that's maybe doing a tangent. There's doing something really. [00:13:17] Speaker B: Would they really say that? [00:13:18] Speaker A: Yeah, yeah, sure. I mean, people thought that those were not really critical for the traditional kind of perception of objects. And those were maybe brought into play when you're doing very specific kinds of attention, maybe during sleep and things like that. A lot of People were not necessarily even acknowledging that those feedback connections could be playing a very dominant role. Predictive coding flips it on the head. It says the goal of the brain, or at least one goal of the brain, is to learn a generative model of the world, an internal model of the world. It's constantly making these predictions or doing essentially hypotheses testing. It's essentially generating hypotheses and it's trying to then match those hypotheses is what's coming in through the census. And when there are deviations or mismatches, those are called prediction errors. And those errors are what are used to then update the hypotheses. So you're sending them back in the feed forward pathways. So the feed forward pathways are actually not carrying the raw signals, but they're carrying the deviations or errors, mismatches. That was very insight, I think. Yeah, that was different from what the conventional thinking was at that time. [00:14:29] Speaker B: Well, okay, so let's stick with that insight for a second. I mean, did you. Was that insights an aha moment for you? [00:14:39] Speaker A: I would say the idea was in the air. Right. So I was at that time reading papers by people like David Mumford, for example, who was also, you know, talking about things like that between the thalamus and the cortex and cortex to other cortical areas. There were people like James Albus, who was talking about in the context of AI, and he had already done work, obviously in the cerebellum that was his main contribution. Neuroscience was the model, the more Albus model of the cerebellum. But he also had really interesting ideas about hierarchies in AI and robotic controllers and things like that. Then finally, if you really go back in time, there was somebody named Donald Mackay, he's actually the father of David Mackay, who wrote that book on AI information theory. He actually proposed, he had a paper called the Epistem Problem for Automata back in the 1950s. And even he proposed this kind of idea that what if you can send error signals from one module to another and then you could then build up this sort of abstract representations at multiple levels of the hierarchy. So what we did was essentially take many of those ideas that have been around and then say, okay, let's actually implement that mathematically. Let's see what comes out of it. And then we showed that some of these things that come out of it, if you interpret neural responses as prediction errors, is that you can explain some of these puzzling effects like end stopping, contextual modulation, orientation, contrast effects, things like that. That would be harder to explain if you just have a feedforward model. This actually gives you a normative explanation in terms of natural images and in terms of natural behaviors. [00:16:14] Speaker B: But even if you look back at the McCulloch Pitts articles, they knew feedback was important and they said it was sort of, maybe you can educate me on this a little bit better. But it was almost like a nod to the fact when they were drawing their little neuron diagrams, McCulloch Pitts neurons, right. One to the other, and then there'd be a feedback loop. And I think they wrote about how it would be important one day perhaps. [00:16:43] Speaker A: Yes, yes. I think they call them loops, right. And they definitely had this idea that loops are pretty important. I think at that time, there were a lot of people interested in characterizing essentially what we call dynamical systems now. But they were trying to characterize the properties of these loopy networks. I think the correspondence between that kind of work, which is very fundamental, rigorous, basic, elementary work to now trying to map that onto anatomy is, I think, the leap. That's what Mumford has suggested, and that's what we also tried to suggest, was that, okay, let's try to map some of those ideas now onto the cortical anatomy, right? The idea of, you know, there's six layers in the cortical area. There's feed forward connections coming into the middle layers, layer four, and then there's feedback coming from the superficial deep layers. And if you look at that, you have the Vanesson, Feldman and Vanesson hierarchy. Right? So can we actually. [00:17:41] Speaker B: Monstrosity. That monstrosity of a diagram. Yeah, so. [00:17:45] Speaker A: So. So the way that we looked at that Feldman Vanison hierarchy was, okay, what if you interpret that as a generative model that evolution came up with for essentially modeling the world? Once you interpret in that sense, then it says, okay, if you have this idea that you can sample from that generative model and generate examples of what the animal faces and its interactions with the environment, then inference becomes this idea of trying to update those estimates. You have those hypotheses you have about the world. I think the key idea there was inference as a fast update of the neural responses at the population in all the different cortical areas and at a slower time scale, using the same mismatch or errors to then update the weights, the learning part, the synaptic plasticity part, that was happening at a much slower time scale. [00:18:39] Speaker B: So your famous paper is the 1999. That must be your most cited paper, am I correct? [00:18:46] Speaker A: It is, yeah. [00:18:48] Speaker B: So you had that and in fact, I've seen you in a talk point to it with optimism for graduate students saying, see, stick to your ideas and eventually perhaps they will come to fruition. [00:19:03] Speaker A: Perhaps the keywords. [00:19:05] Speaker B: Well, perhaps. I mean, maybe. What do you think is the. Oh, gosh, I would say 5%. No, no, 3%. What do you think? [00:19:13] Speaker A: Yeah, so I think it's hard to say. Right. So I would say one of the reasons why that paper, it was published at a time when I think they were still not the kind of techniques you would need to really test some of these predictions, like for example, these prediction error neurons. Can we look at specifically layer two and three, which is where you would expect these prediction error neurons to be, because those are the feed forward connections, the next cortical areas. Similarly, can you look at the deeper layer, five neurons, and look at what responses? Because in the predictive coding model, the deeper layers are the ones that are storing the estimates of state or in a motor kind of responses and so on. So you'd have to have the techniques which were not available at that time. But I think if you fast forward 10 years from then and then start looking at, you know, in the late 2000s, you start to see more citations of that paper. And then now I think it's, you know, there's so many people trying to look at different aspects of that theory with evidence, you know, for some people against. Right. So I think it's become very, very rich area of research. So I think the theory was maybe a little too early for its time at that time. But I think the fact that nature neuroscience, which is where it was published, was willing to publish it. Right. As opposed to saying it's just pure speculation, which they could have done. I think it's credit to them for having taken that particular paper. [00:20:32] Speaker B: Yeah, I mean, Occam's razor, it's theoretically great, right? [00:20:36] Speaker A: Yeah. So it had an elegance to it, which I think was we ourselves didn't quite appreciate as much. And I think Friston actually wrote like a news and view it 10 years, 10 or 20, 20 years after it was published, talking about how that one paper really influenced him as a researcher to really, you know, pick, pick up the, I guess the mantle at that point because we had actually, I mean, I had sort of moved on to more general models on the Bayesian brain hypotheses and belief propagation in Bayesian networks and things like that. But then of course, Friston went on to do the free energy principle based on predictive coding ideas and then more recently active inference and you know, there's a whole bunch of work. [00:21:20] Speaker B: Principle. Yeah, but, but so at the time, and I'm sorry to perseverate on this, but at the time, did you know that it was going to be so influential or not? Not, not know. But did you have some confidence, some, some sense that this is. I know you had a sense that it was a good idea, but I have good ideas every day and they're never good ideas. [00:21:43] Speaker A: Yeah, I think it's. At that time I would say, you know, I was not, you know, too confident that it would actually be something that would inspire people to do experiments. Because there was a news and views that was written on that paper by Koch and Poggio and it's an interesting news and views. You can see them, you know, they code Sherlock Holmes story and then talk about how the fact that the dogs were not barking at night in the Silver Blaze was the story of Sherlock Holmes is an example of predictive coding because the person was not surprising. So then the deduction is that the person who stole, I guess the horse called Silver Blaze. [00:22:24] Speaker B: So that's a very positive review. [00:22:26] Speaker A: Yeah, so it was a positive review, but before that I was told by the editor that the previous person refused to write and use and views. Right. So it's like, oh, there's already people that are, you know, not necessarily going to like this. Right. Maybe it goes against their way of thinking about how the cortex works. But. But, yeah, but, but I think it, like I said, there was not as many citations. There was not like a huge number of people, you know, writing to me and saying, hey, you know, that's a great. So I was like, okay, so we wanted to get it out. It's out there, let's see what happens. And then it took, you know, 10, 15 years right. Before it really caught on. [00:23:03] Speaker B: Caught on. But were there early detractors also? [00:23:07] Speaker A: Yeah, I think there's always been detractors. Right. So there's been people. I mean, even now. Right. So basically a lot of people don't buy into this notion of the feedback really influencing the responses and then the prediction. [00:23:22] Speaker B: How can that even. Oh, well, specifically with the prediction errors, that's a different. Yeah, so feedback is obvious in the. [00:23:28] Speaker A: Brain, but even feedback carrying predictions from a higher area. Right. So even that's not obvious. That is still debated for particular tasks. Right. I think there's really clear evidence now that for sensory motor tasks where there is the animal makes a motor movement and then there's an inference copy that allows you to make a prediction. Right. I think there's a lot of evidence now coming from, you know, David Schneider's lab, from your Keller's lab and so on, showing that, you know, when there's locomotion or there's a four limb pressure of a lever and there's an auditory response, you get a prediction in the auditory cortex. There's a suppression that looks like a prediction error signal there. So they're definitely for the motor. I think you see that if it's purely sensory, which is, I would say, I would argue that's in some sense a purely sensory experiment, typically has a fixation and then you're flashing something. It may not necessarily be the kind of naturalistic behavior that would elicit predictions and prediction errors. Right. So. But there have been some reports about these error like responses in those situations as well. Right. So I think. Yeah, yeah, that's the, that's where I think we. There were initially, you know, there was initially some resistance in terms of the fact that there wasn't at that time, you know, any evidence for feedback conveying predictions and feed forward conveying prediction errors. Except for these contextual effects like these extra classical receptive field effects and even that some people are explaining it just through lateral inhibition. Right. Which could be one way to explain that. [00:24:57] Speaker B: Yeah. So I just want to. So what is the lesson here? It's not that predictive coding is universally accepted still, but now the technology is better and we can have better and better measurements of hypotheses generated by predictive coding framework. So one lesson is keep going grad students, it might get better, but for a. So the jury's still not out, I guess. Right? [00:25:28] Speaker A: Yeah, I would say there's evidence for some parts of the theory. And I think as a theorist, you know, I feel like our job, you know, all the theoreticians out there and our job is to propose theories that are, you know, explicit enough in terms of connecting to the anatomy, which is what? Specific enough that they can be shot down, they can be falsified. We basically want to have the ability to falsify theories and in the process of that, if it generates some new data that then really sparks a new theory and then leads the field forward. I think that's the job. And we should not be afraid to propose these theories as long as they're sensible, they match the existing data. Which is what we were trying to do at that time was saying we have these puzzling responses called end stopping or contextual effects. We have the anatomical evidence saying there's feedback connections. So let's propose this theory and then let's see what happens. Right. If you start doing experiments. So I would say the same thing now, Right. Look at what's out there, what people have collected in terms of the data, and look at the computational aspects also. Right. So what is the brain really trying to do? Let's look at evolution and. Yeah, I mean, that's what I was. [00:26:33] Speaker B: About to ask you because. So do you see predictive coding or the family of predicting coding approaches as some sort of general theory about what the brain is supposed to do? You know, what the brain is doing because, you know, there's the free energy principle, now there's Jeff Hawkins. Everything is prediction. You know, these things go back years, obviously, and everyone points to Helmholtz and there is someone before him and there's. [00:27:04] Speaker A: ChatGPT and Transformers, which are also based on prediction, right? [00:27:07] Speaker B: That's right. And we love them. [00:27:09] Speaker A: Yeah, I would say it's just part of the story. Right. So prediction is important because, you know, animals have to predict to survive. I mean, ultimately it goes down, goes up. It's this idea of, you know, at a certain point in evolution, you know, the brain start building models of the world to be able to predict, to anticipate. Right. And so I think that's an integral part of, if you're learning a model of the world that comes with this benefit of you can use that to generate hypotheses, you can use that to generate predictions. That I think is an integral component, this idea of world models or predictive models of intelligence, either artificial or natural. It gives you a lot of benefits. And it wouldn't be surprising if evolution came up with that way of handling the uncertainties of the world and, you know, to compensate for delays as well, from, you know, your sensors and your muscle senses all the way to the brain. Right. So there's always going to be delays. So if you're too late to react, you're probably going to get killed. Right. If you, if you, you know, also if you have the ability to predict and have a model, then you can do planning, Right. You could essentially go forward in time and then you can do planning and that, and then that leads to abilities that allow you to have much more sophisticated behaviors than if you're just a reactive organism. Right. So I think there are a lot of reasons, at least from a computational point of view, but probably also from a evolutionary biological point of view, to have something like an internal model and prediction that comes with it. [00:28:42] Speaker B: Okay, so Rowland Ballard was in 1999. And then the sensorimotor theory of the cortex. I think I have the name wrong, but I'll get it right and I'll flash it up there. Which you've recently presented, which you've been working on for, I think a long time now. [00:28:59] Speaker A: Not so long. I would say maybe like four years. Like three, four years. [00:29:03] Speaker B: Oh, that's not so long at all in research age. Oh, okay, well, so that's what I wondered. Like, I'm not sure if you flipped in that regard. It sounds like you've somewhat flipped, but if you have, like how that. Who did someone drag you out of perception or did you climb out yourself or. [00:29:27] Speaker A: I think it was multiple things. So, you know, I think it was just this realization that, you know, if you think about what is the purpose of, you know, having a brain, the context of evolution, right? So going back to some of the early, you know, you know, creatures that were trying to, for example, you know, move towards places that have more nutrients because you don't have enough nutrients. [00:29:55] Speaker B: So, but, but, but hang on, let me, let me just stop you there. I'm, I'm so sorry to interrupt, but they're, they're moving to a place with more nutrients because they can already perceive whether there are nutrients there. Right? [00:30:09] Speaker A: So then way to minimize the error. Right? So ultimately you're trying to. [00:30:13] Speaker B: Perception came first. [00:30:15] Speaker A: You need both, right? So you, in order to, you can, you can do, you can do random movements, but you still have some notion of perception because when you taste the nutrients or when you actually sample the nutrients, then that is like perception ready for the caught perception. So I think it's kind of like the sensory and motor part are integrally tied, right? So the motor part allows you to change your location or, you know, sample a different part of your, of your environment. But then if you want to know was that successful or not, right. Then you don't, you don't need sensation, right. Of some sort to know. Okay, what I did was, you know, good for me or bad for me. Right. And that's what we're calling perception in a very, in a courses level. Right. [00:30:52] Speaker B: But so do you think. All right, so I guess we have sensation without brains or perception. Sensation, perception, we could use them interchangeably, although that's dangerous. But let's say light sensors, right? [00:31:06] Speaker A: What do you mean artificial? Artificial sensors? [00:31:08] Speaker B: No, no. Well, you could do artificial, but I meant biological, right. Pre movement. I guess what I'm wondering is like, you know, evolutionarily, why did brains develop? And that is in Some sense a moot point because they are integrated, like you said. But I think you have to perceive before you can move. [00:31:34] Speaker A: No, I think it's. I mean, I don't think it's one or the other. Right. So essentially, I mean, you could say perceive because you're perceiving the fact that maybe you're low in nutrients inside your body. Right. So you're hungry. I mean, you could think of that as a kind of internal perception. Right. And that may drive you to then move. Right. But then when you go out and you actually consume some nutrients, then your internal state may change because you're not perceiving it as, okay, I've consumed food and so I'm more. [00:32:00] Speaker B: Yeah, but if you move without any perception, then you can't, you can't. You don't decide. [00:32:05] Speaker A: Exactly, exactly. So you do need perception. But at the same time, you know, if it's just, I mean, if just a creature that doesn't move at all, but it's just perceiving, I mean, that also doesn't quite make sense. Right. [00:32:16] Speaker B: Does a Venus flytrap move to you. [00:32:20] Speaker A: In the sense that it does have an action. Right. Of grabbing the insect and digesting, but it doesn't move in the sense of creatures moving. Right. But in some sense maybe it does if you consider the movement. But again, it's not a moving organism like the one that we associate with brains. [00:32:42] Speaker B: Right. Yeah. [00:32:47] Speaker A: Go ahead. No, these are great questions to think about, right. In terms of. [00:32:50] Speaker B: Oh, no, I'm sorry, I'm getting a soft. I'm telling you, man, I had a long motorcycle ride and my brain is fried. And so I'm sorry if I'm a. [00:32:57] Speaker A: Little squishy myself, but sensory motor coordination, that's what it is. [00:33:03] Speaker B: Yeah. Back to this, what is it? 25 year divide between these two papers. I'm just going to focus on these two because it's like two of. I think you've published in total probably four papers or something like that. Slightly over on. [00:33:20] Speaker A: Yeah, I guess in total. Sorry. [00:33:25] Speaker B: All right, so. But it felt like a big reveal when you added the motor aspect to the predictive coding. And we just talked at length about how you have come around to thinking that action and motor behavior is important. So can you maybe just give a summary of how you incorporated it and what people can expect if they read this paper? [00:33:52] Speaker A: Yeah. So I think the first observation that really prompted me to go on that whole journey was the fact that there were all these new results coming in showing that even primary sensory areas or quote unquote, sensory areas like V1, S1 or A1 were being influenced by motor activity or actions. Like, there was papers from the Carandini Harris lab, there was papers from David Schneider's lab showing that you can actually get these responses that are motor related, but they're happening in the traditional sensory areas like V1 or A1. Not only that, but then the other interesting observation was anatomically, again, if you look at an anatomy, the layer 5 cells, right? There's like 5A 5B. But the layer 5 cells, even in the primary sensory areas, like V1, A1S1. Right? Somatosensory, cortex. Primary somatosensory. The layer 5 in all of these areas are sending axons to motor, subcortical motor areas. Right? Like as the v1 sends it to superior caliculus, a1 sends to inferior caliculus. [00:34:58] Speaker B: It's all like, completely connected. It's a ridiculous system for an engineer. [00:35:04] Speaker A: I mean, it's not completely connected in the sense that you don't have, like, the superficial cells sending axons to superior colliculus yet. [00:35:13] Speaker B: We don't know that yet. [00:35:14] Speaker A: Well, okay, so there's always a possibility. Right? But I think that was the other interesting anatomical constraint that we were looking at. And then the funny thing there is, okay, it's not just that you're going from layer four to layer two, three, and then to layer five. And then that's it. Right? Now layer five is sending information back to layer two, three, right? There's a loop there between the superficial deep layers. What is that loop doing? If the superficial layers are receiving sensory information from layer four, and then the deeper layers are sending outputs to motor centers, then there is this loop going on between the sensory and motor within each cortical area. That was the first insight, was what if we take some inspiration from reinforcement learning? So we all know about the Markov decision process idea and the fact that there is. [00:36:04] Speaker B: We don't all know about it, even for non scientists. Like the most basic version. Yeah. [00:36:11] Speaker A: So the basic idea with reinforcement learning is that there's a world and there's something called the state that the animal or agent is in. And then when you take an action, that state changes. Right. Basically you have. That's called a Markov assumption, because you're saying the next state only depends on my current state and current action. But that is typically called the forward model or the dynamics of the world. Right. You're saying the world is in this state, or I'm in this state, and then I Take this action, the world changes to this next state. And that if you repeat that, it becomes like a dynamics, like the physics of the world, right? And that is the forward model or the, what we call the generative model of the world. But then reinforcement learning. The goal is to learn a policy which is given a state of the world. What is the best action I can take? That is called a policy. It's a function for any state the animal is in. It's going to say, okay, I want to take this action, because that action is the one that will let me achieve my current task. If you now combine the model which is at the top, and then the policy at the bottom, you have one that is making predictions of what the next state is going to be based on your sensory inputs, and the other that is saying, given that I'm in this state, here's what the action is that's most suitable. Then you can have this, a cycle between prediction, action, prediction, action, prediction, action. That is the general idea. And that's what we suggested as a module, like a sensory motor module for any cortical area. And it's going to operate at its own spatial temporal scale, depending on the kind of inputs it is getting. Every cortical area has its own spatial temporal scale and modality, depending on the kinds of inputs it's getting in, the kinds of outputs it's controlling or potentially controlling. Then the last thing we threw in there was, okay, what about the different hierarchy and the different cortical areas interacting? Why is there a feedback connection or a reciprocal connection? It turns out that from a computational point of view, if you want to have a very rich way of modeling the world, then what you want to do is have the higher level modulate the dynamics of the lower level. You need to change the function being computed at the lower level depending on the task. So if my task is to drive to the grocery store, right? Then you basically want to load up the program for getting into your car, figuring out where you're going, and then driving. So you have all those quote unquote programs that are policies, right? These are these policies that are already being learned that are being plugged in to achieve your current goal. But if your goal is now to go to your friend's house, it'll be similar programs, but the sequence may be different, right? Or if it's, if you, if it's a cook something, right? Again, it's the same idea, right? You load up a different program. So we're suggesting that that kind of, you know, schema or, you know, loading up new programs can be done through top down modulation. Maybe prefrontal cortex loads up in the lower areas, these programs and the corresponding dynamics of what to expect. Right. [00:38:58] Speaker B: All right, so we'll just say it's in the prefrontal cortex. So how does the prefrontal cortex come to be that way? [00:39:05] Speaker A: Yeah. So our hypothesis is that it starts off with all areas operating in terms of their respective spatial and temporal scales. But by the time you get to the prefrontal cortex, you've already converged, you have information that is now operating at a much longer time scale. And the abstraction level has gotten to the point where it's now operating at the level of abstract actions. There could be a population of neurons that are coding the current context or task, which is saying, right now I want to go to the grocery store. That's like a goal that got instantiated at the highest level, but it's maintained until that goal is achieved. And while it's being maintained, it is modulating all the other areas that are involved in achieving that goal. That involves even the basic lower level somatose, like even somatosensory cortex, visual cortex, auditory cortex, they all now have that context from the higher area saying, here's what my current goal is. Now let's see if we can achieve that goal. [00:40:02] Speaker B: And then you get your. I don't know what you would get. My grandmother would get cookies. Definitely at the grocery store, you get those cookies and then how does it switch to the next goal? [00:40:15] Speaker A: So the idea is it's a hierarchy. So as you achieve the sub goals. So basically the idea is that a complex problem, right? A complex task is split up into its sub goals, subtasks, and those are in turn split up into subtasks all the way until you get to the point where you're operating at the level of milliseconds and controlling your muscles. That's happening in the spinal cord loops. The spinal cord is controlling things at a very fast timescale. But then as you go up the hierarchy through the different areas, including the midbrain, and then you go all the way up to the cortex, you start to get longer and longer time scales. Then that's where you start to get these subtasks being achieved. And then those in turn cause a flip to the higher levels and we go to the next goal. So you're going from one goal to the next goal to the next goal. Right. [00:41:01] Speaker B: But how those goals get programmed. [00:41:03] Speaker A: Yeah, so that's all about learning. Right. So you have to learn and that's what, you know, some people call that curriculum learning in the AI literature. But as babies, after we were born, or even before we were born in the wombs, we are starting to learn these, you know, modules of, okay, what happens if I move my hand in this way and my arm in that way? But then once you're born, you're starting to grab objects, you're trying to reach for objects, you're trying to grasp different kinds of things. So you're building up this repertoire of action policies and the corresponding dynamics that go with the action policies. And the idea is, as you go through learning, through your toddler years and so on, you're building up a very rich set of primitives that can eventually be composed for solving more and more complex problems. There's still a lot of unknowns, but I think that's right. [00:41:52] Speaker B: Yeah. I guess what I was asking was, what? So, I mean, I think we've all had the feeling like once you're done with a goal, you're like, well, all right, what now? And you kind of feel lost for a moment until you realize what else you have to do. But what I'm wondering is, like, how those sequence of abstract goals, if you have an abstract goal, how does it switch once you accomplish one? You got to do the next one and then the next one. How do you. How does your prefrontal cortex know? But that's a great question. [00:42:19] Speaker A: That's a wonderful question. I think in the model at least, you start off with some higher level goal, which I guess we give it, or we say, here's the task. Then the way it decomposes, that is, as you go down each higher level abstract action or policy, it generates a function at the lower level that function basically has as the output. Remember, functions are basically state to action mappings. Each of those actions is now another abstract action, which is like a sub goal, right? And then that in turn generates another function at the lower level, all the way down to the spinal cord level, right? And that is what we're saying has to be learned, right? You have to learn each of those sequence sequences, those functions that capture the sequence of sub goals. But we're saying it's the same module replicated at all the levels. And that's one of the things that, you know, for me, you know, Mount Castle's whole idea of, hey, the cortex has something similar across the algorithm, may be similar across different cortical areas. As a theorist, you always have that fascination with those kinds of ideas. The challenge here was, is there Something like that that we can come up with from a computational point of view. Then of course the brain may or may not implement it, but let's try to come up with something like that. It turns out that this active predictive coding idea seems like it is versatile enough to tackle many, many different kinds of problems. But it remains to be seen how well does it map to something that the brain may actually be doing. [00:43:42] Speaker B: But do you believe it? Do you believe Mount Castle? [00:43:47] Speaker A: Yeah, not like exactly. The cortical column is something where everything is. But I believe in this more broader idea that if you take any piece of cortex, a chunk of it, and then you give it even to trained neuroscientists, it may be hard for them to figure out which part of cortex that is. Right. I mean, of course there's motor M1 versus V1, but I think. [00:44:13] Speaker B: But 90% is that is who they're connected to, not what they're doing. Probably. [00:44:17] Speaker A: Exactly. Yeah, yeah, yeah. And I think that's a really interesting finding because, I mean, we would think, you know, that language area should be programmed very differently from vision areas and so on. If you sort of train. Traditionally we think, okay, visual cortex is doing edge detection. Right. But then if you look at auditory cortex, there's no edges, so they can be doing edge detection. You're doing something completely different. Motor cortex must be doing something very different. But it's really intriguing if you think about it in terms of, okay, what if they all have components of this sensory and motor aspects to them? And it's just that in some areas, some got more emphasized than others, but they still have the basic primitives of sensory inference as well as motor control. Right. And they're operating together at multiple levels. [00:45:04] Speaker B: You're a BCI brain computer, first brain computer interface person, an engineer minded person. This is maybe a throwaway question, but given. All right, so what you just described is maybe it's somewhere in the prefrontal cortex there's this higher level, more abstract goal, and then it gets implemented at more at lower and lower levels until it pans out in musculoskeletal system. Where would you, if you were evolution, where would you move next? Like, do we just get more and more abstract? What is, what does that mean, more and more abstract? What would be a more abstract goal than, you know, make. Honoring my great great great great grandmother's legacy by. I don't want to be stereotypical, but by doing something the way that she would do it. Right. [00:46:06] Speaker A: Yeah, yeah. So basically you're saying, okay, what Is the next sort of evolutionary mind. [00:46:11] Speaker B: Yeah. What would be, you know, in 10,000 years. Yeah. How come we can't guess that? [00:46:18] Speaker A: Because we've been very bad at it. We can guess it, but we've been pretty bad in terms of we can't. [00:46:23] Speaker B: We're very good at guessing and being wrong. Sure. [00:46:26] Speaker A: Yeah. So I would say, I think, I mean, in terms of at least human evolution, we see that we're essentially, you know, biologically, we may not have added, you know, new cortical areas or new brain structures, but we have been really making amazing strides in terms of adding tools. Right. And the cultural knowledge and tools. [00:46:46] Speaker B: Perfect segue, by the way, to bci, but I'm sorry to interrupt you. [00:46:50] Speaker A: Yeah, exactly. So basically, I think if you think about it in terms of we basically augmented ourselves, right. So humans augmented themselves originally using tools like rocks. Right. And sort of made these little blades to cut meat and so on. And then we, of course, augmented ourselves in terms of wheels to move faster because our legs can only go so fast. And then, of course, we went beyond that to, you know, we had. We were able to fly with airplanes, I mean, more recently. Right. And then similarly, memory. Right. So memory augmentation. So, you know, we could only hold so much in our memories. But then we started writing. We had language, so we started writing things down, and then we were able to pass it from generation to generation. So that's another big, you know, tool. Right. In terms of cultural knowledge accumulating over time, this is the most beautiful thing. [00:47:39] Speaker B: About language, I think. [00:47:40] Speaker A: Yes. It's an amazing thing which other animals unfortunately don't have. Right. So I think that leads to this idea of, okay, so if it's really tool use. Right. And nowadays we all hang around with these devices in our hands and in our pockets, and basically we've augmented ourselves in terms of knowledge and access to information with these kinds of devices. And so people in the BCI world and brain computer interfaces thing, maybe there is a progression there. If you really take that idea to the extreme, then if the brain itself biologically is limited in terms of its speed and in terms of its memory capacities, then would it make sense to then augment it with artificial memories and artificial processing capacities and artificial communication capacities? And that sort of leads into the whole augmentation side of brain machine interfaces or brain computer interfaces, which is an area that is not something most, I think, academics want to really go too much into, because there's a lot of ethical, ethical applications. So most of us tend to stay on the Medical application side, that's where the funding is and that's where a lot of important contributions have been made and need to be made. But there are companies and people in the tech world that are really interested in that aspect, the augmentation aspect. [00:49:06] Speaker B: Yeah. And uninterested in ethics. So how do you navigate that? [00:49:10] Speaker A: Yeah, so I co direct the center for Neurotechnology here. We were funded by the NSF for about 10 years as part of an engineering research center. And we had a team, a neuroethics team. We still have them. But during the grant, what we did was for all the projects, we embedded a member from the neuroethics team into the engineering team itself. Right. So every engineering team that was developing a brain computer interface application had an ethicist who was giving them active feedback about, okay, you know, what is the end user going to think about this? What are the long term implications? [00:49:41] Speaker B: And the engineer, did your team hate the ethicist? How did that work? I mean, I don't need to ask you how you feel personally about it, because it is a conundrum, I think, from a technological point of view. But that's like embedding a philosopher into a research lab or something. [00:50:01] Speaker A: Like, they work a lot, actually, many of them. [00:50:03] Speaker B: Of course, of course there. Well, that's all, that's all philosophy has left is ethics. Right? No, I'm just kidding. Just kidding. But okay. But yeah, I mean, was the resistance, I mean, you know, no, I think, I don't, I don't want to have to deal with that kind of stuff is what I'm saying. And I admire you for doing it. [00:50:23] Speaker A: No, I think it's important because as engineers, we're really excited about building the next great thing and doing something very novel. But at the same time, the consequences may not be anticipated by engineers as much as people that are more trained to anticipate those consequences. [00:50:39] Speaker B: Yeah, but the consequences are always bad, right? [00:50:43] Speaker A: Not always right. I think there are some constructive things you could do. Like for example, there were actually some interesting observations the ethicists made when they talked to the families of patients that had, for example, deep brain stimulation. Right. And many times a family member provide insights that may not be very obvious. Like they may say, okay, the personality of my partner changed right after they got the DBs. Or they may say, you know, the actual patient may say, you know, sometimes I feel like I've lost my sense of agency because there's this device in there. So then the question becomes, okay, what are the ways in which you could restore that sense. If it gets to the point where the person feels like they've lost their sense of agency, then is it okay to have a kill switch or some other way to at least provide them some relief? So there are ways to have engineering solutions if you know about some of these things that the patients are going through. And I think at least that's a good way to then find ways to address issues before they become too bad to really address. If they're commercialized, the technology gets commercialized, maybe too late at that point in time. So that's one of the hopes. But you're right that many companies may not necessarily look into any of these. [00:52:02] Speaker B: Issues or even academically and research wise. So we're talking about, you mentioned agency. We're talking about concepts that. Not that philosophy has settled on anything ever, but concepts that are very much in the air in terms of like, well, if, I mean my personality has changed, but if I add a BCI or you know, some sort of prosthetic to my brain and then, well, is it the press then is it the fault of the prosthetic or was that a deep seated thing in me or was, you know, so how. Who's responsible? And oh, ethically I don't want to deal with it. [00:52:43] Speaker A: So yeah, yeah, those are great questions that, I mean we've been, you know, talking about them and we've been discussing those and. And I think it gets even more trickier. Just think about it in terms of. Okay, now we have dbs, which is, you know, open loop stimulation, just delivering some constant current. Let's talk about AI. Right. So if you add an AI into the picture that adapts. [00:53:03] Speaker B: Well, hang on, before you do that. So, so this is something that you've done. [00:53:06] Speaker A: Yes. [00:53:07] Speaker B: And are continuing to do so just to get. And I'll introduce this in the beginning, but I mean there's brain net, there's neural co processors. I'm not sure how you would like to describe it, as sort of an introductory idea. But you've done two people, you've done three. I don't know how many people have you gotten up to in terms of collaborating? [00:53:28] Speaker A: Yeah, yeah. So I think the idea there. So we were talking about brain to brain direct communication or brain networks. And we had three people collaborating directly, using their brain signals to solve a problem that was, you know, in that case was just a Tetris video game. Right. [00:53:45] Speaker B: And this is one person gets their. Has their brain signals communicated and stimulated and the other two get to interact. [00:53:56] Speaker A: Right. So the idea here Was, you know, the technology is not there yet in terms of actually doing brain to brain communication. But the idea was, if we take up this challenge of, okay, let's see what we can do with the technology we have now as a way to spur people to talk about this technology so that they can put safeguards now or at least start thinking about safeguards before it gets too far away. Right. So we said, why don't we take technology we have now like eeg? So we all have. Many people have EEG equipment in their lab. So we can use the EEG to decode simple things like, you know, the. If you have like a flashing light, it's called steady state, visually evoke potentials. You can decode the frequency at which you're steering. Right. So if you're staring at two different flashing lights, one of them will trigger, you know, oscillations in your visual cortex. And we can decode your intention. Or if you imagine making your movement, like move your hand. Right. Or imagine moving your hand, you can decode that from the motor cortex. So what we said was, okay, we can use EEG for decoding and let's use tms, so transcranial magnetic stimulation, to deliver information directly to the brain. Right. [00:54:55] Speaker B: Which is just for people listening. TMS is like when there's a coil outside of your brain and there's a pulse aimed at a very small part of your brain, it delivers a non invasive, non damaging pulse in your brain. And then depending on what function or affects you, potentially affects how you think or your behavior. [00:55:19] Speaker A: Yeah. And especially if you deliver it to the back of your head, the visual cortex, and you see what is called a phosphine, which may be a little spot of light or a bar or something like that. And so we basically used like a Morse code kind of idea. So the flash was just conveying binary, a bit of information. Right. And so we said, okay, why don't we use these primitive modes of interacting with the brain? So EEG and tms, and let's build a brain to brain direct brain to brain communication system with that. And this was several years ago, and we did some proof of concepts and published a few papers on that. Just as a way of not necessarily saying, look, we can actually commercialize this technology, because obviously we cannot, it's too bulky. But if you just think about companies that are now developing implants, then you can imagine once they have implants on multiple people, then they can start doing this. If they can stimulate the brain and record from the brain, it's bidirectional bci, bidirectional brain computer interface. Then you can start to do these kinds of experiments and you can start to implement these kinds of systems of recording from one brain, stimulating another brain, decoding and encoding the information in another brain and so on. And that's where we get into augmentation. Right. So now we're using technologies, neurotechnology, for augmenting the, in this case, the communicative and processing capacity of the human brain. [00:56:33] Speaker B: So where are you now right now with these? [00:56:36] Speaker A: Right. So once again, I think so we are not in the business of really like going on the commercial route. So what we're doing is a lot of theory. So what we were talking about is. [00:56:46] Speaker B: That what people usually ask you that in and their question is really like what future human are you creating? Is that their question? [00:56:55] Speaker A: Yeah, I mean some people are very much interested in that, the singularity and things like that. Right. What's that going to be in the future? Of course there's some people who want to merge with AI. So that's a whole other crowd. Right. Who want to actually. [00:57:05] Speaker B: Are you going to. Do you want to merge with AI? [00:57:07] Speaker A: No, not really. [00:57:08] Speaker B: I don't either. Yeah, I don't understand it. [00:57:11] Speaker A: Right. But I think what we're doing is in some sense we're saying, okay, if you really are thinking about processing information from the brain and you know, making sense of it and sending information back to the brain, that's what we call a CO processor for a brain. And that's what we're working on is a brain CO processor, which is this idea that there's a device that is essentially a tool for the brain that can both decode information from the brain and encode information back into the brain. [00:57:38] Speaker B: So what that means is like, so it's listening to the brain and then it goes into an algorithm. And it. Well, in terms of modern AI, it sort of categorizes what that brain process is signaling. I almost said understands it, and then transforms it into a signal to inject into the brain. I'm sorry if these are too low level descriptions, but. [00:58:05] Speaker A: Right. No, I think that's one way to think about it. And of course the way it would have to work would be that core processor has to have AI in it of some form. And so we suggested a neural CO processor which was a neural network, so an artificial neural network. And so what we have there is an interaction between artificial and biological neural networks where both of them are adapting. So it's a CO adaptive system. Right. So and the challenge there is how do they co adapt to really achieve a goal? Right? So, and from a point of view of medical point of view, if you, for example, are trying to replace lost function, like somebody had stroke and you want to be able to restore their movement, then this could be a device that could replace that lost function of the cortical or other subcortical circuitry. And the AI has to train itself and the person now has to command this AI replacement circuit, right. To do what the person wants it to do. So then that becomes an interesting AI problem. Right? So how do you ensure that the communication between the human brain and this artificial device achieves a common goal? Right? And that's what we did with the neural core processor was we said, hey, look, you know, if the goal is, for example, to reach for an object when the person cannot reach the object, but then you can stimulate, let's say, the spinal cord and make the person reach that particular object, then the error is now in the space of visual error, or so there's an error signal. Can you do back propagation, for example? Right. Which is a standard thing people do in AI. Right, but the problem is back propagation here you'd have to do it through your body and your brain, Right? You have to do it through the. [00:59:32] Speaker B: That's not back propagation by definition. [00:59:35] Speaker A: Yeah, but I mean, in terms of ideally, if you want to train your neural network that's delivering the stimulation, right. Then you have to get that error signal somehow. And the error signal, unfortunately, is now in the space of an external task. Right. So what we suggested was, okay, for all of these AIs, they need to have an internal model, this forward model of, in this case, your brain dynamics, right? So how your brain reacts to stimulation. So it's like a forward model of, you know, the action here is stimulation, and then it's going to cause your brain dynamics to move in a particular way. And if you're able to predict that dynamics, then you can come up with the best stimulation to give for achieving a particular goal. Right. So it becomes, again, a control problem. Right. [01:00:17] Speaker B: How imperfect. So, I mean, these are things I just, I'm a neuroscientist, but I ponder these things, you know, like, I mean, one of the most amazing things about the brain is how adaptive it is. And sometimes I think, well, maybe that is the function of the function, as if there was a function, but the adaptability is like almost the answer. And so what I was going to ask is, how wrong can it be for the brain to still be able to deal with it. [01:00:48] Speaker A: Right. [01:00:49] Speaker B: Is that a testable question? How bad can you make the signal relative to how quickly the brain can learn it? That's a terrible, poorly phrased question. [01:00:58] Speaker A: I'm sorry. It's actually a very important question, especially in the sensory stimulation area. So people that are trying to restore the sense of touch, for example, from an out of artificial prosthetic device stimulating the somatosensory cortex. Right. There's a big question there in terms of should we make the stimulation be as natural as possible so we make a sense of touch when you stimulate the somatosensory cortex? Patients typically report, okay, I feel like it's pins and needles, right. It's not quite the nice sense of touch I get when I use my own hand. Right. [01:01:29] Speaker B: And then you say, well, just wait a week or two and it won't be, or something like that. [01:01:33] Speaker A: Some people would say, hey, you're just not used to it. It's a novel signal right now. But if you keep using it, maybe you get used to it and that becomes a natural, you know, sense. [01:01:41] Speaker B: This is like training. This is like trying to convince my children that, oh, don't worry, keep practicing, it'll get easier. You know, in some way it sounds pretty painful. [01:01:49] Speaker A: Yeah, so that's the, that's the painful is a key part there. Because if you are not really delivering, you know, quote unquote naturalistic stimuli, then it may take much, much longer. Right. So I think there's a trade off there in terms of if you're able to speak the language of the brain in some way, right. Basically you're, you're, you're delivering the pulse electrical stimulation or optical stimulation, the way that the brain can parse it more easily, then you may be able to learn much more quickly than it's. If it's an artificial set of pulses like stimuli. But it's something that's interesting problem in. [01:02:21] Speaker B: Terms of natural, then is that just noise or is it like the natural statistics of the world? [01:02:27] Speaker A: It's more the way that the neurons in that particular area are receiving information from other neurons. Right. So for example, if you know that in the population level there's a particular frequencies that are operating at that scale, like beta frequency oscillations or something. So then you can think of delivering the stimulation at some particular phase of that oscillation, or you're trying to synchronize with that oscillation. Maybe that is better than you just delivering it irrespective of what's happening in that brain area. The bottom line is you need to be recording as well as stimulating. You can't just be stimulating the brain area. You need to be recording to know what's happening in that brain area and then change the stimulation according to what's happening at that point in time to consistently deliver intelligible stimulation, but within a. [01:03:13] Speaker B: Window, because that's antithetical to the idea that the brain can handle a bunch of a wide range. Right. So a brain area doesn't get to decide the signal incoming. But what you just said is that you need to do it in a way that it's expecting it. It's a very predictive coding outlook, by the way, otherwise it might not work. [01:03:36] Speaker A: Yeah, it's a way to enable the brain to learn faster. Like I said, it's possible that you could not pay attention, just keep delivering something, and eventually the brain may figure out how to interpret that stimulation pattern. But the question is, how long will that take? It may happen quickly or it may take very, very long, or maybe it'll take forever. Right. So in general, I think it makes sense to have some feedback about what's happening in that brain area. And as you're delivering the stimulation, can. [01:04:07] Speaker B: We partially get to this by doing learning studies and maybe just probing the historical literature and learning studies? So there's a. Modern learning studies are becoming more better and better. And I don't know if there is an answer to the proper way to learn X yet, like the most optimal. I know that there's a lot of people who make a lot of money who say that there is, but I mean, I sort of intuitively think you got to do it. It's got to be. Feel really bad at first, and that's how you're going to eventually learn. But it also depends on the goal. But do you think that we can. Is this an experimentally, behaviorally tractable problem that could help answer these issues? Because you're speaking about the activity in the brain, but then we're eventually talking about behavior. [01:05:01] Speaker A: Yeah, it's an interesting question in terms of how do you study what's happening as the brain is interacting now with this external device that is sending direct input into the brain? Right. [01:05:13] Speaker B: Or it's just the optimal input to if you're going to use these external devices. Right. [01:05:21] Speaker A: And so the neural co processor idea was one way to get at it in terms of saying, okay, if you are able to adapt the neural network that's delivering the stimulation to minimize the error in the external stat space, then you're relieving the brain of having to do a lot of that learning. But of course the brain is already trying to learn, so it will be changing. It's a non stationary system. Right. Which is why we call it a co adaptive system. And that's where it becomes tricky. So if it's only one side is adaptive. Right. Let's say the brain is adaptive but then you have a fixed mapping for your stimulation pattern. Then that the all the burden is on the brain to learn that. Right. But then if you do it where the both the systems are able to adapt, then the challenge is can you make it. You know, they're both optimizing the same cost function or same optimization function. And that's where the interesting engineering challenges. Right. To make it a co. Adaptive. Co evolving system. Right. And that's where I think the future is probably going to go if you are going to use AI for these brain computer interfaces. [01:06:20] Speaker B: You want to do a quick segment of epiphenomenon or causal. [01:06:26] Speaker A: What was that again? [01:06:27] Speaker B: You want to do a quick segment of. Is this an epiphenomenon or is it causal? [01:06:32] Speaker A: Oh, okay. Like a quiz. [01:06:33] Speaker B: Ready? Yeah, yeah. Well not. Yeah, opinion. Ready. Oscillations. [01:06:39] Speaker A: It's. It's. I would say it's causal. Yeah. [01:06:43] Speaker B: Consciousness. [01:06:46] Speaker A: Epiphenomenon. [01:06:47] Speaker B: Oh my God, we're done. I just, I just, I just heard someone say that. [01:06:52] Speaker A: But I'm taking the like the most liberal definition there. [01:06:57] Speaker B: Okay. Those are the only two. I just, I heard someone I respect say consciousness was epiphenomenal and then you were talking about oscillations. [01:07:04] Speaker A: So I just want to be a little bit on that cutting edge there. Let's go with that. [01:07:12] Speaker B: Great. That's staying in by the way. I'm not going to edit that out unless you force me to. [01:07:16] Speaker A: Sure. [01:07:21] Speaker B: I'm looking at the time right now. First of all, what have we missed about the neural. I'm going to point everyone to all of your work but punchy and jumping around. So what have we missed in terms of the neural co. Processing and. Or you know, brain net and the. We haven't used the term telepathy. Right. Which is sort of the coup de grace, the pinnacle. You know what people? The spooky term, right. That these things imply. I mean, is that a goal? Well, I'm going to stop interjecting. What have we missed in terms of what we haven't spoken about with the neural coprocessors? [01:08:06] Speaker A: Yeah. So the neural coprocessor idea, I mean there was relationships to the brain to Brain communication work that we had done early on. Right. And that work, of course, got us, you know, both people that were, you know, quite interested and quite intrigued by it. But also I think a lot of scientists were like, oh, yeah, that's like, you know, just the old using tms, just like frog's legs being stimulated. Right. So we got people saying, you know, why did you even do that? Right. I think the. [01:08:33] Speaker B: Oh, really? [01:08:34] Speaker A: Yeah, yeah. So the. So the main point is it's not that we were claiming that, hey, you know, this is like a grand new discovery or invention. Right. The idea was to say, okay, you know, we have been talking about things like mind to mind, you know, communication or telepathy for a long time in science fiction. Right. And then with BCI really progressing at the rate that it's progressing with companies getting interested in it, this may be a good time to have that conversation. Right. So. And I think one of the interesting things there is, you know, we may actually get something like very primitive forms of brain to brain communication if these companies have, you know, multiple patients with implants. Right. Because it doesn't take much to, you know, send really simple kinds of information from one person's brain to another as soon as you have the ability to stimulate. Right. Which is something I haven't seen yet. But once that starts happening, then I think it's only a small step towards starting some really primitive kinds of. What you may call, like, I mean, I'm sure the companies will claim is telepathy, right? [01:09:34] Speaker B: But they'll be called tele. What will they be called? Because there was the telephone beforehand. What would be a good company name for that? [01:09:42] Speaker A: Right, Right. Yeah. So for this one, it will be like teleneuronics or something like that, right? [01:09:48] Speaker B: Yeah. Why can't we come up with something, Raj? [01:09:52] Speaker A: I'll think of it trademarked first before we can talk about it. Right? No, I'm just kidding. [01:09:57] Speaker B: So I know you have it. You have it in your back pocket and you're not letting. Letting on. [01:10:02] Speaker A: But yeah, I think the unfortunate consequence of some of these kinds of claims by companies and even I think some of the experiments that we've done is that there's a lot of people who are probably schizophrenics who think that they are implanted with these devices already. And so a lot of us get emails every day, right. From people saying, hey, you know, can you help us? We think the government has implanted something in me or there's a chip in. [01:10:31] Speaker B: My brain or the Lord did you know or, yeah. [01:10:37] Speaker A: I mean, people actually even come to campus. Right. So there's people who've driven miles just to come to campus, to our center, to say, you know, hey, the doctor says there's nothing there. Here's an X ray. There's nothing in the X ray. But I firmly believe it's this microscopic thing that's in my brain that's controlling my actions. Right. So, unfortunately, I think that's, you know, you know, the price you have to pay. If you start to push that frontier. There's going to be people who, you know, also think that, you know, that's happening to them, even though we can assure them that that's not really possible at this point in time. Right. And we don't have the technology to do something like that. But that is something we're faced with as well. So it's part of the whole ethical, you know, issue and communication issues with this field. [01:11:19] Speaker B: Yeah. So you said that we can't telepathically communicate yet and we can't stimulate each other's brains, but we do all the time. You and I are doing it right now through language. Yep. And you have an interest in, I don't know where your interest in the Indus script came from and. Or now you've, at some point you learned a ton about language and just how language works and some of the statistical properties that are indicative of what a language is. But. So is this a hobby of yours? It's beyond a hobby now, but how did you get involved in the Indus script and determining. Trying to decipher this thing? [01:12:08] Speaker A: Yeah, I've always been interested in ancient history. And so it was during one of my sabbaticals, I think, my first sabbatical, when I said, okay, I indulge in that, you know, fascination with ancient history. Right. And especially ancient Indian history, because I'm from India and we have the civilization there that's, you know, more than 4,000 years old that we know very little about, partly because the script of that civilization is not yet deciphered. Right. And this was a civilization that was, you know, much bigger than the Egyptian and the Mesopotamian civilization. [01:12:38] Speaker B: Okay, so when was my. So I took. I'm sorry to interrupt, but one of my favorite classes that I took when I was a. An undergraduate was the Origins of Writing. Denise Schmalt Bessel ET taught the course, and we learned all about cuneiform and how early writing might have been brought about by economic exchanges with little clay tablets where people put little marks to show how many goats or cattle that need to be owed or exchanged and stuff. Where does this. And I can't remember how. So what you're talking about is for. There's like 2000 B.C. [01:13:18] Speaker A: Yeah. Before that, I think starting from 2000, you know, probably 2800 B.C. is when they believe the civilization started evolving to the point where it was mature enough and then it ended. Or, you know, the demise of the civilization is thought to be more like 1800 BC like 2000-1800 BC. Right. But the curious thing about them is that they. That was not. Yeah. It was not like a huge amount of time. And there were some precursors, you know, 3000 B.C. and so on, but the prime of the civilization when they were really, you know, in their prime, was probably, you know, 2800-1800 or so, like 1500 BC. Right. And I think the interesting thing about the civilization is the fact that they. I mean, they were actually trading actively with the Mesopotamian civilization. Right. The Sumerians and. And Akkad and those civilizations in the Middle East. And they've found these seals you mentioned, they had these stamp seals so you can stamp information about, typically trade the kind of goods that they may have been stamped on. So I think the intriguing thing there is, of course, we know cuneiform and the kind of history of writing did evolve from the Middle east, and then the Egyptians, of course, followed suit with their hieroglyphics. And it's believed that the Indus civilization also evolved their own set of symbols, about 500 to 700 symbols that are very pictographic in many cases. And unfortunately, there are no long inscriptions. So most of the inscriptions they're found have been on these stamp seals, which are only about five symbols long on average. [01:14:59] Speaker B: Some of them kind of squished in different. [01:15:02] Speaker A: Yeah. So there was like a. There's usually an animal imprint which may have been like a symbol of a clan. Like a unicorn is the most common animal symbol, really? Yes. A unicorn. Yes. But, you know, people in the ancient times thought it originated in India. Right. So unicorns. [01:15:17] Speaker B: I didn't know that. I didn't know that. [01:15:19] Speaker A: Yeah, yeah, yeah. So perhaps it was because of the Indus civilization and their seals. Right. But the seals always had at the very top a set of symbols, a sequence of symbols which were typically, on average, length five symbols long. But there have been no long inscriptions on tablets. And so unfortunately, we cannot apply many of our AI techniques. Right. Because there's only about 6,000 of these. Right. [01:15:43] Speaker B: How do you tokenize that? [01:15:44] Speaker A: Yeah, right. But what we did do, we did do some kind of analysis of it, which is we looked at the entropy, the block entropy of these symbol sequences in terms of which symbol follows which other symbol sequence. And then we show that the block entropies tend to match the block entropies of linguistic scripts, as opposed to DNA, for example, or proteins, which are much more random. They're more flexible. So the entropies are much higher. Music is slightly above language, so music is more flexible, but not as much as DNA or proteins. [01:16:21] Speaker B: What's less flexible? [01:16:22] Speaker A: Least flexible is, you know, computer programming languages like Fortran. Right. It's below. [01:16:28] Speaker B: You're going to say Fortran. [01:16:30] Speaker A: Yeah, that was the original one that they did. Right. So we followed that. But I'm sure it may start to fall. I mean, especially with now more different kinds of programming languages. Naturalistic, naturalistic. You're probably getting closer to the entropy of natural languages. Right. But it's an interesting way to think about it. So what we did was said, okay, here's an undeciphered script. Script. Let's look at all the symbol sequences and see if they fall in any of those ranges. And it turned out that they fall right in the middle of the range of linguistic scripts, which means that perhaps there's at least one piece of evidence that they may have used it to encode language. But we still don't know what language that was. There's a huge political debate in India about whether it was the Indo Aryan language, Sanskrit as the Vedic, you know, civilization, or is it. The South Indians have a Dravidian language family. So some people think it's that language that's encoded in there. [01:17:23] Speaker B: So. So political factions are trying to claim it or something is. [01:17:28] Speaker A: In some sense. Yeah, in some sense there's that. There's also this notion of, you know, Aryan invaders coming to India. People don't like that. Right. Any kind of migration to India, especially if it was these Indo European, you know, invaders coming to India. That was the original colonial interpretation of it. But the genetic evidence seems to indicate there was actually migrations happening from Central Asia into India. And that's possibly how these languages like Sanskrit and many of the North Indian languages have similarities in grammar with the Indo European language family. Compared to the south, which is more Dravidian language family, which is not related to the North Indian language families. So there's some amount of friction there. And even now you'll see people claiming decipherment of the Vedas, right. In the, in the seals. But it stretches. [01:18:18] Speaker B: But you're an expert now, so. So you can judge those things, right? [01:18:22] Speaker A: To the. To. To some extent. Yeah. I mean, I've definitely delved into that for a long time now. So I can. I can look into it. And unfortunately, most of the time you can see that, you know, they're. They're stretching it, and it's just not a plausible decipherment. Right. So I think you're just gonna have to wait for more excavations and maybe a Indus rosetta stone that'll help us decipher that script. [01:18:44] Speaker B: But I doubt we'll find it. Do you think we'll find it? [01:18:47] Speaker A: I mean, it's a difficult task. I mean, it's possible it's already been found and it's just lying in some warehouse in Iraq somewhere in a museum, because they were trading with these folks. So maybe there were some inscriptions that had both cuneiform and the Indus script, but so far nobody has found anything like that. They have found in the scripts with really odd sequences. So it seems like they were using the Indus script to express the language in Iraq. Like a foreign language, maybe in. Oh, yeah. So that's really interesting. [01:19:18] Speaker B: But they're only with, like, five characters. [01:19:20] Speaker A: Yes, five characters. But the symbol sequences are very different from what you'd find in the Indus Valley in the finds there. So they were using it in a different way at least. So I think there are some really intriguing suggestions like that. Right. [01:19:33] Speaker B: So is this going to be a lifelong project for you? [01:19:36] Speaker A: For me, I mean, it's. I think at this point it's, you know, it kind of. It went from being a hobby to more of, you know, really getting drawn into the literature and the history of it. And so I do spend a lot of time reading about it, but. Yeah, but I have no funding for it. And it's, you know, it's more of a sabbatical project. Right. But it keeps me interested in something very different from what I typically do. I sort of look at myself as somebody who's interested in, you know, the past, which is this Indus script. And then the present is like, okay, let's figure the brain out. And then the future is like the BCI AI augmentation. So I like to keep, you know, looking at all of those aspects. Right. And keeps me. Keeps my brain. Brain at least, you know, occupied and interested. [01:20:18] Speaker B: So one obvious question to me then is, I mean, not many people who use or. Sorry, not many people who engineer large language models are linguists or have a sense of what language is or how it may have come about or its early post origins et CETERA or even its origins. I mean, has this affected how you think about large language models? And I'll preface this also by saying that whatever great AI model is working, people are going to say, well, maybe the brain is like that, and then map whatever AI model is currently working to brain processes. And often it works and stuff. So has this affected how you think about large language models? Or how do you situate how you think about large language models? Maybe is a better question. [01:21:14] Speaker A: Yeah, I was quite intrigued. I mean, when the paper came out. The attention is all you need paper. And then also the fact that prediction was a very key aspect of it. Right. It's not. [01:21:25] Speaker B: Yeah. And you must have been elated. [01:21:27] Speaker A: Yeah. So it was like it was basically predicting coding, except for the fact that they were not using the prediction errors to do inference. I mean, there was no update of the internal representation, but the learning was obviously driven by prediction errors. And the fact that there was a hierarchy there as well, so that there's these soft attention layers which is building up these dynamic representations over time and. [01:21:50] Speaker B: You'Re doing sending out the representation to every layer. I mean, it's just massive. [01:21:56] Speaker A: Right, exactly. [01:21:58] Speaker B: But it's sort of in some sense beautiful. Probably to you. [01:22:02] Speaker A: Yeah. It was actually very interesting to know that there's now this artificial system that just purely based on prediction. Right. By predicting the next word, it was able to do so much, not even. [01:22:15] Speaker B: Recurrent, like there's no feedback. Yeah. [01:22:17] Speaker A: It's not a regressive model, but. Yeah, yeah, you're right. But effectively it was like taking into account everything in the past. Well, processing the future and the future and the past were all getting processed simultaneously. Right. But I think the cool thing there was the fact that it was able to really learn these very impressive predictive models and then generate these sequences which seem to be quite in line with the kind of thing we were thinking about for our. In our active predictive coding model. This idea that there's a sensory processing aspect that's constantly predicting the next state based on the previous state and action. But there was no action here. Right. The action was missing. Right. [01:22:58] Speaker B: Action is just output. [01:22:59] Speaker A: It's just like a token. Basically. The previous token may be like being put in or all the tokens in the past are being used to process the current token. Right. But I think that that's also something that's missing. My sense is the controller is missing. Right. So in the active predictive coding architecture, there's a policy network that is controlling the generation of the next input. We have the predictive network and the controller network, and they're closely interacting with each other. And depending on the task, the controller network is feeding in the action to the predictive network, saying, okay, generate this next, because I'm going to take this action. I think we've lost that in the transformer model. I think that's where some of these RL approaches they're applying now to the transformer models. It's kind of getting the controller back in the picture. So I think that's where I see a relationship. And more recently, people have shown that transformer networks implement a special kind of hyper networks, which is what we are using for active predictive coding. So when I said that the higher areas are modulating the lower areas in active predictive coding, right. The higher cortical areas may be modulating the function being computed at the lower areas. That is basically what is called a hyper network in the AI literature, where the vector, an input at the higher level is transformed by a neural network called a hyper network, to another neural network, which is the function. It's basically a whole new network is being generated. And it appears that transformers are doing something like that in a simple way, like a linear hyper network model. So I think if that's true, then I would say there's probably a closer link between what we've suggested and what Transformers are doing. Except we were trying to model the cortical architecture in terms of hierarchy and laminar structure. Transformers don't have that. Right. And they don't have the controller aspect either. [01:24:50] Speaker B: Yeah. So then overall, are transformers. I'm trying to get a sense of your judgment on them. [01:25:04] Speaker A: I would say they're great, but they're not enough. Right. They're great as a principle of how predictive models can really capture a lot of information about the. Essentially the statistics of the world. Right. The physics of the world or the dynamics of the world. They can be used to capture that, but that's not enough for intelligence. Right. So I would say I go back to that notion that people have suggested, Paul Chizzek and Buzaki and others, that ultimately it's all about action and movements. And so if you don't have the controller there telling you what to do, Right. Then you, like, Transformers don't have, like ChatGPT and Transformers, they don't have a sense of agency. Right. And that. What's that? [01:25:46] Speaker B: How do you get that? How do you get a sense of agency? [01:25:48] Speaker A: Well, I think we get it from being able to act in the world and being able to actively use Action as a way to generate language. Right. So, and we get that because we actually act upon the world and we, you know, we use speech. Right. And that's actual motor actions. Right? [01:26:03] Speaker B: Yeah, but squirrels have agency. Right. And they don't have speech. [01:26:07] Speaker A: Yeah, yeah. So you have to have action, you don't have to have speech. But in order to have agency, you need to be able to act upon the world and be able to, you know, cause consequences. You have to be able to see the consequences of your action and then maybe achieve goals. So you start to see, okay, you know, if I take these actions, I achieve my goals and then if I do these actions, I get something else. And that's how you build a sense of agency. [01:26:28] Speaker B: Right. [01:26:28] Speaker A: And for a purely predictive system, which is only predicting, but not really using actions as a way to guide those predictions and get to a particular goal. Right. Then it becomes hard to argue it has agency. I mean, some people do, but I think it's a missing piece. Right. [01:26:45] Speaker B: So you think that we could build agency. [01:26:49] Speaker A: Yeah, I think any system that starts to interact with the world with its own actions and generate goals, I mean, those are all things that you associate with an agent. Right. So I think you do require to have that policy or controller and you know, potentially a model of the world to go along with it. [01:27:10] Speaker B: I have two more questions for you and I promise I'll let you go. [01:27:13] Speaker A: Oh, it's been fun. [01:27:14] Speaker B: Way earlier. I'm sorry. [01:27:16] Speaker A: Oh, it's been fun. So I could probably go on. So. Yeah, go ahead. [01:27:19] Speaker B: Okay. [01:27:20] Speaker A: But you're tired. You just came back. [01:27:22] Speaker B: I have to be up for the. I have to. I'm doing a brain initiative workshop tomorrow where I participating, so I have to like. Yeah, one. So way earlier on, we were talking, when we were talking about active predictive coding or predictive processing in general, you were talking about, you know, back then, back in 1999, there, you know, when we were, when people. I said we were recording single neurons. And that's when I grew up, I recorded single neurons. And the technology wasn't quite there to be able to begin testing the hypotheses that might be issued from a theoretical framework like predictive coding. Since then, we now have, we're starting to get connectomes which everyone always said, like, well, man, once we have a map of every. Of all the connections, we can do everything. And that is starting to come to pass. We now have super high density recording techniques, neuronal recording techniques, so that we can put high density electrodes in different Parts of the brain and get lots and lots of neurons. But between those two. Okay, and then now and we have. Sorry, there's like five things, right? Those two. And then we have AI models, we have high compute, we have lots of more. A lot more. A lot more statistical models that we can employ. That's the same as AI in terms of modeling. Which of those, if any like is. Has been. Do you think is the most important that is happening right now? [01:29:07] Speaker A: That's a trick question. Because I think if I choose one, it's probably going to alienate a whole bunch of people working on the others. I would say we need. [01:29:13] Speaker B: Well, no, no. Okay, I'll ask it differently, I'll ask it differently. Which of those do you like? When you wake up, you're like, oh man, I could use this to do that. In your current. I know, they're all necessary and awesome. [01:29:25] Speaker A: Yes. Yeah, I think, I mean, it changes, right? So for me, the amazing part, I mean, it's a fantastic time to be a theoretician in neuroscience right now because there's so much data, right. And this is a great time to really think about these larger scale theories. So when I was a grad student, there were books and papers by Mumford, There was a book called Large Scale Neuronal Theories of the Brain by Christoph Koch and Joel Davis. And these were all like, at that time, you know, there were these theories, but then the data just was not there to test these theories. But now I think we're at a time where we can actually start to test some of these theories. So I would say, you know, from one day to another. Right. I feel like some days when I'm thinking about a particular theory, I may go into the literature on large scale neural recordings, really try to find, okay, is this supported or not? And what are the things out there? But then the next day I'm actually looking at connectomics to see, okay, is the connectivity and anatomy still, you know, is that there. Right. To support some of these other theories. Right. And of course, AI is a wellspring of new ideas always. I mean, there's so much going on in AI that is hard to keep up. Right. But then of course, there's some gems there that you can pick out and then see if that's relevant to understanding the brain. So I think it just feels like we're at a time where it's this explosion of information and really to make advances, it's almost like you have to train your brain to essentially forage and find the right kinds of information to really build up new theories. It's a tough task, but I think it's actually better than not having any data. When I was a grad student, we didn't really have that kind of data. We had, like you said, single neuron responses and so on. It was very difficult to say, okay, what, what's the system doing? Right. At the systems level now, I think we're getting to the point where we can actively collaborate. I think your generation, the generation that's actually training right now, they're much more, I think, savvy with computational models. They're much more receptive computational models. I think that's really great. That means that these ideas, like predictive coding, hopefully can be tested more quickly and they're not going to lie dormant for 10 years before people start taking an interest in them. So I feel like this is a great time for people. I encourage people in many different areas. They can be computer scientists, they can be people doing AI, they can be people in neuroscience psychology. I think they should all feel comfortable with coming up with theories and computational models. I think we have the training now in many areas, but this is a great time to start thinking big. Think across just one area and one region. Right. [01:32:01] Speaker B: Can you. So there's maybe a paradox here, right? So if you have tons and tons of tools, it may be harder to think big. Like, it may be harder to think theoretically. I'm wondering how you might think to foster someone's theoretical bent or how to think theoretically given this deluge of tools. [01:32:24] Speaker A: Yeah, so that's a great question again. So I think it's something that you, for me, at least, it comes from thinking about it first from a computational point of view, saying, okay, here's a particular problem that, let's see if we think of solving this problem. How would you do that from a normative point of view or from an AI point of view. Right. But then I think you can then say, okay, let's look at some of the data from the brain in terms of where these kinds of problems have been explored. Right. And then that's one way to do it is to start from the computational and then go more deeper and deeper into the neuroscience. The other way is to go bottom up. Right? So you could say, okay, you know, I'm working on these areas, but then, you know, I want to actually go beyond just this one area from a theoretical point of view. Right. Even though I'm just recording from this area. But let's think of it in terms of this area interacting with all the other areas that are connected to it and then the bigger picture of behavior. Right. And how is that going to work? Right. And so maybe look at people working in those other areas as well. Right. It is a hard problem, but I think it's worth, you know, looking at it beyond just. I remember, like, when I was a postdoc, I went to the lab of a very famous neuroscientist. And then, you know, I was doing a postdoc in Sanowski's lab in sag, but I was visiting another laboratory somewhere else. [01:33:44] Speaker B: Okay, right. That's a different Famous nerds. [01:33:46] Speaker A: Yes, yes, it was another. I was. I was doing a postdoc in Terry's lab, but then I went to a different. Visiting a lab. Right. Many different labs were visiting at that time as a theoretician. And then I was asking them, you know, hey, you know, the person was working in V1. And then. So I said, okay, what about V2? Right. I mean, V2 and V1 are connected, you know, shouldn't you be considering the interactions between the two? But the person said, no, no, I'm going to figure out V1 first after I understand V1, and then I'm going to understand V2. Right. But what if you cannot do that? What if you cannot reduce. It's not reducibility in that sense. What if V1's properties are intimately connected with V2 and other structures? Then it becomes really hard to understand just one area. I think that's the challenge we have as neuroscientists, is the brain is a very complex piece of machinery that's been evolving over millennia. And if you use reductionism, then it become. Becomes very hard. Right. To understand what's happening across the brain. At the same time, maybe we don't have the capacity to understand the whole brain, like what's happening everywhere. So then we have to find ways of picking the right abstraction level to understand and hopefully connect those abstraction levels, like from the behavior level all the way down to the molecules. Right. But one person may not be able to do all of that. But hopefully as a community, we can start to understand the brain at multiple levels of understanding. [01:35:13] Speaker B: It's hard. It's hard. And even people who like you, who are like super good at it, you will admit it, it's a little difficult. [01:35:21] Speaker A: Definitely. Yeah. Especially with all the information that's now coming out and trying to. But I don't think we can lose hope. Right. So I'm optimistic that it's going to happen. Right. I don't know if it's in my lifetime, but I think we're on our way. Right. It may be sooner than we think. Right. So I think that we should be optimistic about that. [01:35:38] Speaker B: All right, well, that's a great note to end it on. I meant to mention this up at the top. What I'm going to do when we're done here, I'm going to email you an invitation to join me again, because I know it takes about a year to get you on, so it took a long time to get you on. But man, I really appreciate it. Raj, you coming on. [01:36:01] Speaker A: It was fun. Yeah, it was fun. Thanks again for having me. [01:36:09] Speaker B: Brain Inspired is powered by the Transmitter, an online publication that aims to deliver useful information, insights and tools to build bridges across neuroscience and advanced research. Visit thetransmitter.org to explore the latest neuroscience news and perspectives written by journalists and scientists. If you value Brain Inspired, support it through Patreon to access full length episodes, join our Discord community and even influence who I invite to the podcast. Go to Brain Inspired Co to learn more. The music you're hearing is Little Wing performed by Kyle Donovan. Thank you for your support. See you next time.

Other Episodes

Episode 0

February 07, 2023 01:28:39
Episode Cover

BI 160 Ole Jensen: Rhythms of Cognition

Support the show to get full episodes and join the Discord community. Check out my free video series about what's missing in AI and...

Listen

Episode 0

April 03, 2022 01:17:20
Episode Cover

BI 132 Ila Fiete: A Grid Scaffold for Memory

Announcement: I'm releasing my Neuro-AI course April 10-13, after which it will be closed for some time. Learn more here. Support the show to...

Listen

Episode 0

May 17, 2022 01:34:12
Episode Cover

BI 136 Michel Bitbol and Alex Gomez-Marin: Phenomenology

Support the show to get full episodes and join the Discord community. Check out my short video series about what's missing in AI and...

Listen