(PhysOrg.com) -- Dr. Ben Goertzel is Chairman of Humanity+; CEO of AI software company Novamente LLC and bioinformatics company Biomind LLC; leader of the open-source OpenCog Artificial General Intelligence (AGI) software project; Chief Technology Officer of biopharma firm Genescient Corp.; Director of Engineering of digital media firm Vzillion Inc.; Advisor to the Singularity University and Singularity Institute; Research Professor in the Fujian Key Lab for Brain-Like Intelligent Systems at Xiamen University, China; and general Chair of the Artificial General Intelligence Conference Series. His research work encompasses artificial general intelligence, natural language processing, cognitive science, data mining, machine learning, computational finance, bioinformatics, virtual worlds and gaming and other areas, Dr. Goertzel has published a dozen scientific books, 100+ technical papers, and numerous journalistic articles, and the futurist treatise A Cosmist Manifesto. Before entering the software industry he served as a university faculty in several departments of mathematics, computer science and cognitive science, in the US, Australia and New Zealand.
Dr. Goertzel spoke with Critical Thoughts Stuart Mason Dambrot following his talk at the recent 2011 Transhumanism Meets Design Conference in New York City. His presentation, Designing Minds and Worlds, asked and answered the key questions, How can we design a world (virtual or physical) so that it supports ongoing learning and growth and ethical behavior? How can we design a mind so that it takes advantage of the affordances its world offers? These are fundamental issues that bridge AI, robotics, cyborgics, virtual world and game design, sociology and psychology and other areas. His talk addressed them from a cognitive systems theory perspective and discussed how theyre concretely being confronted in his current work applying the OpenCog Artificial General Intelligence system to control game characters in virtual worlds.
This is the first part of a two-part article. The second part is available at http://phys.org/news/2011-06-dr-ben-goertzel-artificial-intelligence_1.html
SM Dambrot: Were here with Dr. Ben Goertzel, CEO of Novamente, Leader of OpenCog and Chairmen of Humanity+ [at the 2011 Humanity+ Transhumanism Meets Design Conference in New York City]. Thank you so much for your time.
Dr. Goertzel: Its great to be here.
SM Dambrot: In your very interesting talk yesterday, you spoke about the importance of the relationship between minds and worlds. Could you please expound on that a bit in terms of Artificial General Intelligence?
Dr. Goertzel: As an AGI developer this is a very practical issue which initially presents itself in a mundane form but many subtle philosophical and conceptual problems are lurking there. From the beginning, when youre building an AGI system you need that system to do something and most AI history is about building AI systems to solve very particular problems, like planning and scheduling in a military context, or finding documents online in a Google context, playing chess, and so forth. In these cases youre taking a very specific environment a specific set of stimuli --- and some very specific tasks -- and customizing an AI system to do those tasks in that environment, all of which is quite precisely defined. When you start thinking about AGI Artificial General Intelligence in the sense of human-level AI you not only need to think about a broader level of cognitive processes and structures inside the AIs mind, you need to think about a broader set of tasks and environments for the AI system to deal with.
In the ideal case, one could approach human-level AGI by placing a humanoid robot capable of doing everything a human body can do in the everyday human world, and then the environment is taken care of but thats not the situation were confronted with right now. Our current robots are not very competent when compared with the human body. Theyre better in some ways such as withstanding extremes of weather that we cant but by and large they cant move around as freely, they cant grasp things and manipulate objects as well, and so on. Moreover, if you look at the alternatives such as implementing complex objects and environments in virtual and game worlds you encounter a lot of limitations as well.
You can also look at types of environments that are very different from the kinds of environments in which humans are embedded. For example, the Internet is a kind of environment that is immense and has many aspects that the everyday pre-Internet human environment doesnt have: billions of text documents, satellite data from weather satellites, millions of webcams but when you have a world for the AI thats so different from what we humans ordinarily perceive, you start to question whether a AI modeled on human cognitive architecture is really suited for that sort of environment.
Initially the matter of environments and tasks may seem like a trivial issue it may seem that the real problem is creating the artificial mind, and then when thats done, theres the small problem of making the mind do something in some environment. However, the world the environment and the set of tasks that the AI will do is very tightly coupled with what is going on inside the AI system. I therefore think you have to look at both minds and worlds together.
SM Dambrot: What youve just said about minds and worlds reminds me of two things. One is the way living systems evolved that is, species evolve not in a null context, but rather, as you so well put it, tightly coupled to in this case an environment niche; every creatures sensory apparatus is tuned that niche, so the mind and world co-evolve. The other is what you mentioned yesterday when discussing virtual and game worlds that physics engines are not being used in all interactive situations which leads me to ask you what you think will happen once true AGIs are embodied.
Dr. Goertzel: If we want to, we can make the boundary between the virtual and physical worlds pretty thin. Most roboticists work mostly in robot simulators, and a good robot simulator can simulate a great deal of what the robot confronts in the real world. There isnt a good robot simulator for walking out in the field with birds flying overhead, the wind, the rain, and so forth but if youre talking about what occurs within someones house a lot can be accomplished.
Its interesting to see what robot simulators can and cant do. If we were trying to simulate the interior of a kitchen, for example, a robot simulator can deal with the physics of chairs and tables, pots and pans, the oven door, and so forth. Current virtual worlds dont do that particularly well because they only use a physics engine for a certain class of interactions, and generally not for agent-object or agent-agent interactions but these are just conventional simplifications made for the sake of efficiency, and can be overcome fairly straightforwardly if one wants to expend the computational resources on simulating those details of the environment.
If you took the best current robot simulators, most of which are open source, and integrated them with a virtual world, then you could build a very cool massive multiplayer robot simulator. The reason this hasnt happened so far is simply that businesses and research funding agencies arent interested in this. Ive thought a bit about how to motivate work in that regard. One idea is to design a video game that requires physics for example, a robot wars game in players build robot from spare parts, and the robots do battle. You could also make the robots intelligent and bring some AI into it, which if done correctly would lead to the development of an appropriate cognitive infrastructure.
Having said that, going back to the kitchen what would current robot simulators not be able to handle, but would have to be newly programmed? Dirt on the kitchen floor, so that in some areas you could slip more than others; baking, so when you mix flour and sugar and put it in the oven, the chemistry which is beyond what any current physics engine can really do; paper burning in the flame of a gas stove; and so on. The open question is how important these bits and pieces of everyday human life are to the development of an intelligence.
Theres a lot of richness in the everyday human world that little kids are fascinated by fire, cooking, little animals because this is part of the environmental niche that humans adapted to. Even the best robot simulators dont have that much richness, so I think that its an interesting area to explore. I think we should push simulators as far as we can, create robot simulators with virtual worlds, and so forth but at the same time Im interested in proceeding with robotics as well because theres a lot of richness in the real world and we dont yet know how to simulate it.
The other thing you have to be careful of is that most of the work done with robots now completely ignores all this richness -- and Im as guilty of that as anybody, When we use robots in our lab in China do we let the robots roam free in the lab? Not currently. We made a little fenced-off area, we put some toys in it, and we made sure the lighting is OK because the robots were using (Aldebaran Nao robots) cost $15,000 and they have a tendency to fall down. Its annoying when they break you have to send them back to France to get repaired.
So, given the realities of current robot technology we tend to keep the robots in a simplified environment both for their protection, and so that their sensation and actuation will work better. They work, theyre cool, and they pick up certain objects well but not most of those in everyday human life. When we fill the robot lab only with objects they can pick up, were eliminating a lot of the richness and flexibility a small child has.
SM Dambrot: This raises two more questions: Is cultural specificity required for any given AGI, and is it necessary to imbue an AGI with a sense of curiosity?
Dr. Goertzel: Our fascination with fire is an interesting example. You wonder to what extent its driven by pure curiosity versus our actual evolutionary history with fire something thats been going on for millions of years. I think our genome is programmed with reactions to many things in our everyday environment which drive curiosity and fire and cooking are two interesting examples.
Having said that, yes, curiosity is one of the base motivators. Were already using that fact in our OpenCog work. One of the top-level demands, as we call them, of our system is the ability to experience novelty, to discover new things. There are two demands: to discover new things in the world around it and just have the experience of learning new things internally, which can come through external or internal discovery. So weve already programmed things very similar to curiosity as top-level goals of the system. Otherwise you could end up with a boring system that just wanted to get all of its basic needs gratified, and would then just sit there with nothing to do.
SM Dambrot: Thats very interesting especially the internal novelty drive. That seems even more exciting in terms of any type of AGI analogue to human intelligence, because we spend so much time discovery ideas internally.
Dr. Goertzel: Some people more than others its cultural to some extent. I think we as Westerners spend more time intellectually introspecting than do people from Eastern cultures. Being from a Jewish background, I grew up in a culture particularly inclined towards intellectual introspection and meta-meta-meta thinking.
On a technical level, what weve done to inculcate the OpenCog system with a drive for internal novelty and internal learning and curiosity is actually very simple: Its based on information theory and is related to work by Jürgen Schmidhuber and others on the mathematical formulation of surprise. In an information-theoretic sense, OpenCog is always trying to surprise itself.
SM Dambrot: I recall that when Prof. Schmidhuber was discussing Recursive Neural Networks at Singularity Summit 09, he talked about how the system looks for that type of novelty in its bit configurations.
Dr. Goertzel: Thats right and what we do with OpenCog is quite similar to that. These are ideas that I encountered in the 1980s in the domain of music theory, based on Leonard Meyers Emotion and Meaning in Music. He was analyzing classical music Bach, Mozart and so forth and the idea he came up with was that aesthetically good music is all about the surprising fulfillment of expectations, which I thought was an interesting phrase. Now, if something is just surprising its too random, and some modern music can be like that modern classical music in particular. If something is just predictable pop music is often like that, and some classical music seems like that its boring. The best music shows you something new yet it still fulfills the theme in a way that you didnt quite expect to be fulfilled so its even better than if it just fulfilled the theme.
I think thats an important aesthetic in human psychology, and if you look at the goal system of a system like OpenCog, the system is seeking surprise but it also gets some reward from having its expectations fulfilled. If it can do both of those at once then its getting many of its demands fulfilled at the same time, so in principle it should be aesthetically satisfied by the same sorts of things that people are.
This is all at a very vague level, because I dont think that surprise and fulfillment of expectations are the ultimate equation of aesthetics, music theory or anything else. Its an interesting guide, though, and its interesting to see the same principles seem to hold up for human aesthetics in quite refined domains, and also for guiding the motivations of very simple AI systems in video game type worlds.
SM Dambrot: Ive been wondering about materials and the structure of those materials. Do you think its important or even necessary in any way to have something that is patterned on our neocortical structure neurons, axons, synapse, propagation in order to really emulate our cognitive behavior, or not so relevant?
Dr. Goertzel: The first thing I would say is that in my own primary work right now with OpenCog, Im not trying to emulate human cognition in any detail, so for what Im trying to do which is just to make a system thats as smart as a human in vaguely the same sort of ways that humans are, and then ultimately capable of going beyond human intelligence Im almost sure that its not necessary to emulate the cognitive structure of human beings. Now, if you ask a different question lets say I really want to simulate Ben Goertzel and make a robot Ben Goertzel that really acts, thinks, and hopefully feels like the real Ben Goertzel to do that is a different proposition and its less clear to me how far down one needs to go, in terms of emulating neural structure and dynamics.
In principle, of course, one could simulate all the molecules and atoms in my brain in some kind of computer, be it a classical or quantum computer so you wouldnt actually need to get wet and sticky. On the other hand, if you need to go to a really low level of detail, the simulation might be so consumptive of computing power, you might be better off getting wet and sticky with some type of nanobiotech. When you talk about mind uploading, I dont think we know yet how micro or nano we need to get in order to really emulate the mind of a particular person but I see that as a somewhat separate project from AGI, where were trying to emulate human-like human level intelligence that is not an upload of any particular person. Of course if you could upload a person, that would be one path to a human-level AGI its just that its not the path Im pursuing now, not because its uninteresting but I dont know how to progress directly and rapidly on that right now.
I think I know how to build a human-level thinking machine I could be wrong, but at least I have a detailed plan, and I think if you follow this plan for, lets say, a decade, youd get there. In the case of mind uploading, it seems theres a large bottleneck of information capture we dont currently have the brain scanning methods capable of capturing the structure of an individual human brain with high spatial and temporal accuracy at the same time, and because of that we dont have the data to experiment with. So if I were going to work on mind uploading, Id start by trying to design better methods of scanning the brain which is interesting but not what Ive chosen to focus on.
SM Dambrot: Regarding uploading, then, how far down do you feel we might have to go? Is imaging a certain level of structure sufficient? Do we have to capture quantum spin states? I ask because Max More mentioned random quantum tunneling in his talk, suggesting that quantum events may be a factor in cryogenically-preserved neocortical tissue.
Dr. Goertzel: Im almost certain that going down to the level of neurons, synapses and neurotransmitter concentrations will be enough to make a mind upload. When you look at what we know from neuroscience so far -- such as what sorts of neurons are activated during different sorts of memories, the impact that neurotransmitter levels have on thought, and the whole area of cognitive neuroscience -- I think theres a pretty strong case that neurons and glia and the molecules intervening in interactions between these cells and other things on this level are good enough to emulate thought without having to go down to the level of quarks and gluons, or even (as Dr. Stuart Hameroff suggests) the level of the microtubular structure of the cell walls of the neuron. I wouldnt say that I know that for certain, but it would be my guess.
From the perspective of cryogenic preservation, you might as well cover all bases and preserve something, so well that even if our current theories of neuroscience and physics turn out to be wrong, you can still revive the person. So from Max Mores perspective as CEO of Alcor, I think hes right you need to preserve as much as you can, so as not to make any assumptions that might prevent you from reviving someone.
SM Dambrot: Like capturing a photograph in RAW image format
Dr. Goertzel: Yes you want to save more pixels than youll ever need just in case. But from the viewpoint of guiding scientific research, I think its a fair assumption that the levels currently looked at in cognitive neuroscience are good enough.
This is the first part of a two-part article. The second part is available at http://phys.org/news/2011-06-dr-ben-goertzel-artificial-intelligence_1.html
Explore further: Rare early shakespeare compilation found in small French library