This week, we'll strip computer science down to it's components and find out what we should expect to see in the next 5 years. We find out about the thinking behind artificial intelligence, what the future holds for Second Life and how neuroscience can help us build truly intelligent computers. Plus, get your sunglasses out early this year for Kitchen Science where we make an LCD monitor vanish.
In this episode
Can we dig our way from the UK to Australia?
Chris - I think the answer is probably not - purely because of pressure and temperature constraints. Also the fact that the Earth is pretty liquid inside as far as the inner core. Therefore you'd have to contend with huge amounts of pressure, huge amounts of temperature and I don't think we have the energy or the building materials that would be capable of withstanding that. If you think about it the earth has a radius of 5000-6000km. If we're standing on the earth's surface you're feeling one atmosphere of pressure. The atmosphere above you is about 50km high. If you were going to go 5000km to the centre of the Earth you'd therefore have 100 times the greater amount of atmospheric pressure on you so the pressure just from the atmosphere would be so huge that just trying to move through that kind of gas would be like running into a brick wall. Very tricky indeed.
Disappearing monitors - the power of polarisation
Why do we stick to ice?
Chris - I've had a nasty experience because I used to work in a lab where we had lots of -70 freezers. We used to develop various gels and things in those -70 freezers. If you weren't careful and didn't put some gloves on when you went into the freezer and grabbed your developer out then your skin could freeze onto the rack. When you removed your fingers and let go then it left a lovely fingerprint on the thing. That quite literally was a fingerprint because it left the surface layers of skin from your finger. The reason ice is sticky is for that very reason. Ice itself is so cold if you touch it with skin because your body secretes tiny amounts of liquid, sweat, which is a salty fluid onto your skin surface it actually makes your skin stickier. This is why we have it. It's for grip. If you then touch that onto a very cold ice surface the ice then re-freezes the liquid on your finger. Because that liquid is a fluid and it has got into all the nooks and crannies on your finger it then freezes solid and will form a very tight bond between your finger and the frozen surface, the ice. You get stuck to the surface. If it's an ice cube - if it's okay because there's enough heat flowing through your fingers (usually to re-melt that transient freezing) then you can detach yourself. In the case of a -70 freezer or even colder, people in the Antarctic have to be very careful about this kind of thing. It doesn't warm up and you can end up permanently frozen to the surface or you can do quite a bad injury. That's why ice is sticky. You get literally frozen to the spot.
How does Blu-tack work?
Chris - I think there's two aspects to this. One is in the same way as the water gets in to all of the nooks and crannies both in the skin and the surface that you touch - in the same way as when you lick your finger to turn a page - you create more friction between your finger and the surface. The water forms an attachment on that surface. Blu-tack is plastic. In other words it can deform plastically. It gets into the nooks and crannies of the surface you're sticking it to and the surface it's already stuck to which makes it sticky. The other point is that when you press it onto a hard surface the Blu-tack forms a very smooth, flat surface against that hard surface. This excludes air. In order to get the Blu-tack back off the surface you've effectively got to break a vacuum. The atmosphere is helping to hold the Blu-tack stuck onto the surface. That's why I think it's sticky. Ben - That's the same reason why, if you get two sheets of glass and layer water between them and press them together to exclude the air it's really hard to peel apart even though water isn't sticky at all. Chris - Two microscope slides are almost impossible to separate if you put a drop of water and then put the two glasses together. You've got to twist them off each other and you can't pull them apart because the atmosphere is squeezing down on you. Every square metre is feeling a force of ten tonnes: the weight of a London bus. On our bodies every square metre of our body's got ten tonnes of atmospheric pressure on it. That's what's holding your leaflet stuck to the wall with Blu-tak.
23:46 - The Future of Computing
The Future of Computing
with Chris Bishop
Ben - It's hard to think of any aspect of our lives that isn't dominated by computers, and that shows little sign of stopping. Joining us now is Professor Chris Bishop, from Microsoft Research, to tell us about one way that computers may change in the future - they might become more intelligent. But first, just what do we mean by Computer Science?
Chris B - What I mean by computer science is really the set of underlying ideas; the mathematics of concepts that makes digital technology possible. I think it's also worth saying what computer science isn't. To my mind, computer science is not really about how to use computers. Something that we teach our children in school is ICT skills. We teach them how to use computers, how to use software. That's a very important set of skills. That itself is not computer science. It's not the top level, normal use of computers. The next level down is really how to modify computers, how to get them to do new things. That's really about programming and software development. That's something which I think would be great if we taught even more of that in schools. We've barely scratched the surface of that in school curriculum. It would be lovely to teach more of them. Underneath that again is this sort of third level, the foundation of the subject. That's computer science. That's the set of mathematics, the concepts that make it all possible. While computer technology evolves incredibly rapidly and next year you've a whole new set of gadgets and the year after it's another set of gadgets again. The concepts of computer science are much more fundamental. Some of them have been around for half a century and they evolve much more slowly. I think it would be just wonderful if we could show children: in particular a lot more of these basic long-term concepts.
Ben - It's really an understanding of the algorithms, the maths that lie at the very heart - completely regardless of the technology and what you're then doing with it. You need to understand how the numbers work in order to build up these layers that eventually get to a word processor or a shoot 'em up.
Chris B - That's exactly right. A famous computer scientist called Dijkstra once said that, "Computer science is no more about computers than astronomy is about telescopes." The ideas of computer science apply whether you're talking about silicon or whether you've built your computers out of gear wheels or whether you've built it out of DNA or chemistry in some way. These are very general concepts about the limits of computation and the capabilities of computation in the general sense.
Ben - As we've said you're the Chief Research Scientist at Microsoft Research. What is it that you do there? We think of Microsoft as being the people that make Windows that a lot of our computers run on. My own laptop runs on Windows. What do you do?
Chris B - Microsoft research which is about 1% of the whole in terms of people is the basic research components of Microsoft. You can think of us as, in many ways, functioning like academics. Our researchers have a complete freedom to choose the research they do, to publish it where and when they choose and to collaborate fully with academics and so on. At the same time they may also develop ideas that might have commercial relevance. Some of those ideas might be fed into the rest of Microsoft. Now, in Cambridge (which is the European Lab of Microsoft Research) we have about 120 scientists covering a very broad range of fields from many of the mainstream areas of computer science like security and systems and networking and programming language theory and so on. We're increasingly diverse. We have people in machine learning like myself. We have people in vision researchers, even ecologists and biologists these days. One of my main roles as chief research scientist is to promote cross-disciplinary interaction. There's a big danger in this very heterogeneous environment that everybody stays in their own little stove pipe and they don't really talk to each other. I try to find mechanisms to mix people up and get them to understand the ideas and problems in each other's fields. I hope that will promote interaction between the different disciplines because I'm a great believer that a lot of the really new ideas actually emerge at the intersection of different disciplines. Genuinely new ideas in science are actually quite rare but very hard to find. It's slightly less hard to take a good idea from someone else's field and import it into your own field where it may be quite novel and have quite an impact.
Ben - We had quite a good example of this on the show a few weeks ago with Alyssa Goodman who was using the software that's been developed for understanding and visualising MRI scans to look at deep space. It really is true this sort of multi-disciplinarianism is the way science is moving. Going back to something you said you were working on yourself: machine learning. The idea of intelligent machines has been around for a while but what do we really mean? How can we get a machine to learn?
Chris B - The idea of intelligence machines goes right back to the dawn of computing and certainly Alan Turing, the famous Cambridge mathematician who was one of the founders of modern computer science, was very interested in the idea of building intelligent machines. It's obviously enormously difficult problem. Technology in many respects has evolved incredibly quickly. You now have machines that are extremely good at multiplying numbers together, far better than humans. There are other tasks that people and animals are very good at that we tend to characterise as intelligence. Things such as visual recognition: simply looking at a visual scene and understanding what's going on in different objects, understanding what's happening has proved to be immensely difficult to build those capabilities into a computer. Over the years we've been making good progress. There's a long way to go before we have a machine that has genuine intelligence. In fact, although perhaps 30 years ago people sought to build machines that would be intelligent in a high-level sense (machines you could have a conversation with about philosophy or culture) we've sort of abandoned that. It's really just way too difficult. We've taken a bottom-up approach, more of a signal-processing approach. We say let's look at something of a simpler task that machines are not yet good at, let's say people aren't good at, and an example I gave already was recognising objects in the visual world. It's an immensely difficult problem. It appears to be trivial. I can turn around. I can say: well look there's a glass, there's a table, there's a chair, there's a microphone in front of me and so on. I can recognise these objects. Actually what your eye sees is a pattern of light and dark and colour but is miniscule-y variable. No two chairs look the same, the pattern of light in the retina varies as the chair is moved closer or farther away or rotated as light is scattered off other objects. It changes the colour of the light, changes the colour of the chair as this huge variability. Our brain, in think in some way we don't understand yet, mysteriously takes away all that variability and says, no that's still a chair. If we could build a similar capability into machines it would be very powerful technologically. It's an outstanding problem that we're just beginning to make good progress with. It's one of the fast-moving frontiers of machine intelligence at the moment.
Ben - It sounds exciting. It is true that we see a chair but there are so many different varieties that if we had to programme this into a computer it would be an enormous database of things they need to look at. That was Edinburgh University's Professor Chris Bishop. So in the future our computers may more intelligent, but that probably wouldn't stop people getting distracted by games of minesweeper!
31:00 - The Underbelly of Second Life
The Underbelly of Second Life
with Mike Hobbs
Chris - As you may have noticed, here at the Naked Scientists we're fans of Second Life. It's an interesting way to meet people from all over the world and do things like listen to great radio shows too. But the virtual world needs a great deal of technology and artificial intelligence to keep it going, as Meera Senthilingam recently found out when she met Mike Hobbs at a special evening which was held by Anglia Ruskin University as part of the Cambridge Science Festival.
Mike - Second Life is a lot of things to a lot of people. Essentially it's a 3D virtual world that allows users to create images, sounds and activities within an environment that supports this. They key thing that most people end up doing is communicating with other people. It's a way of mediating communication.
Meera - Obviously a lot of complicated technology is involved creating Second Life. Would you be able to summarise just the amount of technology involved in creating such a virtual world.
Mike - The basic unit of Second Life would be a sim or a simulation which is a square area represented on a map. That is essentially produced by a core on a chip. Computer chips these days can have dual cores or quad cores. Many of these chips go together to make up entire racks of computing and so you can find we have tens of thousands of individual areas all stuck together. Each one of those is supported by a core on a chip on a board, in a computer, in a database in a service engine. Everything that you see and everything that you do is bits and bytes somewhere. It's entries in a large number of databases that recall where you've been, what you're doing, what you're trying to see at the moment and what someone else is seeing you do as well. It is quite accurate on your view of the world and someone else's view of the world is kept in sync: quite a technological feat!
Meera - When Second Life first came about it was thought it might just dominate the world. Do you think this is still going to happen with it?
Mike - No, I think Second Life is a part of the internet and has a particular technological base and technological limitations that it actually takes quite a lot more effort to generate these things than a lot of people may first assume. I suspect the air conditioning unit for the servers is taking up as much power as a small village or town. If you scale that up you see there are going to be fundamental limits.
Meera - Your background is in artificial intelligence. To what extent do you think artificial intelligence is being used in Second Life and to what extent could it be used in Second Life?
Mike - It's been used quite a lot because of the fairly powerful level of scripting. There's a language in Second Life that allows you to get objects and avatars to react in certain ways so you can programme interactions which is different to a lot of game environment which are more limited.
Meera - This evening you showed us the possibility for avatars to own dogs and this uses artificial intelligence, doesn't it?
Mike - Yes, the dogs have quite a sophisticated level of artificial intelligence. They have a basic set of actions. The user doesn't see the programming languages around them. They interact with them by asking them to do things like beg/howl and they can also add these actions together to get dogs to do more complicated sequences of actions. The dogs also have a random behaviour mode where they will do something out of their repertoire. If you praise them and tell them they're a good dog it will be more likely to use that particular action.
Meera - These dogs are actually being conditioned in the same way a real dog would.
Mike - Yes, they're being trained. In fact there are dog training classes or more accurately owner training classes. The dogs don't need training. It's the owners that need training on how to interact with the dog.
Meera - If it's possible for people to have dogs what else do you think is possible?
Mike - Certainly other animals, I know there's a big equestrian sector and there are more mundane activities such as vending machines and chat bots that will interact with you and say good morning greeting bots, which can be really annoying. They come up and say, "Hello, how are you - here's a card, read this, do that - can I take you on a tour," and that sort of thing. There are little bits of animated activity that give the impression of an artificial intelligence.
Meera - What would you say the benefits and advantages of using Second Life are?
Mike - Well, they're multiple. On a personal level I think it's a great way of interacting and meeting people because it's very easy to fall into chatting with people. I was wondering around an art exhibition and saw this annoying person actually dressed as a witch flying around the room on a broom. It turned out to be the artist and we had a nice conversation about how inspiration was created and I got a much greater insight than I would do just by looking at the static art. There's the communication side of things and the exploration side. There used to be a representation of Paris. There's certainly a representation of Krakow in Poland and science exhibitions and the National Physical Laboratory as interesting stuff going on.
Meera - Finally, where do you think Second Life is going next? In 5 years what do you think Second Life is going to be like?
Mike - To a certain extent Second Life is ahead of the curve. I think more things will be more like Second Life. I think there are some fundamental limitations as to how big it's going to get. I think the richness and quality of what's going on in it will improve.
37:48 - The Human CPU: from Computer Science to Neuroscience.
The Human CPU: from Computer Science to Neuroscience.
with Demis Hassabis
Chris - We're joined by Demis Hassabis, who traded in a career designing and programming computer games such as Theme Park to become a neuroscientist! Tell us your story. You started off as a computer science student.
Demis - That's right. So my undergrad degree was in Computer Science at Cambridge and I was having a traditional root in computer science then. When I left I set up my own computer games company which I'd also programmed some games like Theme Park before coming up to Cambridge. It was a natural progression for me to do that. I set up my own company in Camden in London and we grew it up to 65-70 people and created several games for big publishers around the world. And then about 3-4 years ago I decided that the games industry was going in a slightly different direction to the way I thought it would do when I first got into games in the early nineties. It was becoming a more big business, big budget sort of Hollywood-type industry where there was less and less room for creativity, I was feeling.
Chris - Do you think the financiers noticed the scale of this market? Previously it was a bit geeky, bit niche, not a very big market - still cost quite a bit. Because obviously, you're saying big teams of people were needed to do this but the profits weren't huge so no one was that interested. Suddenly people realised you can tether a computer game to a Hollywood blockbuster, you can make a fortune.
Demis - Yeah, I think partly the realisation how big a business it was but I think the main problem and the reason that caused the conservatism, if you like, in terms of creative sense was that games became very expensive to make. The graphics got so sophisticated you needed teams of - the TripleA game now would have easily 50 artists on it. The costs of that are just huge. Because it is costing 10 million pounds for one game the money men at these publishers would need to be more sure there was definitely going to be a return. This means linking it in with a Hollywood franchise.
Chris - That obviously was inconsistently compatible with your view of what you wanted to do. So you did the rather unusual thing of take a side-step into neuroscience.
Demis - Yeah, that's right. Although I had been in the games industry for a long while, underlying all that all the games I'd been involved with designing and programming actually composed a lot of AI in those games. Most of the games were big strategy
Chris - Artificial Intelligence for the non-initiated like me.
Demis - That's right. All the games I wrote like Theme Park and Republic and Evil Genius. They all involved simulations. Most of the involved hundred of little computer people coming in and getting involved with the game environment. Most of the games, like Theme Park, involved you manipulating that environment and seeing how these autonomous agents reacted to what you were doing. Those were the kinds of games I found fun to play and they were the kinds of games I found fun to create. But so underlying all of this was my passion. My main passion is actually in artificial intelligence and related to that (as soon as you start thinking about what artificial intelligence is) then you start thinking about - how is it the mind achieves these end-goals?
Chris - You didn't get hooked on your own games, did you?
Demis - No. Actually after you've worked on a game for 3-4 years you're sick of the sight of it in general, even if it's the best game ever!
Chris - It used to take me about ten minutes to get bored of them. Some of the strategy games I think are terrific because they forced you to think in a certain way. I was a very big fan of text games in the early days. Largely computers were rubbish and the only thing they could do was to generate spurious streams of text that you could read laboriously. They taught me amazing language skills. I think it's partly responsible for why I have an incredible memory for text and words and facts. You had to remember effectively a graphical representation in your head of the text environment that was this computer adventure. I think that probably had a huge brain training effect on me.
Demis - Absolutely. I think those early games left a lot more to your imagination in the way a book would, a great book. It exercises you imagination which, of course, with all the latest flashy graphics although it looks very beautiful obviously leaves less need for visualisation and creative powers.
Chris - Now just very briefly, how did you develop an interest in neuroscience? How did you then take those skills that you had in computer programming to start answering important questions about how the brain works?
Demis - I didn't really know much about neuroscience before I did my PhD but I did a lot of reading and it struck me that a computational approach or an understanding of algorithms or the basic computer science that Professor Bishop was talking about is actually a useful approach maybe to take looking at how the brain works. In general most of the people work with and most people are from a life sciences or medical background which, of course, is incredibly useful form an anatomical point of view. There are actually relatively few people in the neurosciences field that look at the brain in a computational way, a machine-like way which can give other insights. It's not a machine, the brain, but a lot of the things it does are relatively machine-like.
Chris - And the problem you solved most recently in a nutshell, what was that?
Demis - My most recent study involved trying to accurately predict where someone was standing in a virtual reality room whilst they were lying in a scanner, just from their brain scans.
Chris - Why is that so difficult a problem to solve? I thought there were cells in the brain that fire off when you go into a certain place. We can just read what that activity is, can't we?
Demis - it's been known for a good 30 years that these experiments were done in rats. There are cells that tell you where a rat is positioned in an environment or a pen. No one really knew what those cells looked like on a population level. If you were to look at a million of those cells at once, which you can't do with a single cell recording from electrodes in rats' brains. What I wondered was if you could have this global eagle's eye view of the whole population of cells that might tell you something different that you couldn't see on the individual cell level.
Chris - Is this supposed to be a model or a representation of what's going on in the real world? Are you trying to understand how, because the person you worked with, Eleanor Maguire very famously worked out how our cab drivers have bigger bits of their brains through navigating round London. Are you trying to basically understand how a cab driver finds his way around London?
Demis - We're trying to understand those findings, really. We know the hippocampus is vital for spatial navigation and spatial memories but we still don't know what it is about the environment that is encoded in those memories and is encoded by those neurons and how and why it's doing that. You can't carry a two tonne brain scanner round with you. The best thing you can do is to record people whilst they're navigating round a virtual, very realistic virtual reality environment.
Chris - When you do this, what did you see and how does this influence our understanding of how the brain tells us how to get from A to B?
Demis - Brain scans work at quite a low resolution compared to single cell recording. When you look at one pixel, if you like, of a brain scan which is called a voxel - a 3D pixel. That contains about 40,000 neurons. For our pattern recognition algorithms you need to classify which location that was related to. It needs several hundred voxels to be active to train a pattern recognition algorithm with. What that shows is the spatial memories must be represented by very large neural populations' codes. Including probably between two and five million neurons.
Chris - Where previously we thought it was just the odd cell that fired off. Now you're saying there's huge populations of cells that join together. They code specific locations.
Demis - That's right. Also it says something about the way that they're clustered and structured. We found that these cells were also clustered quite closely together whereas previous rat literature suggested that perhaps they were random and uniformly distributed across the hippocampus.
45:21 - Chameleon skin foods?
Chameleon skin foods?
We put this to Dr Stephen Juan, Ashley Montagu Fellow for Public Understanding of Human Sciences at the University of Sydney.
There are substances that can turn your skin different colours. One of the most famous ones is carotenemia which is when you eat too many carrots. Your skin can turn yellowish or orange. It's a benign condition, doesn't seem to be related to anything but if you eat too many carrots the beta carotene builds up in your system and you turn into the colour of a carrot: first a little yellow and then a little orange. There was an interesting study in 2006 in Paediatric Dermatology by Royal Liverpool Hospital that showed that carotenemia can come from eating green beans as well. There are other vegetables as well like yams it's been known to happen where the colour can change and some other fruits as well. Yes, you have to be very careful. Of course it's going to show up in lighter skinned people first but it'll happen to anybody. By the way, speaking of carrots the old question is what about carrots improving eyesight? Yes, carrots do improve the eyesight a little bit but only if you are vitamin A deficient to begin with.
How do CDs, DVDs and hard discs store information?
We put this to Chris Bishop, Microsoft Research
That's a good question. CDs and DVDs work in slightly different ways but they have something in common, which is that like all computer storage systems they use binary. All the information's expressed in terms of strings of 1s and 0s or 'on' and 'off.' In a hard disc that's represented magnetically. Each bit is represented by a tiny magnet built into the surface of the hard disc and if it's facing north-up it's a 1 and if it's south-up it's a 0. The head that reads this information can also flip the magnets and so it can write information to the disc. The CD is rather similar. It also has these 1s and 0s but they're represented rather differently. They're represented by little pits. When a disc is written a laser burns little pits into the surface of the disc and another laser can read back those points. If there's a pit there might be a 0 and if there's no pit it might be a 1.
Can you build a mind-control helmet?
We put this to Demis Hassabis of the Institute of Cognitive Neuroscience
Demis - Well at the moment we're very far away from that kind of technology. So far All the technology that is being used in neuroscience at the moment is passively reading the electrical activity the brain is creating itself, rather than actually influencing it the other way. I do know of some rather experimental studies that have been done trying to connect chip board to rat brains and trying to control the way they navigate. As far as I know those are very experimental at the moment. Nothing's really practical that's come out of that yet. Chris - I believe there was a researcher in Europe who showed if you supply or apply an intense magnetic field to the head while they are learning things it helps them remember them better. He was doing this with word learning. He was doing this with students to sleep with a magnetic field round their head. I think it improved their ability to learn, probably because the brain's an electromagnetic organ.
Demis - There may be ways. You could do it neuro-chemically as well with some drugs. There are probably ways of changing and up-regulating entire systems in the brain but that's very different from influencing a specific thought or getting them to specifically think about a key episode, as well.
How will faster computer processors be made in the future?
We put this to Chris Bishop from Microsoft Research
Computer processors have been doubling in speed every two years for the last fifty years or so and it's been driven by something called Moore's Law which says the number of transistors on a processor keeps doubling every two years. As the transistors have got smaller and smaller we've fitted more on the chip. We've started to run into some problems. One of the problems we're hitting already is to do with the heat density that's produced by all these transistors in a tiny space switching on and off. Right now the heat density on a chip is equal to that on a hotplate on a cooker. If we carry on the way we've been going then in ten years' time that heat density will equal that of the surface of the sun. We have to find a new way forward. The main approach that we're adopting is to do what we call parallel computing. Instead of trying to make each individual processor run faster we have several processor'sside-by-side, all working on the problem together. Using that technique of parallel processing we should be able to continue to get processors collectively to run faster and faster and to continue this doubling every two years for a good many years if not decades to come.