Cambridge researchers are pioneering a form of machine learning that starts with only a little prior knowledge and continually learns from the world around it.

This is just like a human would learn. We don鈥檛 start knowing everything. We learn things incrementally, from only a few examples, and we know when we are not yet confident in our understanding

Zoubin Ghahramani

In the centre of the screen is a tiny unicycle. 探花直播animation starts, the unicycle lurches forward and falls. This is trial #1. It鈥檚 now trial #11 and there鈥檚 a change 鈥 an almost imperceptible delay in the fall, perhaps an attempt to right itself before the inevitable crash. 鈥淚t鈥檚 learning from experience,鈥 nods Professor Carl Edward Rasmussen.

After a minute, the unicycle is gently rocking back and forth as it circles on the spot. It鈥檚 figured out how this extremely unstable system works and has mastered its goal. 鈥 探花直播unicycle starts with knowing nothing about what鈥檚 going on 鈥 it鈥檚 only been told that its goal is to stay in the centre in an upright fashion. As it starts falling forwards and backwards, it starts to learn,鈥 explains Rasmussen, who leads the Computational and Biological Learning Lab in the Department of Engineering. 鈥淲e had a real unicycle robot but it was actually quite dangerous 鈥 it was strong 鈥 and so now we use data from the real one to run simulations, and we have a mini version.鈥

Rasmussen uses the self-taught unicycle to demonstrate how a machine can start with very little data and learn dynamically, improving its knowledge every time it receives new information from its environment. 探花直播consequences of adjusting its motorised momentum and balance help the unicycle to learn which moves were important in helping it to stay upright in the centre.

鈥淭his is just like a human would learn,鈥 explains Professor Zoubin Ghahramani, who leads the Machine Learning Group in the Department of Engineering. 鈥淲e don鈥檛 start knowing everything. We learn things incrementally, from only a few examples, and we know when we are not yet confident in our understanding.鈥

Ghahramani鈥檚 team is pioneering a branch of AI called continual machine learning. He explains that many of the current forms of machine learning are based on neural networks and deep learning models that use complex algorithms to find patterns in vast datasets. Common applications include translating phrases into different languages, recognising people and objects in images, and detecting unusual spending on credit cards.

鈥淭hese systems need to be trained on millions of labelled examples, which takes time and a lot of computer memory,鈥 he explains. 鈥淎nd they have flaws. When you test them outside of the data they were trained on they tend to perform poorly. Driverless cars, for instance, may be trained on a huge dataset of images but they might not be able to generalise to foggy conditions.

鈥淲orse than that, the current deep learning systems can sometimes give us confidently wrong answers, and provide limited insight into why they have come to particular decisions. This is what bothers me. It鈥檚 okay to be wrong but it鈥檚 not okay to be confidently wrong.鈥

探花直播key is how you deal with uncertainty 鈥 the uncertainty of messy and missing data, and the uncertainty of predicting what might happen next. 鈥淯ncertainty is not a good thing 鈥 it鈥檚 something you fight, but you can鈥檛 fight it by ignoring it,鈥 says Rasmussen. 鈥淲e are interested in representing the uncertainty.鈥

It turns out that there鈥檚 a mathematical theory that tells you what to do. It was first described by 18th-century English statistician Thomas Bayes. Ghahramani鈥檚 group was one of the earliest adopters in AI of Bayesian probability theory, which describes how the probability of an event occurring (such as staying upright in the centre) is updated as more evidence (such as the decision the unicycle last took before falling over) becomes available.

Dr Richard Turner explains how Bayes鈥 rule handles continual learning: 鈥渢he system takes its prior knowledge, weights it by how accurate it thinks that knowledge is, then combines it with new evidence that is also weighted by its accuracy.

鈥淭his is much more data-efficient than the way a standard neural network works,鈥 he adds. 鈥淣ew information can cause a neural network to forget everything it learned previously 鈥 called catastrophic forgetting 鈥 meaning it needs to look at all of its labelled examples all over again, like relearning the rules and glossary of a language every time you learn a new word.

鈥淥ur system doesn鈥檛 need to revisit all the data it鈥檚 seen before 鈥 just like humans don鈥檛 remember all past experiences; instead we learn a summary and we update it as things go on.鈥 Ghahramani adds: 鈥 探花直播great thing about Bayesian machine learning is the system makes decisions based on evidence 鈥 it鈥檚 sometimes thought of as 鈥榓utomating the scientific method鈥 鈥 and because it鈥檚 based on probability, it can tell us when it鈥檚 outside its comfort zone.鈥

Ghahramani is also Chief Scientist at Uber. He sees a future where machines are continually learning not just individually but as part of a group. 鈥淲hether it鈥檚 companies like Uber optimising supply and demand, or autonomous vehicles alerting each other to what鈥檚 ahead on the road, or robots working together to lift a heavy load 鈥 cooperation, and sometimes competition, in AI will help solve problems across a huge range of industries.鈥

One of the really exciting frontiers is being able to model probable outcomes in the future, as Turner describes. 鈥 探花直播role of uncertainty becomes very clear when we start to talk about forecasting future problems such as climate change.鈥

Turner is working with climate scientists Dr Emily Shuckburgh and Dr Scott Hosking at the British Antarctic Survey to ask whether machine learning techniques can improve understanding of climate change risks in the future.

鈥淲e need to quantify the future risk and impacts of extreme weather at a local scale to inform policy responses to climate change,鈥 explains Shuckburgh. 鈥 探花直播traditional computer simulations of the climate give us a good understanding of the average climate conditions. What we are aiming to do with this work is to combine that knowledge with observational data from satellites and other sources to get a better handle on, for example, the risk of low-probability but high-impact weather events.鈥

鈥淚t鈥檚 actually a fascinating machine learning challenge,鈥 says Turner, who is helping to identify which area of climate modelling is most amenable to using Bayesian probability. 鈥 探花直播data are extremely complex, and sometimes missing and unlabelled. 探花直播uncertainties are rife.鈥 One significant element of uncertainty is the fact that the predictions are based on our future reduction of emissions, the extent of which is as yet unknown.

鈥淎n interesting part of this for policy makers, aside from the forecasting value, is that you can imagine having a machine that continually learns from the consequences of mitigation strategies such as reducing emissions 鈥 or the lack of them 鈥 and adjusts its predictions accordingly,鈥 adds Turner.

What he is describing is a machine that 鈥 like the unicycle 鈥 feeds on uncertainty, learns continuously from the real world, and assesses and then reassesses all possible outcomes. When it comes to climate, however, it鈥檚 also a machine of all possible futures.

Inset image: read more about our AI research in the 探花直播's research magazine;听download听a听pdf;听惫颈别飞听辞苍听.



探花直播text in this work is licensed under a . For image use please see separate credits above.