Go to content

SIEMENS

Research & Development
Technology Press and Innovation Communications

Dr. Ulrich Eberl
Herr Dr. Ulrich Eberl
  • Wittelsbacherplatz 2
  • 80333 Munich
  • Germany
Dr. Ulrich Eberl
Herr Florian Martini
  • Wittelsbacherplatz 2
  • 80333 Munich
  • Germany
pictures

Roundworm 302 neurons

Fruit fly 100,000 neurons

Cockroach 1,000,000 neurons

Octopus 300,000,000 neurons

Human 100,000,000,000 neurons

Elephant 200,000,000,000 neurons

From Biological Systems to Machines, Learning is the key

Biological learning systems run the gamut from the lowly roundworm (Caenorhabditis elegans) with its 300 or so neurons, all the way up to the adult elephant brain, with its 200 billion neurons. Whether they’re located in fruit flies or cockroaches, chimpanzees or dolphins, all neurons do the same thing: they process and transmit information. And the reason for this is the same across the biological board: To avoid danger and maximize success in sustaining and propagating themselves, all organisms must be able to sense the environment, respond to it accordingly, and remember those stimuli that indicate risks and rewards. Learning, in short, is a prerequisite for the survival of individuals and species in the natural world. The same iron law, however, is becoming increasingly applicable to the world of man-made systems. According to Dr. Volker Tresp, one of Siemens’ top machine learning authorities and a computer science professor at Ludwig Maximillian University in Munich, there are three kinds of learning: memorization (such as the ability to remember facts); skills (such as the ability to learn to throw a ball); and abstraction (such as the ability to form rules based on observations). Computers, which are born whizzes in the first area, are rapidly catching on to the other two. Take, for instance, the skill needed to produce a flawlessly even sheet of steel in a given thickness — an area in which Siemens has been a leader for over 20 years. “Here,” says Tresp, “the simplest learning schema is to make a prediction, and then check to see if the output product meets the desired specification.” Confronted with an output requirement for, say, a particularly high grade of steel, an automated rolling mill would take sensor data (composition, strip temperature, etc.) into account, estimate the required pressure based on previously learned information, and then adjust itself accordingly in real time in response to its own output data until it achieved exactly the right pressure to get the desired thickness. “In a neural network-based learning system,” explains Tresp, “this would be achieved by adjusting the relative weight matrix (see diagram) of all the factors that influence a given parameter, such as thickness.”

How Neural Networks Learn

Image open

Neural networked-based learning system (1) based on input information (2) and providing output prediction (3) regarding gas demand over seven days based on a 14-day training phase. Learning is represented in three snapshots from random weighted (4) to partially learned (5), to fully trained (6). Neural networked-based systems have the ability to process huge amounts of input data in order to adjust their output. To accomplish this, such a system must build up a mathematical model that duplicates its real-world counterpart. Such a model is essentially a community of decision units. Collectively, the interaction of the decision units can be represented in the form of a matrix (see inset in each box). Depending on the complexity of the application, hundreds of interaction matrices may be required. Initially, interactions among the decision units are random. Thus, when the system begins its training phase (see time line left), its error level — the difference between expectation and observation — is high (4). Once compared to actual output, the error level is fed back into each matrix (arrows pointing right to each box), thus modifying the internal weights of each decision unit away from randomness and altering each input parameter based on what has been learned (arrows pointing left from each box). Eventually, after thousands of iterations, each of which is designed to reduce the error level, the system learns to describe the entire flow of input information over time in such a way that its output exactly duplicates (6) — and eventually predicts — the behavior of the real world.

close

Beyond memorization and the ability to optimize skills, artificial systems are increasingly being called upon to generalize or abstract the characteristics that make an individual item a member of a group. Optical character recognition (OCR), which has traditionally been used for high-speed postal sorting, is a case in point. Since approximately 1985, when this technology was first developed, accuracy has skyrocketed from single digits to over 95 percent for handwritten Latin alphabets and over 90 percent for Arabic handwriting. In fact, in 2007, Siemens’ ARTread learning system won first place in the International Conference on Document Analysis and Recognition contest for OCR in Arabic. Given OCR technology’s exceptionally high level of reliability, it is beginning to migrate to applications such as automatic license plate recognition and industrial vision (for more, see article (“We Read You Loud and Clear”).

Where is machine learning likely to go from here? Clearly, vast opportunities are emerging as sensors proliferate in power and sheer numbers, making ever more data available locally and through information networks. Learning in the context of networked environments is being pursued in two major projects: Theseus (see Picture of the Future, Spring 2008, “Harvest without End” page 89), where Siemens leads with MEDICO, a project that focuses on the extraction of semantic information from images and texts to enable many new applications designed to support physician workflow, and, second, the European Union’s LarKC project (see Pictures of the Future, Spring 2011, Zettabyte Gold Mine) for the development of scalable querying, reasoning and machine learning approaches for linked data. “Learning with linked information,” says Tresp, “that’s where the excitement is today!”

Arthur F. Pease