Symbol of change for AI development

Symbol of change for AI development
A novel algorithm translates symbolic knowledge into vector spaces to combine deductive reasoning with machine learning. Credit: Maxat Kulmanov

A mathematical framework that bridges the gap between high-level human-readable knowledge and statistical data has been developed by a KAUST team and is expected to improve machine learning.

Humans rely on patterns, labels and order to make sense of the world. We categorize, classify and make links between related things and ideas, creating symbols that we can use to share information. Artificial , on the other hand, is trained most effectively using raw numerical data. How, then, can algorithms make use of our vast store of symbolic knowledge? This is a vexing problem and one that, if cracked, could open an enormous new multidimensional library for machine learning and artificial intelligence.

Robert Hoehndorf, Maxat Kulmanov and their collaborators at KAUST'sComputational Bioscience Research Centerand Halifax University, Canada, have developed a mathematical bridge between these seemingly incompatible forms of information.

"There is a big gap in artificial intelligence research between approaches based on high-level symbolic representations understandable by humans and the subsymbolic approaches used for training artificial neural networks," explains Kulmanov. "Symbolic approaches are built on logical relationships, while subsymbolic approaches rely on statistics and continuous real-numbered vector spaces."

The researchers set out to develop an "embedding" function that maps one mathematical structure to another in a way that preserves some of the features of the first structure.

"Embeddings are used because the second structure may be more suitable for some operations," says Hoehndorf. "In this work, we mapped a formal language, called a description logic, into a real-number vector space, which can be used more easily for , such as computing similarity and performing predictive operations."

Description logics are widely used in biology and biomedicine to describe formalized theories, such as the functions of genes and the terminology used in medical diagnosis.

"Logics, such as description logic, have been the foundation for artificial intelligence systems since the 1960s and have been studied in mathematics for more than 100 years," says Hoehndorf. "Building on this history of research, we created an embedding function that not only projects symbols into a vector space, but also generates algebraic models to capture the semantics of the symbols within description logic."

The key to the team's achievement is linking the embedding to model theory, which made it possible to draw on established knowledge and create the first that preserves semantics.

"Our method is directly applicable to hundreds of formalized theories in biological and and hundreds of biological databases," says Kulmanov. "In the future, we will apply our method to more problems in biology, which we hope will improve biomedical applications of artificial intelligence."


Explore further

Researchers' deep learning algorithm solves Rubik's Cube faster than any human

More information: EL embeddings: Geometric construction of models for the description logic EL++. Conference paper at Proceedings of the Twenty-Eighth International Joint Conferences on Artificial Intelligence (IJCAI-19), 10-16 August 2019, Macao, China
Citation: Symbol of change for AI development (2019, October 21) retrieved 25 February 2020 from https://techxplore.com/news/2019-10-ai.html
This document is subject to copyright. Apart from any fair dealing for the purpose of private study or research, no part may be reproduced without the written permission. The content is provided for information purposes only.
4 shares

Feedback to editors

User comments