Semantic, Symbolic and Interpretable Machine Learning

Semantic, Symbolic and Interpretable Machine Learning

 

Our research is concerned with modelling and analysis methods for complex systems, such as those arising in computer networks, electronic devices and biological organisms. The analysis methods that investigated include simulation and formal verification, with particular emphasis on quantitative verification of probabilistic systems. Our work spans the whole spectrum, from theory, through algorithms to software implementation and applications.

From Function-based to Model-based automated probalistic reasoning for Deep Learning

Machine learning is revolutionising computer science and AI. Much of its success is due to deep neural networks, which have demonstrated outstanding performance in perception tasks such as image classification. Solutions based on deep learning are now being deployed in real-world systems, from virtual personal assistants to self-driving cars. Unfortunately, the black-box nature and instability of deep neural networks is raising concerns about the readiness of this technology. Efforts to address robustness of deep learning are emerging, but are limited to simple properties and function-based perception tasks that learn data associations. While perception is an essential feature of an artificial agent, achieving beneficial collaboration between human and artificial agents requires models of autonomy, inference, decision making, control and coordination that significantly go beyond perception. To address this challenge, this project will capitalise on recent breakthroughs by and will develop a model-based, probabilistic reasoning framework for autonomous agents with cognitive aspects, which supports reasoning about their decisions, agent interactions and inferences that capture cognitive information, in presence of uncertainty and partial observability. The objectives are to develop novel probabilistic verification and synthesis techniques to guarantee safety, robustness and fairness for complex decisions based on machine learning, formulate a comprehensive, compositional game-based modelling framework for reasoning about systems of autonomous agents and their interactions, and evaluate the techniques on a variety of case studies. Addressing these challenges will require a fundamental shift towards Bayesian methods, and development of new, scalable, techniques, which differ from conventional probabilistic verification. If successful, the project will result in major advances in the quest towards provably robust and beneficial AI.