Our aim is to develop fundamental theories and effective methods of machine learning, which have various applications in data analysis. Followings are our ongoing topics.

Machine Learning and Information Geometry

We study the relationship between machine learning and information geometry.

Information geometry enables us to flexibly treat machine learning models in a principled manner. For example, the set of model distributions of a Boltzmann machine, which is a basic generative model in machine learning, can be viewed as a dually-flat manifold, a canonical structure in information geometry.

We are trying to design powerful machine learning models with integrating information geometry and partial order structures, where we can incorporate interactions between variables. Our approach can treat various machine learning tasks such as feature selection, representation learning, and matrix/tensor decomposition.

information geometry

Selected publications:

Part of this project is supported by JST PRESTO "Social Design" Area and JSPS KAKENHI (Grants-in-Aid for Scientific Research).

Machine Learning with Discrete Structure

Discrete structures, such as a graph structure, appear in a wide range of applications including bioinformatics and chemoinformatics. Since most of machine learning methods are designed for continuous variables, how to integrate such discreteness into continuous spaces is of high interest. A representative method is a graph kernel, which measures the similarity between graphs and enables us to apply kernel based methods to graph structured objects.

We are also interested in how to learn discrete structures from unstructured objects. One of approaches is to use Formal Concept Analysis (FCA), which extracts a lattice structure using algebraic property of data.

graph kernel

Selected publications:

Part of this project is supported by JSPS KAKENHI (Grants-in-Aid for Scientific Research).

Significant Pattern Mining

Feature selection of detecting informative variables in a high-dimensional data is a fundamental task in machine learning, while most of existing methods can find only single features or linear interactions of features. We study how to find patterns, that is, multiplicative/combinatorial (potentially higher-order) interactions between variables. We use pattern mining algorithms and integrate statistical tests into them to guarantee statistical significance of detected interactions.

This research topic is based on a joint work with Prof. Dr. Karsten Borgwardt's group in ETH Zürich. Please check for more details.

pattern mining

Selected publications:

Part of this project has been supported by JST PRESTO "Big Data" Area and JSPS KAKENHI (Grants-in-Aid for Scientific Research).

Developing Various Machine Learning Methods

We are also developing efficient and/or effective machine learning methods such as clustering and outlier detection.

outlier detection

Selected publications:

© Mahito Sugiyama 2024