Skip to content

Latest commit

 

History

History
76 lines (53 loc) · 6.03 KB

README.md

File metadata and controls

76 lines (53 loc) · 6.03 KB

MiniGP

MiniGP is a minimalistic Gaussian Process (GP) library focused on regression tasks. It is designed to be simple and easy to understand, super lightweight, and friendly to researchers and developers.

Motivation

Despite that there are many successful GP libraries, such as GPy, GPflow, and GPyTorch we find them difficult to use for beginners and very time-consuming to customize. It will take a lot of time to understand the structure of the library and the GP model, by which time the user (young research student) may give up.

Thus we want to create a simple and easy-to-use GP library that can be used by anyone. MiniGP is designed to be simple and easy to understand. It is a great tool for educational purposes. We also try to make it easy for anyone to use without a lot of background knowledge of GP.

Useful and practical GP Models:

  • CIGP: simple yet accurate multi-output regression model with complexity $O(n^3 d)$ for n training points with d outputs.
  • NeuralKernel: automatic kernel learning with neural network structured kernel.

Installation

We do not have a pip package yet. You can install it by cloning the repository and rune the code for your own purpose. At this stage we think it is better to keep it simple and customizable. It servers as rather demo code than a library, with some useful functions to make computation easier.

To start using MiniGP, you can clone the repository by running the following command:

git clone

You can start by running the Demo.ipynb to have a taste of the library. You can also check the tutorial in the GPmodels_xxx folder to learn how to use the library.

Most models have two version, the API version for direct call and the tutorial version for customized usage. The API version is in the 'core' folder, and the tutorial version is in the 'GPmodels_xxx' folder.

Structure

  • core: This folder contains all the core functions (computing likelihood, matrix inversion, kernels, etc.) for Gaussian Processes (GP). It serves as the backbone of the library. Additionally, it includes API GP models (.py) that are designed to be directly called for research and experimentation. More details can be found in the README file within the core folder.
  • GPmodels_Advance: Advance GP models, including GP with GPU acceleration, and automatic GP.

    • 01_GP&GPU: a GP model leverage GPU acceleration.
    • 01_GP&GPU_GPTutorial: Algorithms for GP leverage GPU acceleration.
    • 02_DynamicModel_GPTutorial: GP dynamic model in Chinese. Original paper
  • GPmodels_Classic: basic GP model and its variation, such as DeepKernel GP, InputWarp GP . It demonstrates how to build a GP model with the GP_CommonCalculation.

    • 01_simpleGP_GPTutorial: simple GP tutorial in both English and Chinese. This is a good starting point for beginners.
    • 01_simpleGP, a basic GP model. It demonstrates how to build a GP model with the GP_CommonCalculation.
    • 02_deepKernelGP, a GP model with deep kernel. Original paper
    • 03_logTransformWarpGP, a GP model with log transform on the target values, this can improve the model performance when the noise does not follow Gaussian distribution.
    • 04_neuralKernelGP, a GP model with neural kernel.
  • GPmodels_MultiOutput: provides tools for implementing Gaussian Process models with multiple outputs.

    • 01_IntrinsicModel: Foundation work for multi-output GP.
    • 02_hogp_GPTutorial_Chinese: high-oreder GP in Chinese. Original paper
  • GPmodels_Sparse: provides tools for implementing sparse Gaussian Process models.

    • 01_sgpr_GPTutorial: A detailed tutorial for Sparse Gaussian Process with variational learning inducing points. Original paper
    • 02_svgp, A demo for implementing mini-batch gradient descent on SVGP allows training a GP with 10k inputs in 2 seconds.
    • 02_svgp_GPTutorial, A detailed tutorial for Stochastic Variational Interference GP model that can allow mini-batch training. Original paper
  • Debug_NaNError_FAQ: Frequently asked questions for GP model. It contains some techniques to solve NaN problem in GP model. More details can be found in the README file in the Debug_NaNError_FAQ folder.

  • Bayesian_Optimization: This folder contains useful tools for Bayesian optimization

    • acq: A Python scripy including several widely used acquisition functions.
    • BO_demo: A demonstration of the process of Bayesian optimization.
  • asset: This folder contains the python scripts for the model comparison and regression test. As well as the result in both .csv and .png format. For more details, please refer to the README.md in the folder.

    • Model_comparison.py: A Python script that compares the performance of different GP models on various synthetic datasets, including periodic, warped, and polynomial. The default models are set as autoGP and its base model vsgp.

    • Regression_test.py: A Python script that tests the accuracy and training speed on different sizes of training sets. The results are stored in result1.csv and result2.csv.

    • result1.csv: The result of the regression test for different training set sizes.

    • result2.csv: The result of the regression test for different numbers of inducing points.

Contribution

We welcome any contribution to this project. Please feel free to open an issue or pull request. We are happy to discuss any ideas with you.