Node-level Graph Regression with Deep Gaussian Process Models

Overview

Node-level Graph Regression with Deep Gaussian Process Models

Prerequests

our implementation is mainly based on tensorflow 1.x and gpflow 1.x:

python 3.x (3.7 tested)
conda install tensorflow-gpu==1.15
pip install keras==2.3.1
pip install gpflow==1.5
pip install gpuinfo

Besides, some basic packages like numpy are also needed. It's maybe easy to wrap the codes for TF2.0 and GPflow2, but it's not tested yet.

Specification

Source code and experiment result are both provided. Unzip two archive files before using experiment notebooks.

Files

  • dgp_graph/: cores codes of the DGPG model.
    • impl_parallel.py: a fast node-level computation parallelized implementation, invoked by all experiments.
    • my_op.py: some custom tensorflow operations used in the implementation.
    • impl.py: a basic loop-based implementation, easy to understand but not practical, leaving just for calibration.
  • data/: datasets.
  • doubly_stochastic_dgp/: codes from repository DGP
  • compatible/: codes to make the DGP source codes compatible with gpflow1.5.
  • gpflow_monitor/: monitoring tool for gpflow models, from this repo.
  • GRN inference: code and data for the GRN inference experiment.
  • demo_city45.ipynb: jupyter notebooks for city45 dataset experiment.
  • experiments.zip: jupyter notebooks for other experiments.
  • results.zip: contains original jupyter notebooks results. (exported as HTML files for archive)
  • run_toy.sh: shell script to run additional experiment.
  • toy_main.py: code for additional experiment (Traditional ML methods and DGPG with linear kernel).
  • ER-0.1.ipynb: example script for analyzing time-varying graph structures.

Experiments

The experiments are based on python src files and demonstrated by jupyter notebooks. The source of an experiment is under directory src/experiments.zip and the corresponding result is exported as a static HTML file stored in the directory results.zip. They are organized by dataset names:

  1. Synthetic Datasets

For theoretical analysis.

  • demo_toy_run1.ipynb

  • demo_toy_run2.ipynb

  • demo_toy_run3.ipynb

  • demo_toy_run4.ipynb

  • demo_toy_run5.ipynb

For graph signal analysis on time-varying graphs.

  • ER-0.05.ipynb

  • ER-0.2.ipynb

  • RWP-0.1.ipynb

  • RWP-0.2.ipynb

  • RWP-0.3.ipynb

  1. Small Datasets
  • demo_city45.ipynb
  • demo_city45_linear.ipynb (linear kernel)
  • demo_city45_baseline.ipynb (traditional regression methods)
  • demo_etex.ipynb
  • demo_etex_linear.ipynb
  • demo_etex_baseline.ipynb
  • demo_fmri.ipynb
  • demo_fmri_linear.ipynb
  • demo_fmri_baseline.ipynb
  1. Large Datasets (traffic flow prediction)
  • LA
    • demo_la_15min.ipynb
    • demo_la_30min.ipynb
    • demo_la_60min.ipynb
  • BAY
    • demo_bay_15min.ipynb
    • demo_bay_30min.ipynb
    • demo_bay_60min.ipynb
Training DiffWave using variational method from Variational Diffusion Models.

Variational DiffWave Training DiffWave using variational method from Variational Diffusion Models. Quick Start python train_distributed.py discrete_10

Chin-Yun Yu 26 Dec 13, 2022
Code for binary and multiclass model change active learning, with spectral truncation implementation.

Model Change Active Learning Paper (To Appear) Python code for doing active learning in graph-based semi-supervised learning (GBSSL) paradigm. Impleme

Kevin Miller 1 Jul 24, 2022
An Unpaired Sketch-to-Photo Translation Model

Unpaired-Sketch-to-Photo-Translation We have released our code at https://github.com/rt219/Unsupervised-Sketch-to-Photo-Synthesis This project is the

38 Oct 28, 2022
Repository for "Exploring Sparsity in Image Super-Resolution for Efficient Inference", CVPR 2021

SMSR Reposity for "Exploring Sparsity in Image Super-Resolution for Efficient Inference" [arXiv] Highlights Locate and skip redundant computation in S

Longguang Wang 225 Dec 26, 2022
FB-tCNN for SSVEP Recognition

FB-tCNN for SSVEP Recognition Here are the codes of the tCNN and FB-tCNN in the paper "Filter Bank Convolutional Neural Network for Short Time-Window

Wenlong Ding 12 Dec 14, 2022
Official Implementation of "LUNAR: Unifying Local Outlier Detection Methods via Graph Neural Networks"

LUNAR Official Implementation of "LUNAR: Unifying Local Outlier Detection Methods via Graph Neural Networks" Adam Goodge, Bryan Hooi, Ng See Kiong and

Adam Goodge 25 Dec 28, 2022
Official codebase for ICLR oral paper Unsupervised Vision-Language Grammar Induction with Shared Structure Modeling

CLIORA This is the official codebase for ICLR oral paper: Unsupervised Vision-Language Grammar Induction with Shared Structure Modeling. We introduce

Bo Wan 32 Dec 23, 2022
Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)

Channel Pruning for Accelerating Very Deep Neural Networks (ICCV'17)

Yihui He 1k Jan 03, 2023
Invertible conditional GANs for image editing

Invertible Conditional GANs This is the implementation of the IcGAN model proposed in our paper: Invertible Conditional GANs for image editing. Novemb

Guim 278 Dec 12, 2022
Rainbow DQN implementation that outperforms the paper's results on 40% of games using 20x less data ๐ŸŒˆ

Rainbow ๐ŸŒˆ An implementation of Rainbow DQN which reaches a median HNS of 205.7 after only 10M frames (the original Rainbow from Hessel et al. 2017 re

Dominik Schmidt 31 Dec 21, 2022
Codes for TIM2021 paper "Anchor-Based Spatio-Temporal Attention 3-D Convolutional Networks for Dynamic 3-D Point Cloud Sequences"

Codes for TIM2021 paper "Anchor-Based Spatio-Temporal Attention 3-D Convolutional Networks for Dynamic 3-D Point Cloud Sequences"

Intelligent Robotics and Machine Vision Lab 4 Jul 19, 2022
Official implementation of Self-supervised Graph Attention Networks (SuperGAT), ICLR 2021.

SuperGAT Official implementation of Self-supervised Graph Attention Networks (SuperGAT). This model is presented at How to Find Your Friendly Neighbor

Dongkwan Kim 127 Dec 28, 2022
Imposter-detector-2022 - HackED 2022 Team 3IQ - 2022 Imposter Detector

HackED 2022 Team 3IQ - 2022 Imposter Detector By Aneeljyot Alagh, Curtis Kan, Jo

Joshua Ji 3 Aug 20, 2022
For encoding a text longer than 512 tokens, for example 800. Set max_pos to 800 during both preprocessing and training.

LongScientificFormer For encoding a text longer than 512 tokens, for example 800. Set max_pos to 800 during both preprocessing and training. Some code

Athar Sefid 6 Nov 02, 2022
The source code of CVPR17 'Generative Face Completion'.

GenerativeFaceCompletion Matcaffe implementation of our CVPR17 paper on face completion. In each panel from left to right: original face, masked input

Yijun Li 313 Oct 18, 2022
Official code for paper "ISNet: Costless and Implicit Image Segmentation for Deep Classifiers, with Application in COVID-19 Detection"

Official code for paper "ISNet: Costless and Implicit Image Segmentation for Deep Classifiers, with Application in COVID-19 Detection". LRPDenseNet.py

Pedro Ricardo Ariel Salvador Bassi 2 Sep 21, 2022
ใ€ŒPyTorch Implementation of AnimeGANv2ใ€ใ‚’็”จใ„ใฆใ€็”Ÿๆˆใ—ใŸ้ก”็”ปๅƒใ‚’ๅ…ƒใฎ็”ปๅƒใซไธŠๆ›ธใใ™ใ‚‹ใƒ‡ใƒข

AnimeGANv2-Face-Overlay-Demo PyTorch Implementation of AnimeGANv2ใ‚’็”จใ„ใฆใ€็”Ÿๆˆใ—ใŸ้ก”็”ปๅƒใ‚’ๅ…ƒใฎ็”ปๅƒใซไธŠๆ›ธใใ™ใ‚‹ใƒ‡ใƒขใงใ™ใ€‚

KazuhitoTakahashi 21 Oct 18, 2022
Automatically download the cwru data set, and then divide it into training data set and test data set

Automatically download the cwru data set, and then divide it into training data set and test data set.่‡ชๅŠจไธ‹่ฝฝcwruๆ•ฐๆฎ้›†๏ผŒ็„ถๅŽๅˆ†่ฎญ็ปƒๆ•ฐๆฎ้›†ๅ’Œๆต‹่ฏ•ๆ•ฐๆฎ้›†

6 Jun 27, 2022
Extreme Dynamic Classifier Chains - XGBoost for Multi-label Classification

Extreme Dynamic Classifier Chains Classifier chains is a key technique in multi-label classification, sinceit allows to consider label dependencies ef

6 Oct 08, 2022