# 1. Installing Maven & Pandas First, please install Java (JDK11) and Python 3 if they are not already. Next, make sure that Maven (for importing JGraphT) and Pandas(for data analysis) are installed. To install Maven on Ubuntu, type the following commands on terminal: sudo apt-get update sudo apt-get install maven For Pandas, type the following: pip3 install pandas ( sudo apt-get install python3-pip if pip is not installed already) # 2. Compilation Type the following to compile this project: mvn compile # 3. Running the code Below is the command for running tests for SNAP(DIMACS) and grid data. java -Xms24G -Xmx48G -Xmn36G -Xss1G -cp $CLASSPATHS shell.TestSNAP (the filename of data; just the name and not the path) (# of tests) (randomization seed) java -Xms32G -Xmx64G -Xmn48G -Xss1G -cp $CLASSPATHS shell.TestGrid (Maximum dimension) (dimension increment) [List of the values for k, space-separated] You may change the randomization seed (vertex selection) to assess reproducibility. (In our experiment, the seed was set to 2021.) For the data, check "src/SNAP(or DIMACS)". Output "test_result.csv" will be saved on "target" directory. Check if 'CLASSPATHS' is set properly. Please refer to " sample.sh " for examples & further details. #4. Obtaining average processing time and diversity First, move to the target directory. Then run get_averages.py python3 get_averages (.csv file name) [list of the values for k, space-separated. Optional parameter.]
Diverse graph algorithms implemented using JGraphT library.
Overview
PyTorch implementation of paper "IBRNet: Learning Multi-View Image-Based Rendering", CVPR 2021.
IBRNet: Learning Multi-View Image-Based Rendering PyTorch implementation of paper "IBRNet: Learning Multi-View Image-Based Rendering", CVPR 2021. IBRN
Code for the paper: Hierarchical Reinforcement Learning With Timed Subgoals, published at NeurIPS 2021
Hierarchical reinforcement learning with Timed Subgoals (HiTS) This repository contains code for reproducing experiments from our paper "Hierarchical
Project for music generation system based on object tracking and CGAN
Project for music generation system based on object tracking and CGAN The project was inspired by MIDINet: A Convolutional Generative Adversarial Netw
Implementation of Memory-Efficient Neural Networks with Multi-Level Generation, ICCV 2021
Memory-Efficient Multi-Level In-Situ Generation (MLG) By Jiaqi Gu, Hanqing Zhu, Chenghao Feng, Mingjie Liu, Zixuan Jiang, Ray T. Chen and David Z. Pan
Face-Recognition-based-Attendance-System - An implementation of Attendance System in python.
Face-Recognition-based-Attendance-System A real time implementation of Attendance System in python. Pre-requisites To understand the implentation of F
Physics-informed Neural Operator for Learning Partial Differential Equation
PINO Physics-informed Neural Operator for Learning Partial Differential Equation Abstract: Machine learning methods have recently shown promise in sol
A scikit-learn-compatible module for estimating prediction intervals.
|Anaconda|_ MAPIE - Model Agnostic Prediction Interval Estimator MAPIE allows you to easily estimate prediction intervals using your favourite sklearn
Physics-Aware Training (PAT) is a method to train real physical systems with backpropagation.
Physics-Aware Training (PAT) is a method to train real physical systems with backpropagation. It was introduced in Wright, Logan G. & Onodera, Tatsuhiro et al. (2021)1 to train Physical Neural Networ
Understanding Convolution for Semantic Segmentation
TuSimple-DUC by Panqu Wang, Pengfei Chen, Ye Yuan, Ding Liu, Zehua Huang, Xiaodi Hou, and Garrison Cottrell. Introduction This repository is for Under
Active and Sample-Efficient Model Evaluation
Active Testing: Sample-Efficient Model Evaluation Hi, good to see you here! 👋 This is code for "Active Testing: Sample-Efficient Model Evaluation". P
A high-performance anchor-free YOLO. Exceeding yolov3~v5 with ONNX, TensorRT, NCNN, and Openvino supported.
YOLOX is an anchor-free version of YOLO, with a simpler design but better performance! It aims to bridge the gap between research and industrial communities. For more details, please refer to our rep
Dense Contrastive Learning (DenseCL) for self-supervised representation learning, CVPR 2021.
Dense Contrastive Learning for Self-Supervised Visual Pre-Training This project hosts the code for implementing the DenseCL algorithm for se
Efficient 3D Backbone Network for Temporal Modeling
VoV3D is an efficient and effective 3D backbone network for temporal modeling implemented on top of PySlowFast. Diverse Temporal Aggregation and
Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement
Decompose to Adapt: Cross-domain Object Detection via Feature Disentanglement In this project, we proposed a Domain Disentanglement Faster-RCNN (DDF)
📚 A collection of all the Deep Learning Metrics that I came across which are not accuracy/loss.
📚 A collection of all the Deep Learning Metrics that I came across which are not accuracy/loss.
Pytorch implementation for our ICCV 2021 paper "TRAR: Routing the Attention Spans in Transformers for Visual Question Answering".
TRAnsformer Routing Networks (TRAR) This is an official implementation for ICCV 2021 paper "TRAR: Routing the Attention Spans in Transformers for Visu
The-Secret-Sharing-Schemes - This interactive script demonstrates the Secret Sharing Schemes algorithm
The-Secret-Sharing-Schemes This interactive script demonstrates the Secret Shari
Code accompanying our NeurIPS 2021 traffic4cast challenge
Traffic forecasting on traffic movie snippets This repo contains all code to reproduce our approach to the IARAI Traffic4cast 2021 challenge. In the c
This is the first released system towards complex meters` detection and recognition, which is implemented by computer vision techniques.
A three-stage detection and recognition pipeline of complex meters in wild This is the first released system towards detection and recognition of comp
This is the repo of the manuscript "Dual-branch Attention-In-Attention Transformer for speech enhancement"
DB-AIAT: A Dual-branch attention-in-attention transformer for single-channel SE