Beginner’s Guide To Text Classification Using PyCaret

Introduction Have you ever solved a Machine Learning problem in just one go? Solving a problem using machine learning isn’t straightforward. It involves various steps to come up with an accurate solution. The process/steps to be followed for solving an ml problem is known as ML Pipeline/ML Cycle. ML Pipeline/ ML Cycle (Credits: https://medium.com/analytics-vidhya/machine-learning-development-life-cycle-dfe88c44222e) As shown in the figure, the Machine Learning pipeline consists of different steps like: Understand Problem Statement, Hypothesis Generation, Exploratory Data Analysis, Data Preprocessing, Feature Engineering, […]

Read more

Neural Semi supervised Learning for Text Classification Under Large Scale Pretraining

Neural-Semi-Supervised-Learning-for-Text-Classification Neural Semi supervised Learning for Text Classification Under Large Scale Pretraining. Download Models and Dataset Datasets and Models are found in the follwing list. Download 3.4M IMDB movie reviews. Save the data at [REVIEWS_PATH].You can download the dataset HERE. Download the vanilla RoBERTa-large model released by HuggingFace. Save the model at [VANILLA_ROBERTA_LARGE_PATH].You can download the model HERE. Download in-domain pretrained models in the paper and save the model at [PRETRAIN_MODELS]. We provide three following models.You can download HERE. init-roberta-base: […]

Read more

An open-source library of algorithms to analyse time series in GPU and CPU

Khiva Khiva is an open-source library of efficient algorithms to analyse time series in GPU and CPU. It can be used to extract insights from one or a group of time series. The large number of available methods allow us to understand the nature of each time series. Based on the results of this analysis, users can reduce dimensionality, find out recurrent motifs or discords, understand the seasonality or trend from a given time series, forecasting and detect anomalies. Khiva […]

Read more

Train in Germany, Test in The USA: Making 3D Object Detectors Generalize

3D_adapt_auto_driving This paper has been accpeted by Conference on Computer Vision and Pattern Recognition (CVPR) 2020. Train in Germany, Test in The USA: Making 3D Object Detectors Generalize by Yan Wang*, Xiangyu Chen*, Yurong You, Li Erran, Bharath Hariharan, Mark Campbell, Kilian Q. Weinberger, Wei-Lun Chao* Dependencies Usage Prepare Datasets (Jupyter notebook) We develop our method on these datasets: Configure dataset_path in config_path.py. Raw datasets will be organized as the following structure: dataset_path/ | kitti/ # KITTI object detection 3D […]

Read more

Image2Reverb: Cross-Model Reverb Impulse Response Synthesis

Image2Reverb Image2Reverb: Cross-Modal Reverb Impulse Response SynthesisNikhil Singh, Jeff Mentch, Jerry Ng, Matthew Beveridge, Iddo Drori Code for the ICCV 2021 paper [arXiv]. Image2Reverb is a method for generating audio impulse responses, to simulate the acoustic reverberation of a given environment, from a 2D image of it. Dependencies Model/Data: PyTorch>=1.7.0 PyTorch Lightning torchvision torchaudio librosa PyRoomAcoustics PIL Eval/Preprocessing: PySoundfile SciPy Scikit-Learn python-acoustics google-images-download matplotlib Resources Model Checkpoint Code Acknowlegdements We borrow and adapt code snippets from GANSynth (and this PyTorch […]

Read more

A-CNN: Annularly Convolutional Neural Networks on Point Clouds

A-CNN Created by Artem Komarichev, Zichun Zhong, Jing Hua from Department of Computer Science, Wayne State University. Introduction Our paper (arXiV) proposes a new approach to define and compute convolution directly on 3D point clouds by the proposed annular convolution. To appear, Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), June 2019 A-CNN usage We provide the code of A-CNN model that was tested with Tensorflow 1.3.0, CUDA 8.0, and python 3.6 on Ubuntu 16.04. We […]

Read more

Improving Unsupervised Image Clustering With Robust Learning

RUC This repo is the PyTorch codes for “Improving Unsupervised Image Clustering With Robust Learning (RUC)” Improving Unsupervised Image Clustering With Robust Learning Sungwon Park, Sungwon Han, Sundong Kim, Danu Kim, Sungkyu Park, Seunghoon Hong, Meeyoung Cha. Highlight RUC is an add-on module to enhance the performance of any off-the-shelf unsupervised learning algorithms. RUC is inspired by robust learning. It first divides clustered data points into clean and noisy set, then refine the clustering results. With RUC, state-of-the-art unsupervised clustering […]

Read more

An Out-of-Distribution Detection Score For Variational Auto-encoder

Likelihood-Regret Official implementation of Likelihood Regret: An Out-of-Distribution Detection Score For Variational Auto-encoder at NeurIPS 2020. Training To train the VAEs, use appropriate arguments and run this command: python train_pixel.py Evaluation To evaluate likelihood regret’s OOD detection performance, run python compute_LR.py To evaluate likelihood ratio, run python test_likelihood_ratio.py To evaluate input complexity, run python test_inputcomplexity.py Above commands will save the numpy arrays containing the OOD scores for in-distribution and OOD samples in specific location, and to compute aucroc score, run […]

Read more

Pixel-level self-paced learning for super-resolution

This is an official implementaion of the paper Pixel-level Self-Paced Learning for Super-Resolution, which has been accepted by ICASSP 2020. [arxiv][PDF] trained model files: Baidu Pan(code: v0be) Requirements This code is forked from thstkdgus35/EDSR-PyTorch. In the light of its README, following libraries are required: Python 3.6+ (Python 3.7.0 in my experiments) PyTorch >= 1.0.0 (1.1.0 in my experiments) numpy skimage imageio matplotlib tqdm Core Parts Detail code can be found in Loss.forward, which can be simplified as: # take L1 […]

Read more

Block-wisely Supervised Neural Architecture Search with Knowledge Distillation

This repository provides the code of our paper: Blockwisely Supervised Neural Architecture Search with Knowledge Distillation. Illustration of DNA. Each cell of the supernet is trained independently to mimic the behavior of the corresponding teacher block. Comparison of model ranking for DNA vs. DARTS, SPOS and MnasNet under two different hyper-parameters. Our Trained Models Usage 1. Requirements Install PyTorch (pytorch.org) Install third-party requirements Download the ImageNet dataset and move validation images to labeled subfolders 2. Searching The code for supernet […]

Read more
1 18 19 20 21 22 51