josipd / Torch Two Sample
Licence: other
A PyTorch library for two-sample tests
Stars: ✭ 170
Labels
Projects that are alternatives of or similar to Torch Two Sample
Photorealistic Style Transfer
High-Resolution Network for Photorealistic Style Transfer
Stars: ✭ 170 (+0%)
Mutual labels: jupyter-notebook
Tf Tutorials
Tutorials for deep learning course here:
Stars: ✭ 169 (-0.59%)
Mutual labels: jupyter-notebook
Audiosignalprocessingforml
Code and slides of my YouTube series called "Audio Signal Proessing for Machine Learning"
Stars: ✭ 169 (-0.59%)
Mutual labels: jupyter-notebook
Deeplearning.ai Andrewng
deeplearning.ai , By Andrew Ng, All slide and notebook + data + solutions and video link
Stars: ✭ 165 (-2.94%)
Mutual labels: jupyter-notebook
Deep Learning Genomics Primer
Contains files for the deep learning in genomics primer.
Stars: ✭ 170 (+0%)
Mutual labels: jupyter-notebook
Colab
Continual Learning tutorials and demo running on Google Colaboratory.
Stars: ✭ 168 (-1.18%)
Mutual labels: jupyter-notebook
Structurenet
StructureNet: Hierarchical Graph Networks for 3D Shape Generation
Stars: ✭ 170 (+0%)
Mutual labels: jupyter-notebook
Projects
TLM Monthly Projects. Join our slack to work on them.
Stars: ✭ 169 (-0.59%)
Mutual labels: jupyter-notebook
Deep Reinforcement Learning Algorithms
31 projects in the framework of Deep Reinforcement Learning algorithms: Q-learning, DQN, PPO, DDPG, TD3, SAC, A2C and others. Each project is provided with a detailed training log.
Stars: ✭ 167 (-1.76%)
Mutual labels: jupyter-notebook
Notebooks
Notebooks using the Hugging Face libraries 🤗
Stars: ✭ 168 (-1.18%)
Mutual labels: jupyter-notebook
Prottrans
ProtTrans is providing state of the art pretrained language models for proteins. ProtTrans was trained on thousands of GPUs from Summit and hundreds of Google TPUs using Transformers Models.
Stars: ✭ 164 (-3.53%)
Mutual labels: jupyter-notebook
Mimic extract
MIMIC-Extract:A Data Extraction, Preprocessing, and Representation Pipeline for MIMIC-III
Stars: ✭ 168 (-1.18%)
Mutual labels: jupyter-notebook
Lungcancerdetection
Use CNN to detect nodules in LIDC dataset.
Stars: ✭ 168 (-1.18%)
Mutual labels: jupyter-notebook
2048 Deep Reinforcement Learning
Trained A Convolutional Neural Network To Play 2048 using Deep-Reinforcement Learning
Stars: ✭ 169 (-0.59%)
Mutual labels: jupyter-notebook
Imagination Augmented Agents
Building Agents with Imagination: pytorch step-by-step implementation
Stars: ✭ 170 (+0%)
Mutual labels: jupyter-notebook
Log Anomaly Detector
Log Anomaly Detection - Machine learning to detect abnormal events logs
Stars: ✭ 169 (-0.59%)
Mutual labels: jupyter-notebook
Gate Decorator Pruning
Code for the NuerIPS'19 paper "Gate Decorator: Global Filter Pruning Method for Accelerating Deep Convolutional Neural Networks"
Stars: ✭ 170 (+0%)
Mutual labels: jupyter-notebook
torch-two-sample
A PyTorch library for differentiable two-sample tests
Description
This package implements a total of six two sample tests:
- The classical Friedman-Rafsky test [FR79].
- The classical k-nearest neighbours (kNN) test [FR83].
- The differentiable Friedman-Rafsky test [DK17].
- The differentiable k-nearest neighbours (kNN) test [DK17].
- The maximum mean discrepancy (MMD) test [GBR+12].
- The energy test [SzekelyR13].
Please refer to the documentation for more information about the project. You can also have a look at the following notebook that showcases how to use the code to train a generative model on MNIST.
Installation
After installing PyTorch, you can install the package with:
python setup.py install
Testing
To run the tests you simply have to run:
python setup.py test
Note that you will need to have Shogun installed for one of the test cases.
Bibliography
- [DK17] J. Djolonga and A. Krause. Learning Implicit Generative Models Using Differentiable Graph Tests. ArXiv e-prints, September 2017. arXiv:1709.01006.
- [FR79] Jerome H Friedman and Lawrence C Rafsky. Multivariate generalizations of the wald-wolfowitz and smirnov two-sample tests. Annals of Statistics, pages 697–717, 1979.
- [FR83] Jerome H Friedman and Lawrence C Rafsky. Graph-theoretic measures of multivariate association and prediction. Annals of Statistics, pages 377–391, 1983.
- [GBR+12] Arthur Gretton, Karsten M Borgwardt, Malte J Rasch, Bernhard Schölkopf, and Alexander Smola. A kernel two-sample test. Journal of Machine Learning Research, 13(Mar):723–773, 2012.
- [SST+12] Kevin Swersky, Ilya Sutskever, Daniel Tarlow, Richard S Zemel, Ruslan R Salakhutdinov, and Ryan P Adams. Cardinality restricted boltzmann machines. In Advances in Neural Information Processing Systems (NIPS), 3293–3301. 2012.
- [SzekelyR13] Gábor J Székely and Maria L Rizzo. Energy statistics: a class of statistics based on distances. Journal of Statistical Planning and Inference, 143(8):1249–1272, 2013.
- [TSZ+12] Daniel Tarlow, Kevin Swersky, Richard S Zemel, Ryan Prescott Adams, and Brendan J Frey. Fast exact inference for recursive cardinality models. Uncertainty in Artificial Intelligence (UAI), 2012.
Note that the project description data, including the texts, logos, images, and/or trademarks,
for each open source project belongs to its rightful owner.
If you wish to add or remove any projects, please contact us at [email protected].
