diff --git a/_data/ecosystem/pted/2021/posters.yaml b/_data/ecosystem/pted/2021/posters.yaml index dd4f3bfc7412..5f5f22524a3a 100644 --- a/_data/ecosystem/pted/2021/posters.yaml +++ b/_data/ecosystem/pted/2021/posters.yaml @@ -10,7 +10,7 @@ are provided as a Torch tensor with a defined gradient. We highlight how this functionality can be used to explore new paradigms in machine learning, including the use of hybrid models for transfer learning. - link: http://www.pennylane.ai + link: http://pennylane.ai poster_link: https://s3.amazonaws.com/assets.pytorch.org/pted2021/posters/K1.png section: K1 thumbnail_link: https://s3.amazonaws.com/assets.pytorch.org/pted2021/posters/thumb-K1.png @@ -321,7 +321,7 @@ supports accelerated mixed precision training. AMD also provides hardware support for the PyTorch community build to help develop and maintain new features. This poster will highlight some of the work that has gone into enabling PyTorch support. - link: www.amd.com/rocm + link: https://www.amd.com/rocm poster_link: https://s3.amazonaws.com/assets.pytorch.org/pted2021/posters/K8.png section: K8 thumbnail_link: https://s3.amazonaws.com/assets.pytorch.org/pted2021/posters/thumb-K8.png diff --git a/_mobile/android.md b/_mobile/android.md index eb25100b8da6..f057a28806ae 100644 --- a/_mobile/android.md +++ b/_mobile/android.md @@ -94,7 +94,7 @@ Tensor inputTensor = TensorImageUtils.bitmapToFloat32Tensor(bitmap, TensorImageUtils.TORCHVISION_NORM_MEAN_RGB, TensorImageUtils.TORCHVISION_NORM_STD_RGB); ``` `org.pytorch.torchvision.TensorImageUtils` is part of `org.pytorch:pytorch_android_torchvision` library. -The `TensorImageUtils#bitmapToFloat32Tensor` method creates tensors in the [torchvision format](https://pytorch.org/docs/stable/torchvision/models.html) using `android.graphics.Bitmap` as a source. +The `TensorImageUtils#bitmapToFloat32Tensor` method creates tensors in the [torchvision format](https://pytorch.org/vision/stable/models.html) using `android.graphics.Bitmap` as a source. > All pre-trained models expect input images normalized in the same way, i.e. mini-batches of 3-channel RGB images of shape (3 x H x W), where H and W are expected to be at least 224. > The images have to be loaded in to a range of `[0, 1]` and then normalized using `mean = [0.485, 0.456, 0.406]` and `std = [0.229, 0.224, 0.225]` diff --git a/_mobile/ios.md b/_mobile/ios.md index fb00da2ebd40..585191cc764f 100644 --- a/_mobile/ios.md +++ b/_mobile/ios.md @@ -23,7 +23,7 @@ HelloWorld is a simple image classification application that demonstrates how to ### Model Preparation -Let's start with model preparation. If you are familiar with PyTorch, you probably should already know how to train and save your model. In case you don't, we are going to use a pre-trained image classification model - [MobileNet v2](https://pytorch.org/hub/pytorch_vision_mobilenet_v2/), which is already packaged in [TorchVision](https://pytorch.org/docs/stable/torchvision/index.html). To install it, run the command below. +Let's start with model preparation. If you are familiar with PyTorch, you probably should already know how to train and save your model. In case you don't, we are going to use a pre-trained image classification model - [MobileNet v2](https://pytorch.org/hub/pytorch_vision_mobilenet_v2/), which is already packaged in [TorchVision](https://pytorch.org/vision/stable/index.html). To install it, run the command below. > We highly recommend following the [Pytorch Github page](https://github.com/pytorch/pytorch) to set up the Python development environment on your local machine. diff --git a/_posts/2018-03-5-tensor-comprehensions.md b/_posts/2018-03-5-tensor-comprehensions.md index a777c076a432..df83ea75dccd 100644 --- a/_posts/2018-03-5-tensor-comprehensions.md +++ b/_posts/2018-03-5-tensor-comprehensions.md @@ -34,7 +34,7 @@ conda install -c pytorch -c tensorcomp tensor_comprehensions At this time we only provide Linux-64 binaries which have been tested on Ubuntu 16.04 and CentOS7. -TC depends on heavyweight C++ projects such as [Halide](http://halide-lang.org/), [Tapir-LLVM](https://github.com/wsmoses/Tapir-LLVM) and [ISL](http://isl.gforge.inria.fr/). Hence, we rely on Anaconda to distribute these dependencies reliably. For the same reason, TC is not available via PyPI. +TC depends on heavyweight C++ projects such as [Halide](http://halide-lang.org/), [Tapir-LLVM](https://github.com/wsmoses/Tapir-LLVM) and ISL. Hence, we rely on Anaconda to distribute these dependencies reliably. For the same reason, TC is not available via PyPI. #### 2. Import the python package @@ -74,8 +74,6 @@ The autotuner is your biggest friend. You generally do not want to use a `tc` fu When the autotuning is running, the current best performance is displayed. If you are satisfied with the current result or you are out of time, stop the tuning procedure by pressing `Ctrl+C`. - - `cache` saves the results of the autotuned kernel search and saves it to the file `fcrelu_100_128_100.tc`. The next time you call the same line of code, it loads the results of the autotuning without recomputing it. The autotuner has a few hyperparameters (just like your ConvNet has learning rate, number of layers, etc.). We pick reasonable defaults, but you can read about using advanced options [here](https://facebookresearch.github.io/TensorComprehensions/framework/pytorch_integration/writing_layers.html#specifying-mapping-options). @@ -146,7 +144,7 @@ Note: the syntax for passing in scalars is subject to change in the next release ## torch.nn layers -We added some sugar-coating around the basic PyTorch integration of TC to make it easy to integrate TC into larger `torch.nn` models by defining the forward and backward TC expressions and taking `Variable` inputs / outputs. Here is an [example](https://github.com/facebookresearch/TensorComprehensions/blob/master/test_python/layers/test_convolution_train.py) of defining a convolution layer with TC. +We added some sugar-coating around the basic PyTorch integration of TC to make it easy to integrate TC into larger `torch.nn` models by defining the forward and backward TC expressions and taking `Variable` inputs / outputs. ## Some essentials that you will miss (we're working on them) @@ -183,12 +181,12 @@ You cannot write this operation in TC: `torch.matmul(...).view(...).mean(...)`. ## Getting Started - [Walk through Tutorial](https://facebookresearch.github.io/TensorComprehensions/framework/pytorch_integration/writing_layers.html) to quickly get started with understanding and using Tensor Comprehensions PyTorch package. -- [Over 20 examples](https://github.com/facebookresearch/TensorComprehensions/tree/master/test_python/layers) of various ML layers with TC, including `avgpool`, `maxpool`, `matmul`, matmul - give output buffers and `batch-matmul`, `convolution`, `strided-convolution`, `batchnorm`, `copy`, `cosine similarity`, `Linear`, `Linear + ReLU`, `group-convolutions`, strided `group-convolutions`, `indexing`, `Embedding` (lookup table), small-mobilenet, `softmax`, `tensordot`, `transpose` +- Over 20 examples of various ML layers with TC, including `avgpool`, `maxpool`, `matmul`, matmul - give output buffers and `batch-matmul`, `convolution`, `strided-convolution`, `batchnorm`, `copy`, `cosine similarity`, `Linear`, `Linear + ReLU`, `group-convolutions`, strided `group-convolutions`, `indexing`, `Embedding` (lookup table), small-mobilenet, `softmax`, `tensordot`, `transpose` - [Detailed docs](https://facebookresearch.github.io/TensorComprehensions/framework/pytorch_integration/getting_started.html) on Tensor Comprehensions and integration with PyTorch. ## Communication -- [Slack](https://tensorcomprehensions.herokuapp.com/): For discussion around framework integration, build support, collaboration, etc. join our slack channel. +- Slack: For discussion around framework integration, build support, collaboration, etc. join our slack channel. - Email: tensorcomp@fb.com - [GitHub](https://github.com/facebookresearch/TensorComprehensions): bug reports, feature requests, install issues, RFCs, thoughts, etc. diff --git a/_posts/2019-05-08-model-serving-in-pyorch.md b/_posts/2019-05-08-model-serving-in-pyorch.md index c25b1c89f7ab..512268e5f198 100644 --- a/_posts/2019-05-08-model-serving-in-pyorch.md +++ b/_posts/2019-05-08-model-serving-in-pyorch.md @@ -52,7 +52,7 @@ If you can't use the cloud or prefer to manage all services using the same techn If you want to manage multiple models within a non-cloud service solution, there are teams developing PyTorch support in model servers like [MLFlow](https://mlflow.org/), [Kubeflow](https://www.kubeflow.org/), and [RedisAI.](https://oss.redislabs.com/redisai/) We're excited to see innovation from multiple teams building OSS model servers, and we'll continue to highlight innovation in the PyTorch ecosystem in the future. -If you can use the cloud for your application, there are several great choices for working with models in the cloud. For AWS Sagemaker, you can start find a guide to [all of the resources from AWS for working with PyTorch](https://docs.aws.amazon.com/sagemaker/latest/dg/pytorch.html), including docs on how to use the [Sagemaker Python SDK](https://sagemaker.readthedocs.io/en/stable/using_pytorch.html). You can also see [some](https://youtu.be/5h1Ot2dPi2E) [talks](https://youtu.be/qc5ZikKw9_w) we've given on using PyTorch on Sagemaker. Finally, if you happen to be using PyTorch via FastAI, then they've written a [really simple guide](https://course.fast.ai/deployment_amzn_sagemaker.html) to getting up and running on Sagemaker. +If you can use the cloud for your application, there are several great choices for working with models in the cloud. For AWS Sagemaker, you can start find a guide to [all of the resources from AWS for working with PyTorch](https://docs.aws.amazon.com/sagemaker/latest/dg/pytorch.html), including docs on how to use the [Sagemaker Python SDK](https://sagemaker.readthedocs.io/en/stable/using_pytorch.html). You can also see [some](https://youtu.be/5h1Ot2dPi2E) [talks](https://youtu.be/qc5ZikKw9_w) we've given on using PyTorch on Sagemaker. Finally, if you happen to be using PyTorch via FastAI, then they've written a really simple guide to getting up and running on Sagemaker. The story is similar across other major clouds. On Google Cloud, you can follow [these instructions](https://cloud.google.com/deep-learning-vm/docs/pytorch_start_instance) to get access to a Deep Learning VM with PyTorch pre-installed. On Microsoft Azure, you have a number of ways to get started from [Azure Machine Learning Service](https://azure.microsoft.com/en-us/services/machine-learning-service/) to [Azure Notebooks](https://notebooks.azure.com/pytorch/projects/tutorials) showing how to use PyTorch. diff --git a/_posts/2019-06-10-towards-reproducible-research-with-pytorch-hub.md b/_posts/2019-06-10-towards-reproducible-research-with-pytorch-hub.md index 3bdd2db84dbe..35a4306d7557 100644 --- a/_posts/2019-06-10-towards-reproducible-research-with-pytorch-hub.md +++ b/_posts/2019-06-10-towards-reproducible-research-with-pytorch-hub.md @@ -106,7 +106,7 @@ Users can list all available entrypoints in a repo using the ```torch.hub.list() 'vgg19_bn'] ``` -Note that PyTorch Hub also allows auxillary entrypoints (other than pretrained models), e.g. ```bertTokenizer``` for preprocessing in the [BERT](https://pytorch.org/hub/huggingface_pytorch-pretrained-bert_bert/) models, to make the user workflow smoother. +Note that PyTorch Hub also allows auxillary entrypoints (other than pretrained models), e.g. ```bertTokenizer``` for preprocessing in the BERT models, to make the user workflow smoother. ### Load a model @@ -164,7 +164,7 @@ forward(input_ids, token_type_ids=None, attention_mask=None, masked_lm_labels=No ... ``` -Have a closer look at the [BERT](https://pytorch.org/hub/huggingface_pytorch-pretrained-bert_bert/) and [DeepLabV3](https://pytorch.org/hub/pytorch_vision_deeplabv3_resnet101/) pages, where you can see how these models can be used once loaded. +Have a closer look at the BERT and [DeepLabV3](https://pytorch.org/hub/pytorch_vision_deeplabv3_resnet101/) pages, where you can see how these models can be used once loaded. ### Other ways to explore diff --git a/_posts/2019-07-18-pytorch-ecosystem.md b/_posts/2019-07-18-pytorch-ecosystem.md index 3b87f2f10d0f..7351cbbd9d4f 100644 --- a/_posts/2019-07-18-pytorch-ecosystem.md +++ b/_posts/2019-07-18-pytorch-ecosystem.md @@ -45,7 +45,7 @@ If you would like to have your project included in the PyTorch ecosystem and fea ## PyTorch Hub for reproducible research | New models -Since [launching](https://pytorch.org/blog/towards-reproducible-research-with-pytorch-hub/) the PyTorch Hub in beta, we’ve received a lot of interest from the community including the contribution of many new models. Some of the latest include [U-Net for Brain MRI](https://pytorch.org/hub/mateuszbuda_brain-segmentation-pytorch_unet/) contributed by researchers at Duke University, [Single Shot Detection](https://pytorch.org/hub/nvidia_deeplearningexamples_ssd/) from NVIDIA and [Transformer-XL](https://pytorch.org/hub/huggingface_pytorch-pretrained-bert_transformerXL/) from HuggingFace. +Since [launching](https://pytorch.org/blog/towards-reproducible-research-with-pytorch-hub/) the PyTorch Hub in beta, we’ve received a lot of interest from the community including the contribution of many new models. Some of the latest include [U-Net for Brain MRI](https://pytorch.org/hub/mateuszbuda_brain-segmentation-pytorch_unet/) contributed by researchers at Duke University, [Single Shot Detection](https://pytorch.org/hub/nvidia_deeplearningexamples_ssd/) from NVIDIA and Transformer-XL from HuggingFace. We’ve seen organic integration of the PyTorch Hub by folks like [paperswithcode](https://paperswithcode.com/), making it even easier for you to try out the state of the art in AI research. In addition, companies like [Seldon](https://github.com/axsaucedo/seldon-core/tree/pytorch_hub/examples/models/pytorchhub) provide production-level support for PyTorch Hub models on top of Kubernetes. diff --git a/_posts/2019-08-08-pytorch-1.2-and-domain-api-release.md b/_posts/2019-08-08-pytorch-1.2-and-domain-api-release.md index 5e8ce05d52f8..bcc30d86963a 100644 --- a/_posts/2019-08-08-pytorch-1.2-and-domain-api-release.md +++ b/_posts/2019-08-08-pytorch-1.2-and-domain-api-release.md @@ -115,9 +115,9 @@ We are excited to see an active community around torchaudio and eager to further ## Torchtext 0.4 with supervised learning datasets -A key focus area of torchtext is to provide the fundamental elements to help accelerate NLP research. This includes easy access to commonly used datasets and basic preprocessing pipelines for working on raw text based data. The torchtext 0.4.0 release includes several popular supervised learning baselines with "one-command" data loading. A [tutorial](https://pytorch.org/tutorials/beginner/text_sentiment_ngrams_tutorial.html) is included to show how to use the new datasets for text classification analysis. We also added and improved on a few functions such as [get_tokenizer](https://pytorch.org/text/data.html?highlight=get_tokenizer#torchtext.data.get_tokenizer) and [build_vocab_from_iterator](https://pytorch.org/text/vocab.html#build-vocab-from-iterator) to make it easier to implement future datasets. Additional examples can be found [here](https://github.com/pytorch/text/tree/master/examples/text_classification). +A key focus area of torchtext is to provide the fundamental elements to help accelerate NLP research. This includes easy access to commonly used datasets and basic preprocessing pipelines for working on raw text based data. The torchtext 0.4.0 release includes several popular supervised learning baselines with "one-command" data loading. A [tutorial](https://pytorch.org/tutorials/beginner/text_sentiment_ngrams_tutorial.html) is included to show how to use the new datasets for text classification analysis. We also added and improved on a few functions such as get_tokenizer and build_vocab_from_iterator to make it easier to implement future datasets. Additional examples can be found [here](https://github.com/pytorch/text/tree/master/examples/text_classification). -Text classification is an important task in Natural Language Processing with many applications, such as sentiment analysis. The new release includes several popular [text classification datasets](https://pytorch.org/text/datasets.html?highlight=textclassification#torchtext.datasets.TextClassificationDataset) for supervised learning including: +Text classification is an important task in Natural Language Processing with many applications, such as sentiment analysis. The new release includes several popular text classification datasets for supervised learning including: * AG_NEWS * SogouNews diff --git a/_posts/2019-12-06-pytorch-adds-new-tools-and-libraries-welcomes-preferred-networks-to-its-community.md b/_posts/2019-12-06-pytorch-adds-new-tools-and-libraries-welcomes-preferred-networks-to-its-community.md index 0a13979349de..d39d84959ce5 100644 --- a/_posts/2019-12-06-pytorch-adds-new-tools-and-libraries-welcomes-preferred-networks-to-its-community.md +++ b/_posts/2019-12-06-pytorch-adds-new-tools-and-libraries-welcomes-preferred-networks-to-its-community.md @@ -39,7 +39,7 @@ Image and video classification are at the core of content understanding. To that * Ease of use - This framework features a modular, flexible design that allows anyone to train machine learning models on top of PyTorch using very simple abstractions. The system also has out-of-the-box integration with AWS on PyTorch Elastic, facilitating research at scale and making it simple to move between research and production. * High performance - Researchers can use the framework to train models such as Resnet50 on ImageNet in as little as 15 minutes. -You can learn more at the [NeurIPS Expo workshop](https://nips.cc/ExpoConferences/2019/schedule?workshop_id=16) on Multi-Modal research to production or get started with the PyTorch Elastic Imagenet example [here](https://github.com/pytorch/elastic/blob/master/examples/imagenet/main.py). +You can learn more at the NeurIPS Expo workshop on Multi-Modal research to production or get started with the PyTorch Elastic Imagenet example [here](https://github.com/pytorch/elastic/blob/master/examples/imagenet/main.py). ## Come see us at NeurIPS @@ -47,13 +47,13 @@ The PyTorch team will be hosting workshops at NeurIPS during the industry expo o We’re also publishing a [paper that details the principles that drove the implementation of PyTorch](https://papers.nips.cc/paper/9015-pytorch-an-imperative-style-high-performance-deep-learning-library) and how they’re reflected in its architecture. -*[Multi-modal Research to Production](https://nips.cc/ExpoConferences/2019/schedule?workshop_id=16)* - This workshop will dive into a number of modalities such as computer vision (large scale image classification and instance segmentation) and Translation and Speech (seq-to-seq Transformers) from the lens of taking cutting edge research to production. Lastly, we will also walk through how to use the latest APIs in PyTorch to take eager mode developed models into graph mode via Torchscript and quantize them for scale production deployment on servers or mobile devices. Libraries used include: +*Multi-modal Research to Production* - This workshop will dive into a number of modalities such as computer vision (large scale image classification and instance segmentation) and Translation and Speech (seq-to-seq Transformers) from the lens of taking cutting edge research to production. Lastly, we will also walk through how to use the latest APIs in PyTorch to take eager mode developed models into graph mode via Torchscript and quantize them for scale production deployment on servers or mobile devices. Libraries used include: * Classification Framework - a newly open sourced PyTorch framework developed by Facebook AI for research on large-scale image and video classification. It allows researchers to quickly prototype and iterate on large distributed training jobs. Models built on the framework can be seamlessly deployed to production. * Detectron2 - the recently released object detection library built by the Facebook AI Research computer vision team. We will articulate the improvements over the previous version including: 1) Support for latest models and new tasks; 2) Increased flexibility, to enable new computer vision research; 3) Maintainable and scalable, to support production use cases. * Fairseq - general purpose sequence-to-sequence library, can be used in many applications, including (unsupervised) translation, summarization, dialog and speech recognition. -*[Responsible and Reproducible AI](https://nips.cc/ExpoConferences/2019/schedule?workshop_id=14)* - This workshop on Responsible and Reproducible AI will dive into important areas that are shaping the future of how we interpret, reproduce research, and build AI with privacy in mind. We will cover major challenges, walk through solutions, and finish each talk with a hands-on tutorial. +*Responsible and Reproducible AI* - This workshop on Responsible and Reproducible AI will dive into important areas that are shaping the future of how we interpret, reproduce research, and build AI with privacy in mind. We will cover major challenges, walk through solutions, and finish each talk with a hands-on tutorial. * Reproducibility: As the number of research papers submitted to arXiv and conferences skyrockets, scaling reproducibility becomes difficult. We must address the following challenges: aid extensibility by standardizing code bases, democratize paper implementation by writing hardware agnostic code, facilitate results validation by documenting “tricks” authors use to make their complex systems function. To offer solutions, we will dive into tool like PyTorch Hub and PyTorch Lightning which are used by some of the top researchers in the world to reproduce the state of the art. * Interpretability: With the increase in model complexity and the resulting lack of transparency, model interpretability methods have become increasingly important. Model understanding is both an active area of research as well as an area of focus for practical applications across industries using machine learning. To get hands on, we will use the recently released Captum library that provides state-of-the-art algorithms to provide researchers and developers with an easy way to understand the importance of neurons/layers and the predictions made by our models.` diff --git a/_posts/2020-07-28-pytorch-feature-classification-changes.md b/_posts/2020-07-28-pytorch-feature-classification-changes.md index 9ff4291513aa..057867a68158 100644 --- a/_posts/2020-07-28-pytorch-feature-classification-changes.md +++ b/_posts/2020-07-28-pytorch-feature-classification-changes.md @@ -42,7 +42,7 @@ Additionally, the following features will be reclassified under this new rubric: 5. [Channels Last Memory Layout](https://pytorch.org/docs/stable/tensor_attributes.html#torch-memory-format): Beta (was Experimental) 6. [Custom C++ Classes](https://pytorch.org/docs/stable/jit.html?highlight=experimental): Beta (was Experimental) 7. [PyTorch Mobile](https://pytorch.org/mobile/home/): Beta (was Experimental) -8. [Java Bindings](https://pytorch.org/docs/stable/packages.html#): Beta (was Experimental) +8. [Java Bindings](https://pytorch.org/docs/stable/index.html): Beta (was Experimental) 9. [Torch.Sparse](https://pytorch.org/docs/stable/sparse.html?highlight=experimental#): Beta (was Experimental) diff --git a/_posts/2020-1-15-pytorch-1-dot-4-released-and-domain-libraries-updated.md b/_posts/2020-1-15-pytorch-1-dot-4-released-and-domain-libraries-updated.md index 2be782f18b47..e55070202d16 100644 --- a/_posts/2020-1-15-pytorch-1-dot-4-released-and-domain-libraries-updated.md +++ b/_posts/2020-1-15-pytorch-1-dot-4-released-and-domain-libraries-updated.md @@ -40,8 +40,6 @@ With the scale of models, such as RoBERTa, continuing to increase into the billi To learn more about the APIs and the design of this feature, see the links below: * [API documentation](https://pytorch.org/docs/stable/rpc.html) -* [Distributed Autograd design doc](https://pytorch.org/docs/stable/notes/distributed_autograd.html) -* [Remote Reference design doc](https://pytorch.org/docs/stable/notes/rref.html) For the full tutorials, see the links below: diff --git a/_posts/2020-10-1-announcing-the-winners-of-the-2020-global-pytorch-summer-hackathon.md b/_posts/2020-10-1-announcing-the-winners-of-the-2020-global-pytorch-summer-hackathon.md index 0bc542fe8484..76c7e05ef8f4 100644 --- a/_posts/2020-10-1-announcing-the-winners-of-the-2020-global-pytorch-summer-hackathon.md +++ b/_posts/2020-10-1-announcing-the-winners-of-the-2020-global-pytorch-summer-hackathon.md @@ -63,7 +63,7 @@ A PyTorch-based automated machine learning (AutoML) solution, carefree-learn pro **3rd Place** - [TorchExpo](https://devpost.com/software/torchexpo) -TorchExpo is a collection of models and extensions that simplifies taking PyTorch from research to production in mobile devices. This library is more than a web and mobile application, and also comes with a Python library. The Python library is available via pip install and it helps researchers convert a state-of-the-art model in TorchScript and ONNX format in just one line. Detailed docs are available [here](https://torchexpo.readthedocs.io/en/latest/). +TorchExpo is a collection of models and extensions that simplifies taking PyTorch from research to production in mobile devices. This library is more than a web and mobile application, and also comes with a Python library. The Python library is available via pip install and it helps researchers convert a state-of-the-art model in TorchScript and ONNX format in just one line. ## Web/Mobile Applications Powered by PyTorch @@ -91,10 +91,7 @@ FairTorch is a fairness library for PyTorch. It lets developers add constraints
diff --git a/_posts/2020-10-27-pytorch-1.7-released.md b/_posts/2020-10-27-pytorch-1.7-released.md index a3679b25e166..766ed9889ef6 100644 --- a/_posts/2020-10-27-pytorch-1.7-released.md +++ b/_posts/2020-10-27-pytorch-1.7-released.md @@ -62,7 +62,6 @@ Note that this is necessary, **but not sufficient**, for determinism **within a See the documentation for ```torch.set_deterministic(bool)``` for the list of affected operations. * [RFC](https://github.com/pytorch/pytorch/issues/15359) -* [Documentation](https://pytorch.org/docs/stable/generated/torch.set_deterministic.html) # Performance & Profiling ## [Beta] Stack traces added to profiler diff --git a/_posts/2020-11-1-pytorch-developer-day-2020.md b/_posts/2020-11-1-pytorch-developer-day-2020.md index ef6cb6ab8144..c68bafaec0d1 100644 --- a/_posts/2020-11-1-pytorch-developer-day-2020.md +++ b/_posts/2020-11-1-pytorch-developer-day-2020.md @@ -16,9 +16,8 @@ For Developer Day, we have an online networking event limited to people composed All talks will be livestreamed and available to the public. * [Livestream event page](https://www.facebook.com/events/802177440559164/) -* [Apply for an invitation to the networking event](https://pytorchdeveloperday.fbreg.com/apply) -Visit the [event website](https://pytorchdeveloperday.fbreg.com/) to learn more. We look forward to welcoming you to PyTorch Developer Day on November 12th! +Visit the event website to learn more. We look forward to welcoming you to PyTorch Developer Day on November 12th! Thank you, diff --git a/_posts/2020-4-21-pytorch-library-updates-new-model-serving-library.md b/_posts/2020-4-21-pytorch-library-updates-new-model-serving-library.md index 69101b8abc09..af49e31a38f7 100644 --- a/_posts/2020-4-21-pytorch-library-updates-new-model-serving-library.md +++ b/_posts/2020-4-21-pytorch-library-updates-new-model-serving-library.md @@ -61,7 +61,7 @@ The torchvision 0.6 release includes updates to datasets, models and a significa * Added `aligned` flag to `RoIAlign` to match Detectron2. * Refactored abstractions for C++ video decoder -See the release full notes [here](https://github.com/pytorch/vision/releases) and full docs can be found [here](https://pytorch.org/docs/stable/torchvision/index.html). +See the release full notes [here](https://github.com/pytorch/vision/releases) and full docs can be found [here](https://pytorch.org/vision/stable/index.html). ### torchtext 0.6 The torchtext 0.6 release includes a number of bug fixes and improvements to documentation. Based on user's feedback, dataset abstractions are currently being redesigned also. Highlights for the release include: diff --git a/_posts/2020-7-28-pytorch-1.6-released.md b/_posts/2020-7-28-pytorch-1.6-released.md index eb07fe53867a..9d1f6442249a 100644 --- a/_posts/2020-7-28-pytorch-1.6-released.md +++ b/_posts/2020-7-28-pytorch-1.6-released.md @@ -101,7 +101,7 @@ torch.distributed.rpc.rpc_sync(...) ``` * Design doc ([Link](https://github.com/pytorch/pytorch/issues/35251)) -* Documentation ([Link](https://pytorch.org/docs/stable/rpc/index.html)) +* Documentation ([Link](https://pytorch.org/docs/stable/)) ## [Beta] DDP+RPC @@ -123,7 +123,7 @@ for data in batch: ``` * DDP+RPC Tutorial ([Link](https://pytorch.org/tutorials/advanced/rpc_ddp_tutorial.html)) -* Documentation ([Link](https://pytorch.org/docs/stable/rpc/index.html)) +* Documentation ([Link](https://pytorch.org/docs/stable/)) * Usage Examples ([Link](https://github.com/pytorch/examples/pull/800)) ## [Beta] RPC - Asynchronous User Functions @@ -147,7 +147,7 @@ ret = rpc.rpc_sync( print(ret) # prints tensor([3., 3.]) ``` -* Tutorial for performant batch RPC using Asynchronous User Functions ([Link](https://github.com/pytorch/tutorials/blob/release/1.6/intermediate_source/rpc_async_execution.rst)) +* Tutorial for performant batch RPC using Asynchronous User Functions * Documentation ([Link](https://pytorch.org/docs/stable/rpc.html#torch.distributed.rpc.functions.async_execution)) * Usage examples ([Link](https://github.com/pytorch/examples/tree/master/distributed/rpc/batch)) diff --git a/_posts/2021-08-23-pytorch-developer-day-2021.md b/_posts/2021-08-23-pytorch-developer-day-2021.md index 0e369ab6354d..6d121fb00cd0 100644 --- a/_posts/2021-08-23-pytorch-developer-day-2021.md +++ b/_posts/2021-08-23-pytorch-developer-day-2021.md @@ -22,12 +22,9 @@ Stay up to date by following us on our social channels: [Twitter](https://twitte On the second day, we’ll be hosting an online poster exhibition on Gather.Town. There will be opportunities to meet the authors and learn more about their PyTorch projects as well as network with the community. This poster and networking event is limited to people composed of PyTorch maintainers and contributors, long-time stakeholders and experts in areas relevant to PyTorch’s future. Conversations from the networking event will strongly shape the future of PyTorch. As such, invitations are required to attend the networking event. -Apply for an invitation to the networking event by clicking [here](https://pytorchdeveloperday.fbreg.com/). ## Call for Content Now Open Submit your poster abstracts today! Please send us the title and brief summary of your project, tools and libraries that could benefit PyTorch researchers in academia and industry, application developers, and ML engineers for consideration. The focus must be on academic papers, machine learning research, or open-source projects related to PyTorch development, Responsible AI or Mobile. Please no sales pitches. **Deadline for submission is September 24, 2021**. -You can submit your poster abstract during your application & registration process [here](https://pytorchdeveloperday.fbreg.com/apply). - -Visit the [event website](https://pytorchdeveloperday.fbreg.com/) for more information and we look forward to having you at PyTorch Developer Day. For any questions about the event, contact [pytorch@fbreg.com](mailto:pytorch@fbreg.com). +Visit the event website for more information and we look forward to having you at PyTorch Developer Day. \ No newline at end of file diff --git a/_posts/2021-10-21-pytorch-1.10-new-library-releases.md b/_posts/2021-10-21-pytorch-1.10-new-library-releases.md index 5d6413570500..8356cb1bc9cf 100644 --- a/_posts/2021-10-21-pytorch-1.10-new-library-releases.md +++ b/_posts/2021-10-21-pytorch-1.10-new-library-releases.md @@ -113,7 +113,7 @@ TorchAudio now adds support for differentiable Minimum Variance Distortionless R >>> # Get the enhanced waveform via iSTFT >>> waveform_enhanced = istft(specgram_enhanced, length=waveform.shape[-1]) ``` -Please refer to the [documentation](https://pytorch.org/audio/0.10.0/transforms.html#mvdr) for more details and try out this feature using the [MVDR tutorial](https://github.com/pytorch/audio/blob/main/examples/beamforming/MVDR_tutorial.ipynb). +Please refer to the [documentation](https://pytorch.org/audio/0.10.0/transforms.html#mvdr) for more details and try out this feature using the MVDR tutorial. ### (Beta) RNN Transducer Loss The RNN transducer (RNNT) loss is part of the RNN transducer pipeline, which is a popular architecture for speech recognition tasks. Recently it has gotten attention for being used in a streaming setting, and has also achieved state-of-the-art WER for the LibriSpeech benchmark. diff --git a/_posts/2021-12-22-introducing-torchvision-new-multi-weight-support-api.md b/_posts/2021-12-22-introducing-torchvision-new-multi-weight-support-api.md index 99280f4d45db..6086188e92e0 100644 --- a/_posts/2021-12-22-introducing-torchvision-new-multi-weight-support-api.md +++ b/_posts/2021-12-22-introducing-torchvision-new-multi-weight-support-api.md @@ -2,7 +2,7 @@ layout: blog_detail title: "Introducing TorchVision’s New Multi-Weight Support API" author: Vasilis Vryniotis -featured-img: "assets/images/torchvision_featured.png" +featured-img: "assets/images/torchvision_featured.jpg" --- TorchVision has a new backwards compatible API for building models with multi-weight support. The new API allows loading different pre-trained weights on the same model variant, keeps track of vital meta-data such as the classification labels and includes the preprocessing transforms necessary for using the models. In this blog post, we plan to review the prototype API, show-case its features and highlight key differences with the existing one. diff --git a/_posts/2021-4-16-ml-models-torchvision-v0.9.md b/_posts/2021-4-16-ml-models-torchvision-v0.9.md index 5a5b3bdf156e..ff4e22f2c7c6 100644 --- a/_posts/2021-4-16-ml-models-torchvision-v0.9.md +++ b/_posts/2021-4-16-ml-models-torchvision-v0.9.md @@ -7,7 +7,7 @@ author: Team PyTorch TorchVision v0.9 has been [released](https://github.com/pytorch/vision/releases) and it is packed with numerous new Machine Learning models and features, speed improvements and bug fixes. In this blog post, we provide a quick overview of the newly introduced ML models and discuss their key features and characteristics. ### Classification -* **MobileNetV3 Large & Small:** These two classification models are optimized for Mobile use-cases and are used as backbones on other Computer Vision tasks. The implementation of the new [MobileNetV3 architecture](https://github.com/pytorch/vision/blob/master/torchvision/models/mobilenetv3.py) supports the Large & Small variants and the depth multiplier parameter as described in the [original paper](https://arxiv.org/pdf/1905.02244.pdf). We offer pre-trained weights on ImageNet for both Large and Small networks with depth multiplier 1.0 and resolution 224x224. Our previous [training recipes](https://github.com/pytorch/vision/tree/master/references/classification#mobilenetv3-large--small) have been updated and can be used to easily train the models from scratch (shoutout to Ross Wightman for inspiring some of our [training configuration](https://rwightman.github.io/pytorch-image-models/training_hparam_examples/#mobilenetv3-large-100-75766-top-1-92542-top-5)). The Large variant offers a [competitive accuracy](https://github.com/pytorch/vision/blob/master/docs/source/models.rst#classification) comparing to ResNet50 while being over 6x faster on CPU, meaning that it is a good candidate for applications where speed is important. For applications where speed is critical, one can sacrifice further accuracy for speed and use the Small variant which is 15x faster than ResNet50. +* **MobileNetV3 Large & Small:** These two classification models are optimized for Mobile use-cases and are used as backbones on other Computer Vision tasks. The implementation of the new [MobileNetV3 architecture](https://github.com/pytorch/vision/blob/master/torchvision/models/mobilenetv3.py) supports the Large & Small variants and the depth multiplier parameter as described in the [original paper](https://arxiv.org/pdf/1905.02244.pdf). We offer pre-trained weights on ImageNet for both Large and Small networks with depth multiplier 1.0 and resolution 224x224. Our previous [training recipes](https://github.com/pytorch/vision/tree/master/references/classification#mobilenetv3-large--small) have been updated and can be used to easily train the models from scratch (shoutout to Ross Wightman for inspiring some of our training configuration). The Large variant offers a [competitive accuracy](https://github.com/pytorch/vision/blob/master/docs/source/models.rst#classification) comparing to ResNet50 while being over 6x faster on CPU, meaning that it is a good candidate for applications where speed is important. For applications where speed is critical, one can sacrifice further accuracy for speed and use the Small variant which is 15x faster than ResNet50. * **Quantized MobileNetV3 Large:** The quantized version of MobilNetV3 Large reduces the number of parameters by 45% and it is roughly 2.5x faster than the non-quantized version while remaining competitive in [terms of accuracy](https://github.com/pytorch/vision/blob/master/docs/source/models.rst#quantized-models). It was fitted on ImageNet using Quantization Aware Training by iterating on the non-quantized version and it can be trained from scratch using the existing [reference scripts](https://github.com/pytorch/vision/tree/master/references/classification#quantized). diff --git a/_posts/2021-5-26-torchvision-mobilenet-v3-implementation.md b/_posts/2021-5-26-torchvision-mobilenet-v3-implementation.md index 6496c42a1806..2dfe3bba49c1 100644 --- a/_posts/2021-5-26-torchvision-mobilenet-v3-implementation.md +++ b/_posts/2021-5-26-torchvision-mobilenet-v3-implementation.md @@ -81,7 +81,7 @@ Another important detail is that though PyTorch’s and TensorFlow’s RMSProp i **Increasing our accuracy by tuning hyperparameters & improving our training recipe** -After configuring the optimizer to achieve fast and stable training, we turned into optimizing the accuracy of the model. There are a few techniques that helped us achieve this. First of all, to avoid overfitting we augmented out data using the AutoAugment algorithm, followed by RandomErasing. Additionally we tuned parameters such as the weight decay using cross validation. We also found beneficial to perform [weight averaging](https://github.com/pytorch/vision/blob/674e8140042c2a3cbb1eb9ebad1fa49501599130/references/classification/utils.py#L259) across different epoch checkpoints after the end of the training. Finally, though not used in our published training recipe, we found that using Label Smoothing, Stochastic Depth and LR noise injection improve the overall accuracy by over [1.5 points](https://rwightman.github.io/pytorch-image-models/training_hparam_examples/#mobilenetv3-large-100-75766-top-1-92542-top-5). +After configuring the optimizer to achieve fast and stable training, we turned into optimizing the accuracy of the model. There are a few techniques that helped us achieve this. First of all, to avoid overfitting we augmented out data using the AutoAugment algorithm, followed by RandomErasing. Additionally we tuned parameters such as the weight decay using cross validation. We also found beneficial to perform [weight averaging](https://github.com/pytorch/vision/blob/674e8140042c2a3cbb1eb9ebad1fa49501599130/references/classification/utils.py#L259) across different epoch checkpoints after the end of the training. Finally, though not used in our published training recipe, we found that using Label Smoothing, Stochastic Depth and LR noise injection improve the overall accuracy by over 1.5 points. The graph and table depict a simplified summary of the most important iterations for improving the accuracy of the MobileNetV3 Large variant. Note that the actual number of iterations done while training the model was significantly larger and that the progress in accuracy was not always monotonically increasing. Also note that the Y-axis of the graph starts from 70% instead from 0% to make the difference between iterations more visible: diff --git a/_posts/2021-6-15-pytorch-1.9-new-library-releases.md b/_posts/2021-6-15-pytorch-1.9-new-library-releases.md index eda8ac7b4318..6ed505185db8 100644 --- a/_posts/2021-6-15-pytorch-1.9-new-library-releases.md +++ b/_posts/2021-6-15-pytorch-1.9-new-library-releases.md @@ -150,7 +150,7 @@ We have: For more details, see [the documentation](https://pytorch.org/audio/0.9.0/transforms.html#resample). ### (Prototype) RNN Transducer Loss -The RNN transducer loss is used in training RNN transducer models, which is a popular architecture for speech recognition tasks. The prototype loss in torchaudio currently supports autograd, torchscript, float16 and float32, and can also be run on both CPU and CUDA. For more details, please refer to [the documentation](https://pytorch.org/audio/master/rnnt_loss.html). +The RNN transducer loss is used in training RNN transducer models, which is a popular architecture for speech recognition tasks. The prototype loss in torchaudio currently supports autograd, torchscript, float16 and float32, and can also be run on both CPU and CUDA. For more details, please refer to [the documentation](https://pytorch.org/audio/stable/index.html). # TorchText 0.10.0 diff --git a/_posts/2021-6-15-pytorch-1.9-released.md b/_posts/2021-6-15-pytorch-1.9-released.md index 71c3f80e5bc4..9394a036efca 100644 --- a/_posts/2021-6-15-pytorch-1.9-released.md +++ b/_posts/2021-6-15-pytorch-1.9-released.md @@ -91,7 +91,7 @@ We are releasing a new video app based on [PyTorch Video](https://pytorchvideo.o ### (Beta) TorchElastic is now part of core -[TorchElastic](https://github.com/pytorch/pytorch/issues/50621), which was open sourced over a year ago in the [pytorch/elastic](https://github.com/pytorch/elastic) github repository, is a runner and coordinator for PyTorch worker processes. Since then, it has been adopted by various distributed torch use-cases: 1) [deepspeech.pytorch](https://medium.com/pytorch/training-deepspeech-using-torchelastic-ad013539682) 2) [pytorch-lightning](https://pytorch-lightning.readthedocs.io/en/stable/advanced/multi_gpu.html#torchelastic) 3) [Kubernetes CRD](https://github.com/pytorch/elastic/blob/master/kubernetes/README.md). Now, it is part of PyTorch core. +[TorchElastic](https://github.com/pytorch/pytorch/issues/50621), which was open sourced over a year ago in the [pytorch/elastic](https://github.com/pytorch/elastic) github repository, is a runner and coordinator for PyTorch worker processes. Since then, it has been adopted by various distributed torch use-cases: 1) [deepspeech.pytorch](https://medium.com/pytorch/training-deepspeech-using-torchelastic-ad013539682) 2) pytorch-lightning 3) [Kubernetes CRD](https://github.com/pytorch/elastic/blob/master/kubernetes/README.md). Now, it is part of PyTorch core. As its name suggests, the core function of TorcheElastic is to gracefully handle scaling events. A notable corollary of elasticity is that peer discovery and rank assignment are built into TorchElastic enabling users to run distributed training on preemptible instances without requiring a gang scheduler. As a side note, [etcd](https://etcd.io/) used to be a hard dependency of TorchElastic. With the upstream, this is no longer the case since we have added a “standalone” rendezvous based on c10d::Store. For more details, refer to the [documentation](https://pytorch.org/docs/1.9.0/distributed.elastic.html). diff --git a/_posts/2021-6-16-torchvision-ssd-implementation.md b/_posts/2021-6-16-torchvision-ssd-implementation.md index 40ce299b1ef0..3f55188b4847 100644 --- a/_posts/2021-6-16-torchvision-ssd-implementation.md +++ b/_posts/2021-6-16-torchvision-ssd-implementation.md @@ -2,7 +2,6 @@ layout: blog_detail title: 'Everything You Need To Know About Torchvision’s SSD Implementation' author: Vasilis Vryniotis -featured-img: 'assets/images/prediction-examples.png' --- In TorchVision v0.10, we’ve released two new Object Detection models based on the SSD architecture. Our plan is to cover the key implementation details of the algorithms along with information on how they were trained in a two-part article. diff --git a/_posts/2021-8-18-pipetransformer-automated-elastic-pipelining.md b/_posts/2021-8-18-pipetransformer-automated-elastic-pipelining.md index a682c3a13382..02c73d77541b 100644 --- a/_posts/2021-8-18-pipetransformer-automated-elastic-pipelining.md +++ b/_posts/2021-8-18-pipetransformer-automated-elastic-pipelining.md @@ -70,13 +70,13 @@ Finally, we have also developed open-source flexible APIs for PipeTransformer, w Suppose we aim to train a massive model in a distributed training system where the hybrid of pipelined model parallelism and data parallelism is used to target scenarios where either the memory of a single GPU device cannot hold the model, or if loaded, the batch size is small enough to avoid running out of memory. More specifically, we define our settings as follows: -Training task and model definition. We train Transformer models (e.g., Vision Transformer, BERT on large-scale image or text datasets. The Transformer model
-Figure 7. Pipeline Bubble: , and
denote forward, backward, and the optimizer update of micro-batch
on device
, respectively. The total bubble size in each iteration is
times per micro-batch forward and backward cost.
+Figure 7. Pipeline Bubble: F_{d,b}, and U_d" denote forward, backward, and the optimizer update of micro-batch b on device d, respectively. The total bubble size in each iteration is K-1 times per micro-batch forward and backward cost.
@@ -258,7 +258,7 @@ This section first summarizes experiment setups and then evaluates PipeTransform
-Figure 10. Tuning in Freezing Algorithm
+Figure 10. Tuning \alpha in Freezing Algorithm
Gify: Jump to Source using Visual Studio Code Plug In UI
-For how to optimize batch size performance, check out the step-by-step tutorial [here](https://opendatascience.com/optimizing-pytorch-performance-batch-size-with-pytorch-profiler/). PyTorch Profiler is also integrated with [PyTorch Lightning](https://pytorch-lightning.readthedocs.io/en/stable/advanced/profiler.html#pytorch-profiling) and you can simply launch your lightning training jobs with --```trainer.profiler=pytorch``` flag to generate the traces. Check out an example [here](https://github.com/PyTorchLightning/pytorch-lightning/blob/master/pl_examples/basic_examples/profiler_example.py). +For how to optimize batch size performance, check out the step-by-step tutorial [here](https://opendatascience.com/optimizing-pytorch-performance-batch-size-with-pytorch-profiler/). PyTorch Profiler is also integrated with PyTorch Lightning and you can simply launch your lightning training jobs with --```trainer.profiler=pytorch``` flag to generate the traces. ## What’s Next for the PyTorch Profiler? You just saw how PyTorch Profiler can help optimize a model. You can now try the Profiler by ```pip install torch-tb-profiler``` to optimize your PyTorch model. diff --git a/_posts/2022-10-13-scaling-pytorch-models-on-cloud-tpus-with-fsdp.md b/_posts/2022-10-13-scaling-pytorch-models-on-cloud-tpus-with-fsdp.md index fe852ce4cf70..4f07564cbc86 100644 --- a/_posts/2022-10-13-scaling-pytorch-models-on-cloud-tpus-with-fsdp.md +++ b/_posts/2022-10-13-scaling-pytorch-models-on-cloud-tpus-with-fsdp.md @@ -36,7 +36,7 @@ Wrapping an `nn.Module` instance with `XlaFullyShardedDataParallel` enables the **Model checkpoint saving and loading** for models and optimizers can be done like before by saving and loading their `.state_dict()`. Meanwhile, each training process should save its own checkpoint file of the sharded model parameters and optimizer states, and load the checkpoint file for the corresponding rank when resuming (regardless of ZeRO-2 or ZeRO-3, i.e. nested wrapping or not). A command line tool and a Python interface are provided to consolidate the sharded model checkpoint files together into a full/unshareded model checkpoint file. -[**Gradient checkpointing**](https://spell.ml/blog/gradient-checkpointing-pytorch-YGypLBAAACEAefHs) (also referred to as "activation checkpointing" or "rematerialization") is another common technique for model scaling and can be used in conjunction with FSDP. We provide `checkpoint_module`, a wrapper function over a given `nn.Module` instance for gradient checkpointing (based on `torch_xla.utils.checkpoint.checkpoint`). +**Gradient checkpointing** (also referred to as "activation checkpointing" or "rematerialization") is another common technique for model scaling and can be used in conjunction with FSDP. We provide `checkpoint_module`, a wrapper function over a given `nn.Module` instance for gradient checkpointing (based on `torch_xla.utils.checkpoint.checkpoint`). The MNIST and ImageNet examples below provide illustrative usages of (plain or nested) FSDP, saving and consolidation of model checkpoints, as well as gradient checkpointing. diff --git a/_posts/2022-10-28-new-library-updates-in-pytorch-1.13.md b/_posts/2022-10-28-new-library-updates-in-pytorch-1.13.md index 3c28b3aa7449..4882d1538079 100644 --- a/_posts/2022-10-28-new-library-updates-in-pytorch-1.13.md +++ b/_posts/2022-10-28-new-library-updates-in-pytorch-1.13.md @@ -110,7 +110,7 @@ In this release, we further consolidated the API for `DataLoader2` and a [detail We extended our support to load data from additional cloud storage providers via DataPipes, now covering AWS, Google Cloud Storage, and Azure. A [tutorial is also available](https://pytorch.org/data/0.5/tutorial.html#working-with-cloud-storage-providers). We are open to feedback and feature requests. -We also performed a simple benchmark, comparing the performance of data loading from AWS S3 and attached volume on an AWS EC2 instance. The results are [visible here](https://github.com/pytorch/data/blob/gh/NivekT/100/head/benchmarks/cloud/aws_s3_results.md). +We also performed a simple benchmark, comparing the performance of data loading from AWS S3 and attached volume on an AWS EC2 instance. ### torch::deploy (Beta) @@ -154,7 +154,7 @@ torch::deploy now has basic support for aarch64 Linux systems. TorchEval is a library built for users who want highly performant implementations of common metrics to evaluate machine learning models. It also provides an easy to use interface for building custom metrics with the same toolkit. Building your metrics with TorchEval makes running distributed training loops with [torch.distributed](https://pytorch.org/docs/stable/distributed.html) a breeze. -Learn more with our [docs](https://pytorch.org/torcheval), see our [examples](https://pytorch.org/torcheval/metric_example.html), or check out our [GitHub repo](http://github.com/pytorch/torcheval). +Learn more with our [docs](https://pytorch.org/torcheval), see our [examples](https://pytorch.org/torcheval/stable/metric_example.html), or check out our [GitHub repo](http://github.com/pytorch/torcheval). ### TorchMultimodal Release (Beta) diff --git a/_posts/2022-11-22-effective-multi-objective-nueral-architecture.md b/_posts/2022-11-22-effective-multi-objective-nueral-architecture.md index 96d3ba38da25..ab3643a4873b 100644 --- a/_posts/2022-11-22-effective-multi-objective-nueral-architecture.md +++ b/_posts/2022-11-22-effective-multi-objective-nueral-architecture.md @@ -108,7 +108,7 @@ Ax has a number of other advanced capabilities that we did not discuss in our tu ### Early Stopping -When evaluating a new candidate configuration, partial learning curves are typically available while the NN training job is running. We can use the information contained in the partial curves to identify under-performing trials to stop early in order to free up computational resources for more promising candidates. While not demonstrated in the above tutorial, Ax supports early stopping out-of-the-box - see our [early stopping tutorial](https://ax.dev/versions/latest/tutorials/early_stopping/early_stopping.html) for more details. +When evaluating a new candidate configuration, partial learning curves are typically available while the NN training job is running. We can use the information contained in the partial curves to identify under-performing trials to stop early in order to free up computational resources for more promising candidates. While not demonstrated in the above tutorial, Ax supports early stopping out-of-the-box. ### High-dimensional search spaces diff --git a/_posts/2022-11-28-optimizing-production-pytorch-performance-with-graph-transformations.md b/_posts/2022-11-28-optimizing-production-pytorch-performance-with-graph-transformations.md index ad7593b253c4..93fda1037bd4 100644 --- a/_posts/2022-11-28-optimizing-production-pytorch-performance-with-graph-transformations.md +++ b/_posts/2022-11-28-optimizing-production-pytorch-performance-with-graph-transformations.md @@ -2,7 +2,6 @@ layout: blog_detail title: "Optimizing Production PyTorch Models’ Performance with Graph Transformations" author: Jade Nie, CK Luk, Xiaodong Wang, Jackie (Jiaqi) Xu -featured-img: "assets/images/blog1-3b.png" --- ## 1. Introduction diff --git a/_posts/2022-2-8-quantization-in-practice.md b/_posts/2022-2-8-quantization-in-practice.md index b95b6f4f7608..43c9aeb1f73f 100644 --- a/_posts/2022-2-8-quantization-in-practice.md +++ b/_posts/2022-2-8-quantization-in-practice.md @@ -45,9 +45,9 @@ where [