PyTorch(@PyTorch) 's Twitter Profileg
PyTorch done

@PyTorch

Tensors and neural networks in Python with strong hardware acceleration. PyTorch is an open source project at the Linux Foundation. #PyTorchFoundation

ID:776585502606721024

linkhttp://pytorch.org calendar_today16-09-2016 00:56:26

1,6K Tweets

272,0K Followers

70 Following

PyTorch(@PyTorch) 's Twitter Profile Photo

Cloud TPUs are powerful hardware for building large models & have enabled many research successes. We enable scaling up PyTorch models to 10B+ parameters on Cloud TPU with a new Fully Sharded Data Parallel (FSDP) interface in PyTorch/XLA.

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Interested in joining us at the PyTorch Conference on Dec. 2nd, 2022 in New Orleans? Apply now for an in-person spot at our flagship event: pytorchconference22.splashthat.com

account_circle
Francisco Massa(@fvsmassa) 's Twitter Profile Photo

Do you need fast and memory-efficient transformers which are easy to install?
I'm happy to share that xFormers now ships precompiled conda packages for PyTorch 1.12.1 and CUDA 11.3/11.6 (Linux-only for now). github.com/facebookresear…

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Interested in becoming a PyTorch contributor? 🇺🇦Viacheslav Kovalevskyi🇺🇦 walks you through the process, starting with why, sharing the pre-requirements, outlining ways to find issues to work on, and more. Learn about this process: bit.ly/3EzpWxd

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

In the final video of this DDP series, we put it all together and train a GPT-like model across multiple GPUs and nodes.

Suraj Subramanian walks you through best practices and structuring your project for fault-tolerant distributed training: bit.ly/3ytkGHC

In the final video of this DDP series, we put it all together and train a GPT-like model across multiple GPUs and nodes. @subramen walks you through best practices and structuring your project for fault-tolerant distributed training: bit.ly/3ytkGHC
account_circle
Zachary DeVito(@Zachary_DeVito) 's Twitter Profile Photo

Want to understand where your PyTorch program is allocating GPU memory, or why it ran out?

We recently added an experimental API for visualizing allocated GPU memory in PyTorch using flamegraphs. Check it out in newer nightly builds. zdevito.github.io/2022/08/16/mem…

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Maxed out all GPUs on your machine but need more? Add more machines to the mix!

In DDP video 5, we extend our multi-GPU training script to running on multiple nodes, with almost no code changes!

Suraj Subramanian shows you how to run it on a cluster via SLURM: bit.ly/3MeQwh0

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Some ML jobs fail due to code bugs, others due to (cloud) matrix glitches. Only need to restart the job? torchrun is your friend.

Watch Suraj Subramanian explain how training jobs can be fault-tolerant w/ torchrun & watch your script auto bounce back to life. bit.ly/3RJ7chY

Some ML jobs fail due to code bugs, others due to (cloud) matrix glitches. Only need to restart the job? torchrun is your friend. Watch @subramen explain how training jobs can be fault-tolerant w/ torchrun & watch your script auto bounce back to life. bit.ly/3RJ7chY
account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Watch Soumith Chintala and Yann LeCun discuss accomplishments made in PyTorch and . What are your favorite recent breakthroughs in machine learning ? Share below, and watch Soumith and Yann’s full conversation here: bit.ly/3D62ZkM

account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Using multiple GPUs to train your model can greatly reduce training time!

In video 3 of the DDP series, we migrate a single-GPU training job to run on 4 GPUs, while Suraj Subramanian explains distributed training concepts in PyTorch code.

Watch the tutorial: bit.ly/3EiApgw

Using multiple GPUs to train your model can greatly reduce training time! In video 3 of the DDP series, we migrate a single-GPU training job to run on 4 GPUs, while @subramen explains distributed training concepts in PyTorch code. Watch the tutorial: bit.ly/3EiApgw
account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

If your training jobs are using DataParallel, consider using DDP!

Watch the video to learn what DDP is doing under all those function calls, and how it can help more than using DP for multi-GPU training

bit.ly/3SSqDG3

If your training jobs are using DataParallel, consider using DDP! Watch the video to learn what DDP is doing under all those function calls, and how it can help more than using DP for multi-GPU training bit.ly/3SSqDG3
account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

Learn about distributed training in PyTorch. In this tutorial, Suraj Subramanian walks you through training your models on a single GPU -> multiple GPUs -> training LMs on multiple machines in less than an hour.

Watch the tutorial: bit.ly/3Sxq2K7

Learn about distributed training in PyTorch. In this tutorial, @subramen walks you through training your models on a single GPU -> multiple GPUs -> training LMs on multiple machines in less than an hour. Watch the tutorial: bit.ly/3Sxq2K7
account_circle
PyTorch(@PyTorch) 's Twitter Profile Photo

PyTorch Contributors and Maintainers are hosting a Bay Area Meetup on Oct. 12th at 6pm PT. Join to connect with our community and learn about becoming a contributor. Hosted by 🇺🇦Viacheslav Kovalevskyi🇺🇦. RSVP: bit.ly/3y8NqFK

PyTorch Contributors and Maintainers are hosting a Bay Area Meetup on Oct. 12th at 6pm PT. Join to connect with our community and learn about becoming a contributor. Hosted by @b0noi. RSVP: bit.ly/3y8NqFK
account_circle
Nouamane Tazi(@Nouamanetazi) 's Twitter Profile Photo

Stable Diffusion in the diffusers library became x3 times faster thanks to a set of optimizations tips, some of which require minimal code changes, making it the fastest implementation of Stable Diffusion (afaik)! What are these optimizations?
A thread 🧵 (1/n)

Stable Diffusion in the diffusers library became x3 times faster thanks to a set of optimizations tips, some of which require minimal code changes, making it the fastest implementation of Stable Diffusion (afaik)! What are these optimizations? A thread 🧵 (1/n)
account_circle
Linden Li(@lindensli) 's Twitter Profile Photo

Abhi Venigalla (aveni.eth) and I turned Andrej Karpathy’s minGPT into a GPT-3 quality model with 30 billion parameters—projected to cost only $450k to train. The code to do so is public: it's easily readable and can be launched on however many GPUs you want.

Here’s how:

@abhi_venigalla and I turned @karpathy’s minGPT into a GPT-3 quality model with 30 billion parameters—projected to cost only $450k to train. The code to do so is public: it's easily readable and can be launched on however many GPUs you want. Here’s how:
account_circle
MosaicML(@MosaicML) 's Twitter Profile Photo

We have exciting news! In our latest and greatest LLM blog, we show how MosaicML Cloud can help you train LLMs from 1B - 70B parameters, and for the first time, publish transparent times + costs for doing so. It's a lot cheaper than you think! (1/9)

mosaicml.com/blog/gpt-3-qua…

account_circle