PyTorch
@PyTorch
Tensors and neural networks in Python with strong hardware acceleration. PyTorch is an open source project at the Linux Foundation. #PyTorchFoundation
ID:776585502606721024
http://pytorch.org 16-09-2016 00:56:26
1,6K Tweets
272,0K Followers
70 Following
Interested in becoming a PyTorch contributor? 🇺🇦Viacheslav Kovalevskyi🇺🇦 walks you through the process, starting with why, sharing the pre-requirements, outlining ways to find issues to work on, and more. Learn about this process: bit.ly/3EzpWxd
In the final video of this DDP series, we put it all together and train a GPT-like model across multiple GPUs and nodes.
Suraj Subramanian walks you through best practices and structuring your project for fault-tolerant distributed training: bit.ly/3ytkGHC
Maxed out all GPUs on your machine but need more? Add more machines to the mix!
In DDP video 5, we extend our multi-GPU training script to running on multiple nodes, with almost no code changes!
Suraj Subramanian shows you how to run it on a cluster via SLURM: bit.ly/3MeQwh0
Some ML jobs fail due to code bugs, others due to (cloud) matrix glitches. Only need to restart the job? torchrun is your friend.
Watch Suraj Subramanian explain how training jobs can be fault-tolerant w/ torchrun & watch your script auto bounce back to life. bit.ly/3RJ7chY
Watch Soumith Chintala and Yann LeCun discuss accomplishments made in PyTorch and #AI . What are your favorite recent breakthroughs in machine learning #ML ? Share below, and watch Soumith and Yann’s full conversation here: bit.ly/3D62ZkM
Using multiple GPUs to train your model can greatly reduce training time!
In video 3 of the DDP series, we migrate a single-GPU training job to run on 4 GPUs, while Suraj Subramanian explains distributed training concepts in PyTorch code.
Watch the tutorial: bit.ly/3EiApgw
Learn about distributed training in PyTorch. In this tutorial, Suraj Subramanian walks you through training your models on a single GPU -> multiple GPUs -> training LMs on multiple machines in less than an hour.
Watch the tutorial: bit.ly/3Sxq2K7
PyTorch Contributors and Maintainers are hosting a Bay Area Meetup on Oct. 12th at 6pm PT. Join to connect with our community and learn about becoming a contributor. Hosted by 🇺🇦Viacheslav Kovalevskyi🇺🇦. RSVP: bit.ly/3y8NqFK
Abhi Venigalla (aveni.eth) and I turned Andrej Karpathy’s minGPT into a GPT-3 quality model with 30 billion parameters—projected to cost only $450k to train. The code to do so is public: it's easily readable and can be launched on however many GPUs you want.
Here’s how: