82 bölüm

The PyTorch Developer Podcast is a place for the PyTorch dev team to do bite sized (10-20 min) topics about all sorts of internal development topics in PyTorch.

PyTorch Developer Podcast Edward Yang, Team PyTorch

    • Teknoloji
    • 5,0 • 1 Oy

The PyTorch Developer Podcast is a place for the PyTorch dev team to do bite sized (10-20 min) topics about all sorts of internal development topics in PyTorch.

    TORCH_TRACE and tlparse

    TORCH_TRACE and tlparse

    TORCH_TRACE and tlparse are a structured log and log parser for PyTorch 2. It gives useful information about what code was compiled and what the intermediate build products look like.

    • 15 dk.
    Higher order operators

    Higher order operators

    Higher order operators are a special form of operators in torch.ops which have relaxed input argument requirements: in particular, they can accept any form of argument, including Python callables. Their name is based off of their most common use case, which is to represent higher order functions like control flow operators. However, they are also used to implement other variants of basic operators and can also be used to smuggle in Python data that is quite unusual. They are implemented using a Python dispatcher.

    • 17 dk.
    Inductor - Post-grad FX passes

    Inductor - Post-grad FX passes

    The post-grad FX passes in Inductor run after AOTAutograd has functionalized and normalized the input program into separate forward/backward graphs. As such, they generally can assume that the graph in question is functionalized, except for some mutations to inputs at the end of the graph. At the end of post-grad passes, there are special passes that reintroduce mutation into the graph before going into the rest of Inductor lowering which is generally aware of passes. The post-grad FX passes are varied but are typically domain specific passes making local changes to specific parts of the graph.

    • 24 dk.
    CUDA graph trees

    CUDA graph trees

    CUDA graph trees are the internal implementation of CUDA graphs used in PT2 when you say mode="reduce-overhead". Their primary innovation is that they allow the reuse of memory across multiple CUDA graphs, as long as they form a tree structure of potential paths you can go down with the CUDA graph. This greatly reduced the memory usage of CUDA graphs in PT2. There are some operational implications to using CUDA graphs which are described in the podcast.

    • 20 dk.
    Min-cut partitioner

    Min-cut partitioner

    The min-cut partitioner makes decisions about what to save for backwards when splitting the forward and backwards graph from the joint graph traced by AOTAutograd. Crucially, it doesn't actually do a "split"; instead, it is deciding how much of the joint graph should be used for backwards. I also talk about the backward retracing problem.

    • 15 dk.
    AOTInductor

    AOTInductor

    AOTInductor is a feature in PyTorch that lets you export an inference model into a self-contained dynamic library, which can subsequently be loaded and used to run optimized inference. It is aimed primarily at CUDA and CPU inference applications, for situations when your model export once to be exported once while your runtime may still get continuous updates. One of the big underlying organizing principles is a limited ABI which does not include libtorch, which allows these libraries to stay stable over updates to the runtime. There are many export-like use cases you might be interested in using AOTInductor for, and some of the pieces should be useful, but AOTInductor does not necessarily solve them.

    • 17 dk.

Müşteri Yorumları

5,0/5
1 Oy

1 Oy

En Popüler Teknoloji Podcast’leri

Barış Özcan ile 111 Hz
Podbee Media
Acquired
Ben Gilbert and David Rosenthal
Haftalık Gündem Değerlendirmesi
TeknoSeyir
TED Radio Hour
NPR
Lex Fridman Podcast
Lex Fridman
Farklı Düşün
Seyfeddin Başsaraç, Mert Bulan

Beğenebilirsiniz

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al
Alessio + swyx
The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)
Sam Charrington
Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)
Practical AI: Machine Learning, Data Science
Changelog Media
The Changelog: Software Development, Open Source
Changelog Media
Python Bytes
Michael Kennedy and Brian Okken