Liczba odcinków: 82

The PyTorch Developer Podcast is a place for the PyTorch dev team to do bite sized (10-20 min) topics about all sorts of internal development topics in PyTorch.

PyTorch Developer Podcast Edward Yang, Team PyTorch

    • Technologie

The PyTorch Developer Podcast is a place for the PyTorch dev team to do bite sized (10-20 min) topics about all sorts of internal development topics in PyTorch.

    TORCH_TRACE and tlparse

    TORCH_TRACE and tlparse

    TORCH_TRACE and tlparse are a structured log and log parser for PyTorch 2. It gives useful information about what code was compiled and what the intermediate build products look like.

    • 15 min
    Higher order operators

    Higher order operators

    Higher order operators are a special form of operators in torch.ops which have relaxed input argument requirements: in particular, they can accept any form of argument, including Python callables. Their name is based off of their most common use case, which is to represent higher order functions like control flow operators. However, they are also used to implement other variants of basic operators and can also be used to smuggle in Python data that is quite unusual. They are implemented using a Python dispatcher.

    • 17 min
    Inductor - Post-grad FX passes

    Inductor - Post-grad FX passes

    The post-grad FX passes in Inductor run after AOTAutograd has functionalized and normalized the input program into separate forward/backward graphs. As such, they generally can assume that the graph in question is functionalized, except for some mutations to inputs at the end of the graph. At the end of post-grad passes, there are special passes that reintroduce mutation into the graph before going into the rest of Inductor lowering which is generally aware of passes. The post-grad FX passes are varied but are typically domain specific passes making local changes to specific parts of the graph.

    • 24 min
    CUDA graph trees

    CUDA graph trees

    CUDA graph trees are the internal implementation of CUDA graphs used in PT2 when you say mode="reduce-overhead". Their primary innovation is that they allow the reuse of memory across multiple CUDA graphs, as long as they form a tree structure of potential paths you can go down with the CUDA graph. This greatly reduced the memory usage of CUDA graphs in PT2. There are some operational implications to using CUDA graphs which are described in the podcast.

    • 20 min
    Min-cut partitioner

    Min-cut partitioner

    The min-cut partitioner makes decisions about what to save for backwards when splitting the forward and backwards graph from the joint graph traced by AOTAutograd. Crucially, it doesn't actually do a "split"; instead, it is deciding how much of the joint graph should be used for backwards. I also talk about the backward retracing problem.

    • 15 min
    AOTInductor

    AOTInductor

    AOTInductor is a feature in PyTorch that lets you export an inference model into a self-contained dynamic library, which can subsequently be loaded and used to run optimized inference. It is aimed primarily at CUDA and CPU inference applications, for situations when your model export once to be exported once while your runtime may still get continuous updates. One of the big underlying organizing principles is a limited ABI which does not include libtorch, which allows these libraries to stay stable over updates to the runtime. There are many export-like use cases you might be interested in using AOTInductor for, and some of the pieces should be useful, but AOTInductor does not necessarily solve them.

    • 17 min

Najpopularniejsze podcasty w kategorii Technologie

Podcast o technologii
Kanał o technologii
Lex Fridman Podcast
Lex Fridman
Techstorie - rozmowy o technologiach
TOK FM - Sylwia Czubkowska, Joanna Sosnowska
Mateusz Chrobok Bezpieczeństwo, Startupy i Sztuczna Inteligencja
Mateusz Chrobok
Nadgryzieni - Rozmowy (nie tylko) o Tech
Wojtek Pietrusiewicz
Opanuj.AI Podcast
Opanuj.AI Podcast

Możesz również polubić

Latent Space: The AI Engineer Podcast — Practitioners talking LLMs, CodeGen, Agents, Multimodality, AI UX, GPU Infra and al
Alessio + swyx
Machine Learning Street Talk (MLST)
Machine Learning Street Talk (MLST)
Practical AI: Machine Learning, Data Science
Changelog Media
The AI Podcast
NVIDIA
Last Week in AI
Skynet Today
Talk Python To Me
Michael Kennedy (@mkennedy)