TVM Monthly - August 2025

As discussed by the TVM PMC, our goal is to provide a monthly summary of the project so users and developers can get a better understanding of the goings on of the TVM community.

Feedback and suggestions are welcomed so that we can further improve these updates.

RFCs

None


We continue to improve FFI, Relax .

BugFix

  • #18244 - [Fix] Update FlashInfer JIT header lookup
  • #18237 - [FFI]Fix type_traits on DataType after SmallStr update
  • #18219 - [Fix] Resolve deadlock in PopenPoolExecutor and LocalBuilder
  • #18210 - [Bug] Fix core dump in InferLayoutRMSNorm and fix typo
  • #18208 - [FFI][Fix] Update datatype registry calls to the new paths
  • #18190 - [Fix] Codegen fix for relax cutlass

CI

  • #18227 - Exit the build for AbortException

Docs

  • #18256 - [FFI]Wheel Packaging

Frontend

  • #18207 - [Fix][ONNX] No precision widening for numpy binary operations
  • #18209 - [ONNX][Fix] Update Resize to accept ShapeExpr

FFI

  • #18254 - [FFI][ABI] ABI Updates to for future metadata and complex ordering
  • #18249 - [FFI][CMAKE] Revert cmake libbacktrace URL and update submodule
  • #18246 - [FFI][CMAKE] Add missing download path for libbacktrace
  • #18234 - [FFI] Misc fixup for windows
  • #18233 - [FFI] Robustify the pyproject setup
  • #18226 - [FFI][REFACTOR] Establish tvm_ffi python module
  • #18221 - [FFI] Fix JSON parser/writer for the fast-math flag
  • #18218 - [FFI][REFACTOR] Cleanup API locations
  • #18217 - [FFI] AudoDLPack compatible with torch stream context
  • #18216 - [FFI][REFACTOR] Establish Stream Context in ffi
  • #18214 - [FFI][REFACTOR] Establish ffi.Module in python
  • #18213 - [FFI] Formalize ffi.Module
  • #18212 - [FFI] Make JSON Parser/Write fastmath safe
  • #18205 - [FFI][REFATOR] Cleanup entry function to redirect
  • #18200 - [FFI][REFACTOR] Update Map ABI to enable flexible smallMap switch
  • #18198 - [FFI][REFACTOR] Move Downcast out of ffi for now
  • #18192 - [FFI] Phase out ObjectPath in favor of AccessPath
  • #18191 - [FFI][REFACTOR] Refactor AccessPath to enable full tree repr
  • #18189 - [FFI][REFACTOR] Phase out getattr based attribute handling
  • #18188 - [FFI][REFACTOR] Migrate the Save/Load JSON to the new reflection
  • #18187 - [FFI][EXTRA] Serialization To/From JSONGraph
  • #18186 - [FFI] Lightweight json parser/writer
  • #18185 - [FFI] Introduce small string/bytes
  • #18184 - [FFI][REFACTOR] Hide StringObj/BytesObj into details
  • #18183 - [FFI][REFACTOR] Cleanup to align to latest ffi
  • #18181 - [REFACTOR] Upgrade NestedMsg to use new ffi::Any mechanism

LLVM

  • #18245 - [MSWIN][CI] Fix LLVM module build with latest CI update
  • #18232 - [Fix] Do not emit debuginfo on vscale or other unknown types
  • #18204 - Fixes up to the latest LLVM21
  • #18202 - [CPPTEST] Small fixes for LLVM >= 20

ROCm

  • #18225 - Minor fixes for latest refactor

Relax

  • #18229 - Add Python function support and BasePyModule for PyTorch integration
  • #18242 - ONNX frontend using relax softplus operator
  • #18180 - [ONNX] Parse ONNX Upsample to Relax resize2d
  • #18179 - Support Relax Operator PReLU

TIR

  • #18199 - Fix host/device function check for build

cuda & cutlass & tensorrt

  • #18238 - [CUTLASS] Fix CUTLASS kernel compilation

Misc

  • #18236 - upgrade cutlass v4.2.0 supporting cuda 13
  • #18251 - [Python] Complete Python packaging with scikit-build-core
  • #18248 - [Python] Update version.py to bump pyproject.toml automatically
  • #18239 - [Build] Migrate Python packaging to pyproject.toml with scikit-build-core
  • #18222 - [NVSHMEM] Fix compatibility with CUDA code without nvshmem use
  • #18220 - [Thrust] Fix getting CUDA stream
  • #18211 - [TARGET]add target for nvidia rtx 5060ti
  • #18206 - [CODEGEN][REFACTOR] tir.call_llvm_intrin to remove nargs
  • #18197 - [REFACTOR] Update data type rewriter to enable recursive rewrite in Any
  • #18193 - Bump cutlass_fpA_intB_gemm to latest commit