|
|
|||
|
||||
OverviewBuild production grade ML compilers with MLIR, from TensorFlow and PyTorch graphs to fast GPU, CPU, and embedded executables Machine learning teams struggle to turn research models into efficient binaries across diverse hardware. Toolchains are fragmented, passes are opaque, and small changes can break performance or correctness. This book gives you a clear path. You get a practical workflow that starts with readable IR, enforces graph invariants with strong verifiers, and lowers to portable or vendor specific code that you can ship with confidence. Design solid operators using ODS and traits, add verifiers and builders that keep graphs legal, and attach interfaces that unlock tiling, fusion, and bufferization Import TensorFlow with StableHLO and VHLO, use the TFLite and TF bridges, and keep portability with TOSA when you need framework neutral flows Capture PyTorch programs with Torch MLIR, decompose to arith tensor and linalg, and manage distinct training and inference paths without forking pipelines Apply shape reasoning with the Shape dialect, handle static and dynamic ranks, and wire in inference that feeds downstream transforms Run post training quantization with the Quant dialect, carry scales and zero points correctly, and build calibration aware dequant pipelines Bufferize tensors with One Shot Bufferize, control function boundaries, model effects precisely, and validate lifetimes with ownership based deallocation Tune memory with MemRef layout maps, alignment and packing, and pick layouts that suit accelerators without losing legality Generate GPU code with GPU and NVGPU dialects, target NVVM or ROCDL, and use vector and tensor core paths that map to real intrinsics Target SPIR V for Vulkan environments with capability gating, or generate portable C and C++ for microcontrollers with EmitC JIT with ExecutionEngine and JitRunner, or use IREE end to end for compilation and runtime on mobile, desktop, and server Drive performance with tiling fusion and vectorization in Linalg and Vector, add autotuning hooks, and apply the Sparse Tensor dialect for structured sparsity Profile with remarks counters and traces, then lock down stability with lit and FileCheck, mlir reduce, bytecode, and dialect versioning Work through complete case studies, TensorFlow ResNet to CUDA with NVGPU and NVVM, PyTorch Transformer to ROCm with ROCDL, quantized MobileNet to EmitC for Cortex M, and sparse attention to SPIR V for Vulkan This is a code heavy guide with labeled MLIR Python C++ Shell and TableGen listings, you can copy pipelines and schedules directly into your builds to stand up real projects. Grab your copy today Full Product DetailsAuthor: Anik RaoPublisher: Independently Published Imprint: Independently Published Dimensions: Width: 17.80cm , Height: 1.30cm , Length: 25.40cm Weight: 0.445kg ISBN: 9798272164148Pages: 252 Publication Date: 29 October 2025 Audience: General/trade , General Format: Paperback Publisher's Status: Active Availability: Available To Order We have confirmation that this item is in stock with the supplier. It will be ordered in for you and dispatched immediately. Table of ContentsReviewsAuthor InformationTab Content 6Author Website:Countries AvailableAll regions |
||||