Skip to content

[FEA] Add cuTensorMapEncodeTiled to CudaHostAdapter #1566

@drisspg

Description

@drisspg

Summary

We are currently working on integrating a fp8 scaled matmul kernel written using Cutlass into PyTorch. PyTorch has the constraint that it can be linked against the cuda driver api. There is one symbol/direct call to a cuda driver api cuTensorMapEncodeTiled that is causing issues.

We have a temporary workaround here: pytorch/pytorch#125204 (comment)

There was a suggestion to add this symbol to CudaHostAdapter so as to add one more layer of indirection. This would greatly aid in PyTorch in its utilization of Cutlass.

Metadata

Metadata

Assignees

No one assigned

    Type

    No type

    Projects

    No projects

    Milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions