Skip to content

Conversation

@jd7-tr
Copy link
Contributor

@jd7-tr jd7-tr commented May 6, 2025

Summary:

Context

  • Currently torchrec IR serializer can't handle variable batch KJT use case.
  • To support VBE KJT, the stride_per_key_per_rank field needs to be flattened as a variable in the pytree flatten spec for a VBE KJT to be unflattened correctly by`torch.export.
  • Currently stride_per_key_per_rank is a List. To flatten the stride_per_key_per_rank info as a variable we have to add a new tensor field for it.

Ref

Differential Revision: D74207283

@facebook-github-bot facebook-github-bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label May 6, 2025
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D74207283

jd7-tr added 2 commits May 6, 2025 18:15
…s for VBE KJTs. (meta-pytorch#2949)

Summary:

Update the `_maybe_compute_stride_kjt` logic to calculate stride based off of `inverse_indices` for VBE KJTs.

Currently, stride of VBE KJT with `stride_per_key_per_rank` is calculated as the max "stride per key". This is different from the batch size of the EBC output KeyedTensor which is based off of inverse_indices. This causes issues in IR module serialization: debug doc.

Differential Revision: D74273083
…ch.export (meta-pytorch#2950)

Summary:

# Context
* Currently torchrec IR serializer can't handle variable batch KJT use case.
* To support VBE KJT, the `stride_per_key_per_rank` field needs to be flattened as a variable in the pytree flatten spec for a VBE KJT to be unflattened correctly by`torch.export`.
* Currently `stride_per_key_per_rank` is a List. To flatten the `stride_per_key_per_rank`  info as a variable we have to add a new tensor field for it.

# Ref

Differential Revision: D74207283
@jd7-tr jd7-tr force-pushed the export-D74207283 branch from cdf8953 to 96ee8d5 Compare May 7, 2025 01:15
@facebook-github-bot
Copy link
Contributor

This pull request was exported from Phabricator. Differential Revision: D74207283

jd7-tr added a commit to jd7-tr/torchrec that referenced this pull request May 7, 2025
…ch.export (meta-pytorch#2950)

Summary:

# Context
* Currently torchrec IR serializer can't handle variable batch KJT use case.
* To support VBE KJT, the `stride_per_key_per_rank` field needs to be flattened as a variable in the pytree flatten spec for a VBE KJT to be unflattened correctly by`torch.export`.
* Currently `stride_per_key_per_rank` is a List. To flatten the `stride_per_key_per_rank`  info as a variable we have to add a new tensor field for it.

# Ref

Differential Revision: D74207283
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants