Skip to content

[ET Device Support] Parse device info from serialized tensor in tensor_parser#18328

Merged
Gasoonjia merged 9 commits intogh/gasoonjia/143/basefrom
gh/gasoonjia/143/head
Apr 17, 2026
Merged

[ET Device Support] Parse device info from serialized tensor in tensor_parser#18328
Gasoonjia merged 9 commits intogh/gasoonjia/143/basefrom
gh/gasoonjia/143/head

Conversation

@Gasoonjia
Copy link
Copy Markdown
Contributor

@Gasoonjia Gasoonjia commented Mar 19, 2026

Stack from ghstack (oldest at bottom):

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: D97199497

…r_parser

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
@pytorch-bot
Copy link
Copy Markdown

pytorch-bot Bot commented Mar 19, 2026

🔗 Helpful Links

🧪 See artifacts and rendered test results at hud.pytorch.org/pr/pytorch/executorch/18328

Note: Links to docs will display an error until the docs builds have been completed.

❗ 1 Active SEVs

There are 1 currently active SEVs. If your PR is affected, please view them below:

✅ You can merge normally! (2 Unrelated Failures)

As of commit 6deab13 with merge base 81bc830 (image):

FLAKY - The following job failed but was likely due to flakiness present on trunk:

BROKEN TRUNK - The following job failed but was present on the merge base:

👉 Rebase onto the `viable/strict` branch to avoid these failures

This comment was automatically generated by Dr. CI and updates every 15 minutes.

@github-actions
Copy link
Copy Markdown

This PR needs a release notes: label

If your change should be included in the release notes (i.e. would users of this library care about this change?), please use a label starting with release notes:. This helps us keep track and include your important work in the next release notes.

To add a label, you can comment to pytorchbot, for example
@pytorchbot label "release notes: none"

For more information, see
https://github.com/pytorch/pytorch/wiki/PyTorch-AutoLabel-Bot#why-categorize-for-release-notes-and-how-does-it-work.

…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
Gasoonjia added a commit that referenced this pull request Apr 13, 2026
…r_parser

Pull Request resolved: #18328

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、
ghstack-source-id: 366667637
@exported-using-ghexport

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
Gasoonjia added a commit that referenced this pull request Apr 14, 2026
…ensorSpecs (#18078)

Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at
bottom):
* #18080
* #18328
* #18079
* __->__ #18078

Add end-to-end device type annotation support from export to runtime.
Currently we only support one device per graph

The overall pipeline is:
a. Partitioner use `compile_spec` to determine which device the
partitoned blob is runing on
b. after lowered partitioned graph to backend, the new-introed
propagate_device_pass will annotate the input and output tensors of
delegate blob as target device.

Differential Revision:
[D95842511](https://our.internmc.facebook.com/intern/diff/D95842511/)
Gasoonjia added a commit that referenced this pull request Apr 14, 2026
…ized Tensor (#18079)

Stack from [ghstack](https://github.com/ezyang/ghstack) (oldest at
bottom):
* #18080
* #18328
* __->__ #18079
* #18078

Propagate device information from `TensorSpec.device` (set by
`PropagateDevicePass`) to
the serialized `schema.Tensor` in the emitted PTE file, to make runtime
further aware of it.

Differential Revision:
[D95899706](https://our.internmc.facebook.com/intern/diff/D95899706/)
…or in tensor_parser"

Parse device info (device_type, device_index) from the serialized ExtraTensorInfo in .pte files into TensorImpl at runtime.
When a tensor's extra_tensor_info contains device annotations (e.g., CUDA), the tensor parser now reads and propagates them to the TensorImpl constructor. Tensors without extra_tensor_info default to CPU/0 for backward compatibility with older PTE files.、

Differential Revision: [D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)

[ghstack-poisoned]
@Gasoonjia Gasoonjia merged commit c72f072 into gh/gasoonjia/143/base Apr 17, 2026
161 of 164 checks passed
@Gasoonjia Gasoonjia deleted the gh/gasoonjia/143/head branch April 17, 2026 02:14
@Gasoonjia Gasoonjia temporarily deployed to cherry-pick-bot April 17, 2026 02:14 — with GitHub Actions Inactive
Gasoonjia added a commit that referenced this pull request Apr 17, 2026
…r_parser (#18966)

This PR was created by the merge bot to help merge the original PR into
the main branch.
ghstack PR number: #18328 by
@Gasoonjia
^ Please use this as the source of truth for the PR details, comments,
and reviews
ghstack PR base:
https://github.com/pytorch/executorch/tree/gh/gasoonjia/143/base
ghstack PR head:
https://github.com/pytorch/executorch/tree/gh/gasoonjia/143/head
Merge bot PR base: https://github.com/pytorch/executorch/tree/main
Merge bot PR head:
https://github.com/pytorch/executorch/tree/gh/gasoonjia/143/orig
Differential Revision:
[D97199497](https://our.internmc.facebook.com/intern/diff/D97199497/)
@diff-train-skip-merge

Co-authored-by: gasoonjia <gasoonjia@icloud.com>
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants