Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Ep Context Model generated with external data is still dependent on the same data file #23358

Open
BoarQing opened this issue Jan 14, 2025 · 2 comments
Assignees
Labels
ep:VitisAI issues related to Vitis AI execution provider

Comments

@BoarQing
Copy link
Contributor

Describe the issue

We found when the Ep Context Model is generated from a model with an external data, if an OP is not at the provider domain, it would still read to the original data file. This is undesirable since the external data is quite big.

Maybe MSFT can trim the external data for the new Ep Context Model to only contains those OPs' weights?
Or Maybe MSFT can just embed those weights into the new Ep Context Model?

To reproduce

  1. Generate an Ep context model on a model with external data.
  2. Delete the external data.
  3. Inferencing the Ep context model would crash for not finding the external data.

Urgency

It is important for MSFT release.

Platform

Windows

OS Version

Windows11

ONNX Runtime Installation

Built from Source

ONNX Runtime Version or Commit ID

main

ONNX Runtime API

C++

Architecture

X64

Execution Provider

Vitis AI

Execution Provider Library Version

No response

@github-actions github-actions bot added the ep:VitisAI issues related to Vitis AI execution provider label Jan 14, 2025
@BoarQing
Copy link
Contributor Author

@snnn @jywu-msft @HectorSVC

@BoarQing
Copy link
Contributor Author

#23374 A temporary solution to move all the tensor into the .onnx.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
ep:VitisAI issues related to Vitis AI execution provider
Projects
None yet
Development

No branches or pull requests

2 participants