Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

ValidateSyncInputTensors tensor_data is empty #1241

Open
2 of 4 tasks
xinsu626 opened this issue Aug 12, 2024 · 1 comment
Open
2 of 4 tasks

ValidateSyncInputTensors tensor_data is empty #1241

xinsu626 opened this issue Aug 12, 2024 · 1 comment
Labels
bug Something isn't working

Comments

@xinsu626
Copy link

xinsu626 commented Aug 12, 2024

System Info

Docker image: pytorch-installer-2.3.1:1.17.0-417
optimum-habana: main branch

Information

  • The official example scripts
  • My own modified scripts

Tasks

  • An officially supported task in the examples folder (such as GLUE/SQuAD, ...)
  • My own task or dataset (give details below)

Reproduction

When I ran the llama3.1-70b-instruct model for inference and set self.generation_config.ignore_eos to False, I encountered the following error.

[rank0]:   File "/usr/local/lib/python3.10/dist-packages/torch/utils/_contextlib.py", line 115, in decorate_context
[rank0]:     return func(*args, **kwargs)
[rank0]:   File "/usr/local/lib/python3.10/dist-packages/optimum/habana/transformers/generation/utils.py", line 1287, in generate
[rank0]:     result = self._sample(
[rank0]:   File "/usr/local/lib/python3.10/dist-packages/optimum/habana/transformers/generation/utils.py", line 2360, in _sample
[rank0]:     unfinished_sequences = unfinished_sequences & ~stopping_criteria(
[rank0]:   File "/usr/local/lib/python3.10/dist-packages/optimum/habana/transformers/generation/stopping_criteria.py", line 113, in gaudi_StoppingCriteriaList_call
[rank0]:     is_done = is_done | criteria(input_ids, scores, **kwargs)
[rank0]:   File "/usr/local/lib/python3.10/dist-packages/optimum/habana/transformers/generation/stopping_criteria.py", line 84, in gaudi_EosTokenCriteria_call
[rank0]:     is_done = torch.isin(input_ids[:, token_idx - 1], self.eos_token_id)
[rank0]: RuntimeError: [Rank:0] FATAL ERROR :: MODULE:PT_BRIDGE Exception in Launch thread...
[rank0]: Check $HABANA_LOGS/ for details[Rank:0] FATAL ERROR :: MODULE:PT_LAZY Error, ValidateSyncInputTensors tensor_data is empty. Tensorid:2006352 QueueStatus:ThreadPool m_tasks size: 1 irValue:id_6403519_module/model/79/hpu__input

Expected behavior

No errors during model inference.

@xinsu626 xinsu626 added the bug Something isn't working label Aug 12, 2024
@regisss
Copy link
Collaborator

regisss commented Oct 20, 2024

@xinsu626 Can you share the command you executed please?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
bug Something isn't working
Projects
None yet
Development

No branches or pull requests

2 participants