Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Handle long contexts better #192

Closed
pseudotensor opened this issue May 28, 2023 · 1 comment
Closed

Handle long contexts better #192

pseudotensor opened this issue May 28, 2023 · 1 comment
Assignees
Labels
priority/blocker Priority: issue is blocking development or release process

Comments

@pseudotensor
Copy link
Collaborator

While had added critical protection against failure when inputs too long, still if prompt happens to be too long, tokenization will do truncation=True for a max_length in some way (left or right or whatever), and can drop prompting of : etc.

Right now form prompt, then tokenize. But need to tokenize first ourselves (even for h2oai_pipeline), and remove non-human non-bot parts using split or similar, remove some stuff to get token length down, and then join back up.

Else missing human truncated out confuses model that just sees : and will repeat answers etc with another :. The output is cleaned from that, but stopping doesn't stop from that so calculations go in background and lag user.

@pseudotensor pseudotensor self-assigned this May 29, 2023
@pseudotensor pseudotensor added the priority/blocker Priority: issue is blocking development or release process label May 29, 2023
pseudotensor added a commit that referenced this issue May 30, 2023
…default. But need to deal with Issue #192, output lost for wizard case if prompting not right
pseudotensor added a commit that referenced this issue May 30, 2023
@pseudotensor
Copy link
Collaborator Author

#255 helps for h2oai_pipeline.py use. Should solve 99% of issues, but still fudge w.r.t. prompt itself.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
priority/blocker Priority: issue is blocking development or release process
Projects
None yet
Development

No branches or pull requests

1 participant