Promptize attempts to solve the issues with context limits when working with AI systems. It allows a user to add an attribute to their struct called "Promptize" with a "chunkable" attribute on a specific field, along with "model" and "max_chunks" attributes.
If the prompt exceeds the allowable token size for your model, then the "chunkable" field will be chunked up into equal sized parts that fit within the context limit.
The fields from the original struct are repeated in a vec of a new version of that struct but with the chunked fields being broken up accordingly.
It's up to the caller to call openai multiple times for each prompt that is returned as a result of chunking.
use promptize::Promptize;
#[derive(Promptize, Debug, serde::Serialize)]
pub struct SomePrompt {
system_prompt: String,
user_prompt: String,
filename: String,
#[chunkable]
large_content: String,
#[model]
model: String,
#[max_chunks]
max_chunks: i32
}
fn main() -> Result<()> {
// Completion model is your chunked request struct
// Prompt is your original struct that was built
let (completion_model, prompt) = create_prompt()?;
if completion_model.len > 1 {
// do something now with your chunked model, like loop and call an LLM
}
Ok(())
}
fn create_prompt() -> Result<(Vec<Vec<ChatCompletionRequest>>, SomePrompt)> {
let prompt = SomePrompt::builder()
.system_prompt("System prompt here".to_string())
.user_prompt("User prompt here".to_string())
.filename("huge_file.rs".to_string())
.large_content("some huge amount of content here".to_string())
.model("gpt-4o")
.max_chunks(5)
.build()?;
Ok(prompt.build_prompt()?)
}
Your cargo.toml needs to have these two dependencies:
promptize = { version ="0.2.0", git = "https://github.com/danwritecode/promptize-ai-rs" }
promptize_internals = { version ="0.1.0", git = "https://github.com/danwritecode/promptize-internals" }
Perhaps in the future I can eliminate the need for the internals.
- Only 1 field can be "chunkable"
- A chunkable field has a limit to how many times it can be chunked to fit within the context limit
- There are some dependencies like tiktoken-rs, serde, anyhow
- I chose to use tiktoken's prompt struct. It's up to you to map this back to whatever struct you're using
- It's up to the caller to handle the response and call open-ai multiple times for each prompt