2000 chunk size

This commit is contained in:
Cyberes 2023-10-16 18:01:17 -06:00
parent 19a193b792
commit 20047fa0e4
1 changed files with 2 additions and 2 deletions

View File

@ -22,8 +22,8 @@ def tokenize(prompt: str, backend_url: str) -> int:
tokenizer = tiktoken.get_encoding("cl100k_base")
# Split the prompt into 1000 character chunks
chunk_size = 1000
# Split the prompt into 2000 character chunks
chunk_size = 2000
chunks = [prompt[i:i + chunk_size] for i in range(0, len(prompt), chunk_size)]
# Define a function to send a chunk to the server