Skip to content

Question - Batch Processing #718

Open
@virentakia

Description

@virentakia

Problem

I'm processing a batch of approximately 100 prompts, each ranging from 1,200 to 14,000 tokens in length. Given that the input context length must be specified during model initialisation, I'm considering two options:

  1. Initialize the model with the maximum token length of 14,000 tokens.
    OR
  2. Instantiate the model each time with the required context length.

Are there other avenues I should be exploring?

Solution

Need to know best possible options

Alternatives

No response

Explanation

Need to know best possible options

Examples

No response

Additional context

No response

Acknowledgements

  • I have looked for similar requests before submitting this one.
  • I understand that the developers have lives and my issue will be answered when possible.
  • I understand the developers of this program are human, and I will make my requests politely.

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Type

    No type

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions