Skip to content

Context Size + extra 128 tokens #1729

@Nabokov86

Description

@Nabokov86

Hey! I've noticed something a bit weird with the context size. It seems to be running 128 tokens larger than what I'm setting, even when I use the --noshift --nofastforward flags. This isn't happening with llamacpp, so I thought I'd bring it up.

--contextsize 8192 --noshift --nofastforward
llama_context: n_ctx_per_seq (8320) < n_ctx_train (131072) -- the full capacity of the model will not be utilized
Image

The GUI just shows the value I set. Any chance I can actually run the model with just the context size I'm specifying? I'm guessing it might be something to do with context shifting?

Thanks!

Metadata

Metadata

Assignees

No one assigned

    Labels

    No labels
    No labels

    Projects

    No projects

    Milestone

    No milestone

    Relationships

    None yet

    Development

    No branches or pull requests

    Issue actions