DefaultSamplingPipeline
Namespace: LLama.Sampling
An implementation of ISamplePipeline which mimics the default llama.cpp sampling
1 |
|
Inheritance Object → BaseSamplingPipeline → DefaultSamplingPipeline
Implements ISamplingPipeline, IDisposable
Properties
LogitBias
Bias values to add to certain logits
1 |
|
Property Value
RepeatPenalty
Repetition penalty, as described in https://arxiv.org/abs/1909.05858
1 |
|
Property Value
AlphaFrequency
Frequency penalty as described by OpenAI: https://platform.openai.com/docs/api-reference/chat/create
Number between -2.0 and 2.0. Positive values penalize new tokens based on their existing frequency in the text
so far, decreasing the model's likelihood to repeat the same line verbatim.
1 |
|
Property Value
AlphaPresence
Presence penalty as described by OpenAI: https://platform.openai.com/docs/api-reference/chat/create
Number between -2.0 and 2.0. Positive values penalize new tokens based on whether they appear in the
text so far, increasing the model's likelihood to talk about new topics.
1 |
|
Property Value
Temperature
Temperature to apply (higher temperature is more "creative")
1 |
|
Property Value
TopK
Number of tokens to keep in TopK sampling
1 |
|
Property Value
TailFreeZ
Z value for tail free sampling
1 |
|
Property Value
TypicalP
P value for locally typical sampling
1 |
|
Property Value
TopP
P value for TopP sampling
1 |
|
Property Value
MinP
P value for MinP sampling
1 |
|
Property Value
PenalizeNewline
Whether the newline value should be protected from being modified by logit bias and repeat penalty
1 |
|
Property Value
Grammar
Grammar to constrain valid tokens
1 |
|
Property Value
Constructors
DefaultSamplingPipeline()
1 |
|
Methods
ProcessLogits(SafeLLamaContextHandle, Span<Single>, ReadOnlySpan<LLamaToken>)
1 |
|
Parameters
logits
Span<Single>
lastTokens
ReadOnlySpan<LLamaToken>
ProcessTokenDataArray(SafeLLamaContextHandle, LLamaTokenDataArray, ReadOnlySpan<LLamaToken>)
1 |
|
Parameters
candidates
LLamaTokenDataArray
lastTokens
ReadOnlySpan<LLamaToken>
Returns
Accept(SafeLLamaContextHandle, LLamaToken)
1 |
|
Parameters
token
LLamaToken
Clone()
1 |
|