Optional
fields: GooglePaLMChatInputModel Name to use
Note: The format must follow the pattern - models/{model}
Model Name to use
Note: The format must follow the pattern - models/{model}
Alias for model
Optional
apiGoogle Palm API key to use
Optional
temperatureControls the randomness of the output.
Values can range from [0.0,1.0], inclusive. A value closer to 1.0 will produce responses that are more varied and creative, while a value closer to 0.0 will typically result in less surprising responses from the model.
Note: The default value varies by model
Optional
topKTop-k changes how the model selects tokens for output.
A top-k of 1 means the selected token is the most probable among all tokens in the model’s vocabulary (also called greedy decoding), while a top-k of 3 means that the next token is selected from among the 3 most probable tokens (using temperature).
Note: The default value varies by model
Optional
topPTop-p changes how the model selects tokens for output.
Tokens are selected from most probable to least until the sum of their probabilities equals the top-p value.
For example, if tokens A, B, and C have a probability of .3, .2, and .1 and the top-p value is .5, then the model will select either A or B as the next token (using temperature).
Note: The default value varies by model
Generated using TypeDoc
⚠️ Deprecated ⚠️
This feature is deprecated and will be removed in the future.
It is not recommended for use.
A class that wraps the Google Palm chat model.
Example