Generate text completions using state-of-the-art language models
API key authentication. Get your API key from the Gravix Layer Dashboard.
Model identifier
"meta-llama/llama-3.1-8b-instruct"
Prompt to complete
Maximum tokens to generate
Sampling temperature
0 <= x <= 2Nucleus sampling
0 <= x <= 1Number of completions
Whether to stream the response
Include top N log probabilities
Echo the prompt in response
Stop sequences
Presence penalty
-2 <= x <= 2Frequency penalty
-2 <= x <= 2Generate N completions, return best
Modify token probabilities
User identifier
Completion response