Skip to main content
POST
/
v1
/
inference
/
chat
/
completions
curl -X POST https://api.gravixlayer.com/v1/inference/chat/completions \
-H 'Content-Type: application/json' \
-H 'Authorization: Bearer $GRAVIXLAYER_API_KEY' \
-d '{\n "model": "meta-llama/llama-3.1-8b-instruct",\n "messages": [\n {\n "role": "user",\n "content": "Hello! Tell me about AI."\n }\n ]\n }'
{
  "data": {
    "id": "<string>",
    "object": "chat.completion",
    "created": 123,
    "model": "<string>",
    "choices": [
      {
        "message": {
          "role": "system",
          "content": "<string>"
        },
        "index": 123,
        "finish_reason": "<string>"
      }
    ]
  }
}

Authorizations

Authorization
string
header
required

API key authentication. Get your API key from the Gravix Layer Dashboard.

Body

application/json
model
string
required

Model identifier

Example:

"meta-llama/llama-3.1-8b-instruct"

messages
object[]
required

Messages to generate the completion from

temperature
number
default:1

Sampling temperature

Required range: 0 <= x <= 2
top_p
number
default:1

Nucleus sampling

Required range: 0 <= x <= 1
n
integer
default:1

Number of completions to generate

stream
boolean
default:false

Whether to stream the response

stop

Stop sequences

max_tokens
integer | null

Maximum tokens to generate

presence_penalty
number
default:0

Presence penalty

Required range: -2 <= x <= 2
frequency_penalty
number
default:0

Frequency penalty

Required range: -2 <= x <= 2
logit_bias
object

Modify token probabilities

user
string | null

User identifier

Response

Chat completion response

data
object