Perform LLM text completion inference.

POST {{baseUrl}}/api/v2/cortex/inference:complete

Perform LLM text completion inference, similar to snowflake.cortex.Complete.

Request Body

{"model"=>"<string>", "messages"=>[{"content"=>"<string>"}], "stream"=>false, "temperature"=>0, "top_p"=>1, "max_output_tokens"=>4096}

HEADERS

KeyDatatypeRequiredDescription
Content-Typestring
Acceptstring

RESPONSES

status: OK

{&quot;choices&quot;:[{&quot;message&quot;:{&quot;content&quot;:&quot;\u003cstring\u003e&quot;}},{&quot;message&quot;:{&quot;content&quot;:&quot;\u003cstring\u003e&quot;}}],&quot;usage&quot;:{&quot;prompt_tokens&quot;:&quot;\u003cinteger\u003e&quot;,&quot;completion_tokens&quot;:&quot;\u003cinteger\u003e&quot;,&quot;total_tokens&quot;:&quot;\u003cinteger\u003e&quot;}}