Skip to content
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -151,7 +151,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json'\
>> -H 'content-type: application/json' \
>> -d '{
Expand All @@ -160,6 +160,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> { "content": "You are a helpful assistant that help users rank different things. You always answer in JSON format.", "role": "system" },
>> { "content": "What are the top 3 most popular programming languages ?", "role": "user" }
>> ],
>> "model": "Meta-Llama-3_3-70B-Instruct",
>> "response_format": {
>> "type":"json_schema",
>> "json_schema": {
Expand Down Expand Up @@ -258,14 +259,15 @@ The following code samples provide a simple example on how to specify a JSON sch
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_schema',
>> json_schema: jsonSchema
Expand Down Expand Up @@ -327,7 +329,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -375,7 +377,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json' \
>> -H 'content-type: application/json' \
>> -d '{
Expand Down Expand Up @@ -413,14 +415,15 @@ The following code samples provide a simple example on how to use the legacy JSO
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_object',
>> },
Expand Down Expand Up @@ -487,7 +490,7 @@ class LanguageRankings(BaseModel):
# Initialise the client
api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
openai_client = openai.OpenAI(
base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
api_key=api_key
)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -151,7 +151,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json'\
>> -H 'content-type: application/json' \
>> -d '{
Expand All @@ -160,6 +160,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> { "content": "You are a helpful assistant that help users rank different things. You always answer in JSON format.", "role": "system" },
>> { "content": "What are the top 3 most popular programming languages ?", "role": "user" }
>> ],
>> "model": "Meta-Llama-3_3-70B-Instruct",
>> "response_format": {
>> "type":"json_schema",
>> "json_schema": {
Expand Down Expand Up @@ -258,14 +259,15 @@ The following code samples provide a simple example on how to specify a JSON sch
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_schema',
>> json_schema: jsonSchema
Expand Down Expand Up @@ -327,7 +329,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -375,7 +377,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json' \
>> -H 'content-type: application/json' \
>> -d '{
Expand Down Expand Up @@ -413,14 +415,15 @@ The following code samples provide a simple example on how to use the legacy JSO
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_object',
>> },
Expand Down Expand Up @@ -487,7 +490,7 @@ class LanguageRankings(BaseModel):
# Initialise the client
api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
openai_client = openai.OpenAI(
base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
api_key=api_key
)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -151,7 +151,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json'\
>> -H 'content-type: application/json' \
>> -d '{
Expand All @@ -160,6 +160,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> { "content": "You are a helpful assistant that help users rank different things. You always answer in JSON format.", "role": "system" },
>> { "content": "What are the top 3 most popular programming languages ?", "role": "user" }
>> ],
>> "model": "Meta-Llama-3_3-70B-Instruct",
>> "response_format": {
>> "type":"json_schema",
>> "json_schema": {
Expand Down Expand Up @@ -258,14 +259,15 @@ The following code samples provide a simple example on how to specify a JSON sch
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_schema',
>> json_schema: jsonSchema
Expand Down Expand Up @@ -327,7 +329,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -375,7 +377,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json' \
>> -H 'content-type: application/json' \
>> -d '{
Expand Down Expand Up @@ -413,14 +415,15 @@ The following code samples provide a simple example on how to use the legacy JSO
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_object',
>> },
Expand Down Expand Up @@ -487,7 +490,7 @@ class LanguageRankings(BaseModel):
# Initialise the client
api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
openai_client = openai.OpenAI(
base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
api_key=api_key
)

Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -151,7 +151,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json'\
>> -H 'content-type: application/json' \
>> -d '{
Expand All @@ -160,6 +160,7 @@ The following code samples provide a simple example on how to specify a JSON sch
>> { "content": "You are a helpful assistant that help users rank different things. You always answer in JSON format.", "role": "system" },
>> { "content": "What are the top 3 most popular programming languages ?", "role": "user" }
>> ],
>> "model": "Meta-Llama-3_3-70B-Instruct",
>> "response_format": {
>> "type":"json_schema",
>> "json_schema": {
Expand Down Expand Up @@ -258,14 +259,15 @@ The following code samples provide a simple example on how to specify a JSON sch
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_schema',
>> json_schema: jsonSchema
Expand Down Expand Up @@ -327,7 +329,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> # Initialise the client
>> api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> openai_client = openai.OpenAI(
>> base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
>> base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
>> api_key=api_key
>> )
>>
Expand Down Expand Up @@ -375,7 +377,7 @@ The following code samples provide a simple example on how to use the legacy JSO
>> Input query:
>>
>> ```sh
>> curl -X POST "https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions" \
>> curl -X POST "https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions" \
>> -H 'accept: application/json' \
>> -H 'content-type: application/json' \
>> -d '{
Expand Down Expand Up @@ -413,14 +415,15 @@ The following code samples provide a simple example on how to use the legacy JSO
>> // Initialise the client
>> const apiKey = process.env.AI_ENDPOINT_API_KEY; // Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
>> const options = {
>> url: 'https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1/chat/completions',
>> url: 'https://oai.endpoints.kepler.ai.cloud.ovh.net/v1/chat/completions',
>> headers: {
>> 'Content-Type': 'application/json',
>> 'Authorization': `Bearer ${apiKey}`
>> },
>> json: true,
>> body: {
>> messages: messages,
>> model: 'Meta-Llama-3_3-70B-Instruct',
>> response_format: {
>> type: 'json_object',
>> },
Expand Down Expand Up @@ -487,7 +490,7 @@ class LanguageRankings(BaseModel):
# Initialise the client
api_key = os.environ['AI_ENDPOINT_API_KEY'] # Assuming your API key is available in this environment variable (export AI_ENDPOINT_API_KEY='your_api_key')
openai_client = openai.OpenAI(
base_url='https://llama-3-3-70b-instruct.endpoints.kepler.ai.cloud.ovh.net/api/openai_compat/v1',
base_url='https://oai.endpoints.kepler.ai.cloud.ovh.net/v1',
api_key=api_key
)

Expand Down
Loading