Skip to content

Commit cb2fecc

Browse files
committed
wip
1 parent 8f467a1 commit cb2fecc

1 file changed

Lines changed: 39 additions & 26 deletions

File tree

openapi.yaml

Lines changed: 39 additions & 26 deletions
Original file line numberDiff line numberDiff line change
@@ -3213,41 +3213,41 @@ paths:
32133213
type: classify
32143214
parameters:
32153215
judge:
3216-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3217-
system_template: "You are a helpful assistant that classifies text."
3218-
labels: ["Toxic", "Non-Toxic"]
3219-
pass_labels: ["Non-Toxic"]
3220-
model_to_evaluate: "output_column"
3221-
input_data_file_path: "file-abcd-1234"
3216+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3217+
system_template: 'You are a helpful assistant that classifies text.'
3218+
labels: ['Toxic', 'Non-Toxic']
3219+
pass_labels: ['Non-Toxic']
3220+
model_to_evaluate: 'output_column'
3221+
input_data_file_path: 'file-abcd-1234'
32223222
score:
32233223
summary: Score evaluation example
32243224
value:
32253225
type: score
32263226
parameters:
32273227
judge:
3228-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3229-
system_template: "You are a helpful assistant that scores responses."
3228+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3229+
system_template: 'You are a helpful assistant that scores responses.'
32303230
min_score: 0
32313231
max_score: 10
32323232
pass_threshold: 7
32333233
model_to_evaluate:
3234-
model_name: "meta-llama/Llama-3-8B-Instruct-Turbo"
3234+
model_name: 'meta-llama/Llama-3-8B-Instruct-Turbo'
32353235
max_tokens: 512
32363236
temperature: 0.7
3237-
system_template: "You are a helpful assistant."
3238-
input_template: "Classify: {prompt}"
3239-
input_data_file_path: "file-1234-abcd"
3237+
system_template: 'You are a helpful assistant.'
3238+
input_template: 'Classify: {prompt}'
3239+
input_data_file_path: 'file-1234-abcd'
32403240
compare:
32413241
summary: Compare evaluation example
32423242
value:
32433243
type: compare
32443244
parameters:
32453245
judge:
3246-
model_name: "meta-llama/Llama-3-70B-Instruct-Turbo"
3247-
system_template: "You are a helpful assistant that compares responses."
3248-
model_a: "response_a"
3249-
model_b: "response_b"
3250-
input_data_file_path: "file-1234-abcd"
3246+
model_name: 'meta-llama/Llama-3-70B-Instruct-Turbo'
3247+
system_template: 'You are a helpful assistant that compares responses.'
3248+
model_a: 'response_a'
3249+
model_b: 'response_b'
3250+
input_data_file_path: 'file-1234-abcd'
32513251
responses:
32523252
'200':
32533253
description: Successful response
@@ -3367,7 +3367,8 @@ paths:
33673367
properties:
33683368
status:
33693369
type: string
3370-
enum: [pending, queued, running, completed, error, user_error]
3370+
enum:
3371+
[pending, queued, running, completed, error, user_error]
33713372
example: completed
33723373
results:
33733374
oneOf:
@@ -3416,7 +3417,15 @@ paths:
34163417
properties:
34173418
status:
34183419
type: string
3419-
enum: [completed, error, running, queued, user_error, inference_error]
3420+
enum:
3421+
[
3422+
completed,
3423+
error,
3424+
running,
3425+
queued,
3426+
user_error,
3427+
inference_error,
3428+
]
34203429
description: The new status for the job
34213430
results:
34223431
type: object
@@ -4130,6 +4139,11 @@ components:
41304139
type: string
41314140
description: The name of the moderation model used to validate tokens. Choose from the available moderation models found [here](https://docs.together.ai/docs/inference-models#moderation-models).
41324141
example: 'safety_model_name'
4142+
reasoning_effort:
4143+
type: string
4144+
enum: ['low', 'medium', 'high']
4145+
description: Controls the level of reasoning effort the model should apply when generating responses. Higher values may result in more thoughtful and detailed responses but may take longer to generate.
4146+
example: 'medium'
41334147

41344148
ChatCompletionMessageParam:
41354149
oneOf:
@@ -4563,7 +4577,7 @@ components:
45634577
maxItems: 2
45644578
description: Controls level of timestamp detail in verbose_json. Only used when response_format is verbose_json. Can be a single granularity or an array to get multiple levels.
45654579
default: segment
4566-
example: ["word", "segment"]
4580+
example: ['word', 'segment']
45674581

45684582
AudioTranscriptionResponse:
45694583
oneOf:
@@ -4728,7 +4742,7 @@ components:
47284742
maxItems: 2
47294743
description: Controls level of timestamp detail in verbose_json. Only used when response_format is verbose_json. Can be a single granularity or an array to get multiple levels.
47304744
default: segment
4731-
example: ["word", "segment"]
4745+
example: ['word', 'segment']
47324746

47334747
AudioTranslationResponse:
47344748
oneOf:
@@ -6609,14 +6623,14 @@ components:
66096623
type: string
66106624
minItems: 2
66116625
description: List of possible classification labels
6612-
example: ["yes", "no"]
6626+
example: ['yes', 'no']
66136627
pass_labels:
66146628
type: array
66156629
items:
66166630
type: string
66176631
minItems: 1
66186632
description: List of labels that are considered passing
6619-
example: ["yes"]
6633+
example: ['yes']
66206634
model_to_evaluate:
66216635
$ref: '#/components/schemas/EvaluationModelOrString'
66226636
input_data_file_path:
@@ -6734,7 +6748,7 @@ components:
67346748
workflow_id:
67356749
type: string
67366750
description: The ID of the created evaluation job
6737-
example: "eval-1234-1244513"
6751+
example: 'eval-1234-1244513'
67386752
status:
67396753
type: string
67406754
enum: [pending]
@@ -6843,7 +6857,6 @@ components:
68436857
description: JSON string representing label counts
68446858
example: '{"yes": 10, "no": 0}'
68456859

6846-
68476860
EvaluationScoreResults:
68486861
type: object
68496862
properties:
@@ -6904,4 +6917,4 @@ components:
69046917
example: 0
69056918
result_file_id:
69066919
type: string
6907-
description: Data File ID
6920+
description: Data File ID

0 commit comments

Comments
 (0)