1
1
lockVersion: 2.0.0
2
2
id: 2d045ec7-2ebb-4f4d-ad25-40953b132161
3
3
management:
4
- docChecksum: 82cf79b2dee6811d91e2912113c21d3a
4
+ docChecksum: 3ded9717d39dd1d0880fef10c713583c
5
5
docVersion: 1.0.0
6
6
speakeasyVersion: 1.568.2
7
7
generationVersion: 2.634.2
8
- releaseVersion: 1.9.2
9
- configChecksum: 1ca921f44508650d65ccf46783910ff3
8
+ releaseVersion: 1.9.3
9
+ configChecksum: 0f65a9bdd8df5ae03eaaaea3ab055bf1
10
10
repoURL: https://github.com/mistralai/client-python.git
11
11
installationURL: https://github.com/mistralai/client-python.git
12
12
published: true
13
13
features:
14
14
python:
15
15
additionalDependencies: 1.0.0
16
+ additionalProperties: 1.0.1
16
17
constsAndDefaults: 1.0.5
17
18
core: 5.19.3
18
19
customCodeRegions: 0.1.1
@@ -89,6 +90,10 @@ generatedFiles:
89
90
- docs/models/assistantmessage.md
90
91
- docs/models/assistantmessagecontent.md
91
92
- docs/models/assistantmessagerole.md
93
+ - docs/models/audiochunk.md
94
+ - docs/models/audiochunktype.md
95
+ - docs/models/audiotranscriptionrequest.md
96
+ - docs/models/audiotranscriptionrequeststream.md
92
97
- docs/models/basemodelcard.md
93
98
- docs/models/batcherror.md
94
99
- docs/models/batchjobin.md
@@ -320,6 +325,7 @@ generatedFiles:
320
325
- docs/models/messageoutputeventrole.md
321
326
- docs/models/messageoutputeventtype.md
322
327
- docs/models/messages.md
328
+ - docs/models/metadata.md
323
329
- docs/models/metricout.md
324
330
- docs/models/mistralpromptmode.md
325
331
- docs/models/modelcapabilities.md
@@ -375,6 +381,10 @@ generatedFiles:
375
381
- docs/models/systemmessagecontent.md
376
382
- docs/models/textchunk.md
377
383
- docs/models/textchunktype.md
384
+ - docs/models/thinkchunk.md
385
+ - docs/models/thinkchunktype.md
386
+ - docs/models/thinking.md
387
+ - docs/models/timestampgranularity.md
378
388
- docs/models/tool.md
379
389
- docs/models/toolcall.md
380
390
- docs/models/toolchoice.md
@@ -398,6 +408,15 @@ generatedFiles:
398
408
- docs/models/tools.md
399
409
- docs/models/tooltypes.md
400
410
- docs/models/trainingfile.md
411
+ - docs/models/transcriptionresponse.md
412
+ - docs/models/transcriptionsegmentchunk.md
413
+ - docs/models/transcriptionstreamdone.md
414
+ - docs/models/transcriptionstreamevents.md
415
+ - docs/models/transcriptionstreameventsdata.md
416
+ - docs/models/transcriptionstreameventtypes.md
417
+ - docs/models/transcriptionstreamlanguage.md
418
+ - docs/models/transcriptionstreamsegmentdelta.md
419
+ - docs/models/transcriptionstreamtextdelta.md
401
420
- docs/models/two.md
402
421
- docs/models/type.md
403
422
- docs/models/unarchiveftmodelout.md
@@ -420,6 +439,7 @@ generatedFiles:
420
439
- docs/models/websearchtooltype.md
421
440
- docs/sdks/accesses/README.md
422
441
- docs/sdks/agents/README.md
442
+ - docs/sdks/audio/README.md
423
443
- docs/sdks/batch/README.md
424
444
- docs/sdks/beta/README.md
425
445
- docs/sdks/chat/README.md
@@ -437,6 +457,7 @@ generatedFiles:
437
457
- docs/sdks/mistraljobs/README.md
438
458
- docs/sdks/models/README.md
439
459
- docs/sdks/ocr/README.md
460
+ - docs/sdks/transcriptions/README.md
440
461
- poetry.toml
441
462
- py.typed
442
463
- scripts/prepare_readme.py
@@ -448,6 +469,7 @@ generatedFiles:
448
469
- src/mistralai/_version.py
449
470
- src/mistralai/accesses.py
450
471
- src/mistralai/agents.py
472
+ - src/mistralai/audio.py
451
473
- src/mistralai/basesdk.py
452
474
- src/mistralai/batch.py
453
475
- src/mistralai/beta.py
@@ -489,6 +511,9 @@ generatedFiles:
489
511
- src/mistralai/models/apiendpoint.py
490
512
- src/mistralai/models/archiveftmodelout.py
491
513
- src/mistralai/models/assistantmessage.py
514
+ - src/mistralai/models/audiochunk.py
515
+ - src/mistralai/models/audiotranscriptionrequest.py
516
+ - src/mistralai/models/audiotranscriptionrequeststream.py
492
517
- src/mistralai/models/basemodelcard.py
493
518
- src/mistralai/models/batcherror.py
494
519
- src/mistralai/models/batchjobin.py
@@ -668,6 +693,8 @@ generatedFiles:
668
693
- src/mistralai/models/ssetypes.py
669
694
- src/mistralai/models/systemmessage.py
670
695
- src/mistralai/models/textchunk.py
696
+ - src/mistralai/models/thinkchunk.py
697
+ - src/mistralai/models/timestampgranularity.py
671
698
- src/mistralai/models/tool.py
672
699
- src/mistralai/models/toolcall.py
673
700
- src/mistralai/models/toolchoice.py
@@ -681,6 +708,14 @@ generatedFiles:
681
708
- src/mistralai/models/toolreferencechunk.py
682
709
- src/mistralai/models/tooltypes.py
683
710
- src/mistralai/models/trainingfile.py
711
+ - src/mistralai/models/transcriptionresponse.py
712
+ - src/mistralai/models/transcriptionsegmentchunk.py
713
+ - src/mistralai/models/transcriptionstreamdone.py
714
+ - src/mistralai/models/transcriptionstreamevents.py
715
+ - src/mistralai/models/transcriptionstreameventtypes.py
716
+ - src/mistralai/models/transcriptionstreamlanguage.py
717
+ - src/mistralai/models/transcriptionstreamsegmentdelta.py
718
+ - src/mistralai/models/transcriptionstreamtextdelta.py
684
719
- src/mistralai/models/unarchiveftmodelout.py
685
720
- src/mistralai/models/updateftmodelin.py
686
721
- src/mistralai/models/uploadfileout.py
@@ -696,6 +731,7 @@ generatedFiles:
696
731
- src/mistralai/py.typed
697
732
- src/mistralai/sdk.py
698
733
- src/mistralai/sdkconfiguration.py
734
+ - src/mistralai/transcriptions.py
699
735
- src/mistralai/types/__init__.py
700
736
- src/mistralai/types/basemodel.py
701
737
- src/mistralai/utils/__init__.py
@@ -751,7 +787,7 @@ examples:
751
787
application/json: {}
752
788
responses:
753
789
"200":
754
- application/json: {"id": "<id>", "object": "model", "created": 124166, "owned_by": "<value>", "root": "<value>", "archived": true, "capabilities": {"completion_chat": true, "completion_fim": false, "function_calling": false, "fine_tuning": false, "classification": false}, "max_context_length": 32768, "job": "c4f8ef9a-6612-4f49-88fa-a80eb8116e46", "model_type": "completion"}
790
+ application/json: {"id": "<id>", "object": "model", "created": 124166, "owned_by": "<value>", "workspace_id": "<id>", " root": "<value>", "root_version ": "<value>", "archived": true, "capabilities": {"completion_chat": true, "completion_fim": false, "function_calling": false, "fine_tuning": false, "classification": false}, "max_context_length": 32768, "job": "c4f8ef9a-6612-4f49-88fa-a80eb8116e46", "model_type": "completion"}
755
791
jobs_api_routes_fine_tuning_archive_fine_tuned_model:
756
792
speakeasy-default-jobs-api-routes-fine-tuning-archive-fine-tuned-model:
757
793
parameters:
@@ -1027,33 +1063,33 @@ examples:
1027
1063
jobs_api_routes_batch_create_batch_job:
1028
1064
speakeasy-default-jobs-api-routes-batch-create-batch-job:
1029
1065
requestBody:
1030
- application/json: {"input_files": ["fe3343a2-3b8d-404b-ba32-a78dede2614a"], "endpoint": "/v1/moderations", "model": "Altima", " timeout_hours": 24}
1066
+ application/json: {"input_files": ["fe3343a2-3b8d-404b-ba32-a78dede2614a"], "endpoint": "/v1/moderations", "timeout_hours": 24}
1031
1067
responses:
1032
1068
"200":
1033
- application/json: {"id": "<id>", "object": "batch", "input_files": ["7b2553d8-e17f-4df5-a862-a1678f6b5271", "8c618d9f-7d82-42ba-a284-d57d84f50a58", "c042f996-e842-441d-ae47-4e0850334e41"], "endpoint": "<value>", "model": "Taurus", " errors": [{"message": "<value>", "count": 1}], "status": "SUCCESS", "created_at": 395527, "total_requests": 166919, "completed_requests": 258552, "succeeded_requests": 480980, "failed_requests": 684176}
1069
+ application/json: {"id": "<id>", "object": "batch", "input_files": ["7b2553d8-e17f-4df5-a862-a1678f6b5271", "8c618d9f-7d82-42ba-a284-d57d84f50a58", "c042f996-e842-441d-ae47-4e0850334e41"], "endpoint": "<value>", "errors": [{"message": "<value>", "count": 1}], "status": "SUCCESS", "created_at": 395527, "total_requests": 166919, "completed_requests": 258552, "succeeded_requests": 480980, "failed_requests": 684176}
1034
1070
jobs_api_routes_batch_get_batch_job:
1035
1071
speakeasy-default-jobs-api-routes-batch-get-batch-job:
1036
1072
parameters:
1037
1073
path:
1038
1074
job_id: "4017dc9f-b629-42f4-9700-8c681b9e7f0f"
1039
1075
responses:
1040
1076
"200":
1041
- application/json: {"id": "<id>", "object": "batch", "input_files": ["11b83f16-f2f9-4de4-a81f-203fff419c99"], "endpoint": "<value>", "model": "Accord", " errors": [], "status": "TIMEOUT_EXCEEDED", "created_at": 900958, "total_requests": 458292, "completed_requests": 184893, "succeeded_requests": 104800, "failed_requests": 836210}
1077
+ application/json: {"id": "<id>", "object": "batch", "input_files": ["11b83f16-f2f9-4de4-a81f-203fff419c99"], "endpoint": "<value>", "errors": [], "status": "TIMEOUT_EXCEEDED", "created_at": 900958, "total_requests": 458292, "completed_requests": 184893, "succeeded_requests": 104800, "failed_requests": 836210}
1042
1078
jobs_api_routes_batch_cancel_batch_job:
1043
1079
speakeasy-default-jobs-api-routes-batch-cancel-batch-job:
1044
1080
parameters:
1045
1081
path:
1046
1082
job_id: "4fb29d1c-535b-4f0a-a1cb-2167f86da569"
1047
1083
responses:
1048
1084
"200":
1049
- application/json: {"id": "<id>", "object": "batch", "input_files": ["8fd9d88a-66be-43fd-a816-ba509ca3ca85"], "endpoint": "<value>", "model": "PT Cruiser", " errors": [], "status": "TIMEOUT_EXCEEDED", "created_at": 608251, "total_requests": 12693, "completed_requests": 203340, "succeeded_requests": 189291, "failed_requests": 969057}
1085
+ application/json: {"id": "<id>", "object": "batch", "input_files": ["8fd9d88a-66be-43fd-a816-ba509ca3ca85"], "endpoint": "<value>", "errors": [], "status": "TIMEOUT_EXCEEDED", "created_at": 608251, "total_requests": 12693, "completed_requests": 203340, "succeeded_requests": 189291, "failed_requests": 969057}
1050
1086
chat_completion_v1_chat_completions_post:
1051
1087
speakeasy-default-chat-completion-v1-chat-completions-post:
1052
1088
requestBody:
1053
1089
application/json: {"model": "mistral-small-latest", "stream": false, "messages": [{"content": "Who is the best French painter? Answer in one short sentence.", "role": "user"}]}
1054
1090
responses:
1055
1091
"200":
1056
- application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 16 , "completion_tokens": 34 , "total_tokens": 50 }, "created": 1702256327, "choices": []}
1092
+ application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 0 , "completion_tokens": 0 , "total_tokens": 0 }, "created": 1702256327, "choices": []}
1057
1093
"422":
1058
1094
application/json: {}
1059
1095
stream_chat:
@@ -1069,7 +1105,7 @@ examples:
1069
1105
application/json: {"model": "codestral-2405", "top_p": 1, "stream": false, "prompt": "def", "suffix": "return a+b"}
1070
1106
responses:
1071
1107
"200":
1072
- application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "codestral-latest", "usage": {"prompt_tokens": 16 , "completion_tokens": 34 , "total_tokens": 50 }, "created": 1702256327, "choices": []}
1108
+ application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "codestral-latest", "usage": {"prompt_tokens": 0 , "completion_tokens": 0 , "total_tokens": 0 }, "created": 1702256327, "choices": []}
1073
1109
"422":
1074
1110
application/json: {}
1075
1111
stream_fim:
@@ -1085,7 +1121,7 @@ examples:
1085
1121
application/json: {"stream": false, "messages": [{"content": "Who is the best French painter? Answer in one short sentence.", "role": "user"}], "agent_id": "<id>"}
1086
1122
responses:
1087
1123
"200":
1088
- application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 16 , "completion_tokens": 34 , "total_tokens": 50 }, "created": 1702256327, "choices": []}
1124
+ application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 0 , "completion_tokens": 0 , "total_tokens": 0 }, "created": 1702256327, "choices": []}
1089
1125
"422":
1090
1126
application/json: {}
1091
1127
stream_agents:
@@ -1101,7 +1137,7 @@ examples:
1101
1137
application/json: {"model": "mistral-embed", "input": ["Embed this sentence.", "As well as this one."]}
1102
1138
responses:
1103
1139
"200":
1104
- application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 16 , "completion_tokens": 34 , "total_tokens": 50 }, "data": [{"object": "embedding", "embedding": [0.1, 0.2, 0.3], "index": 0}, {"object": "embedding", "embedding": [0.4, 0.5, 0.6], "index": 1}]}
1140
+ application/json: {"id": "cmpl-e5cc70bb28c444948073e77776eb30ef", "object": "chat.completion", "model": "mistral-small-latest", "usage": {"prompt_tokens": 0 , "completion_tokens": 0 , "total_tokens": 0 }, "data": [{"object": "embedding", "embedding": [0.1, 0.2, 0.3], "index": 0}, {"object": "embedding", "embedding": [0.4, 0.5, 0.6], "index": 1}]}
1105
1141
"422":
1106
1142
application/json: {}
1107
1143
moderations_v1_moderations_post:
@@ -1342,5 +1378,16 @@ examples:
1342
1378
application/json: {"library_id": "7f9c6af4-e362-4cf1-9363-0409d51c2dfa", "org_id": "6b2cac3a-b29c-4d8f-bebb-0db06ec1bf97", "role": "<value>", "share_with_type": "<value>", "share_with_uuid": "618c78f1-41ca-45c3-8ef2-7d78898c7061"}
1343
1379
"422":
1344
1380
application/json: {}
1381
+ audio_api_v1_transcriptions_post:
1382
+ speakeasy-default-audio-api-v1-transcriptions-post:
1383
+ requestBody:
1384
+ multipart/form-data: {"model": "Model X", "stream": false}
1385
+ responses:
1386
+ "200":
1387
+ application/json: {"model": "Beetle", "text": "<value>", "usage": {"prompt_tokens": 0, "completion_tokens": 0, "total_tokens": 0}, "language": "<value>"}
1388
+ audio_api_v1_transcriptions_post_stream:
1389
+ speakeasy-default-audio-api-v1-transcriptions-post-stream:
1390
+ requestBody:
1391
+ multipart/form-data: {"model": "Camry", "stream": true}
1345
1392
examplesVersion: 1.0.2
1346
1393
generatedTests: {}
0 commit comments