diff --git a/.fern/metadata.json b/.fern/metadata.json index c9ecd919..c55971bb 100644 --- a/.fern/metadata.json +++ b/.fern/metadata.json @@ -1,7 +1,7 @@ { - "cliVersion": "3.77.1", + "cliVersion": "4.46.0", "generatorName": "fernapi/fern-python-sdk", - "generatorVersion": "4.57.2", + "generatorVersion": "4.62.0", "generatorConfig": { "client": { "class_name": "BaseClient", @@ -16,5 +16,6 @@ "skip_validation": true } }, - "sdkVersion": "6.0.1" + "originGitCommit": "879c76c78827f323e425c1640f76a6e50d6c68d3", + "sdkVersion": "6.0.2" } \ No newline at end of file diff --git a/.fernignore b/.fernignore index 1086ff6d..19521f0c 100644 --- a/.fernignore +++ b/.fernignore @@ -4,36 +4,22 @@ # This file is manually maintained and should not be regenerated src/deepgram/client.py -# WireMock mappings: removed duplicate empty-body /v1/listen stub that causes -# non-deterministic matching failures -wiremock/wiremock-mappings.json - -# Wire test with manual fix: transcribe_file() requires request=bytes parameter -tests/wire/test_listen_v1_media.py - # WebSocket socket clients: -# - Optional message parameter defaults for send_flush, send_close, send_clear, -# send_finalize, send_close_stream, send_keep_alive -# - construct_type instead of parse_obj_as (skip_validation for unknown WS messages) -# - except Exception (broad catch for custom transports) +# - construct_type keyword args fix (generator uses positional, function requires keyword-only) +# - except Exception broad catch (supports custom transports, generator narrows to WebSocketException) # - _sanitize_numeric_types in agent socket client (float→int for API) -src/deepgram/speak/v1/socket_client.py +# [temporarily frozen — generator bugs in construct_type call convention and exception handling] +src/deepgram/agent/v1/socket_client.py src/deepgram/listen/v1/socket_client.py src/deepgram/listen/v2/socket_client.py -src/deepgram/agent/v1/socket_client.py +src/deepgram/speak/v1/socket_client.py # Type files with manual int type corrections (Fern generates float for speaker/channel/num_words) +# [temporarily frozen — waiting on internal-api-specs#205] src/deepgram/types/listen_v1response_results_utterances_item.py src/deepgram/types/listen_v1response_results_utterances_item_words_item.py src/deepgram/types/listen_v1response_results_channels_item_alternatives_item_paragraphs_paragraphs_item.py -# Redact type with Union[str, Sequence[str]] support (Fern narrows to Union[Literal, Any]) -src/deepgram/types/listen_v1redact.py - -# Listen client files with Union[str, Sequence[str]] array param support -src/deepgram/listen/v1/client.py -src/deepgram/listen/v2/client.py - # Hand-written custom tests tests/custom/test_text_builder.py tests/custom/test_transport.py @@ -75,4 +61,4 @@ AGENTS.md # Folders to ignore .github docs -examples \ No newline at end of file +examples diff --git a/poetry.lock b/poetry.lock index 4d721e91..a981f03d 100644 --- a/poetry.lock +++ b/poetry.lock @@ -38,135 +38,151 @@ trio = ["trio (>=0.26.1)"] [[package]] name = "certifi" -version = "2026.1.4" +version = "2026.2.25" description = "Python package for providing Mozilla's CA Bundle." optional = false python-versions = ">=3.7" files = [ - {file = "certifi-2026.1.4-py3-none-any.whl", hash = "sha256:9943707519e4add1115f44c2bc244f782c0249876bf51b6599fee1ffbedd685c"}, - {file = "certifi-2026.1.4.tar.gz", hash = "sha256:ac726dd470482006e014ad384921ed6438c457018f4b3d204aea4281258b2120"}, + {file = "certifi-2026.2.25-py3-none-any.whl", hash = "sha256:027692e4402ad994f1c42e52a4997a9763c646b73e4096e4d5d6db8af1d6f0fa"}, + {file = "certifi-2026.2.25.tar.gz", hash = "sha256:e887ab5cee78ea814d3472169153c2d12cd43b14bd03329a39a9c6e2e80bfba7"}, ] [[package]] name = "charset-normalizer" -version = "3.4.4" +version = "3.4.6" description = "The Real First Universal Charset Detector. Open, modern and actively maintained alternative to Chardet." optional = false python-versions = ">=3.7" files = [ - {file = "charset_normalizer-3.4.4-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:e824f1492727fa856dd6eda4f7cee25f8518a12f3c4a56a74e8095695089cf6d"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:4bd5d4137d500351a30687c2d3971758aac9a19208fc110ccb9d7188fbe709e8"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:027f6de494925c0ab2a55eab46ae5129951638a49a34d87f4c3eda90f696b4ad"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f820802628d2694cb7e56db99213f930856014862f3fd943d290ea8438d07ca8"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:798d75d81754988d2565bff1b97ba5a44411867c0cf32b77a7e8f8d84796b10d"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9d1bb833febdff5c8927f922386db610b49db6e0d4f4ee29601d71e7c2694313"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:9cd98cdc06614a2f768d2b7286d66805f94c48cde050acdbbb7db2600ab3197e"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:077fbb858e903c73f6c9db43374fd213b0b6a778106bc7032446a8e8b5b38b93"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:244bfb999c71b35de57821b8ea746b24e863398194a4014e4c76adc2bbdfeff0"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:64b55f9dce520635f018f907ff1b0df1fdc31f2795a922fb49dd14fbcdf48c84"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:faa3a41b2b66b6e50f84ae4a68c64fcd0c44355741c6374813a800cd6695db9e"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:6515f3182dbe4ea06ced2d9e8666d97b46ef4c75e326b79bb624110f122551db"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:cc00f04ed596e9dc0da42ed17ac5e596c6ccba999ba6bd92b0e0aef2f170f2d6"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-win32.whl", hash = "sha256:f34be2938726fc13801220747472850852fe6b1ea75869a048d6f896838c896f"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-win_amd64.whl", hash = "sha256:a61900df84c667873b292c3de315a786dd8dac506704dea57bc957bd31e22c7d"}, - {file = "charset_normalizer-3.4.4-cp310-cp310-win_arm64.whl", hash = "sha256:cead0978fc57397645f12578bfd2d5ea9138ea0fac82b2f63f7f7c6877986a69"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:6e1fcf0720908f200cd21aa4e6750a48ff6ce4afe7ff5a79a90d5ed8a08296f8"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:5f819d5fe9234f9f82d75bdfa9aef3a3d72c4d24a6e57aeaebba32a704553aa0"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:a59cb51917aa591b1c4e6a43c132f0cdc3c76dbad6155df4e28ee626cc77a0a3"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:8ef3c867360f88ac904fd3f5e1f902f13307af9052646963ee08ff4f131adafc"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d9e45d7faa48ee908174d8fe84854479ef838fc6a705c9315372eacbc2f02897"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:840c25fb618a231545cbab0564a799f101b63b9901f2569faecd6b222ac72381"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:ca5862d5b3928c4940729dacc329aa9102900382fea192fc5e52eb69d6093815"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:d9c7f57c3d666a53421049053eaacdd14bbd0a528e2186fcb2e672effd053bb0"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:277e970e750505ed74c832b4bf75dac7476262ee2a013f5574dd49075879e161"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:31fd66405eaf47bb62e8cd575dc621c56c668f27d46a61d975a249930dd5e2a4"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:0d3d8f15c07f86e9ff82319b3d9ef6f4bf907608f53fe9d92b28ea9ae3d1fd89"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:9f7fcd74d410a36883701fafa2482a6af2ff5ba96b9a620e9e0721e28ead5569"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:ebf3e58c7ec8a8bed6d66a75d7fb37b55e5015b03ceae72a8e7c74495551e224"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-win32.whl", hash = "sha256:eecbc200c7fd5ddb9a7f16c7decb07b566c29fa2161a16cf67b8d068bd21690a"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-win_amd64.whl", hash = "sha256:5ae497466c7901d54b639cf42d5b8c1b6a4fead55215500d2f486d34db48d016"}, - {file = "charset_normalizer-3.4.4-cp311-cp311-win_arm64.whl", hash = "sha256:65e2befcd84bc6f37095f5961e68a6f077bf44946771354a28ad434c2cce0ae1"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:0a98e6759f854bd25a58a73fa88833fba3b7c491169f86ce1180c948ab3fd394"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:b5b290ccc2a263e8d185130284f8501e3e36c5e02750fc6b6bdeb2e9e96f1e25"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:74bb723680f9f7a6234dcf67aea57e708ec1fbdf5699fb91dfd6f511b0a320ef"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f1e34719c6ed0b92f418c7c780480b26b5d9c50349e9a9af7d76bf757530350d"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:2437418e20515acec67d86e12bf70056a33abdacb5cb1655042f6538d6b085a8"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:11d694519d7f29d6cd09f6ac70028dba10f92f6cdd059096db198c283794ac86"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:ac1c4a689edcc530fc9d9aa11f5774b9e2f33f9a0c6a57864e90908f5208d30a"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:21d142cc6c0ec30d2efee5068ca36c128a30b0f2c53c1c07bd78cb6bc1d3be5f"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:5dbe56a36425d26d6cfb40ce79c314a2e4dd6211d51d6d2191c00bed34f354cc"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:5bfbb1b9acf3334612667b61bd3002196fe2a1eb4dd74d247e0f2a4d50ec9bbf"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:d055ec1e26e441f6187acf818b73564e6e6282709e9bcb5b63f5b23068356a15"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:af2d8c67d8e573d6de5bc30cdb27e9b95e49115cd9baad5ddbd1a6207aaa82a9"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:780236ac706e66881f3b7f2f32dfe90507a09e67d1d454c762cf642e6e1586e0"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-win32.whl", hash = "sha256:5833d2c39d8896e4e19b689ffc198f08ea58116bee26dea51e362ecc7cd3ed26"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-win_amd64.whl", hash = "sha256:a79cfe37875f822425b89a82333404539ae63dbdddf97f84dcbc3d339aae9525"}, - {file = "charset_normalizer-3.4.4-cp312-cp312-win_arm64.whl", hash = "sha256:376bec83a63b8021bb5c8ea75e21c4ccb86e7e45ca4eb81146091b56599b80c3"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:e1f185f86a6f3403aa2420e815904c67b2f9ebc443f045edd0de921108345794"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:6b39f987ae8ccdf0d2642338faf2abb1862340facc796048b604ef14919e55ed"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:3162d5d8ce1bb98dd51af660f2121c55d0fa541b46dff7bb9b9f86ea1d87de72"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:81d5eb2a312700f4ecaa977a8235b634ce853200e828fbadf3a9c50bab278328"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5bd2293095d766545ec1a8f612559f6b40abc0eb18bb2f5d1171872d34036ede"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a8a8b89589086a25749f471e6a900d3f662d1d3b6e2e59dcecf787b1cc3a1894"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:bc7637e2f80d8530ee4a78e878bce464f70087ce73cf7c1caf142416923b98f1"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:f8bf04158c6b607d747e93949aa60618b61312fe647a6369f88ce2ff16043490"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:554af85e960429cf30784dd47447d5125aaa3b99a6f0683589dbd27e2f45da44"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:74018750915ee7ad843a774364e13a3db91682f26142baddf775342c3f5b1133"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:c0463276121fdee9c49b98908b3a89c39be45d86d1dbaa22957e38f6321d4ce3"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:362d61fd13843997c1c446760ef36f240cf81d3ebf74ac62652aebaf7838561e"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:9a26f18905b8dd5d685d6d07b0cdf98a79f3c7a918906af7cc143ea2e164c8bc"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-win32.whl", hash = "sha256:9b35f4c90079ff2e2edc5b26c0c77925e5d2d255c42c74fdb70fb49b172726ac"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-win_amd64.whl", hash = "sha256:b435cba5f4f750aa6c0a0d92c541fb79f69a387c91e61f1795227e4ed9cece14"}, - {file = "charset_normalizer-3.4.4-cp313-cp313-win_arm64.whl", hash = "sha256:542d2cee80be6f80247095cc36c418f7bddd14f4a6de45af91dfad36d817bba2"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-macosx_10_13_universal2.whl", hash = "sha256:da3326d9e65ef63a817ecbcc0df6e94463713b754fe293eaa03da99befb9a5bd"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:8af65f14dc14a79b924524b1e7fffe304517b2bff5a58bf64f30b98bbc5079eb"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:74664978bb272435107de04e36db5a9735e78232b85b77d45cfb38f758efd33e"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:752944c7ffbfdd10c074dc58ec2d5a8a4cd9493b314d367c14d24c17684ddd14"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d1f13550535ad8cff21b8d757a3257963e951d96e20ec82ab44bc64aeb62a191"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:ecaae4149d99b1c9e7b88bb03e3221956f68fd6d50be2ef061b2381b61d20838"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:cb6254dc36b47a990e59e1068afacdcd02958bdcce30bb50cc1700a8b9d624a6"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:c8ae8a0f02f57a6e61203a31428fa1d677cbe50c93622b4149d5c0f319c1d19e"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:47cc91b2f4dd2833fddaedd2893006b0106129d4b94fdb6af1f4ce5a9965577c"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:82004af6c302b5d3ab2cfc4cc5f29db16123b1a8417f2e25f9066f91d4411090"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:2b7d8f6c26245217bd2ad053761201e9f9680f8ce52f0fcd8d0755aeae5b2152"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:799a7a5e4fb2d5898c60b640fd4981d6a25f1c11790935a44ce38c54e985f828"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:99ae2cffebb06e6c22bdc25801d7b30f503cc87dbd283479e7b606f70aff57ec"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-win32.whl", hash = "sha256:f9d332f8c2a2fcbffe1378594431458ddbef721c1769d78e2cbc06280d8155f9"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-win_amd64.whl", hash = "sha256:8a6562c3700cce886c5be75ade4a5db4214fda19fede41d9792d100288d8f94c"}, - {file = "charset_normalizer-3.4.4-cp314-cp314-win_arm64.whl", hash = "sha256:de00632ca48df9daf77a2c65a484531649261ec9f25489917f09e455cb09ddb2"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:ce8a0633f41a967713a59c4139d29110c07e826d131a316b50ce11b1d79b4f84"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:eaabd426fe94daf8fd157c32e571c85cb12e66692f15516a83a03264b08d06c3"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:c4ef880e27901b6cc782f1b95f82da9313c0eb95c3af699103088fa0ac3ce9ac"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:2aaba3b0819274cc41757a1da876f810a3e4d7b6eb25699253a4effef9e8e4af"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:778d2e08eda00f4256d7f672ca9fef386071c9202f5e4607920b86d7803387f2"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f155a433c2ec037d4e8df17d18922c3a0d9b3232a396690f17175d2946f0218d"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:a8bf8d0f749c5757af2142fe7903a9df1d2e8aa3841559b2bad34b08d0e2bcf3"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:194f08cbb32dc406d6e1aea671a68be0823673db2832b38405deba2fb0d88f63"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_armv7l.whl", hash = "sha256:6aee717dcfead04c6eb1ce3bd29ac1e22663cdea57f943c87d1eab9a025438d7"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_ppc64le.whl", hash = "sha256:cd4b7ca9984e5e7985c12bc60a6f173f3c958eae74f3ef6624bb6b26e2abbae4"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_riscv64.whl", hash = "sha256:b7cf1017d601aa35e6bb650b6ad28652c9cd78ee6caff19f3c28d03e1c80acbf"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_s390x.whl", hash = "sha256:e912091979546adf63357d7e2ccff9b44f026c075aeaf25a52d0e95ad2281074"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:5cb4d72eea50c8868f5288b7f7f33ed276118325c1dfd3957089f6b519e1382a"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-win32.whl", hash = "sha256:837c2ce8c5a65a2035be9b3569c684358dfbf109fd3b6969630a87535495ceaa"}, - {file = "charset_normalizer-3.4.4-cp38-cp38-win_amd64.whl", hash = "sha256:44c2a8734b333e0578090c4cd6b16f275e07aa6614ca8715e6c038e865e70576"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:a9768c477b9d7bd54bc0c86dbaebdec6f03306675526c9927c0e8a04e8f94af9"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1bee1e43c28aa63cb16e5c14e582580546b08e535299b8b6158a7c9c768a1f3d"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux2014_armv7l.manylinux_2_17_armv7l.manylinux_2_31_armv7l.whl", hash = "sha256:fd44c878ea55ba351104cb93cc85e74916eb8fa440ca7903e57575e97394f608"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:0f04b14ffe5fdc8c4933862d8306109a2c51e0704acfa35d51598eb45a1e89fc"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:cd09d08005f958f370f539f186d10aec3377d55b9eeb0d796025d4886119d76e"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:4fe7859a4e3e8457458e2ff592f15ccb02f3da787fcd31e0183879c3ad4692a1"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:fa09f53c465e532f4d3db095e0c55b615f010ad81803d383195b6b5ca6cbf5f3"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:7fa17817dc5625de8a027cb8b26d9fefa3ea28c8253929b8d6649e705d2835b6"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:5947809c8a2417be3267efc979c47d76a079758166f7d43ef5ae8e9f92751f88"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:4902828217069c3c5c71094537a8e623f5d097858ac6ca8252f7b4d10b7560f1"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_riscv64.whl", hash = "sha256:7c308f7e26e4363d79df40ca5b2be1c6ba9f02bdbccfed5abddb7859a6ce72cf"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:2c9d3c380143a1fedbff95a312aa798578371eb29da42106a29019368a475318"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:cb01158d8b88ee68f15949894ccc6712278243d95f344770fa7593fa2d94410c"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-win32.whl", hash = "sha256:2677acec1a2f8ef614c6888b5b4ae4060cc184174a938ed4e8ef690e15d3e505"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-win_amd64.whl", hash = "sha256:f8e160feb2aed042cd657a72acc0b481212ed28b1b9a95c0cee1621b524e1966"}, - {file = "charset_normalizer-3.4.4-cp39-cp39-win_arm64.whl", hash = "sha256:b5d84d37db046c5ca74ee7bb47dd6cbc13f80665fdde3e8040bdd3fb015ecb50"}, - {file = "charset_normalizer-3.4.4-py3-none-any.whl", hash = "sha256:7a32c560861a02ff789ad905a2fe94e3f840803362c84fecf1851cb4cf3dc37f"}, - {file = "charset_normalizer-3.4.4.tar.gz", hash = "sha256:94537985111c35f28720e43603b8e7b43a6ecfb2ce1d3058bbe955b73404e21a"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-macosx_10_9_universal2.whl", hash = "sha256:2e1d8ca8611099001949d1cdfaefc510cf0f212484fe7c565f735b68c78c3c95"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:e25369dc110d58ddf29b949377a93e0716d72a24f62bad72b2b39f155949c1fd"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:259695e2ccc253feb2a016303543d691825e920917e31f894ca1a687982b1de4"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:dda86aba335c902b6149a02a55b38e96287157e609200811837678214ba2b1db"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:51fb3c322c81d20567019778cb5a4a6f2dc1c200b886bc0d636238e364848c89"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux_2_31_armv7l.whl", hash = "sha256:4482481cb0572180b6fd976a4d5c72a30263e98564da68b86ec91f0fe35e8565"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:39f5068d35621da2881271e5c3205125cc456f54e9030d3f723288c873a71bf9"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_aarch64.whl", hash = "sha256:8bea55c4eef25b0b19a0337dc4e3f9a15b00d569c77211fa8cde38684f234fb7"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_armv7l.whl", hash = "sha256:f0cdaecd4c953bfae0b6bb64910aaaca5a424ad9c72d85cb88417bb9814f7550"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_ppc64le.whl", hash = "sha256:150b8ce8e830eb7ccb029ec9ca36022f756986aaaa7956aad6d9ec90089338c0"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_riscv64.whl", hash = "sha256:e68c14b04827dd76dcbd1aeea9e604e3e4b78322d8faf2f8132c7138efa340a8"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_s390x.whl", hash = "sha256:3778fd7d7cd04ae8f54651f4a7a0bd6e39a0cf20f801720a4c21d80e9b7ad6b0"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-musllinux_1_2_x86_64.whl", hash = "sha256:dad6e0f2e481fffdcf776d10ebee25e0ef89f16d691f1e5dee4b586375fdc64b"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-win32.whl", hash = "sha256:74a2e659c7ecbc73562e2a15e05039f1e22c75b7c7618b4b574a3ea9118d1557"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-win_amd64.whl", hash = "sha256:aa9cccf4a44b9b62d8ba8b4dd06c649ba683e4bf04eea606d2e94cfc2d6ff4d6"}, + {file = "charset_normalizer-3.4.6-cp310-cp310-win_arm64.whl", hash = "sha256:e985a16ff513596f217cee86c21371b8cd011c0f6f056d0920aa2d926c544058"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-macosx_10_9_universal2.whl", hash = "sha256:82060f995ab5003a2d6e0f4ad29065b7672b6593c8c63559beefe5b443242c3e"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:60c74963d8350241a79cb8feea80e54d518f72c26db618862a8f53e5023deaf9"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f6e4333fb15c83f7d1482a76d45a0818897b3d33f00efd215528ff7c51b8e35d"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:bc72863f4d9aba2e8fd9085e63548a324ba706d2ea2c83b260da08a59b9482de"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:9cc4fc6c196d6a8b76629a70ddfcd4635a6898756e2d9cac5565cf0654605d73"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux_2_31_armv7l.whl", hash = "sha256:0c173ce3a681f309f31b87125fecec7a5d1347261ea11ebbb856fa6006b23c8c"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:c907cdc8109f6c619e6254212e794d6548373cc40e1ec75e6e3823d9135d29cc"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:404a1e552cf5b675a87f0651f8b79f5f1e6fd100ee88dc612f89aa16abd4486f"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_armv7l.whl", hash = "sha256:e3c701e954abf6fc03a49f7c579cc80c2c6cc52525340ca3186c41d3f33482ef"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_ppc64le.whl", hash = "sha256:7a6967aaf043bceabab5412ed6bd6bd26603dae84d5cb75bf8d9a74a4959d398"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_riscv64.whl", hash = "sha256:5feb91325bbceade6afab43eb3b508c63ee53579fe896c77137ded51c6b6958e"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_s390x.whl", hash = "sha256:f820f24b09e3e779fe84c3c456cb4108a7aa639b0d1f02c28046e11bfcd088ed"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:b35b200d6a71b9839a46b9b7fff66b6638bb52fc9658aa58796b0326595d3021"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-win32.whl", hash = "sha256:9ca4c0b502ab399ef89248a2c84c54954f77a070f28e546a85e91da627d1301e"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-win_amd64.whl", hash = "sha256:a9e68c9d88823b274cf1e72f28cb5dc89c990edf430b0bfd3e2fb0785bfeabf4"}, + {file = "charset_normalizer-3.4.6-cp311-cp311-win_arm64.whl", hash = "sha256:97d0235baafca5f2b09cf332cc275f021e694e8362c6bb9c96fc9a0eb74fc316"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-macosx_10_13_universal2.whl", hash = "sha256:2ef7fedc7a6ecbe99969cd09632516738a97eeb8bd7258bf8a0f23114c057dab"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:a4ea868bc28109052790eb2b52a9ab33f3aa7adc02f96673526ff47419490e21"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:836ab36280f21fc1a03c99cd05c6b7af70d2697e374c7af0b61ed271401a72a2"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:f1ce721c8a7dfec21fcbdfe04e8f68174183cf4e8188e0645e92aa23985c57ff"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0e28d62a8fc7a1fa411c43bd65e346f3bce9716dc51b897fbe930c5987b402d5"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux_2_31_armv7l.whl", hash = "sha256:530d548084c4a9f7a16ed4a294d459b4f229db50df689bfe92027452452943a0"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:30f445ae60aad5e1f8bdbb3108e39f6fbc09f4ea16c815c66578878325f8f15a"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:ac2393c73378fea4e52aa56285a3d64be50f1a12395afef9cce47772f60334c2"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_armv7l.whl", hash = "sha256:90ca27cd8da8118b18a52d5f547859cc1f8354a00cd1e8e5120df3e30d6279e5"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_ppc64le.whl", hash = "sha256:8e5a94886bedca0f9b78fecd6afb6629142fd2605aa70a125d49f4edc6037ee6"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_riscv64.whl", hash = "sha256:695f5c2823691a25f17bc5d5ffe79fa90972cc34b002ac6c843bb8a1720e950d"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_s390x.whl", hash = "sha256:231d4da14bcd9301310faf492051bee27df11f2bc7549bc0bb41fef11b82daa2"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:a056d1ad2633548ca18ffa2f85c202cfb48b68615129143915b8dc72a806a923"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-win32.whl", hash = "sha256:c2274ca724536f173122f36c98ce188fd24ce3dad886ec2b7af859518ce008a4"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-win_amd64.whl", hash = "sha256:c8ae56368f8cc97c7e40a7ee18e1cedaf8e780cd8bc5ed5ac8b81f238614facb"}, + {file = "charset_normalizer-3.4.6-cp312-cp312-win_arm64.whl", hash = "sha256:899d28f422116b08be5118ef350c292b36fc15ec2daeb9ea987c89281c7bb5c4"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-macosx_10_13_universal2.whl", hash = "sha256:11afb56037cbc4b1555a34dd69151e8e069bee82e613a73bef6e714ce733585f"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:423fb7e748a08f854a08a222b983f4df1912b1daedce51a72bd24fe8f26a1843"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:d73beaac5e90173ac3deb9928a74763a6d230f494e4bfb422c217a0ad8e629bf"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:d60377dce4511655582e300dc1e5a5f24ba0cb229005a1d5c8d0cb72bb758ab8"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:530e8cebeea0d76bdcf93357aa5e41336f48c3dc709ac52da2bb167c5b8271d9"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux_2_31_armv7l.whl", hash = "sha256:a26611d9987b230566f24a0a125f17fe0de6a6aff9f25c9f564aaa2721a5fb88"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:34315ff4fc374b285ad7f4a0bf7dcbfe769e1b104230d40f49f700d4ab6bbd84"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:5f8ddd609f9e1af8c7bd6e2aca279c931aefecd148a14402d4e368f3171769fd"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_armv7l.whl", hash = "sha256:80d0a5615143c0b3225e5e3ef22c8d5d51f3f72ce0ea6fb84c943546c7b25b6c"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_ppc64le.whl", hash = "sha256:92734d4d8d187a354a556626c221cd1a892a4e0802ccb2af432a1d85ec012194"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_riscv64.whl", hash = "sha256:613f19aa6e082cf96e17e3ffd89383343d0d589abda756b7764cf78361fd41dc"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_s390x.whl", hash = "sha256:2b1a63e8224e401cafe7739f77efd3f9e7f5f2026bda4aead8e59afab537784f"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:6cceb5473417d28edd20c6c984ab6fee6c6267d38d906823ebfe20b03d607dc2"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-win32.whl", hash = "sha256:d7de2637729c67d67cf87614b566626057e95c303bc0a55ffe391f5205e7003d"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-win_amd64.whl", hash = "sha256:572d7c822caf521f0525ba1bce1a622a0b85cf47ffbdae6c9c19e3b5ac3c4389"}, + {file = "charset_normalizer-3.4.6-cp313-cp313-win_arm64.whl", hash = "sha256:a4474d924a47185a06411e0064b803c68be044be2d60e50e8bddcc2649957c1f"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-macosx_10_15_universal2.whl", hash = "sha256:9cc6e6d9e571d2f863fa77700701dae73ed5f78881efc8b3f9a4398772ff53e8"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ef5960d965e67165d75b7c7ffc60a83ec5abfc5c11b764ec13ea54fbef8b4421"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:b3694e3f87f8ac7ce279d4355645b3c878d24d1424581b46282f24b92f5a4ae2"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:5d11595abf8dd942a77883a39d81433739b287b6aa71620f15164f8096221b30"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:7bda6eebafd42133efdca535b04ccb338ab29467b3f7bf79569883676fc628db"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux_2_31_armv7l.whl", hash = "sha256:bbc8c8650c6e51041ad1be191742b8b421d05bbd3410f43fa2a00c8db87678e8"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:22c6f0c2fbc31e76c3b8a86fba1a56eda6166e238c29cdd3d14befdb4a4e4815"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:7edbed096e4a4798710ed6bc75dcaa2a21b68b6c356553ac4823c3658d53743a"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_armv7l.whl", hash = "sha256:7f9019c9cb613f084481bd6a100b12e1547cf2efe362d873c2e31e4035a6fa43"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_ppc64le.whl", hash = "sha256:58c948d0d086229efc484fe2f30c2d382c86720f55cd9bc33591774348ad44e0"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_riscv64.whl", hash = "sha256:419a9d91bd238052642a51938af8ac05da5b3343becde08d5cdeab9046df9ee1"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_s390x.whl", hash = "sha256:5273b9f0b5835ff0350c0828faea623c68bfa65b792720c453e22b25cc72930f"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:0e901eb1049fdb80f5bd11ed5ea1e498ec423102f7a9b9e4645d5b8204ff2815"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-win32.whl", hash = "sha256:b4ff1d35e8c5bd078be89349b6f3a845128e685e751b6ea1169cf2160b344c4d"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-win_amd64.whl", hash = "sha256:74119174722c4349af9708993118581686f343adc1c8c9c007d59be90d077f3f"}, + {file = "charset_normalizer-3.4.6-cp314-cp314-win_arm64.whl", hash = "sha256:e5bcc1a1ae744e0bb59641171ae53743760130600da8db48cbb6e4918e186e4e"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-macosx_10_15_universal2.whl", hash = "sha256:ad8faf8df23f0378c6d527d8b0b15ea4a2e23c89376877c598c4870d1b2c7866"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f5ea69428fa1b49573eef0cc44a1d43bebd45ad0c611eb7d7eac760c7ae771bc"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:06a7e86163334edfc5d20fe104db92fcd666e5a5df0977cb5680a506fe26cc8e"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e1f6e2f00a6b8edb562826e4632e26d063ac10307e80f7461f7de3ad8ef3f077"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:95b52c68d64c1878818687a473a10547b3292e82b6f6fe483808fb1468e2f52f"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux_2_31_armv7l.whl", hash = "sha256:7504e9b7dc05f99a9bbb4525c67a2c155073b44d720470a148b34166a69c054e"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:172985e4ff804a7ad08eebec0a1640ece87ba5041d565fff23c8f99c1f389484"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:4be9f4830ba8741527693848403e2c457c16e499100963ec711b1c6f2049b7c7"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_armv7l.whl", hash = "sha256:79090741d842f564b1b2827c0b82d846405b744d31e84f18d7a7b41c20e473ff"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_ppc64le.whl", hash = "sha256:87725cfb1a4f1f8c2fc9890ae2f42094120f4b44db9360be5d99a4c6b0e03a9e"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_riscv64.whl", hash = "sha256:fcce033e4021347d80ed9c66dcf1e7b1546319834b74445f561d2e2221de5659"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_s390x.whl", hash = "sha256:ca0276464d148c72defa8bb4390cce01b4a0e425f3b50d1435aa6d7a18107602"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:197c1a244a274bb016dd8b79204850144ef77fe81c5b797dc389327adb552407"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-win32.whl", hash = "sha256:2a24157fa36980478dd1770b585c0f30d19e18f4fb0c47c13aa568f871718579"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-win_amd64.whl", hash = "sha256:cd5e2801c89992ed8c0a3f0293ae83c159a60d9a5d685005383ef4caca77f2c4"}, + {file = "charset_normalizer-3.4.6-cp314-cp314t-win_arm64.whl", hash = "sha256:47955475ac79cc504ef2704b192364e51d0d473ad452caedd0002605f780101c"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-macosx_10_9_universal2.whl", hash = "sha256:659a1e1b500fac8f2779dd9e1570464e012f43e580371470b45277a27baa7532"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:f61aa92e4aad0be58eb6eb4e0c21acf32cf8065f4b2cae5665da756c4ceef982"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:f50498891691e0864dc3da965f340fada0771f6142a378083dc4608f4ea513e2"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:bf625105bb9eef28a56a943fec8c8a98aeb80e7d7db99bd3c388137e6eb2d237"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:2bd9d128ef93637a5d7a6af25363cf5dec3fa21cf80e68055aad627f280e8afa"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux_2_31_armv7l.whl", hash = "sha256:d08ec48f0a1c48d75d0356cea971921848fb620fdeba805b28f937e90691209f"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:1ed80ff870ca6de33f4d953fda4d55654b9a2b340ff39ab32fa3adbcd718f264"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_aarch64.whl", hash = "sha256:f98059e4fcd3e3e4e2d632b7cf81c2faae96c43c60b569e9c621468082f1d104"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_armv7l.whl", hash = "sha256:ab30e5e3e706e3063bc6de96b118688cb10396b70bb9864a430f67df98c61ecc"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_ppc64le.whl", hash = "sha256:d5f5d1e9def3405f60e3ca8232d56f35c98fb7bf581efcc60051ebf53cb8b611"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_riscv64.whl", hash = "sha256:461598cd852bfa5a61b09cae2b1c02e2efcd166ee5516e243d540ac24bfa68a7"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_s390x.whl", hash = "sha256:71be7e0e01753a89cf024abf7ecb6bca2c81738ead80d43004d9b5e3f1244e64"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-musllinux_1_2_x86_64.whl", hash = "sha256:df01808ee470038c3f8dc4f48620df7225c49c2d6639e38f96e6d6ac6e6f7b0e"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-win32.whl", hash = "sha256:69dd852c2f0ad631b8b60cfbe25a28c0058a894de5abb566619c205ce0550eae"}, + {file = "charset_normalizer-3.4.6-cp38-cp38-win_amd64.whl", hash = "sha256:517ad0e93394ac532745129ceabdf2696b609ec9f87863d337140317ebce1c14"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-macosx_10_9_universal2.whl", hash = "sha256:31215157227939b4fb3d740cd23fe27be0439afef67b785a1eb78a3ae69cba9e"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ecbbd45615a6885fe3240eb9db73b9e62518b611850fdf8ab08bd56de7ad2b17"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux2014_ppc64le.manylinux_2_17_ppc64le.manylinux_2_28_ppc64le.whl", hash = "sha256:c45a03a4c69820a399f1dda9e1d8fbf3562eda46e7720458180302021b08f778"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux2014_s390x.manylinux_2_17_s390x.manylinux_2_28_s390x.whl", hash = "sha256:e8aeb10fcbe92767f0fa69ad5a72deca50d0dca07fbde97848997d778a50c9fe"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:54fae94be3d75f3e573c9a1b5402dc593de19377013c9a0e4285e3d402dd3a2a"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux_2_31_armv7l.whl", hash = "sha256:2f7fdd9b6e6c529d6a2501a2d36b240109e78a8ceaef5687cfcfa2bbe671d297"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-manylinux_2_31_riscv64.manylinux_2_39_riscv64.whl", hash = "sha256:4d1d02209e06550bdaef34af58e041ad71b88e624f5d825519da3a3308e22687"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_aarch64.whl", hash = "sha256:8bc5f0687d796c05b1e28ab0d38a50e6309906ee09375dd3aff6a9c09dd6e8f4"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_armv7l.whl", hash = "sha256:ee4ec14bc1680d6b0afab9aea2ef27e26d2024f18b24a2d7155a52b60da7e833"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_ppc64le.whl", hash = "sha256:d1a2ee9c1499fc8f86f4521f27a973c914b211ffa87322f4ee33bb35392da2c5"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_riscv64.whl", hash = "sha256:48696db7f18afb80a068821504296eb0787d9ce239b91ca15059d1d3eaacf13b"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_s390x.whl", hash = "sha256:4f41da960b196ea355357285ad1316a00099f22d0929fe168343b99b254729c9"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-musllinux_1_2_x86_64.whl", hash = "sha256:802168e03fba8bbc5ce0d866d589e4b1ca751d06edee69f7f3a19c5a9fe6b597"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-win32.whl", hash = "sha256:8761ac29b6c81574724322a554605608a9960769ea83d2c73e396f3df896ad54"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-win_amd64.whl", hash = "sha256:1cf0a70018692f85172348fe06d3a4b63f94ecb055e13a00c644d368eb82e5b8"}, + {file = "charset_normalizer-3.4.6-cp39-cp39-win_arm64.whl", hash = "sha256:3516bbb8d42169de9e61b8520cbeeeb716f12f4ecfe3fd30a9919aa16c806ca8"}, + {file = "charset_normalizer-3.4.6-py3-none-any.whl", hash = "sha256:947cf925bc916d90adba35a64c82aace04fa39b46b52d4630ece166655905a69"}, + {file = "charset_normalizer-3.4.6.tar.gz", hash = "sha256:1ae6b62897110aa7c79ea2f5dd38d1abca6db663687c0b1ad9aed6f6bae3d9d6"}, ] [[package]] @@ -660,58 +676,58 @@ files = [ [[package]] name = "tomli" -version = "2.4.0" +version = "2.4.1" description = "A lil' TOML parser" optional = false python-versions = ">=3.8" files = [ - {file = "tomli-2.4.0-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:b5ef256a3fd497d4973c11bf142e9ed78b150d36f5773f1ca6088c230ffc5867"}, - {file = "tomli-2.4.0-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:5572e41282d5268eb09a697c89a7bee84fae66511f87533a6f88bd2f7b652da9"}, - {file = "tomli-2.4.0-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:551e321c6ba03b55676970b47cb1b73f14a0a4dce6a3e1a9458fd6d921d72e95"}, - {file = "tomli-2.4.0-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5e3f639a7a8f10069d0e15408c0b96a2a828cfdec6fca05296ebcdcc28ca7c76"}, - {file = "tomli-2.4.0-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:1b168f2731796b045128c45982d3a4874057626da0e2ef1fdd722848b741361d"}, - {file = "tomli-2.4.0-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:133e93646ec4300d651839d382d63edff11d8978be23da4cc106f5a18b7d0576"}, - {file = "tomli-2.4.0-cp311-cp311-win32.whl", hash = "sha256:b6c78bdf37764092d369722d9946cb65b8767bfa4110f902a1b2542d8d173c8a"}, - {file = "tomli-2.4.0-cp311-cp311-win_amd64.whl", hash = "sha256:d3d1654e11d724760cdb37a3d7691f0be9db5fbdaef59c9f532aabf87006dbaa"}, - {file = "tomli-2.4.0-cp311-cp311-win_arm64.whl", hash = "sha256:cae9c19ed12d4e8f3ebf46d1a75090e4c0dc16271c5bce1c833ac168f08fb614"}, - {file = "tomli-2.4.0-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:920b1de295e72887bafa3ad9f7a792f811847d57ea6b1215154030cf131f16b1"}, - {file = "tomli-2.4.0-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:7d6d9a4aee98fac3eab4952ad1d73aee87359452d1c086b5ceb43ed02ddb16b8"}, - {file = "tomli-2.4.0-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:36b9d05b51e65b254ea6c2585b59d2c4cb91c8a3d91d0ed0f17591a29aaea54a"}, - {file = "tomli-2.4.0-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1c8a885b370751837c029ef9bc014f27d80840e48bac415f3412e6593bbc18c1"}, - {file = "tomli-2.4.0-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:8768715ffc41f0008abe25d808c20c3d990f42b6e2e58305d5da280ae7d1fa3b"}, - {file = "tomli-2.4.0-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:7b438885858efd5be02a9a133caf5812b8776ee0c969fea02c45e8e3f296ba51"}, - {file = "tomli-2.4.0-cp312-cp312-win32.whl", hash = "sha256:0408e3de5ec77cc7f81960c362543cbbd91ef883e3138e81b729fc3eea5b9729"}, - {file = "tomli-2.4.0-cp312-cp312-win_amd64.whl", hash = "sha256:685306e2cc7da35be4ee914fd34ab801a6acacb061b6a7abca922aaf9ad368da"}, - {file = "tomli-2.4.0-cp312-cp312-win_arm64.whl", hash = "sha256:5aa48d7c2356055feef06a43611fc401a07337d5b006be13a30f6c58f869e3c3"}, - {file = "tomli-2.4.0-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:84d081fbc252d1b6a982e1870660e7330fb8f90f676f6e78b052ad4e64714bf0"}, - {file = "tomli-2.4.0-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:9a08144fa4cba33db5255f9b74f0b89888622109bd2776148f2597447f92a94e"}, - {file = "tomli-2.4.0-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c73add4bb52a206fd0c0723432db123c0c75c280cbd67174dd9d2db228ebb1b4"}, - {file = "tomli-2.4.0-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:1fb2945cbe303b1419e2706e711b7113da57b7db31ee378d08712d678a34e51e"}, - {file = "tomli-2.4.0-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:bbb1b10aa643d973366dc2cb1ad94f99c1726a02343d43cbc011edbfac579e7c"}, - {file = "tomli-2.4.0-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:4cbcb367d44a1f0c2be408758b43e1ffb5308abe0ea222897d6bfc8e8281ef2f"}, - {file = "tomli-2.4.0-cp313-cp313-win32.whl", hash = "sha256:7d49c66a7d5e56ac959cb6fc583aff0651094ec071ba9ad43df785abc2320d86"}, - {file = "tomli-2.4.0-cp313-cp313-win_amd64.whl", hash = "sha256:3cf226acb51d8f1c394c1b310e0e0e61fecdd7adcb78d01e294ac297dd2e7f87"}, - {file = "tomli-2.4.0-cp313-cp313-win_arm64.whl", hash = "sha256:d20b797a5c1ad80c516e41bc1fb0443ddb5006e9aaa7bda2d71978346aeb9132"}, - {file = "tomli-2.4.0-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:26ab906a1eb794cd4e103691daa23d95c6919cc2fa9160000ac02370cc9dd3f6"}, - {file = "tomli-2.4.0-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:20cedb4ee43278bc4f2fee6cb50daec836959aadaf948db5172e776dd3d993fc"}, - {file = "tomli-2.4.0-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:39b0b5d1b6dd03684b3fb276407ebed7090bbec989fa55838c98560c01113b66"}, - {file = "tomli-2.4.0-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:a26d7ff68dfdb9f87a016ecfd1e1c2bacbe3108f4e0f8bcd2228ef9a766c787d"}, - {file = "tomli-2.4.0-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:20ffd184fb1df76a66e34bd1b36b4a4641bd2b82954befa32fe8163e79f1a702"}, - {file = "tomli-2.4.0-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:75c2f8bbddf170e8effc98f5e9084a8751f8174ea6ccf4fca5398436e0320bc8"}, - {file = "tomli-2.4.0-cp314-cp314-win32.whl", hash = "sha256:31d556d079d72db7c584c0627ff3a24c5d3fb4f730221d3444f3efb1b2514776"}, - {file = "tomli-2.4.0-cp314-cp314-win_amd64.whl", hash = "sha256:43e685b9b2341681907759cf3a04e14d7104b3580f808cfde1dfdb60ada85475"}, - {file = "tomli-2.4.0-cp314-cp314-win_arm64.whl", hash = "sha256:3d895d56bd3f82ddd6faaff993c275efc2ff38e52322ea264122d72729dca2b2"}, - {file = "tomli-2.4.0-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:5b5807f3999fb66776dbce568cc9a828544244a8eb84b84b9bafc080c99597b9"}, - {file = "tomli-2.4.0-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:c084ad935abe686bd9c898e62a02a19abfc9760b5a79bc29644463eaf2840cb0"}, - {file = "tomli-2.4.0-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:0f2e3955efea4d1cfbcb87bc321e00dc08d2bcb737fd1d5e398af111d86db5df"}, - {file = "tomli-2.4.0-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:0e0fe8a0b8312acf3a88077a0802565cb09ee34107813bba1c7cd591fa6cfc8d"}, - {file = "tomli-2.4.0-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:413540dce94673591859c4c6f794dfeaa845e98bf35d72ed59636f869ef9f86f"}, - {file = "tomli-2.4.0-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:0dc56fef0e2c1c470aeac5b6ca8cc7b640bb93e92d9803ddaf9ea03e198f5b0b"}, - {file = "tomli-2.4.0-cp314-cp314t-win32.whl", hash = "sha256:d878f2a6707cc9d53a1be1414bbb419e629c3d6e67f69230217bb663e76b5087"}, - {file = "tomli-2.4.0-cp314-cp314t-win_amd64.whl", hash = "sha256:2add28aacc7425117ff6364fe9e06a183bb0251b03f986df0e78e974047571fd"}, - {file = "tomli-2.4.0-cp314-cp314t-win_arm64.whl", hash = "sha256:2b1e3b80e1d5e52e40e9b924ec43d81570f0e7d09d11081b797bc4692765a3d4"}, - {file = "tomli-2.4.0-py3-none-any.whl", hash = "sha256:1f776e7d669ebceb01dee46484485f43a4048746235e683bcdffacdf1fb4785a"}, - {file = "tomli-2.4.0.tar.gz", hash = "sha256:aa89c3f6c277dd275d8e243ad24f3b5e701491a860d5121f2cdd399fbb31fc9c"}, + {file = "tomli-2.4.1-cp311-cp311-macosx_10_9_x86_64.whl", hash = "sha256:f8f0fc26ec2cc2b965b7a3b87cd19c5c6b8c5e5f436b984e85f486d652285c30"}, + {file = "tomli-2.4.1-cp311-cp311-macosx_11_0_arm64.whl", hash = "sha256:4ab97e64ccda8756376892c53a72bd1f964e519c77236368527f758fbc36a53a"}, + {file = "tomli-2.4.1-cp311-cp311-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:96481a5786729fd470164b47cdb3e0e58062a496f455ee41b4403be77cb5a076"}, + {file = "tomli-2.4.1-cp311-cp311-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:5a881ab208c0baf688221f8cecc5401bd291d67e38a1ac884d6736cbcd8247e9"}, + {file = "tomli-2.4.1-cp311-cp311-musllinux_1_2_aarch64.whl", hash = "sha256:47149d5bd38761ac8be13a84864bf0b7b70bc051806bc3669ab1cbc56216b23c"}, + {file = "tomli-2.4.1-cp311-cp311-musllinux_1_2_x86_64.whl", hash = "sha256:ec9bfaf3ad2df51ace80688143a6a4ebc09a248f6ff781a9945e51937008fcbc"}, + {file = "tomli-2.4.1-cp311-cp311-win32.whl", hash = "sha256:ff2983983d34813c1aeb0fa89091e76c3a22889ee83ab27c5eeb45100560c049"}, + {file = "tomli-2.4.1-cp311-cp311-win_amd64.whl", hash = "sha256:5ee18d9ebdb417e384b58fe414e8d6af9f4e7a0ae761519fb50f721de398dd4e"}, + {file = "tomli-2.4.1-cp311-cp311-win_arm64.whl", hash = "sha256:c2541745709bad0264b7d4705ad453b76ccd191e64aa6f0fc66b69a293a45ece"}, + {file = "tomli-2.4.1-cp312-cp312-macosx_10_13_x86_64.whl", hash = "sha256:c742f741d58a28940ce01d58f0ab2ea3ced8b12402f162f4d534dfe18ba1cd6a"}, + {file = "tomli-2.4.1-cp312-cp312-macosx_11_0_arm64.whl", hash = "sha256:7f86fd587c4ed9dd76f318225e7d9b29cfc5a9d43de44e5754db8d1128487085"}, + {file = "tomli-2.4.1-cp312-cp312-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:ff18e6a727ee0ab0388507b89d1bc6a22b138d1e2fa56d1ad494586d61d2eae9"}, + {file = "tomli-2.4.1-cp312-cp312-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:136443dbd7e1dee43c68ac2694fde36b2849865fa258d39bf822c10e8068eac5"}, + {file = "tomli-2.4.1-cp312-cp312-musllinux_1_2_aarch64.whl", hash = "sha256:5e262d41726bc187e69af7825504c933b6794dc3fbd5945e41a79bb14c31f585"}, + {file = "tomli-2.4.1-cp312-cp312-musllinux_1_2_x86_64.whl", hash = "sha256:5cb41aa38891e073ee49d55fbc7839cfdb2bc0e600add13874d048c94aadddd1"}, + {file = "tomli-2.4.1-cp312-cp312-win32.whl", hash = "sha256:da25dc3563bff5965356133435b757a795a17b17d01dbc0f42fb32447ddfd917"}, + {file = "tomli-2.4.1-cp312-cp312-win_amd64.whl", hash = "sha256:52c8ef851d9a240f11a88c003eacb03c31fc1c9c4ec64a99a0f922b93874fda9"}, + {file = "tomli-2.4.1-cp312-cp312-win_arm64.whl", hash = "sha256:f758f1b9299d059cc3f6546ae2af89670cb1c4d48ea29c3cacc4fe7de3058257"}, + {file = "tomli-2.4.1-cp313-cp313-macosx_10_13_x86_64.whl", hash = "sha256:36d2bd2ad5fb9eaddba5226aa02c8ec3fa4f192631e347b3ed28186d43be6b54"}, + {file = "tomli-2.4.1-cp313-cp313-macosx_11_0_arm64.whl", hash = "sha256:eb0dc4e38e6a1fd579e5d50369aa2e10acfc9cace504579b2faabb478e76941a"}, + {file = "tomli-2.4.1-cp313-cp313-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:c7f2c7f2b9ca6bdeef8f0fa897f8e05085923eb091721675170254cbc5b02897"}, + {file = "tomli-2.4.1-cp313-cp313-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:f3c6818a1a86dd6dca7ddcaaf76947d5ba31aecc28cb1b67009a5877c9a64f3f"}, + {file = "tomli-2.4.1-cp313-cp313-musllinux_1_2_aarch64.whl", hash = "sha256:d312ef37c91508b0ab2cee7da26ec0b3ed2f03ce12bd87a588d771ae15dcf82d"}, + {file = "tomli-2.4.1-cp313-cp313-musllinux_1_2_x86_64.whl", hash = "sha256:51529d40e3ca50046d7606fa99ce3956a617f9b36380da3b7f0dd3dd28e68cb5"}, + {file = "tomli-2.4.1-cp313-cp313-win32.whl", hash = "sha256:2190f2e9dd7508d2a90ded5ed369255980a1bcdd58e52f7fe24b8162bf9fedbd"}, + {file = "tomli-2.4.1-cp313-cp313-win_amd64.whl", hash = "sha256:8d65a2fbf9d2f8352685bc1364177ee3923d6baf5e7f43ea4959d7d8bc326a36"}, + {file = "tomli-2.4.1-cp313-cp313-win_arm64.whl", hash = "sha256:4b605484e43cdc43f0954ddae319fb75f04cc10dd80d830540060ee7cd0243cd"}, + {file = "tomli-2.4.1-cp314-cp314-macosx_10_15_x86_64.whl", hash = "sha256:fd0409a3653af6c147209d267a0e4243f0ae46b011aa978b1080359fddc9b6cf"}, + {file = "tomli-2.4.1-cp314-cp314-macosx_11_0_arm64.whl", hash = "sha256:a120733b01c45e9a0c34aeef92bf0cf1d56cfe81ed9d47d562f9ed591a9828ac"}, + {file = "tomli-2.4.1-cp314-cp314-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:559db847dc486944896521f68d8190be1c9e719fced785720d2216fe7022b662"}, + {file = "tomli-2.4.1-cp314-cp314-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:01f520d4f53ef97964a240a035ec2a869fe1a37dde002b57ebc4417a27ccd853"}, + {file = "tomli-2.4.1-cp314-cp314-musllinux_1_2_aarch64.whl", hash = "sha256:7f94b27a62cfad8496c8d2513e1a222dd446f095fca8987fceef261225538a15"}, + {file = "tomli-2.4.1-cp314-cp314-musllinux_1_2_x86_64.whl", hash = "sha256:ede3e6487c5ef5d28634ba3f31f989030ad6af71edfb0055cbbd14189ff240ba"}, + {file = "tomli-2.4.1-cp314-cp314-win32.whl", hash = "sha256:3d48a93ee1c9b79c04bb38772ee1b64dcf18ff43085896ea460ca8dec96f35f6"}, + {file = "tomli-2.4.1-cp314-cp314-win_amd64.whl", hash = "sha256:88dceee75c2c63af144e456745e10101eb67361050196b0b6af5d717254dddf7"}, + {file = "tomli-2.4.1-cp314-cp314-win_arm64.whl", hash = "sha256:b8c198f8c1805dc42708689ed6864951fd2494f924149d3e4bce7710f8eb5232"}, + {file = "tomli-2.4.1-cp314-cp314t-macosx_10_15_x86_64.whl", hash = "sha256:d4d8fe59808a54658fcc0160ecfb1b30f9089906c50b23bcb4c69eddc19ec2b4"}, + {file = "tomli-2.4.1-cp314-cp314t-macosx_11_0_arm64.whl", hash = "sha256:7008df2e7655c495dd12d2a4ad038ff878d4ca4b81fccaf82b714e07eae4402c"}, + {file = "tomli-2.4.1-cp314-cp314t-manylinux2014_aarch64.manylinux_2_17_aarch64.manylinux_2_28_aarch64.whl", hash = "sha256:1d8591993e228b0c930c4bb0db464bdad97b3289fb981255d6c9a41aedc84b2d"}, + {file = "tomli-2.4.1-cp314-cp314t-manylinux2014_x86_64.manylinux_2_17_x86_64.manylinux_2_28_x86_64.whl", hash = "sha256:734e20b57ba95624ecf1841e72b53f6e186355e216e5412de414e3c51e5e3c41"}, + {file = "tomli-2.4.1-cp314-cp314t-musllinux_1_2_aarch64.whl", hash = "sha256:8a650c2dbafa08d42e51ba0b62740dae4ecb9338eefa093aa5c78ceb546fcd5c"}, + {file = "tomli-2.4.1-cp314-cp314t-musllinux_1_2_x86_64.whl", hash = "sha256:504aa796fe0569bb43171066009ead363de03675276d2d121ac1a4572397870f"}, + {file = "tomli-2.4.1-cp314-cp314t-win32.whl", hash = "sha256:b1d22e6e9387bf4739fbe23bfa80e93f6b0373a7f1b96c6227c32bef95a4d7a8"}, + {file = "tomli-2.4.1-cp314-cp314t-win_amd64.whl", hash = "sha256:2c1c351919aca02858f740c6d33adea0c5deea37f9ecca1cc1ef9e884a619d26"}, + {file = "tomli-2.4.1-cp314-cp314t-win_arm64.whl", hash = "sha256:eab21f45c7f66c13f2a9e0e1535309cee140182a9cdae1e041d02e47291e8396"}, + {file = "tomli-2.4.1-py3-none-any.whl", hash = "sha256:0d85819802132122da43cb86656f8d1f8c6587d54ae7dcaf30e90533028b49fe"}, + {file = "tomli-2.4.1.tar.gz", hash = "sha256:7c7e1a961a0b2f2472c1ac5b69affa0ae1132c39adcb67aba98568702b9cc23f"}, ] [[package]] diff --git a/pyproject.toml b/pyproject.toml index a11551e5..fd0c51a6 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -4,7 +4,7 @@ dynamic = ["version"] [tool.poetry] name = "deepgram-sdk" -version = "6.0.1" +version = "6.0.2" description = "" readme = "README.md" authors = [] diff --git a/src/deepgram/__init__.py b/src/deepgram/__init__.py index 3e34a4a7..20dd61db 100644 --- a/src/deepgram/__init__.py +++ b/src/deepgram/__init__.py @@ -10,31 +10,27 @@ AgentThinkModelsV1Response, AgentThinkModelsV1ResponseModelsItem, AgentThinkModelsV1ResponseModelsItemId, + AgentThinkModelsV1ResponseModelsItemIdProvider, AgentThinkModelsV1ResponseModelsItemOne, AgentThinkModelsV1ResponseModelsItemOneId, + AgentThinkModelsV1ResponseModelsItemOneProvider, AgentThinkModelsV1ResponseModelsItemThree, + AgentThinkModelsV1ResponseModelsItemThreeId, + AgentThinkModelsV1ResponseModelsItemThreeProvider, AgentThinkModelsV1ResponseModelsItemTwo, AgentThinkModelsV1ResponseModelsItemTwoId, + AgentThinkModelsV1ResponseModelsItemTwoProvider, AgentThinkModelsV1ResponseModelsItemZero, AgentThinkModelsV1ResponseModelsItemZeroId, + AgentThinkModelsV1ResponseModelsItemZeroProvider, Anthropic, - AnthropicThinkProviderModel, AwsBedrockThinkProvider, - AwsBedrockThinkProviderCredentials, - AwsBedrockThinkProviderCredentialsType, - AwsBedrockThinkProviderModel, AwsPollySpeakProvider, - AwsPollySpeakProviderCredentials, - AwsPollySpeakProviderCredentialsType, - AwsPollySpeakProviderEngine, - AwsPollySpeakProviderVoice, BillingBreakdownV1Response, BillingBreakdownV1ResponseResolution, BillingBreakdownV1ResponseResultsItem, BillingBreakdownV1ResponseResultsItemGrouping, Cartesia, - CartesiaSpeakProviderModelId, - CartesiaSpeakProviderVoice, CreateKeyV1RequestOne, CreateKeyV1Response, CreateProjectDistributionCredentialsV1Response, @@ -42,13 +38,11 @@ CreateProjectDistributionCredentialsV1ResponseMember, CreateProjectInviteV1Response, Deepgram, - DeepgramSpeakProviderModel, DeleteProjectInviteV1Response, DeleteProjectKeyV1Response, DeleteProjectMemberV1Response, DeleteProjectV1Response, ElevenLabsSpeakProvider, - ElevenLabsSpeakProviderModelId, ErrorResponse, ErrorResponseLegacyError, ErrorResponseModernError, @@ -68,7 +62,6 @@ GetProjectRequestV1Response, GetProjectV1Response, Google, - GoogleThinkProviderModel, GrantV1Response, Groq, LeaveProjectV1Response, @@ -161,10 +154,7 @@ ListenV2SampleRate, ListenV2Tag, OpenAiSpeakProvider, - OpenAiSpeakProviderModel, - OpenAiSpeakProviderVoice, OpenAiThinkProvider, - OpenAiThinkProviderModel, ProjectRequestResponse, ReadV1Request, ReadV1RequestText, @@ -194,29 +184,12 @@ SharedTopicsResultsTopicsSegmentsItem, SharedTopicsResultsTopicsSegmentsItemTopicsItem, SpeakSettingsV1, - SpeakSettingsV1Endpoint, - SpeakSettingsV1Provider, - SpeakSettingsV1Provider_AwsPolly, - SpeakSettingsV1Provider_Cartesia, - SpeakSettingsV1Provider_Deepgram, - SpeakSettingsV1Provider_ElevenLabs, - SpeakSettingsV1Provider_OpenAi, SpeakV1Encoding, SpeakV1MipOptOut, SpeakV1Model, SpeakV1Response, SpeakV1SampleRate, ThinkSettingsV1, - ThinkSettingsV1ContextLength, - ThinkSettingsV1Endpoint, - ThinkSettingsV1FunctionsItem, - ThinkSettingsV1FunctionsItemEndpoint, - ThinkSettingsV1Provider, - ThinkSettingsV1Provider_Anthropic, - ThinkSettingsV1Provider_AwsBedrock, - ThinkSettingsV1Provider_Google, - ThinkSettingsV1Provider_Groq, - ThinkSettingsV1Provider_OpenAi, UpdateProjectMemberScopesV1Response, UpdateProjectV1Response, UsageBreakdownV1Response, @@ -240,28 +213,19 @@ AgentThinkModelsV1ResponseModelsItemTwoParams, AgentThinkModelsV1ResponseModelsItemZeroParams, AgentThinkModelsV1ResponseParams, - AnthropicParams, - AwsBedrockThinkProviderCredentialsParams, - AwsBedrockThinkProviderParams, - AwsPollySpeakProviderCredentialsParams, - AwsPollySpeakProviderParams, BillingBreakdownV1ResponseParams, BillingBreakdownV1ResponseResolutionParams, BillingBreakdownV1ResponseResultsItemGroupingParams, BillingBreakdownV1ResponseResultsItemParams, - CartesiaParams, - CartesiaSpeakProviderVoiceParams, CreateKeyV1ResponseParams, CreateProjectDistributionCredentialsV1ResponseDistributionCredentialsParams, CreateProjectDistributionCredentialsV1ResponseMemberParams, CreateProjectDistributionCredentialsV1ResponseParams, CreateProjectInviteV1ResponseParams, - DeepgramParams, DeleteProjectInviteV1ResponseParams, DeleteProjectKeyV1ResponseParams, DeleteProjectMemberV1ResponseParams, DeleteProjectV1ResponseParams, - ElevenLabsSpeakProviderParams, ErrorResponseLegacyErrorParams, ErrorResponseModernErrorParams, ErrorResponseParams, @@ -279,9 +243,7 @@ GetProjectKeyV1ResponseParams, GetProjectRequestV1ResponseParams, GetProjectV1ResponseParams, - GoogleParams, GrantV1ResponseParams, - GroqParams, LeaveProjectV1ResponseParams, ListBillingFieldsV1ResponseParams, ListModelsV1ResponseParams, @@ -333,8 +295,6 @@ ListenV1ResponseResultsUtterancesItemWordsItemParams, ListenV1ResponseResultsUtterancesParams, ListenV2KeytermParams, - OpenAiSpeakProviderParams, - OpenAiThinkProviderParams, ProjectRequestResponseParams, ReadV1RequestParams, ReadV1RequestTextParams, @@ -363,25 +323,7 @@ SharedTopicsResultsTopicsParams, SharedTopicsResultsTopicsSegmentsItemParams, SharedTopicsResultsTopicsSegmentsItemTopicsItemParams, - SpeakSettingsV1EndpointParams, - SpeakSettingsV1Params, - SpeakSettingsV1ProviderParams, - SpeakSettingsV1Provider_AwsPollyParams, - SpeakSettingsV1Provider_CartesiaParams, - SpeakSettingsV1Provider_DeepgramParams, - SpeakSettingsV1Provider_ElevenLabsParams, - SpeakSettingsV1Provider_OpenAiParams, - ThinkSettingsV1ContextLengthParams, - ThinkSettingsV1EndpointParams, - ThinkSettingsV1FunctionsItemEndpointParams, - ThinkSettingsV1FunctionsItemParams, ThinkSettingsV1Params, - ThinkSettingsV1ProviderParams, - ThinkSettingsV1Provider_AnthropicParams, - ThinkSettingsV1Provider_AwsBedrockParams, - ThinkSettingsV1Provider_GoogleParams, - ThinkSettingsV1Provider_GroqParams, - ThinkSettingsV1Provider_OpenAiParams, UpdateProjectMemberScopesV1ResponseParams, UpdateProjectV1ResponseParams, UsageBreakdownV1ResponseParams, @@ -399,36 +341,29 @@ "AgentThinkModelsV1ResponseModelsItem": ".types", "AgentThinkModelsV1ResponseModelsItemId": ".types", "AgentThinkModelsV1ResponseModelsItemIdParams": ".requests", + "AgentThinkModelsV1ResponseModelsItemIdProvider": ".types", "AgentThinkModelsV1ResponseModelsItemOne": ".types", "AgentThinkModelsV1ResponseModelsItemOneId": ".types", "AgentThinkModelsV1ResponseModelsItemOneParams": ".requests", + "AgentThinkModelsV1ResponseModelsItemOneProvider": ".types", "AgentThinkModelsV1ResponseModelsItemParams": ".requests", "AgentThinkModelsV1ResponseModelsItemThree": ".types", + "AgentThinkModelsV1ResponseModelsItemThreeId": ".types", "AgentThinkModelsV1ResponseModelsItemThreeParams": ".requests", + "AgentThinkModelsV1ResponseModelsItemThreeProvider": ".types", "AgentThinkModelsV1ResponseModelsItemTwo": ".types", "AgentThinkModelsV1ResponseModelsItemTwoId": ".types", "AgentThinkModelsV1ResponseModelsItemTwoParams": ".requests", + "AgentThinkModelsV1ResponseModelsItemTwoProvider": ".types", "AgentThinkModelsV1ResponseModelsItemZero": ".types", "AgentThinkModelsV1ResponseModelsItemZeroId": ".types", "AgentThinkModelsV1ResponseModelsItemZeroParams": ".requests", + "AgentThinkModelsV1ResponseModelsItemZeroProvider": ".types", "AgentThinkModelsV1ResponseParams": ".requests", "Anthropic": ".types", - "AnthropicParams": ".requests", - "AnthropicThinkProviderModel": ".types", "AsyncDeepgramClient": ".client", "AwsBedrockThinkProvider": ".types", - "AwsBedrockThinkProviderCredentials": ".types", - "AwsBedrockThinkProviderCredentialsParams": ".requests", - "AwsBedrockThinkProviderCredentialsType": ".types", - "AwsBedrockThinkProviderModel": ".types", - "AwsBedrockThinkProviderParams": ".requests", "AwsPollySpeakProvider": ".types", - "AwsPollySpeakProviderCredentials": ".types", - "AwsPollySpeakProviderCredentialsParams": ".requests", - "AwsPollySpeakProviderCredentialsType": ".types", - "AwsPollySpeakProviderEngine": ".types", - "AwsPollySpeakProviderParams": ".requests", - "AwsPollySpeakProviderVoice": ".types", "BadRequestError": ".errors", "BillingBreakdownV1Response": ".types", "BillingBreakdownV1ResponseParams": ".requests", @@ -439,10 +374,6 @@ "BillingBreakdownV1ResponseResultsItemGroupingParams": ".requests", "BillingBreakdownV1ResponseResultsItemParams": ".requests", "Cartesia": ".types", - "CartesiaParams": ".requests", - "CartesiaSpeakProviderModelId": ".types", - "CartesiaSpeakProviderVoice": ".types", - "CartesiaSpeakProviderVoiceParams": ".requests", "CreateKeyV1RequestOne": ".types", "CreateKeyV1Response": ".types", "CreateKeyV1ResponseParams": ".requests", @@ -457,8 +388,6 @@ "Deepgram": ".types", "DeepgramClient": ".client", "DeepgramClientEnvironment": ".environment", - "DeepgramParams": ".requests", - "DeepgramSpeakProviderModel": ".types", "DeleteProjectInviteV1Response": ".types", "DeleteProjectInviteV1ResponseParams": ".requests", "DeleteProjectKeyV1Response": ".types", @@ -468,8 +397,6 @@ "DeleteProjectV1Response": ".types", "DeleteProjectV1ResponseParams": ".requests", "ElevenLabsSpeakProvider": ".types", - "ElevenLabsSpeakProviderModelId": ".types", - "ElevenLabsSpeakProviderParams": ".requests", "ErrorResponse": ".types", "ErrorResponseLegacyError": ".types", "ErrorResponseLegacyErrorParams": ".requests", @@ -506,12 +433,9 @@ "GetProjectV1Response": ".types", "GetProjectV1ResponseParams": ".requests", "Google": ".types", - "GoogleParams": ".requests", - "GoogleThinkProviderModel": ".types", "GrantV1Response": ".types", "GrantV1ResponseParams": ".requests", "Groq": ".types", - "GroqParams": ".requests", "LeaveProjectV1Response": ".types", "LeaveProjectV1ResponseParams": ".requests", "ListBillingFieldsV1Response": ".types", @@ -653,12 +577,7 @@ "ListenV2SampleRate": ".types", "ListenV2Tag": ".types", "OpenAiSpeakProvider": ".types", - "OpenAiSpeakProviderModel": ".types", - "OpenAiSpeakProviderParams": ".requests", - "OpenAiSpeakProviderVoice": ".types", "OpenAiThinkProvider": ".types", - "OpenAiThinkProviderModel": ".types", - "OpenAiThinkProviderParams": ".requests", "ProjectRequestResponse": ".types", "ProjectRequestResponseParams": ".requests", "ReadV1Request": ".types", @@ -716,48 +635,13 @@ "SharedTopicsResultsTopicsSegmentsItemTopicsItem": ".types", "SharedTopicsResultsTopicsSegmentsItemTopicsItemParams": ".requests", "SpeakSettingsV1": ".types", - "SpeakSettingsV1Endpoint": ".types", - "SpeakSettingsV1EndpointParams": ".requests", - "SpeakSettingsV1Params": ".requests", - "SpeakSettingsV1Provider": ".types", - "SpeakSettingsV1ProviderParams": ".requests", - "SpeakSettingsV1Provider_AwsPolly": ".types", - "SpeakSettingsV1Provider_AwsPollyParams": ".requests", - "SpeakSettingsV1Provider_Cartesia": ".types", - "SpeakSettingsV1Provider_CartesiaParams": ".requests", - "SpeakSettingsV1Provider_Deepgram": ".types", - "SpeakSettingsV1Provider_DeepgramParams": ".requests", - "SpeakSettingsV1Provider_ElevenLabs": ".types", - "SpeakSettingsV1Provider_ElevenLabsParams": ".requests", - "SpeakSettingsV1Provider_OpenAi": ".types", - "SpeakSettingsV1Provider_OpenAiParams": ".requests", "SpeakV1Encoding": ".types", "SpeakV1MipOptOut": ".types", "SpeakV1Model": ".types", "SpeakV1Response": ".types", "SpeakV1SampleRate": ".types", "ThinkSettingsV1": ".types", - "ThinkSettingsV1ContextLength": ".types", - "ThinkSettingsV1ContextLengthParams": ".requests", - "ThinkSettingsV1Endpoint": ".types", - "ThinkSettingsV1EndpointParams": ".requests", - "ThinkSettingsV1FunctionsItem": ".types", - "ThinkSettingsV1FunctionsItemEndpoint": ".types", - "ThinkSettingsV1FunctionsItemEndpointParams": ".requests", - "ThinkSettingsV1FunctionsItemParams": ".requests", "ThinkSettingsV1Params": ".requests", - "ThinkSettingsV1Provider": ".types", - "ThinkSettingsV1ProviderParams": ".requests", - "ThinkSettingsV1Provider_Anthropic": ".types", - "ThinkSettingsV1Provider_AnthropicParams": ".requests", - "ThinkSettingsV1Provider_AwsBedrock": ".types", - "ThinkSettingsV1Provider_AwsBedrockParams": ".requests", - "ThinkSettingsV1Provider_Google": ".types", - "ThinkSettingsV1Provider_GoogleParams": ".requests", - "ThinkSettingsV1Provider_Groq": ".types", - "ThinkSettingsV1Provider_GroqParams": ".requests", - "ThinkSettingsV1Provider_OpenAi": ".types", - "ThinkSettingsV1Provider_OpenAiParams": ".requests", "UpdateProjectMemberScopesV1Response": ".types", "UpdateProjectMemberScopesV1ResponseParams": ".requests", "UpdateProjectV1Response": ".types", @@ -815,36 +699,29 @@ def __dir__(): "AgentThinkModelsV1ResponseModelsItem", "AgentThinkModelsV1ResponseModelsItemId", "AgentThinkModelsV1ResponseModelsItemIdParams", + "AgentThinkModelsV1ResponseModelsItemIdProvider", "AgentThinkModelsV1ResponseModelsItemOne", "AgentThinkModelsV1ResponseModelsItemOneId", "AgentThinkModelsV1ResponseModelsItemOneParams", + "AgentThinkModelsV1ResponseModelsItemOneProvider", "AgentThinkModelsV1ResponseModelsItemParams", "AgentThinkModelsV1ResponseModelsItemThree", + "AgentThinkModelsV1ResponseModelsItemThreeId", "AgentThinkModelsV1ResponseModelsItemThreeParams", + "AgentThinkModelsV1ResponseModelsItemThreeProvider", "AgentThinkModelsV1ResponseModelsItemTwo", "AgentThinkModelsV1ResponseModelsItemTwoId", "AgentThinkModelsV1ResponseModelsItemTwoParams", + "AgentThinkModelsV1ResponseModelsItemTwoProvider", "AgentThinkModelsV1ResponseModelsItemZero", "AgentThinkModelsV1ResponseModelsItemZeroId", "AgentThinkModelsV1ResponseModelsItemZeroParams", + "AgentThinkModelsV1ResponseModelsItemZeroProvider", "AgentThinkModelsV1ResponseParams", "Anthropic", - "AnthropicParams", - "AnthropicThinkProviderModel", "AsyncDeepgramClient", "AwsBedrockThinkProvider", - "AwsBedrockThinkProviderCredentials", - "AwsBedrockThinkProviderCredentialsParams", - "AwsBedrockThinkProviderCredentialsType", - "AwsBedrockThinkProviderModel", - "AwsBedrockThinkProviderParams", "AwsPollySpeakProvider", - "AwsPollySpeakProviderCredentials", - "AwsPollySpeakProviderCredentialsParams", - "AwsPollySpeakProviderCredentialsType", - "AwsPollySpeakProviderEngine", - "AwsPollySpeakProviderParams", - "AwsPollySpeakProviderVoice", "BadRequestError", "BillingBreakdownV1Response", "BillingBreakdownV1ResponseParams", @@ -855,10 +732,6 @@ def __dir__(): "BillingBreakdownV1ResponseResultsItemGroupingParams", "BillingBreakdownV1ResponseResultsItemParams", "Cartesia", - "CartesiaParams", - "CartesiaSpeakProviderModelId", - "CartesiaSpeakProviderVoice", - "CartesiaSpeakProviderVoiceParams", "CreateKeyV1RequestOne", "CreateKeyV1Response", "CreateKeyV1ResponseParams", @@ -873,8 +746,6 @@ def __dir__(): "Deepgram", "DeepgramClient", "DeepgramClientEnvironment", - "DeepgramParams", - "DeepgramSpeakProviderModel", "DeleteProjectInviteV1Response", "DeleteProjectInviteV1ResponseParams", "DeleteProjectKeyV1Response", @@ -884,8 +755,6 @@ def __dir__(): "DeleteProjectV1Response", "DeleteProjectV1ResponseParams", "ElevenLabsSpeakProvider", - "ElevenLabsSpeakProviderModelId", - "ElevenLabsSpeakProviderParams", "ErrorResponse", "ErrorResponseLegacyError", "ErrorResponseLegacyErrorParams", @@ -922,12 +791,9 @@ def __dir__(): "GetProjectV1Response", "GetProjectV1ResponseParams", "Google", - "GoogleParams", - "GoogleThinkProviderModel", "GrantV1Response", "GrantV1ResponseParams", "Groq", - "GroqParams", "LeaveProjectV1Response", "LeaveProjectV1ResponseParams", "ListBillingFieldsV1Response", @@ -1069,12 +935,7 @@ def __dir__(): "ListenV2SampleRate", "ListenV2Tag", "OpenAiSpeakProvider", - "OpenAiSpeakProviderModel", - "OpenAiSpeakProviderParams", - "OpenAiSpeakProviderVoice", "OpenAiThinkProvider", - "OpenAiThinkProviderModel", - "OpenAiThinkProviderParams", "ProjectRequestResponse", "ProjectRequestResponseParams", "ReadV1Request", @@ -1132,48 +993,13 @@ def __dir__(): "SharedTopicsResultsTopicsSegmentsItemTopicsItem", "SharedTopicsResultsTopicsSegmentsItemTopicsItemParams", "SpeakSettingsV1", - "SpeakSettingsV1Endpoint", - "SpeakSettingsV1EndpointParams", - "SpeakSettingsV1Params", - "SpeakSettingsV1Provider", - "SpeakSettingsV1ProviderParams", - "SpeakSettingsV1Provider_AwsPolly", - "SpeakSettingsV1Provider_AwsPollyParams", - "SpeakSettingsV1Provider_Cartesia", - "SpeakSettingsV1Provider_CartesiaParams", - "SpeakSettingsV1Provider_Deepgram", - "SpeakSettingsV1Provider_DeepgramParams", - "SpeakSettingsV1Provider_ElevenLabs", - "SpeakSettingsV1Provider_ElevenLabsParams", - "SpeakSettingsV1Provider_OpenAi", - "SpeakSettingsV1Provider_OpenAiParams", "SpeakV1Encoding", "SpeakV1MipOptOut", "SpeakV1Model", "SpeakV1Response", "SpeakV1SampleRate", "ThinkSettingsV1", - "ThinkSettingsV1ContextLength", - "ThinkSettingsV1ContextLengthParams", - "ThinkSettingsV1Endpoint", - "ThinkSettingsV1EndpointParams", - "ThinkSettingsV1FunctionsItem", - "ThinkSettingsV1FunctionsItemEndpoint", - "ThinkSettingsV1FunctionsItemEndpointParams", - "ThinkSettingsV1FunctionsItemParams", "ThinkSettingsV1Params", - "ThinkSettingsV1Provider", - "ThinkSettingsV1ProviderParams", - "ThinkSettingsV1Provider_Anthropic", - "ThinkSettingsV1Provider_AnthropicParams", - "ThinkSettingsV1Provider_AwsBedrock", - "ThinkSettingsV1Provider_AwsBedrockParams", - "ThinkSettingsV1Provider_Google", - "ThinkSettingsV1Provider_GoogleParams", - "ThinkSettingsV1Provider_Groq", - "ThinkSettingsV1Provider_GroqParams", - "ThinkSettingsV1Provider_OpenAi", - "ThinkSettingsV1Provider_OpenAiParams", "UpdateProjectMemberScopesV1Response", "UpdateProjectMemberScopesV1ResponseParams", "UpdateProjectV1Response", diff --git a/src/deepgram/agent/__init__.py b/src/deepgram/agent/__init__.py index 30038648..5b5014f4 100644 --- a/src/deepgram/agent/__init__.py +++ b/src/deepgram/agent/__init__.py @@ -10,33 +10,46 @@ from .v1 import ( AgentV1AgentAudioDone, AgentV1AgentAudioDoneParams, + AgentV1AgentAudioDoneType, AgentV1AgentStartedSpeaking, AgentV1AgentStartedSpeakingParams, + AgentV1AgentStartedSpeakingType, AgentV1AgentThinking, AgentV1AgentThinkingParams, + AgentV1AgentThinkingType, AgentV1ConversationText, AgentV1ConversationTextParams, AgentV1ConversationTextRole, + AgentV1ConversationTextType, AgentV1Error, AgentV1ErrorParams, + AgentV1ErrorType, AgentV1FunctionCallRequest, AgentV1FunctionCallRequestFunctionsItem, AgentV1FunctionCallRequestFunctionsItemParams, AgentV1FunctionCallRequestParams, + AgentV1FunctionCallRequestType, AgentV1InjectAgentMessage, AgentV1InjectAgentMessageParams, + AgentV1InjectAgentMessageType, AgentV1InjectUserMessage, AgentV1InjectUserMessageParams, + AgentV1InjectUserMessageType, AgentV1InjectionRefused, AgentV1InjectionRefusedParams, + AgentV1InjectionRefusedType, AgentV1KeepAlive, AgentV1KeepAliveParams, + AgentV1KeepAliveType, AgentV1PromptUpdated, AgentV1PromptUpdatedParams, + AgentV1PromptUpdatedType, AgentV1ReceiveFunctionCallResponse, AgentV1ReceiveFunctionCallResponseParams, + AgentV1ReceiveFunctionCallResponseType, AgentV1SendFunctionCallResponse, AgentV1SendFunctionCallResponseParams, + AgentV1SendFunctionCallResponseType, AgentV1Settings, AgentV1SettingsAgent, AgentV1SettingsAgentContext, @@ -44,10 +57,12 @@ AgentV1SettingsAgentContextMessagesItemContent, AgentV1SettingsAgentContextMessagesItemContentParams, AgentV1SettingsAgentContextMessagesItemContentRole, + AgentV1SettingsAgentContextMessagesItemContentType, AgentV1SettingsAgentContextMessagesItemFunctionCalls, AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem, AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams, AgentV1SettingsAgentContextMessagesItemFunctionCallsParams, + AgentV1SettingsAgentContextMessagesItemFunctionCallsType, AgentV1SettingsAgentContextMessagesItemParams, AgentV1SettingsAgentContextParams, AgentV1SettingsAgentListen, @@ -56,19 +71,106 @@ AgentV1SettingsAgentListenProviderParams, AgentV1SettingsAgentListenProviderV1, AgentV1SettingsAgentListenProviderV1Params, + AgentV1SettingsAgentListenProviderV1Type, AgentV1SettingsAgentListenProviderV2, AgentV1SettingsAgentListenProviderV2Params, + AgentV1SettingsAgentListenProviderV2Type, AgentV1SettingsAgentListenProvider_V1, AgentV1SettingsAgentListenProvider_V1Params, AgentV1SettingsAgentListenProvider_V2, AgentV1SettingsAgentListenProvider_V2Params, AgentV1SettingsAgentParams, AgentV1SettingsAgentSpeak, + AgentV1SettingsAgentSpeakEndpoint, + AgentV1SettingsAgentSpeakEndpointEndpoint, + AgentV1SettingsAgentSpeakEndpointEndpointParams, + AgentV1SettingsAgentSpeakEndpointParams, + AgentV1SettingsAgentSpeakEndpointProvider, + AgentV1SettingsAgentSpeakEndpointProviderAwsPolly, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams, + AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel, + AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion, + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId, + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion, + AgentV1SettingsAgentSpeakEndpointProviderOpenAi, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, + AgentV1SettingsAgentSpeakEndpointProviderParams, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakEndpointProvider_Cartesia, + AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams, + AgentV1SettingsAgentSpeakEndpointProvider_Deepgram, + AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAi, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams, + AgentV1SettingsAgentSpeakOneItem, + AgentV1SettingsAgentSpeakOneItemEndpoint, + AgentV1SettingsAgentSpeakOneItemEndpointParams, + AgentV1SettingsAgentSpeakOneItemParams, + AgentV1SettingsAgentSpeakOneItemProvider, + AgentV1SettingsAgentSpeakOneItemProviderAwsPolly, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, + AgentV1SettingsAgentSpeakOneItemProviderOpenAi, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, + AgentV1SettingsAgentSpeakOneItemProviderParams, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakOneItemProvider_Cartesia, + AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams, + AgentV1SettingsAgentSpeakOneItemProvider_Deepgram, + AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAi, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams, AgentV1SettingsAgentSpeakParams, AgentV1SettingsAgentThink, + AgentV1SettingsAgentThinkOneItem, + AgentV1SettingsAgentThinkOneItemContextLength, + AgentV1SettingsAgentThinkOneItemContextLengthParams, + AgentV1SettingsAgentThinkOneItemEndpoint, + AgentV1SettingsAgentThinkOneItemEndpointParams, + AgentV1SettingsAgentThinkOneItemFunctionsItem, + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint, + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams, + AgentV1SettingsAgentThinkOneItemFunctionsItemParams, + AgentV1SettingsAgentThinkOneItemParams, + AgentV1SettingsAgentThinkOneItemProvider, + AgentV1SettingsAgentThinkOneItemProviderParams, AgentV1SettingsAgentThinkParams, AgentV1SettingsApplied, AgentV1SettingsAppliedParams, + AgentV1SettingsAppliedType, AgentV1SettingsAudio, AgentV1SettingsAudioInput, AgentV1SettingsAudioInputEncoding, @@ -80,49 +182,162 @@ AgentV1SettingsFlags, AgentV1SettingsFlagsParams, AgentV1SettingsParams, + AgentV1SettingsType, AgentV1SpeakUpdated, AgentV1SpeakUpdatedParams, + AgentV1SpeakUpdatedType, AgentV1UpdatePrompt, AgentV1UpdatePromptParams, + AgentV1UpdatePromptType, AgentV1UpdateSpeak, AgentV1UpdateSpeakParams, + AgentV1UpdateSpeakSpeak, + AgentV1UpdateSpeakSpeakEndpoint, + AgentV1UpdateSpeakSpeakEndpointEndpoint, + AgentV1UpdateSpeakSpeakEndpointEndpointParams, + AgentV1UpdateSpeakSpeakEndpointParams, + AgentV1UpdateSpeakSpeakEndpointProvider, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, + AgentV1UpdateSpeakSpeakEndpointProviderCartesia, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgram, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAi, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, + AgentV1UpdateSpeakSpeakEndpointProviderParams, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia, + AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram, + AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakOneItem, + AgentV1UpdateSpeakSpeakOneItemEndpoint, + AgentV1UpdateSpeakSpeakOneItemEndpointParams, + AgentV1UpdateSpeakSpeakOneItemParams, + AgentV1UpdateSpeakSpeakOneItemProvider, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, + AgentV1UpdateSpeakSpeakOneItemProviderCartesia, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgram, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAi, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, + AgentV1UpdateSpeakSpeakOneItemProviderParams, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia, + AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram, + AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakParams, + AgentV1UpdateSpeakType, AgentV1UserStartedSpeaking, AgentV1UserStartedSpeakingParams, + AgentV1UserStartedSpeakingType, AgentV1Warning, AgentV1WarningParams, + AgentV1WarningType, AgentV1Welcome, AgentV1WelcomeParams, + AgentV1WelcomeType, + Cartesia, + CartesiaParams, + Deepgram, + DeepgramParams, + ElevenLabs, + ElevenLabsParams, + Max, ) _dynamic_imports: typing.Dict[str, str] = { "AgentV1AgentAudioDone": ".v1", "AgentV1AgentAudioDoneParams": ".v1", + "AgentV1AgentAudioDoneType": ".v1", "AgentV1AgentStartedSpeaking": ".v1", "AgentV1AgentStartedSpeakingParams": ".v1", + "AgentV1AgentStartedSpeakingType": ".v1", "AgentV1AgentThinking": ".v1", "AgentV1AgentThinkingParams": ".v1", + "AgentV1AgentThinkingType": ".v1", "AgentV1ConversationText": ".v1", "AgentV1ConversationTextParams": ".v1", "AgentV1ConversationTextRole": ".v1", + "AgentV1ConversationTextType": ".v1", "AgentV1Error": ".v1", "AgentV1ErrorParams": ".v1", + "AgentV1ErrorType": ".v1", "AgentV1FunctionCallRequest": ".v1", "AgentV1FunctionCallRequestFunctionsItem": ".v1", "AgentV1FunctionCallRequestFunctionsItemParams": ".v1", "AgentV1FunctionCallRequestParams": ".v1", + "AgentV1FunctionCallRequestType": ".v1", "AgentV1InjectAgentMessage": ".v1", "AgentV1InjectAgentMessageParams": ".v1", + "AgentV1InjectAgentMessageType": ".v1", "AgentV1InjectUserMessage": ".v1", "AgentV1InjectUserMessageParams": ".v1", + "AgentV1InjectUserMessageType": ".v1", "AgentV1InjectionRefused": ".v1", "AgentV1InjectionRefusedParams": ".v1", + "AgentV1InjectionRefusedType": ".v1", "AgentV1KeepAlive": ".v1", "AgentV1KeepAliveParams": ".v1", + "AgentV1KeepAliveType": ".v1", "AgentV1PromptUpdated": ".v1", "AgentV1PromptUpdatedParams": ".v1", + "AgentV1PromptUpdatedType": ".v1", "AgentV1ReceiveFunctionCallResponse": ".v1", "AgentV1ReceiveFunctionCallResponseParams": ".v1", + "AgentV1ReceiveFunctionCallResponseType": ".v1", "AgentV1SendFunctionCallResponse": ".v1", "AgentV1SendFunctionCallResponseParams": ".v1", + "AgentV1SendFunctionCallResponseType": ".v1", "AgentV1Settings": ".v1", "AgentV1SettingsAgent": ".v1", "AgentV1SettingsAgentContext": ".v1", @@ -130,10 +345,12 @@ "AgentV1SettingsAgentContextMessagesItemContent": ".v1", "AgentV1SettingsAgentContextMessagesItemContentParams": ".v1", "AgentV1SettingsAgentContextMessagesItemContentRole": ".v1", + "AgentV1SettingsAgentContextMessagesItemContentType": ".v1", "AgentV1SettingsAgentContextMessagesItemFunctionCalls": ".v1", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem": ".v1", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams": ".v1", "AgentV1SettingsAgentContextMessagesItemFunctionCallsParams": ".v1", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType": ".v1", "AgentV1SettingsAgentContextMessagesItemParams": ".v1", "AgentV1SettingsAgentContextParams": ".v1", "AgentV1SettingsAgentListen": ".v1", @@ -142,19 +359,106 @@ "AgentV1SettingsAgentListenProviderParams": ".v1", "AgentV1SettingsAgentListenProviderV1": ".v1", "AgentV1SettingsAgentListenProviderV1Params": ".v1", + "AgentV1SettingsAgentListenProviderV1Type": ".v1", "AgentV1SettingsAgentListenProviderV2": ".v1", "AgentV1SettingsAgentListenProviderV2Params": ".v1", + "AgentV1SettingsAgentListenProviderV2Type": ".v1", "AgentV1SettingsAgentListenProvider_V1": ".v1", "AgentV1SettingsAgentListenProvider_V1Params": ".v1", "AgentV1SettingsAgentListenProvider_V2": ".v1", "AgentV1SettingsAgentListenProvider_V2Params": ".v1", "AgentV1SettingsAgentParams": ".v1", "AgentV1SettingsAgentSpeak": ".v1", + "AgentV1SettingsAgentSpeakEndpoint": ".v1", + "AgentV1SettingsAgentSpeakEndpointEndpoint": ".v1", + "AgentV1SettingsAgentSpeakEndpointEndpointParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice": ".v1", + "AgentV1SettingsAgentSpeakEndpointProviderParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi": ".v1", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams": ".v1", + "AgentV1SettingsAgentSpeakOneItem": ".v1", + "AgentV1SettingsAgentSpeakOneItemEndpoint": ".v1", + "AgentV1SettingsAgentSpeakOneItemEndpointParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice": ".v1", + "AgentV1SettingsAgentSpeakOneItemProviderParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi": ".v1", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams": ".v1", "AgentV1SettingsAgentSpeakParams": ".v1", "AgentV1SettingsAgentThink": ".v1", + "AgentV1SettingsAgentThinkOneItem": ".v1", + "AgentV1SettingsAgentThinkOneItemContextLength": ".v1", + "AgentV1SettingsAgentThinkOneItemContextLengthParams": ".v1", + "AgentV1SettingsAgentThinkOneItemEndpoint": ".v1", + "AgentV1SettingsAgentThinkOneItemEndpointParams": ".v1", + "AgentV1SettingsAgentThinkOneItemFunctionsItem": ".v1", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint": ".v1", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams": ".v1", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams": ".v1", + "AgentV1SettingsAgentThinkOneItemParams": ".v1", + "AgentV1SettingsAgentThinkOneItemProvider": ".v1", + "AgentV1SettingsAgentThinkOneItemProviderParams": ".v1", "AgentV1SettingsAgentThinkParams": ".v1", "AgentV1SettingsApplied": ".v1", "AgentV1SettingsAppliedParams": ".v1", + "AgentV1SettingsAppliedType": ".v1", "AgentV1SettingsAudio": ".v1", "AgentV1SettingsAudioInput": ".v1", "AgentV1SettingsAudioInputEncoding": ".v1", @@ -166,18 +470,118 @@ "AgentV1SettingsFlags": ".v1", "AgentV1SettingsFlagsParams": ".v1", "AgentV1SettingsParams": ".v1", + "AgentV1SettingsType": ".v1", "AgentV1SpeakUpdated": ".v1", "AgentV1SpeakUpdatedParams": ".v1", + "AgentV1SpeakUpdatedType": ".v1", "AgentV1UpdatePrompt": ".v1", "AgentV1UpdatePromptParams": ".v1", + "AgentV1UpdatePromptType": ".v1", "AgentV1UpdateSpeak": ".v1", "AgentV1UpdateSpeakParams": ".v1", + "AgentV1UpdateSpeakSpeak": ".v1", + "AgentV1UpdateSpeakSpeakEndpoint": ".v1", + "AgentV1UpdateSpeakSpeakEndpointEndpoint": ".v1", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProviderParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi": ".v1", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItem": ".v1", + "AgentV1UpdateSpeakSpeakOneItemEndpoint": ".v1", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProviderParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi": ".v1", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams": ".v1", + "AgentV1UpdateSpeakSpeakParams": ".v1", + "AgentV1UpdateSpeakType": ".v1", "AgentV1UserStartedSpeaking": ".v1", "AgentV1UserStartedSpeakingParams": ".v1", + "AgentV1UserStartedSpeakingType": ".v1", "AgentV1Warning": ".v1", "AgentV1WarningParams": ".v1", + "AgentV1WarningType": ".v1", "AgentV1Welcome": ".v1", "AgentV1WelcomeParams": ".v1", + "AgentV1WelcomeType": ".v1", + "Cartesia": ".v1", + "CartesiaParams": ".v1", + "Deepgram": ".v1", + "DeepgramParams": ".v1", + "ElevenLabs": ".v1", + "ElevenLabsParams": ".v1", + "Max": ".v1", "v1": ".v1", } @@ -206,33 +610,46 @@ def __dir__(): __all__ = [ "AgentV1AgentAudioDone", "AgentV1AgentAudioDoneParams", + "AgentV1AgentAudioDoneType", "AgentV1AgentStartedSpeaking", "AgentV1AgentStartedSpeakingParams", + "AgentV1AgentStartedSpeakingType", "AgentV1AgentThinking", "AgentV1AgentThinkingParams", + "AgentV1AgentThinkingType", "AgentV1ConversationText", "AgentV1ConversationTextParams", "AgentV1ConversationTextRole", + "AgentV1ConversationTextType", "AgentV1Error", "AgentV1ErrorParams", + "AgentV1ErrorType", "AgentV1FunctionCallRequest", "AgentV1FunctionCallRequestFunctionsItem", "AgentV1FunctionCallRequestFunctionsItemParams", "AgentV1FunctionCallRequestParams", + "AgentV1FunctionCallRequestType", "AgentV1InjectAgentMessage", "AgentV1InjectAgentMessageParams", + "AgentV1InjectAgentMessageType", "AgentV1InjectUserMessage", "AgentV1InjectUserMessageParams", + "AgentV1InjectUserMessageType", "AgentV1InjectionRefused", "AgentV1InjectionRefusedParams", + "AgentV1InjectionRefusedType", "AgentV1KeepAlive", "AgentV1KeepAliveParams", + "AgentV1KeepAliveType", "AgentV1PromptUpdated", "AgentV1PromptUpdatedParams", + "AgentV1PromptUpdatedType", "AgentV1ReceiveFunctionCallResponse", "AgentV1ReceiveFunctionCallResponseParams", + "AgentV1ReceiveFunctionCallResponseType", "AgentV1SendFunctionCallResponse", "AgentV1SendFunctionCallResponseParams", + "AgentV1SendFunctionCallResponseType", "AgentV1Settings", "AgentV1SettingsAgent", "AgentV1SettingsAgentContext", @@ -240,10 +657,12 @@ def __dir__(): "AgentV1SettingsAgentContextMessagesItemContent", "AgentV1SettingsAgentContextMessagesItemContentParams", "AgentV1SettingsAgentContextMessagesItemContentRole", + "AgentV1SettingsAgentContextMessagesItemContentType", "AgentV1SettingsAgentContextMessagesItemFunctionCalls", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams", "AgentV1SettingsAgentContextMessagesItemFunctionCallsParams", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType", "AgentV1SettingsAgentContextMessagesItemParams", "AgentV1SettingsAgentContextParams", "AgentV1SettingsAgentListen", @@ -252,19 +671,106 @@ def __dir__(): "AgentV1SettingsAgentListenProviderParams", "AgentV1SettingsAgentListenProviderV1", "AgentV1SettingsAgentListenProviderV1Params", + "AgentV1SettingsAgentListenProviderV1Type", "AgentV1SettingsAgentListenProviderV2", "AgentV1SettingsAgentListenProviderV2Params", + "AgentV1SettingsAgentListenProviderV2Type", "AgentV1SettingsAgentListenProvider_V1", "AgentV1SettingsAgentListenProvider_V1Params", "AgentV1SettingsAgentListenProvider_V2", "AgentV1SettingsAgentListenProvider_V2Params", "AgentV1SettingsAgentParams", "AgentV1SettingsAgentSpeak", + "AgentV1SettingsAgentSpeakEndpoint", + "AgentV1SettingsAgentSpeakEndpointEndpoint", + "AgentV1SettingsAgentSpeakEndpointEndpointParams", + "AgentV1SettingsAgentSpeakEndpointParams", + "AgentV1SettingsAgentSpeakEndpointProvider", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakEndpointProviderParams", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams", + "AgentV1SettingsAgentSpeakOneItem", + "AgentV1SettingsAgentSpeakOneItemEndpoint", + "AgentV1SettingsAgentSpeakOneItemEndpointParams", + "AgentV1SettingsAgentSpeakOneItemParams", + "AgentV1SettingsAgentSpeakOneItemProvider", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakOneItemProviderParams", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams", "AgentV1SettingsAgentSpeakParams", "AgentV1SettingsAgentThink", + "AgentV1SettingsAgentThinkOneItem", + "AgentV1SettingsAgentThinkOneItemContextLength", + "AgentV1SettingsAgentThinkOneItemContextLengthParams", + "AgentV1SettingsAgentThinkOneItemEndpoint", + "AgentV1SettingsAgentThinkOneItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItem", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams", + "AgentV1SettingsAgentThinkOneItemParams", + "AgentV1SettingsAgentThinkOneItemProvider", + "AgentV1SettingsAgentThinkOneItemProviderParams", "AgentV1SettingsAgentThinkParams", "AgentV1SettingsApplied", "AgentV1SettingsAppliedParams", + "AgentV1SettingsAppliedType", "AgentV1SettingsAudio", "AgentV1SettingsAudioInput", "AgentV1SettingsAudioInputEncoding", @@ -276,17 +782,117 @@ def __dir__(): "AgentV1SettingsFlags", "AgentV1SettingsFlagsParams", "AgentV1SettingsParams", + "AgentV1SettingsType", "AgentV1SpeakUpdated", "AgentV1SpeakUpdatedParams", + "AgentV1SpeakUpdatedType", "AgentV1UpdatePrompt", "AgentV1UpdatePromptParams", + "AgentV1UpdatePromptType", "AgentV1UpdateSpeak", "AgentV1UpdateSpeakParams", + "AgentV1UpdateSpeakSpeak", + "AgentV1UpdateSpeakSpeakEndpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointProvider", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakOneItem", + "AgentV1UpdateSpeakSpeakOneItemEndpoint", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams", + "AgentV1UpdateSpeakSpeakOneItemParams", + "AgentV1UpdateSpeakSpeakOneItemProvider", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakParams", + "AgentV1UpdateSpeakType", "AgentV1UserStartedSpeaking", "AgentV1UserStartedSpeakingParams", + "AgentV1UserStartedSpeakingType", "AgentV1Warning", "AgentV1WarningParams", + "AgentV1WarningType", "AgentV1Welcome", "AgentV1WelcomeParams", + "AgentV1WelcomeType", + "Cartesia", + "CartesiaParams", + "Deepgram", + "DeepgramParams", + "ElevenLabs", + "ElevenLabsParams", + "Max", "v1", ] diff --git a/src/deepgram/agent/v1/__init__.py b/src/deepgram/agent/v1/__init__.py index b7347cb8..c4fe1e2e 100644 --- a/src/deepgram/agent/v1/__init__.py +++ b/src/deepgram/agent/v1/__init__.py @@ -8,49 +8,187 @@ if typing.TYPE_CHECKING: from .types import ( AgentV1AgentAudioDone, + AgentV1AgentAudioDoneType, AgentV1AgentStartedSpeaking, + AgentV1AgentStartedSpeakingType, AgentV1AgentThinking, + AgentV1AgentThinkingType, AgentV1ConversationText, AgentV1ConversationTextRole, + AgentV1ConversationTextType, AgentV1Error, + AgentV1ErrorType, AgentV1FunctionCallRequest, AgentV1FunctionCallRequestFunctionsItem, + AgentV1FunctionCallRequestType, AgentV1InjectAgentMessage, + AgentV1InjectAgentMessageType, AgentV1InjectUserMessage, + AgentV1InjectUserMessageType, AgentV1InjectionRefused, + AgentV1InjectionRefusedType, AgentV1KeepAlive, + AgentV1KeepAliveType, AgentV1PromptUpdated, + AgentV1PromptUpdatedType, AgentV1ReceiveFunctionCallResponse, + AgentV1ReceiveFunctionCallResponseType, AgentV1SendFunctionCallResponse, + AgentV1SendFunctionCallResponseType, AgentV1Settings, AgentV1SettingsAgent, AgentV1SettingsAgentContext, AgentV1SettingsAgentContextMessagesItem, AgentV1SettingsAgentContextMessagesItemContent, AgentV1SettingsAgentContextMessagesItemContentRole, + AgentV1SettingsAgentContextMessagesItemContentType, AgentV1SettingsAgentContextMessagesItemFunctionCalls, AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem, + AgentV1SettingsAgentContextMessagesItemFunctionCallsType, AgentV1SettingsAgentListen, AgentV1SettingsAgentListenProvider, AgentV1SettingsAgentListenProviderV1, + AgentV1SettingsAgentListenProviderV1Type, AgentV1SettingsAgentListenProviderV2, + AgentV1SettingsAgentListenProviderV2Type, AgentV1SettingsAgentListenProvider_V1, AgentV1SettingsAgentListenProvider_V2, AgentV1SettingsAgentSpeak, + AgentV1SettingsAgentSpeakEndpoint, + AgentV1SettingsAgentSpeakEndpointEndpoint, + AgentV1SettingsAgentSpeakEndpointProvider, + AgentV1SettingsAgentSpeakEndpointProviderAwsPolly, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice, + AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel, + AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion, + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId, + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion, + AgentV1SettingsAgentSpeakEndpointProviderOpenAi, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly, + AgentV1SettingsAgentSpeakEndpointProvider_Cartesia, + AgentV1SettingsAgentSpeakEndpointProvider_Deepgram, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAi, + AgentV1SettingsAgentSpeakOneItem, + AgentV1SettingsAgentSpeakOneItemEndpoint, + AgentV1SettingsAgentSpeakOneItemProvider, + AgentV1SettingsAgentSpeakOneItemProviderAwsPolly, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, + AgentV1SettingsAgentSpeakOneItemProviderOpenAi, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly, + AgentV1SettingsAgentSpeakOneItemProvider_Cartesia, + AgentV1SettingsAgentSpeakOneItemProvider_Deepgram, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAi, AgentV1SettingsAgentThink, + AgentV1SettingsAgentThinkOneItem, + AgentV1SettingsAgentThinkOneItemContextLength, + AgentV1SettingsAgentThinkOneItemEndpoint, + AgentV1SettingsAgentThinkOneItemFunctionsItem, + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint, + AgentV1SettingsAgentThinkOneItemProvider, AgentV1SettingsApplied, + AgentV1SettingsAppliedType, AgentV1SettingsAudio, AgentV1SettingsAudioInput, AgentV1SettingsAudioInputEncoding, AgentV1SettingsAudioOutput, AgentV1SettingsAudioOutputEncoding, AgentV1SettingsFlags, + AgentV1SettingsType, AgentV1SpeakUpdated, + AgentV1SpeakUpdatedType, AgentV1UpdatePrompt, + AgentV1UpdatePromptType, AgentV1UpdateSpeak, + AgentV1UpdateSpeakSpeak, + AgentV1UpdateSpeakSpeakEndpoint, + AgentV1UpdateSpeakSpeakEndpointEndpoint, + AgentV1UpdateSpeakSpeakEndpointProvider, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, + AgentV1UpdateSpeakSpeakEndpointProviderCartesia, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgram, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAi, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly, + AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia, + AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi, + AgentV1UpdateSpeakSpeakOneItem, + AgentV1UpdateSpeakSpeakOneItemEndpoint, + AgentV1UpdateSpeakSpeakOneItemProvider, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, + AgentV1UpdateSpeakSpeakOneItemProviderCartesia, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgram, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAi, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly, + AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia, + AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi, + AgentV1UpdateSpeakType, AgentV1UserStartedSpeaking, + AgentV1UserStartedSpeakingType, AgentV1Warning, + AgentV1WarningType, AgentV1Welcome, + AgentV1WelcomeType, + Cartesia, + Deepgram, + ElevenLabs, + Max, ) from . import settings from .requests import ( @@ -80,7 +218,37 @@ AgentV1SettingsAgentListenProvider_V1Params, AgentV1SettingsAgentListenProvider_V2Params, AgentV1SettingsAgentParams, + AgentV1SettingsAgentSpeakEndpointEndpointParams, + AgentV1SettingsAgentSpeakEndpointParams, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams, + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams, + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams, + AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams, + AgentV1SettingsAgentSpeakEndpointProviderParams, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams, + AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams, + AgentV1SettingsAgentSpeakOneItemEndpointParams, + AgentV1SettingsAgentSpeakOneItemParams, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams, + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams, + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, + AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams, + AgentV1SettingsAgentSpeakOneItemProviderParams, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams, + AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams, AgentV1SettingsAgentSpeakParams, + AgentV1SettingsAgentThinkOneItemContextLengthParams, + AgentV1SettingsAgentThinkOneItemEndpointParams, + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams, + AgentV1SettingsAgentThinkOneItemFunctionsItemParams, + AgentV1SettingsAgentThinkOneItemParams, + AgentV1SettingsAgentThinkOneItemProviderParams, AgentV1SettingsAgentThinkParams, AgentV1SettingsAppliedParams, AgentV1SettingsAudioInputParams, @@ -91,40 +259,87 @@ AgentV1SpeakUpdatedParams, AgentV1UpdatePromptParams, AgentV1UpdateSpeakParams, + AgentV1UpdateSpeakSpeakEndpointEndpointParams, + AgentV1UpdateSpeakSpeakEndpointParams, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams, + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams, + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams, + AgentV1UpdateSpeakSpeakEndpointProviderParams, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakOneItemEndpointParams, + AgentV1UpdateSpeakSpeakOneItemParams, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams, + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams, + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams, + AgentV1UpdateSpeakSpeakOneItemProviderParams, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakParams, AgentV1UserStartedSpeakingParams, AgentV1WarningParams, AgentV1WelcomeParams, + CartesiaParams, + DeepgramParams, + ElevenLabsParams, ) _dynamic_imports: typing.Dict[str, str] = { "AgentV1AgentAudioDone": ".types", "AgentV1AgentAudioDoneParams": ".requests", + "AgentV1AgentAudioDoneType": ".types", "AgentV1AgentStartedSpeaking": ".types", "AgentV1AgentStartedSpeakingParams": ".requests", + "AgentV1AgentStartedSpeakingType": ".types", "AgentV1AgentThinking": ".types", "AgentV1AgentThinkingParams": ".requests", + "AgentV1AgentThinkingType": ".types", "AgentV1ConversationText": ".types", "AgentV1ConversationTextParams": ".requests", "AgentV1ConversationTextRole": ".types", + "AgentV1ConversationTextType": ".types", "AgentV1Error": ".types", "AgentV1ErrorParams": ".requests", + "AgentV1ErrorType": ".types", "AgentV1FunctionCallRequest": ".types", "AgentV1FunctionCallRequestFunctionsItem": ".types", "AgentV1FunctionCallRequestFunctionsItemParams": ".requests", "AgentV1FunctionCallRequestParams": ".requests", + "AgentV1FunctionCallRequestType": ".types", "AgentV1InjectAgentMessage": ".types", "AgentV1InjectAgentMessageParams": ".requests", + "AgentV1InjectAgentMessageType": ".types", "AgentV1InjectUserMessage": ".types", "AgentV1InjectUserMessageParams": ".requests", + "AgentV1InjectUserMessageType": ".types", "AgentV1InjectionRefused": ".types", "AgentV1InjectionRefusedParams": ".requests", + "AgentV1InjectionRefusedType": ".types", "AgentV1KeepAlive": ".types", "AgentV1KeepAliveParams": ".requests", + "AgentV1KeepAliveType": ".types", "AgentV1PromptUpdated": ".types", "AgentV1PromptUpdatedParams": ".requests", + "AgentV1PromptUpdatedType": ".types", "AgentV1ReceiveFunctionCallResponse": ".types", "AgentV1ReceiveFunctionCallResponseParams": ".requests", + "AgentV1ReceiveFunctionCallResponseType": ".types", "AgentV1SendFunctionCallResponse": ".types", "AgentV1SendFunctionCallResponseParams": ".requests", + "AgentV1SendFunctionCallResponseType": ".types", "AgentV1Settings": ".types", "AgentV1SettingsAgent": ".types", "AgentV1SettingsAgentContext": ".types", @@ -132,10 +347,12 @@ "AgentV1SettingsAgentContextMessagesItemContent": ".types", "AgentV1SettingsAgentContextMessagesItemContentParams": ".requests", "AgentV1SettingsAgentContextMessagesItemContentRole": ".types", + "AgentV1SettingsAgentContextMessagesItemContentType": ".types", "AgentV1SettingsAgentContextMessagesItemFunctionCalls": ".types", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem": ".types", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams": ".requests", "AgentV1SettingsAgentContextMessagesItemFunctionCallsParams": ".requests", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType": ".types", "AgentV1SettingsAgentContextMessagesItemParams": ".requests", "AgentV1SettingsAgentContextParams": ".requests", "AgentV1SettingsAgentListen": ".types", @@ -144,19 +361,106 @@ "AgentV1SettingsAgentListenProviderParams": ".requests", "AgentV1SettingsAgentListenProviderV1": ".types", "AgentV1SettingsAgentListenProviderV1Params": ".requests", + "AgentV1SettingsAgentListenProviderV1Type": ".types", "AgentV1SettingsAgentListenProviderV2": ".types", "AgentV1SettingsAgentListenProviderV2Params": ".requests", + "AgentV1SettingsAgentListenProviderV2Type": ".types", "AgentV1SettingsAgentListenProvider_V1": ".types", "AgentV1SettingsAgentListenProvider_V1Params": ".requests", "AgentV1SettingsAgentListenProvider_V2": ".types", "AgentV1SettingsAgentListenProvider_V2Params": ".requests", "AgentV1SettingsAgentParams": ".requests", "AgentV1SettingsAgentSpeak": ".types", + "AgentV1SettingsAgentSpeakEndpoint": ".types", + "AgentV1SettingsAgentSpeakEndpointEndpoint": ".types", + "AgentV1SettingsAgentSpeakEndpointEndpointParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice": ".types", + "AgentV1SettingsAgentSpeakEndpointProviderParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly": ".types", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia": ".types", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram": ".types", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs": ".types", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams": ".requests", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi": ".types", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams": ".requests", + "AgentV1SettingsAgentSpeakOneItem": ".types", + "AgentV1SettingsAgentSpeakOneItemEndpoint": ".types", + "AgentV1SettingsAgentSpeakOneItemEndpointParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice": ".types", + "AgentV1SettingsAgentSpeakOneItemProviderParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly": ".types", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia": ".types", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram": ".types", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs": ".types", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams": ".requests", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi": ".types", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams": ".requests", "AgentV1SettingsAgentSpeakParams": ".requests", "AgentV1SettingsAgentThink": ".types", + "AgentV1SettingsAgentThinkOneItem": ".types", + "AgentV1SettingsAgentThinkOneItemContextLength": ".types", + "AgentV1SettingsAgentThinkOneItemContextLengthParams": ".requests", + "AgentV1SettingsAgentThinkOneItemEndpoint": ".types", + "AgentV1SettingsAgentThinkOneItemEndpointParams": ".requests", + "AgentV1SettingsAgentThinkOneItemFunctionsItem": ".types", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint": ".types", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams": ".requests", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams": ".requests", + "AgentV1SettingsAgentThinkOneItemParams": ".requests", + "AgentV1SettingsAgentThinkOneItemProvider": ".types", + "AgentV1SettingsAgentThinkOneItemProviderParams": ".requests", "AgentV1SettingsAgentThinkParams": ".requests", "AgentV1SettingsApplied": ".types", "AgentV1SettingsAppliedParams": ".requests", + "AgentV1SettingsAppliedType": ".types", "AgentV1SettingsAudio": ".types", "AgentV1SettingsAudioInput": ".types", "AgentV1SettingsAudioInputEncoding": ".types", @@ -168,18 +472,118 @@ "AgentV1SettingsFlags": ".types", "AgentV1SettingsFlagsParams": ".requests", "AgentV1SettingsParams": ".requests", + "AgentV1SettingsType": ".types", "AgentV1SpeakUpdated": ".types", "AgentV1SpeakUpdatedParams": ".requests", + "AgentV1SpeakUpdatedType": ".types", "AgentV1UpdatePrompt": ".types", "AgentV1UpdatePromptParams": ".requests", + "AgentV1UpdatePromptType": ".types", "AgentV1UpdateSpeak": ".types", "AgentV1UpdateSpeakParams": ".requests", + "AgentV1UpdateSpeakSpeak": ".types", + "AgentV1UpdateSpeakSpeakEndpoint": ".types", + "AgentV1UpdateSpeakSpeakEndpointEndpoint": ".types", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice": ".types", + "AgentV1UpdateSpeakSpeakEndpointProviderParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly": ".types", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia": ".types", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram": ".types", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs": ".types", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams": ".requests", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi": ".types", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItem": ".types", + "AgentV1UpdateSpeakSpeakOneItemEndpoint": ".types", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice": ".types", + "AgentV1UpdateSpeakSpeakOneItemProviderParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly": ".types", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia": ".types", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram": ".types", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs": ".types", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams": ".requests", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi": ".types", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams": ".requests", + "AgentV1UpdateSpeakSpeakParams": ".requests", + "AgentV1UpdateSpeakType": ".types", "AgentV1UserStartedSpeaking": ".types", "AgentV1UserStartedSpeakingParams": ".requests", + "AgentV1UserStartedSpeakingType": ".types", "AgentV1Warning": ".types", "AgentV1WarningParams": ".requests", + "AgentV1WarningType": ".types", "AgentV1Welcome": ".types", "AgentV1WelcomeParams": ".requests", + "AgentV1WelcomeType": ".types", + "Cartesia": ".types", + "CartesiaParams": ".requests", + "Deepgram": ".types", + "DeepgramParams": ".requests", + "ElevenLabs": ".types", + "ElevenLabsParams": ".requests", + "Max": ".types", "settings": ".settings", } @@ -208,33 +612,46 @@ def __dir__(): __all__ = [ "AgentV1AgentAudioDone", "AgentV1AgentAudioDoneParams", + "AgentV1AgentAudioDoneType", "AgentV1AgentStartedSpeaking", "AgentV1AgentStartedSpeakingParams", + "AgentV1AgentStartedSpeakingType", "AgentV1AgentThinking", "AgentV1AgentThinkingParams", + "AgentV1AgentThinkingType", "AgentV1ConversationText", "AgentV1ConversationTextParams", "AgentV1ConversationTextRole", + "AgentV1ConversationTextType", "AgentV1Error", "AgentV1ErrorParams", + "AgentV1ErrorType", "AgentV1FunctionCallRequest", "AgentV1FunctionCallRequestFunctionsItem", "AgentV1FunctionCallRequestFunctionsItemParams", "AgentV1FunctionCallRequestParams", + "AgentV1FunctionCallRequestType", "AgentV1InjectAgentMessage", "AgentV1InjectAgentMessageParams", + "AgentV1InjectAgentMessageType", "AgentV1InjectUserMessage", "AgentV1InjectUserMessageParams", + "AgentV1InjectUserMessageType", "AgentV1InjectionRefused", "AgentV1InjectionRefusedParams", + "AgentV1InjectionRefusedType", "AgentV1KeepAlive", "AgentV1KeepAliveParams", + "AgentV1KeepAliveType", "AgentV1PromptUpdated", "AgentV1PromptUpdatedParams", + "AgentV1PromptUpdatedType", "AgentV1ReceiveFunctionCallResponse", "AgentV1ReceiveFunctionCallResponseParams", + "AgentV1ReceiveFunctionCallResponseType", "AgentV1SendFunctionCallResponse", "AgentV1SendFunctionCallResponseParams", + "AgentV1SendFunctionCallResponseType", "AgentV1Settings", "AgentV1SettingsAgent", "AgentV1SettingsAgentContext", @@ -242,10 +659,12 @@ def __dir__(): "AgentV1SettingsAgentContextMessagesItemContent", "AgentV1SettingsAgentContextMessagesItemContentParams", "AgentV1SettingsAgentContextMessagesItemContentRole", + "AgentV1SettingsAgentContextMessagesItemContentType", "AgentV1SettingsAgentContextMessagesItemFunctionCalls", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams", "AgentV1SettingsAgentContextMessagesItemFunctionCallsParams", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType", "AgentV1SettingsAgentContextMessagesItemParams", "AgentV1SettingsAgentContextParams", "AgentV1SettingsAgentListen", @@ -254,19 +673,106 @@ def __dir__(): "AgentV1SettingsAgentListenProviderParams", "AgentV1SettingsAgentListenProviderV1", "AgentV1SettingsAgentListenProviderV1Params", + "AgentV1SettingsAgentListenProviderV1Type", "AgentV1SettingsAgentListenProviderV2", "AgentV1SettingsAgentListenProviderV2Params", + "AgentV1SettingsAgentListenProviderV2Type", "AgentV1SettingsAgentListenProvider_V1", "AgentV1SettingsAgentListenProvider_V1Params", "AgentV1SettingsAgentListenProvider_V2", "AgentV1SettingsAgentListenProvider_V2Params", "AgentV1SettingsAgentParams", "AgentV1SettingsAgentSpeak", + "AgentV1SettingsAgentSpeakEndpoint", + "AgentV1SettingsAgentSpeakEndpointEndpoint", + "AgentV1SettingsAgentSpeakEndpointEndpointParams", + "AgentV1SettingsAgentSpeakEndpointParams", + "AgentV1SettingsAgentSpeakEndpointProvider", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakEndpointProviderParams", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams", + "AgentV1SettingsAgentSpeakOneItem", + "AgentV1SettingsAgentSpeakOneItemEndpoint", + "AgentV1SettingsAgentSpeakOneItemEndpointParams", + "AgentV1SettingsAgentSpeakOneItemParams", + "AgentV1SettingsAgentSpeakOneItemProvider", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakOneItemProviderParams", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams", "AgentV1SettingsAgentSpeakParams", "AgentV1SettingsAgentThink", + "AgentV1SettingsAgentThinkOneItem", + "AgentV1SettingsAgentThinkOneItemContextLength", + "AgentV1SettingsAgentThinkOneItemContextLengthParams", + "AgentV1SettingsAgentThinkOneItemEndpoint", + "AgentV1SettingsAgentThinkOneItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItem", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams", + "AgentV1SettingsAgentThinkOneItemParams", + "AgentV1SettingsAgentThinkOneItemProvider", + "AgentV1SettingsAgentThinkOneItemProviderParams", "AgentV1SettingsAgentThinkParams", "AgentV1SettingsApplied", "AgentV1SettingsAppliedParams", + "AgentV1SettingsAppliedType", "AgentV1SettingsAudio", "AgentV1SettingsAudioInput", "AgentV1SettingsAudioInputEncoding", @@ -278,17 +784,117 @@ def __dir__(): "AgentV1SettingsFlags", "AgentV1SettingsFlagsParams", "AgentV1SettingsParams", + "AgentV1SettingsType", "AgentV1SpeakUpdated", "AgentV1SpeakUpdatedParams", + "AgentV1SpeakUpdatedType", "AgentV1UpdatePrompt", "AgentV1UpdatePromptParams", + "AgentV1UpdatePromptType", "AgentV1UpdateSpeak", "AgentV1UpdateSpeakParams", + "AgentV1UpdateSpeakSpeak", + "AgentV1UpdateSpeakSpeakEndpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointProvider", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakOneItem", + "AgentV1UpdateSpeakSpeakOneItemEndpoint", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams", + "AgentV1UpdateSpeakSpeakOneItemParams", + "AgentV1UpdateSpeakSpeakOneItemProvider", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakParams", + "AgentV1UpdateSpeakType", "AgentV1UserStartedSpeaking", "AgentV1UserStartedSpeakingParams", + "AgentV1UserStartedSpeakingType", "AgentV1Warning", "AgentV1WarningParams", + "AgentV1WarningType", "AgentV1Welcome", "AgentV1WelcomeParams", + "AgentV1WelcomeType", + "Cartesia", + "CartesiaParams", + "Deepgram", + "DeepgramParams", + "ElevenLabs", + "ElevenLabsParams", + "Max", "settings", ] diff --git a/src/deepgram/agent/v1/client.py b/src/deepgram/agent/v1/client.py index 6f7d6514..e1a53d78 100644 --- a/src/deepgram/agent/v1/client.py +++ b/src/deepgram/agent/v1/client.py @@ -53,9 +53,9 @@ def connect( Parameters ---------- authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -138,9 +138,9 @@ async def connect( Parameters ---------- authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/agent/v1/raw_client.py b/src/deepgram/agent/v1/raw_client.py index ab89d67e..10f2330c 100644 --- a/src/deepgram/agent/v1/raw_client.py +++ b/src/deepgram/agent/v1/raw_client.py @@ -34,9 +34,9 @@ def connect( Parameters ---------- authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -98,9 +98,9 @@ async def connect( Parameters ---------- authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/agent/v1/requests/__init__.py b/src/deepgram/agent/v1/requests/__init__.py index 9708fd24..c3ace7bb 100644 --- a/src/deepgram/agent/v1/requests/__init__.py +++ b/src/deepgram/agent/v1/requests/__init__.py @@ -42,7 +42,63 @@ from .agent_v1settings_agent_listen_provider_v1 import AgentV1SettingsAgentListenProviderV1Params from .agent_v1settings_agent_listen_provider_v2 import AgentV1SettingsAgentListenProviderV2Params from .agent_v1settings_agent_speak import AgentV1SettingsAgentSpeakParams + from .agent_v1settings_agent_speak_endpoint import AgentV1SettingsAgentSpeakEndpointParams + from .agent_v1settings_agent_speak_endpoint_endpoint import AgentV1SettingsAgentSpeakEndpointEndpointParams + from .agent_v1settings_agent_speak_endpoint_provider import ( + AgentV1SettingsAgentSpeakEndpointProviderParams, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams, + AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams, + ) + from .agent_v1settings_agent_speak_endpoint_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams, + ) + from .agent_v1settings_agent_speak_endpoint_provider_open_ai import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams, + ) + from .agent_v1settings_agent_speak_one_item import AgentV1SettingsAgentSpeakOneItemParams + from .agent_v1settings_agent_speak_one_item_endpoint import AgentV1SettingsAgentSpeakOneItemEndpointParams + from .agent_v1settings_agent_speak_one_item_provider import ( + AgentV1SettingsAgentSpeakOneItemProviderParams, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams, + AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams, + AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams, + ) + from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, + ) + from .agent_v1settings_agent_speak_one_item_provider_open_ai import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams, + ) from .agent_v1settings_agent_think import AgentV1SettingsAgentThinkParams + from .agent_v1settings_agent_think_one_item import AgentV1SettingsAgentThinkOneItemParams + from .agent_v1settings_agent_think_one_item_context_length import ( + AgentV1SettingsAgentThinkOneItemContextLengthParams, + ) + from .agent_v1settings_agent_think_one_item_endpoint import AgentV1SettingsAgentThinkOneItemEndpointParams + from .agent_v1settings_agent_think_one_item_functions_item import ( + AgentV1SettingsAgentThinkOneItemFunctionsItemParams, + ) + from .agent_v1settings_agent_think_one_item_functions_item_endpoint import ( + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams, + ) + from .agent_v1settings_agent_think_one_item_provider import AgentV1SettingsAgentThinkOneItemProviderParams from .agent_v1settings_applied import AgentV1SettingsAppliedParams from .agent_v1settings_audio import AgentV1SettingsAudioParams from .agent_v1settings_audio_input import AgentV1SettingsAudioInputParams @@ -51,9 +107,73 @@ from .agent_v1speak_updated import AgentV1SpeakUpdatedParams from .agent_v1update_prompt import AgentV1UpdatePromptParams from .agent_v1update_speak import AgentV1UpdateSpeakParams + from .agent_v1update_speak_speak import AgentV1UpdateSpeakSpeakParams + from .agent_v1update_speak_speak_endpoint import AgentV1UpdateSpeakSpeakEndpointParams + from .agent_v1update_speak_speak_endpoint_endpoint import AgentV1UpdateSpeakSpeakEndpointEndpointParams + from .agent_v1update_speak_speak_endpoint_provider import ( + AgentV1UpdateSpeakSpeakEndpointProviderParams, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_cartesia import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_deepgram import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_eleven_labs import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams, + ) + from .agent_v1update_speak_speak_endpoint_provider_open_ai import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams, + ) + from .agent_v1update_speak_speak_one_item import AgentV1UpdateSpeakSpeakOneItemParams + from .agent_v1update_speak_speak_one_item_endpoint import AgentV1UpdateSpeakSpeakOneItemEndpointParams + from .agent_v1update_speak_speak_one_item_provider import ( + AgentV1UpdateSpeakSpeakOneItemProviderParams, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams, + AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams, + ) + from .agent_v1update_speak_speak_one_item_provider_cartesia import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams, + ) + from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams, + ) + from .agent_v1update_speak_speak_one_item_provider_deepgram import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams, + ) + from .agent_v1update_speak_speak_one_item_provider_eleven_labs import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams, + ) + from .agent_v1update_speak_speak_one_item_provider_open_ai import AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams from .agent_v1user_started_speaking import AgentV1UserStartedSpeakingParams from .agent_v1warning import AgentV1WarningParams from .agent_v1welcome import AgentV1WelcomeParams + from .cartesia import CartesiaParams + from .deepgram import DeepgramParams + from .eleven_labs import ElevenLabsParams _dynamic_imports: typing.Dict[str, str] = { "AgentV1AgentAudioDoneParams": ".agent_v1agent_audio_done", "AgentV1AgentStartedSpeakingParams": ".agent_v1agent_started_speaking", @@ -81,7 +201,37 @@ "AgentV1SettingsAgentListenProvider_V1Params": ".agent_v1settings_agent_listen_provider", "AgentV1SettingsAgentListenProvider_V2Params": ".agent_v1settings_agent_listen_provider", "AgentV1SettingsAgentParams": ".agent_v1settings_agent", + "AgentV1SettingsAgentSpeakEndpointEndpointParams": ".agent_v1settings_agent_speak_endpoint_endpoint", + "AgentV1SettingsAgentSpeakEndpointParams": ".agent_v1settings_agent_speak_endpoint", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams": ".agent_v1settings_agent_speak_endpoint_provider_cartesia_voice", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams": ".agent_v1settings_agent_speak_endpoint_provider_open_ai", + "AgentV1SettingsAgentSpeakEndpointProviderParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakOneItemEndpointParams": ".agent_v1settings_agent_speak_one_item_endpoint", + "AgentV1SettingsAgentSpeakOneItemParams": ".agent_v1settings_agent_speak_one_item", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams": ".agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams": ".agent_v1settings_agent_speak_one_item_provider_aws_polly", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams": ".agent_v1settings_agent_speak_one_item_provider_cartesia_voice", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams": ".agent_v1settings_agent_speak_one_item_provider_open_ai", + "AgentV1SettingsAgentSpeakOneItemProviderParams": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams": ".agent_v1settings_agent_speak_one_item_provider", "AgentV1SettingsAgentSpeakParams": ".agent_v1settings_agent_speak", + "AgentV1SettingsAgentThinkOneItemContextLengthParams": ".agent_v1settings_agent_think_one_item_context_length", + "AgentV1SettingsAgentThinkOneItemEndpointParams": ".agent_v1settings_agent_think_one_item_endpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams": ".agent_v1settings_agent_think_one_item_functions_item_endpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams": ".agent_v1settings_agent_think_one_item_functions_item", + "AgentV1SettingsAgentThinkOneItemParams": ".agent_v1settings_agent_think_one_item", + "AgentV1SettingsAgentThinkOneItemProviderParams": ".agent_v1settings_agent_think_one_item_provider", "AgentV1SettingsAgentThinkParams": ".agent_v1settings_agent_think", "AgentV1SettingsAppliedParams": ".agent_v1settings_applied", "AgentV1SettingsAudioInputParams": ".agent_v1settings_audio_input", @@ -92,9 +242,43 @@ "AgentV1SpeakUpdatedParams": ".agent_v1speak_updated", "AgentV1UpdatePromptParams": ".agent_v1update_prompt", "AgentV1UpdateSpeakParams": ".agent_v1update_speak", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams": ".agent_v1update_speak_speak_endpoint_endpoint", + "AgentV1UpdateSpeakSpeakEndpointParams": ".agent_v1update_speak_speak_endpoint", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams": ".agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams": ".agent_v1update_speak_speak_endpoint_provider_aws_polly", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams": ".agent_v1update_speak_speak_endpoint_provider_cartesia", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams": ".agent_v1update_speak_speak_endpoint_provider_cartesia_voice", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams": ".agent_v1update_speak_speak_endpoint_provider_deepgram", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams": ".agent_v1update_speak_speak_endpoint_provider_eleven_labs", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams": ".agent_v1update_speak_speak_endpoint_provider_open_ai", + "AgentV1UpdateSpeakSpeakEndpointProviderParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams": ".agent_v1update_speak_speak_one_item_endpoint", + "AgentV1UpdateSpeakSpeakOneItemParams": ".agent_v1update_speak_speak_one_item", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams": ".agent_v1update_speak_speak_one_item_provider_aws_polly_credentials", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams": ".agent_v1update_speak_speak_one_item_provider_aws_polly", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams": ".agent_v1update_speak_speak_one_item_provider_cartesia", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams": ".agent_v1update_speak_speak_one_item_provider_cartesia_voice", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams": ".agent_v1update_speak_speak_one_item_provider_deepgram", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams": ".agent_v1update_speak_speak_one_item_provider_eleven_labs", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams": ".agent_v1update_speak_speak_one_item_provider_open_ai", + "AgentV1UpdateSpeakSpeakOneItemProviderParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakParams": ".agent_v1update_speak_speak", "AgentV1UserStartedSpeakingParams": ".agent_v1user_started_speaking", "AgentV1WarningParams": ".agent_v1warning", "AgentV1WelcomeParams": ".agent_v1welcome", + "CartesiaParams": ".cartesia", + "DeepgramParams": ".deepgram", + "ElevenLabsParams": ".eleven_labs", } @@ -146,7 +330,37 @@ def __dir__(): "AgentV1SettingsAgentListenProvider_V1Params", "AgentV1SettingsAgentListenProvider_V2Params", "AgentV1SettingsAgentParams", + "AgentV1SettingsAgentSpeakEndpointEndpointParams", + "AgentV1SettingsAgentSpeakEndpointParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams", + "AgentV1SettingsAgentSpeakEndpointProviderParams", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams", + "AgentV1SettingsAgentSpeakOneItemEndpointParams", + "AgentV1SettingsAgentSpeakOneItemParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams", + "AgentV1SettingsAgentSpeakOneItemProviderParams", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams", + "AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams", + "AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams", "AgentV1SettingsAgentSpeakParams", + "AgentV1SettingsAgentThinkOneItemContextLengthParams", + "AgentV1SettingsAgentThinkOneItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams", + "AgentV1SettingsAgentThinkOneItemFunctionsItemParams", + "AgentV1SettingsAgentThinkOneItemParams", + "AgentV1SettingsAgentThinkOneItemProviderParams", "AgentV1SettingsAgentThinkParams", "AgentV1SettingsAppliedParams", "AgentV1SettingsAudioInputParams", @@ -157,7 +371,41 @@ def __dir__(): "AgentV1SpeakUpdatedParams", "AgentV1UpdatePromptParams", "AgentV1UpdateSpeakParams", + "AgentV1UpdateSpeakSpeakEndpointEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakEndpointProviderParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakOneItemEndpointParams", + "AgentV1UpdateSpeakSpeakOneItemParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams", + "AgentV1UpdateSpeakSpeakOneItemProviderParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams", + "AgentV1UpdateSpeakSpeakParams", "AgentV1UserStartedSpeakingParams", "AgentV1WarningParams", "AgentV1WelcomeParams", + "CartesiaParams", + "DeepgramParams", + "ElevenLabsParams", ] diff --git a/src/deepgram/agent/v1/requests/agent_v1agent_audio_done.py b/src/deepgram/agent/v1/requests/agent_v1agent_audio_done.py index 43b4f013..017b2c89 100644 --- a/src/deepgram/agent/v1/requests/agent_v1agent_audio_done.py +++ b/src/deepgram/agent/v1/requests/agent_v1agent_audio_done.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1agent_audio_done_type import AgentV1AgentAudioDoneType class AgentV1AgentAudioDoneParams(typing_extensions.TypedDict): - type: typing.Literal["AgentAudioDone"] + type: AgentV1AgentAudioDoneType """ Message type identifier indicating the agent has finished sending audio """ diff --git a/src/deepgram/agent/v1/requests/agent_v1agent_started_speaking.py b/src/deepgram/agent/v1/requests/agent_v1agent_started_speaking.py index 39861c94..85d820c3 100644 --- a/src/deepgram/agent/v1/requests/agent_v1agent_started_speaking.py +++ b/src/deepgram/agent/v1/requests/agent_v1agent_started_speaking.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1agent_started_speaking_type import AgentV1AgentStartedSpeakingType class AgentV1AgentStartedSpeakingParams(typing_extensions.TypedDict): - type: typing.Literal["AgentStartedSpeaking"] + type: AgentV1AgentStartedSpeakingType """ Message type identifier for agent started speaking """ diff --git a/src/deepgram/agent/v1/requests/agent_v1agent_thinking.py b/src/deepgram/agent/v1/requests/agent_v1agent_thinking.py index 13434cbc..70bb9041 100644 --- a/src/deepgram/agent/v1/requests/agent_v1agent_thinking.py +++ b/src/deepgram/agent/v1/requests/agent_v1agent_thinking.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1agent_thinking_type import AgentV1AgentThinkingType class AgentV1AgentThinkingParams(typing_extensions.TypedDict): - type: typing.Literal["AgentThinking"] + type: AgentV1AgentThinkingType """ Message type identifier for agent thinking """ diff --git a/src/deepgram/agent/v1/requests/agent_v1conversation_text.py b/src/deepgram/agent/v1/requests/agent_v1conversation_text.py index ea0601e3..40534694 100644 --- a/src/deepgram/agent/v1/requests/agent_v1conversation_text.py +++ b/src/deepgram/agent/v1/requests/agent_v1conversation_text.py @@ -1,13 +1,12 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions from ..types.agent_v1conversation_text_role import AgentV1ConversationTextRole +from ..types.agent_v1conversation_text_type import AgentV1ConversationTextType class AgentV1ConversationTextParams(typing_extensions.TypedDict): - type: typing.Literal["ConversationText"] + type: AgentV1ConversationTextType """ Message type identifier for conversation text """ diff --git a/src/deepgram/agent/v1/requests/agent_v1error.py b/src/deepgram/agent/v1/requests/agent_v1error.py index 23547cb7..f85ab797 100644 --- a/src/deepgram/agent/v1/requests/agent_v1error.py +++ b/src/deepgram/agent/v1/requests/agent_v1error.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1error_type import AgentV1ErrorType class AgentV1ErrorParams(typing_extensions.TypedDict): - type: typing.Literal["Error"] + type: AgentV1ErrorType """ Message type identifier for error responses """ diff --git a/src/deepgram/agent/v1/requests/agent_v1function_call_request.py b/src/deepgram/agent/v1/requests/agent_v1function_call_request.py index b00cc6d4..f65aac14 100644 --- a/src/deepgram/agent/v1/requests/agent_v1function_call_request.py +++ b/src/deepgram/agent/v1/requests/agent_v1function_call_request.py @@ -3,11 +3,12 @@ import typing import typing_extensions +from ..types.agent_v1function_call_request_type import AgentV1FunctionCallRequestType from .agent_v1function_call_request_functions_item import AgentV1FunctionCallRequestFunctionsItemParams class AgentV1FunctionCallRequestParams(typing_extensions.TypedDict): - type: typing.Literal["FunctionCallRequest"] + type: AgentV1FunctionCallRequestType """ Message type identifier for function call requests """ diff --git a/src/deepgram/agent/v1/requests/agent_v1inject_agent_message.py b/src/deepgram/agent/v1/requests/agent_v1inject_agent_message.py index 8fb718bd..8ef4156d 100644 --- a/src/deepgram/agent/v1/requests/agent_v1inject_agent_message.py +++ b/src/deepgram/agent/v1/requests/agent_v1inject_agent_message.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1inject_agent_message_type import AgentV1InjectAgentMessageType class AgentV1InjectAgentMessageParams(typing_extensions.TypedDict): - type: typing.Literal["InjectAgentMessage"] + type: AgentV1InjectAgentMessageType """ Message type identifier for injecting an agent message """ diff --git a/src/deepgram/agent/v1/requests/agent_v1inject_user_message.py b/src/deepgram/agent/v1/requests/agent_v1inject_user_message.py index 86583a81..ac62ca64 100644 --- a/src/deepgram/agent/v1/requests/agent_v1inject_user_message.py +++ b/src/deepgram/agent/v1/requests/agent_v1inject_user_message.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1inject_user_message_type import AgentV1InjectUserMessageType class AgentV1InjectUserMessageParams(typing_extensions.TypedDict): - type: typing.Literal["InjectUserMessage"] + type: AgentV1InjectUserMessageType """ Message type identifier for injecting a user message """ diff --git a/src/deepgram/agent/v1/requests/agent_v1injection_refused.py b/src/deepgram/agent/v1/requests/agent_v1injection_refused.py index e19f3241..639766fe 100644 --- a/src/deepgram/agent/v1/requests/agent_v1injection_refused.py +++ b/src/deepgram/agent/v1/requests/agent_v1injection_refused.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1injection_refused_type import AgentV1InjectionRefusedType class AgentV1InjectionRefusedParams(typing_extensions.TypedDict): - type: typing.Literal["InjectionRefused"] + type: AgentV1InjectionRefusedType """ Message type identifier for injection refused """ diff --git a/src/deepgram/agent/v1/requests/agent_v1keep_alive.py b/src/deepgram/agent/v1/requests/agent_v1keep_alive.py index 125eb8ae..6326cc3f 100644 --- a/src/deepgram/agent/v1/requests/agent_v1keep_alive.py +++ b/src/deepgram/agent/v1/requests/agent_v1keep_alive.py @@ -1,8 +1,7 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1keep_alive_type import AgentV1KeepAliveType class AgentV1KeepAliveParams(typing_extensions.TypedDict): @@ -10,7 +9,7 @@ class AgentV1KeepAliveParams(typing_extensions.TypedDict): Send a control message to the agent """ - type: typing.Literal["KeepAlive"] + type: AgentV1KeepAliveType """ Message type identifier """ diff --git a/src/deepgram/agent/v1/requests/agent_v1prompt_updated.py b/src/deepgram/agent/v1/requests/agent_v1prompt_updated.py index 40d5a426..7d471c2a 100644 --- a/src/deepgram/agent/v1/requests/agent_v1prompt_updated.py +++ b/src/deepgram/agent/v1/requests/agent_v1prompt_updated.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1prompt_updated_type import AgentV1PromptUpdatedType class AgentV1PromptUpdatedParams(typing_extensions.TypedDict): - type: typing.Literal["PromptUpdated"] + type: AgentV1PromptUpdatedType """ Message type identifier for prompt update confirmation """ diff --git a/src/deepgram/agent/v1/requests/agent_v1receive_function_call_response.py b/src/deepgram/agent/v1/requests/agent_v1receive_function_call_response.py index f26fda70..988bafbf 100644 --- a/src/deepgram/agent/v1/requests/agent_v1receive_function_call_response.py +++ b/src/deepgram/agent/v1/requests/agent_v1receive_function_call_response.py @@ -1,8 +1,7 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1receive_function_call_response_type import AgentV1ReceiveFunctionCallResponseType class AgentV1ReceiveFunctionCallResponseParams(typing_extensions.TypedDict): @@ -18,7 +17,7 @@ class AgentV1ReceiveFunctionCallResponseParams(typing_extensions.TypedDict): interface for function call responses regardless of execution location. """ - type: typing.Literal["FunctionCallResponse"] + type: AgentV1ReceiveFunctionCallResponseType """ Message type identifier for function call responses """ diff --git a/src/deepgram/agent/v1/requests/agent_v1send_function_call_response.py b/src/deepgram/agent/v1/requests/agent_v1send_function_call_response.py index 8d75edeb..87d13ae3 100644 --- a/src/deepgram/agent/v1/requests/agent_v1send_function_call_response.py +++ b/src/deepgram/agent/v1/requests/agent_v1send_function_call_response.py @@ -1,8 +1,7 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1send_function_call_response_type import AgentV1SendFunctionCallResponseType class AgentV1SendFunctionCallResponseParams(typing_extensions.TypedDict): @@ -18,7 +17,7 @@ class AgentV1SendFunctionCallResponseParams(typing_extensions.TypedDict): interface for function call responses regardless of execution location. """ - type: typing.Literal["FunctionCallResponse"] + type: AgentV1SendFunctionCallResponseType """ Message type identifier for function call responses """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings.py b/src/deepgram/agent/v1/requests/agent_v1settings.py index 2f748dcd..a1da08ec 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings.py @@ -3,13 +3,14 @@ import typing import typing_extensions +from ..types.agent_v1settings_type import AgentV1SettingsType from .agent_v1settings_agent import AgentV1SettingsAgentParams from .agent_v1settings_audio import AgentV1SettingsAudioParams from .agent_v1settings_flags import AgentV1SettingsFlagsParams class AgentV1SettingsParams(typing_extensions.TypedDict): - type: typing.Literal["Settings"] + type: AgentV1SettingsType tags: typing_extensions.NotRequired[typing.Sequence[str]] """ Tags to associate with the request diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_content.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_content.py index 1a541ffc..cbb225de 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_content.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_content.py @@ -1,11 +1,12 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions from ..types.agent_v1settings_agent_context_messages_item_content_role import ( AgentV1SettingsAgentContextMessagesItemContentRole, ) +from ..types.agent_v1settings_agent_context_messages_item_content_type import ( + AgentV1SettingsAgentContextMessagesItemContentType, +) class AgentV1SettingsAgentContextMessagesItemContentParams(typing_extensions.TypedDict): @@ -13,7 +14,7 @@ class AgentV1SettingsAgentContextMessagesItemContentParams(typing_extensions.Typ Conversation text as part of the conversation history """ - type: typing.Literal["History"] + type: AgentV1SettingsAgentContextMessagesItemContentType """ Message type identifier for conversation text """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_function_calls.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_function_calls.py index cdc5733c..2a5ebd8f 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_function_calls.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_context_messages_item_function_calls.py @@ -3,6 +3,9 @@ import typing import typing_extensions +from ..types.agent_v1settings_agent_context_messages_item_function_calls_type import ( + AgentV1SettingsAgentContextMessagesItemFunctionCallsType, +) from .agent_v1settings_agent_context_messages_item_function_calls_function_calls_item import ( AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams, ) @@ -13,7 +16,7 @@ class AgentV1SettingsAgentContextMessagesItemFunctionCallsParams(typing_extensio Client-side or server-side function call request and response as part of the conversation history """ - type: typing.Literal["History"] + type: AgentV1SettingsAgentContextMessagesItemFunctionCallsType function_calls: typing.Sequence[AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItemParams] """ List of function call objects diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider.py index 1b5b47e8..b5fdec2e 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider.py @@ -5,11 +5,13 @@ import typing import typing_extensions +from ..types.agent_v1settings_agent_listen_provider_v1type import AgentV1SettingsAgentListenProviderV1Type +from ..types.agent_v1settings_agent_listen_provider_v2type import AgentV1SettingsAgentListenProviderV2Type class AgentV1SettingsAgentListenProvider_V1Params(typing_extensions.TypedDict): version: typing.Literal["v1"] - type: typing.Literal["deepgram"] + type: AgentV1SettingsAgentListenProviderV1Type model: typing_extensions.NotRequired[str] language: typing_extensions.NotRequired[str] keyterms: typing_extensions.NotRequired[typing.Sequence[str]] @@ -18,7 +20,7 @@ class AgentV1SettingsAgentListenProvider_V1Params(typing_extensions.TypedDict): class AgentV1SettingsAgentListenProvider_V2Params(typing_extensions.TypedDict): version: typing.Literal["v2"] - type: typing.Literal["deepgram"] + type: AgentV1SettingsAgentListenProviderV2Type model: str keyterms: typing_extensions.NotRequired[typing.Sequence[str]] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v1.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v1.py index 0fd4e61d..93cb859d 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v1.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v1.py @@ -3,10 +3,11 @@ import typing import typing_extensions +from ..types.agent_v1settings_agent_listen_provider_v1type import AgentV1SettingsAgentListenProviderV1Type class AgentV1SettingsAgentListenProviderV1Params(typing_extensions.TypedDict): - type: typing.Literal["deepgram"] + type: AgentV1SettingsAgentListenProviderV1Type """ Provider type for speech-to-text """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v2.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v2.py index 0e0e5c5a..46d2ac99 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v2.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_listen_provider_v2.py @@ -3,10 +3,11 @@ import typing import typing_extensions +from ..types.agent_v1settings_agent_listen_provider_v2type import AgentV1SettingsAgentListenProviderV2Type class AgentV1SettingsAgentListenProviderV2Params(typing_extensions.TypedDict): - type: typing.Literal["deepgram"] + type: AgentV1SettingsAgentListenProviderV2Type """ Provider type for speech-to-text """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak.py index f6fa232a..3ae1f7c6 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak.py @@ -2,6 +2,9 @@ import typing -from ....requests.speak_settings_v1 import SpeakSettingsV1Params +from .agent_v1settings_agent_speak_endpoint import AgentV1SettingsAgentSpeakEndpointParams +from .agent_v1settings_agent_speak_one_item import AgentV1SettingsAgentSpeakOneItemParams -AgentV1SettingsAgentSpeakParams = typing.Union[SpeakSettingsV1Params, typing.Sequence[SpeakSettingsV1Params]] +AgentV1SettingsAgentSpeakParams = typing.Union[ + AgentV1SettingsAgentSpeakEndpointParams, typing.Sequence[AgentV1SettingsAgentSpeakOneItemParams] +] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint.py new file mode 100644 index 00000000..d90614be --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint.py @@ -0,0 +1,14 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from .agent_v1settings_agent_speak_endpoint_endpoint import AgentV1SettingsAgentSpeakEndpointEndpointParams +from .agent_v1settings_agent_speak_endpoint_provider import AgentV1SettingsAgentSpeakEndpointProviderParams + + +class AgentV1SettingsAgentSpeakEndpointParams(typing_extensions.TypedDict): + provider: AgentV1SettingsAgentSpeakEndpointProviderParams + endpoint: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakEndpointEndpointParams] + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_endpoint.py new file mode 100644 index 00000000..604b9ac5 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_endpoint.py @@ -0,0 +1,19 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import typing_extensions + + +class AgentV1SettingsAgentSpeakEndpointEndpointParams(typing_extensions.TypedDict): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing_extensions.NotRequired[str] + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing_extensions.NotRequired[typing.Dict[str, str]] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider.py new file mode 100644 index 00000000..b786735a --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider.py @@ -0,0 +1,93 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import typing_extensions +from ..types.agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, +) + + +class AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams(typing_extensions.TypedDict): + type: typing.Literal["deepgram"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel + + +class AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams(typing_extensions.TypedDict): + type: typing.Literal["eleven_labs"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId + language: typing_extensions.NotRequired[str] + language_code: typing_extensions.NotRequired[str] + + +class AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams(typing_extensions.TypedDict): + type: typing.Literal["cartesia"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + + +class AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams(typing_extensions.TypedDict): + type: typing.Literal["open_ai"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion] + model: AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel + voice: AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice + + +class AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams(typing_extensions.TypedDict): + type: typing.Literal["aws_polly"] + voice: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice + language: str + language_code: typing_extensions.NotRequired[str] + engine: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams + + +AgentV1SettingsAgentSpeakEndpointProviderParams = typing.Union[ + AgentV1SettingsAgentSpeakEndpointProvider_DeepgramParams, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakEndpointProvider_CartesiaParams, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAiParams, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPollyParams, +] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py new file mode 100644 index 00000000..7335b00b --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py @@ -0,0 +1,32 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderAwsPollyParams(typing_extensions.TypedDict): + voice: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice + """ + AWS Polly voice name + """ + + language: str + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ + + engine: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py new file mode 100644 index 00000000..97ad74b1 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py @@ -0,0 +1,17 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsParams(typing_extensions.TypedDict): + type: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing_extensions.NotRequired[str] + """ + Required for STS only + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py new file mode 100644 index 00000000..51bd279a --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py @@ -0,0 +1,15 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions + + +class AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoiceParams(typing_extensions.TypedDict): + mode: str + """ + Cartesia voice mode + """ + + id: str + """ + Cartesia voice ID + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_open_ai.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_open_ai.py new file mode 100644 index 00000000..43538f4e --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_endpoint_provider_open_ai.py @@ -0,0 +1,29 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, +) +from ..types.agent_v1settings_agent_speak_endpoint_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderOpenAiParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion] + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel + """ + OpenAI TTS model + """ + + voice: AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice + """ + OpenAI voice + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item.py new file mode 100644 index 00000000..c892106c --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item.py @@ -0,0 +1,14 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from .agent_v1settings_agent_speak_one_item_endpoint import AgentV1SettingsAgentSpeakOneItemEndpointParams +from .agent_v1settings_agent_speak_one_item_provider import AgentV1SettingsAgentSpeakOneItemProviderParams + + +class AgentV1SettingsAgentSpeakOneItemParams(typing_extensions.TypedDict): + provider: AgentV1SettingsAgentSpeakOneItemProviderParams + endpoint: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemEndpointParams] + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_endpoint.py new file mode 100644 index 00000000..8c6af538 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_endpoint.py @@ -0,0 +1,19 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import typing_extensions + + +class AgentV1SettingsAgentSpeakOneItemEndpointParams(typing_extensions.TypedDict): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing_extensions.NotRequired[str] + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing_extensions.NotRequired[typing.Dict[str, str]] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider.py new file mode 100644 index 00000000..f9254fa7 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider.py @@ -0,0 +1,93 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, +) + + +class AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams(typing_extensions.TypedDict): + type: typing.Literal["deepgram"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel + + +class AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams(typing_extensions.TypedDict): + type: typing.Literal["eleven_labs"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId + language: typing_extensions.NotRequired[str] + language_code: typing_extensions.NotRequired[str] + + +class AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams(typing_extensions.TypedDict): + type: typing.Literal["cartesia"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + + +class AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams(typing_extensions.TypedDict): + type: typing.Literal["open_ai"] + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion] + model: AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel + voice: AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice + + +class AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams(typing_extensions.TypedDict): + type: typing.Literal["aws_polly"] + voice: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice + language: str + language_code: typing_extensions.NotRequired[str] + engine: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams + + +AgentV1SettingsAgentSpeakOneItemProviderParams = typing.Union[ + AgentV1SettingsAgentSpeakOneItemProvider_DeepgramParams, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabsParams, + AgentV1SettingsAgentSpeakOneItemProvider_CartesiaParams, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAiParams, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPollyParams, +] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly.py new file mode 100644 index 00000000..4e175186 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly.py @@ -0,0 +1,32 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderAwsPollyParams(typing_extensions.TypedDict): + voice: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice + """ + AWS Polly voice name + """ + + language: str + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ + + engine: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py new file mode 100644 index 00000000..7d94447a --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py @@ -0,0 +1,17 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsParams(typing_extensions.TypedDict): + type: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing_extensions.NotRequired[str] + """ + Required for STS only + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py new file mode 100644 index 00000000..b2765075 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py @@ -0,0 +1,15 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions + + +class AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams(typing_extensions.TypedDict): + mode: str + """ + Cartesia voice mode + """ + + id: str + """ + Cartesia voice ID + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_open_ai.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_open_ai.py new file mode 100644 index 00000000..364a3e80 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_speak_one_item_provider_open_ai.py @@ -0,0 +1,29 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderOpenAiParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion] + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel + """ + OpenAI TTS model + """ + + voice: AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice + """ + OpenAI voice + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_think.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think.py index 4e1e9533..f7e29e1a 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_agent_think.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think.py @@ -3,5 +3,8 @@ import typing from ....requests.think_settings_v1 import ThinkSettingsV1Params +from .agent_v1settings_agent_think_one_item import AgentV1SettingsAgentThinkOneItemParams -AgentV1SettingsAgentThinkParams = typing.Union[ThinkSettingsV1Params, typing.Sequence[ThinkSettingsV1Params]] +AgentV1SettingsAgentThinkParams = typing.Union[ + ThinkSettingsV1Params, typing.Sequence[AgentV1SettingsAgentThinkOneItemParams] +] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item.py new file mode 100644 index 00000000..e5102568 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item.py @@ -0,0 +1,24 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import typing_extensions +from .agent_v1settings_agent_think_one_item_context_length import AgentV1SettingsAgentThinkOneItemContextLengthParams +from .agent_v1settings_agent_think_one_item_endpoint import AgentV1SettingsAgentThinkOneItemEndpointParams +from .agent_v1settings_agent_think_one_item_functions_item import AgentV1SettingsAgentThinkOneItemFunctionsItemParams +from .agent_v1settings_agent_think_one_item_provider import AgentV1SettingsAgentThinkOneItemProviderParams + + +class AgentV1SettingsAgentThinkOneItemParams(typing_extensions.TypedDict): + provider: AgentV1SettingsAgentThinkOneItemProviderParams + endpoint: typing_extensions.NotRequired[AgentV1SettingsAgentThinkOneItemEndpointParams] + """ + Optional for non-Deepgram LLM providers. When present, must include url field and headers object + """ + + functions: typing_extensions.NotRequired[typing.Sequence[AgentV1SettingsAgentThinkOneItemFunctionsItemParams]] + prompt: typing_extensions.NotRequired[str] + context_length: typing_extensions.NotRequired[AgentV1SettingsAgentThinkOneItemContextLengthParams] + """ + Specifies the number of characters retained in context between user messages, agent responses, and function calls. This setting is only configurable when a custom think endpoint is used + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_context_length.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_context_length.py new file mode 100644 index 00000000..2709524b --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_context_length.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from ..types.max import Max + +AgentV1SettingsAgentThinkOneItemContextLengthParams = typing.Union[Max, float] diff --git a/src/deepgram/requests/think_settings_v1endpoint.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_endpoint.py similarity index 83% rename from src/deepgram/requests/think_settings_v1endpoint.py rename to src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_endpoint.py index 38e2f34a..3aa4f9ef 100644 --- a/src/deepgram/requests/think_settings_v1endpoint.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_endpoint.py @@ -5,7 +5,7 @@ import typing_extensions -class ThinkSettingsV1EndpointParams(typing_extensions.TypedDict): +class AgentV1SettingsAgentThinkOneItemEndpointParams(typing_extensions.TypedDict): """ Optional for non-Deepgram LLM providers. When present, must include url field and headers object """ diff --git a/src/deepgram/requests/think_settings_v1functions_item.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item.py similarity index 59% rename from src/deepgram/requests/think_settings_v1functions_item.py rename to src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item.py index 2d8212c5..c917ae3d 100644 --- a/src/deepgram/requests/think_settings_v1functions_item.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item.py @@ -3,10 +3,12 @@ import typing import typing_extensions -from .think_settings_v1functions_item_endpoint import ThinkSettingsV1FunctionsItemEndpointParams +from .agent_v1settings_agent_think_one_item_functions_item_endpoint import ( + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams, +) -class ThinkSettingsV1FunctionsItemParams(typing_extensions.TypedDict): +class AgentV1SettingsAgentThinkOneItemFunctionsItemParams(typing_extensions.TypedDict): name: typing_extensions.NotRequired[str] """ Function name @@ -22,7 +24,7 @@ class ThinkSettingsV1FunctionsItemParams(typing_extensions.TypedDict): Function parameters """ - endpoint: typing_extensions.NotRequired[ThinkSettingsV1FunctionsItemEndpointParams] + endpoint: typing_extensions.NotRequired[AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams] """ The Function endpoint to call. if not passed, function is called client-side """ diff --git a/src/deepgram/requests/think_settings_v1functions_item_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item_endpoint.py similarity index 81% rename from src/deepgram/requests/think_settings_v1functions_item_endpoint.py rename to src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item_endpoint.py index 3e3ae16e..b56b4d6a 100644 --- a/src/deepgram/requests/think_settings_v1functions_item_endpoint.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_functions_item_endpoint.py @@ -5,7 +5,7 @@ import typing_extensions -class ThinkSettingsV1FunctionsItemEndpointParams(typing_extensions.TypedDict): +class AgentV1SettingsAgentThinkOneItemFunctionsItemEndpointParams(typing_extensions.TypedDict): """ The Function endpoint to call. if not passed, function is called client-side """ diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_provider.py b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_provider.py new file mode 100644 index 00000000..9725da31 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1settings_agent_think_one_item_provider.py @@ -0,0 +1,13 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from ....types.anthropic import Anthropic +from ....types.aws_bedrock_think_provider import AwsBedrockThinkProvider +from ....types.google import Google +from ....types.groq import Groq +from ....types.open_ai_think_provider import OpenAiThinkProvider + +AgentV1SettingsAgentThinkOneItemProviderParams = typing.Union[ + OpenAiThinkProvider, AwsBedrockThinkProvider, Anthropic, Google, Groq +] diff --git a/src/deepgram/agent/v1/requests/agent_v1settings_applied.py b/src/deepgram/agent/v1/requests/agent_v1settings_applied.py index 32bca304..e4502c84 100644 --- a/src/deepgram/agent/v1/requests/agent_v1settings_applied.py +++ b/src/deepgram/agent/v1/requests/agent_v1settings_applied.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1settings_applied_type import AgentV1SettingsAppliedType class AgentV1SettingsAppliedParams(typing_extensions.TypedDict): - type: typing.Literal["SettingsApplied"] + type: AgentV1SettingsAppliedType """ Message type identifier for settings applied confirmation """ diff --git a/src/deepgram/agent/v1/requests/agent_v1speak_updated.py b/src/deepgram/agent/v1/requests/agent_v1speak_updated.py index 908d6639..c7cc28da 100644 --- a/src/deepgram/agent/v1/requests/agent_v1speak_updated.py +++ b/src/deepgram/agent/v1/requests/agent_v1speak_updated.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1speak_updated_type import AgentV1SpeakUpdatedType class AgentV1SpeakUpdatedParams(typing_extensions.TypedDict): - type: typing.Literal["SpeakUpdated"] + type: AgentV1SpeakUpdatedType """ Message type identifier for speak update confirmation """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_prompt.py b/src/deepgram/agent/v1/requests/agent_v1update_prompt.py index 8f363a56..bbe1eef0 100644 --- a/src/deepgram/agent/v1/requests/agent_v1update_prompt.py +++ b/src/deepgram/agent/v1/requests/agent_v1update_prompt.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1update_prompt_type import AgentV1UpdatePromptType class AgentV1UpdatePromptParams(typing_extensions.TypedDict): - type: typing.Literal["UpdatePrompt"] + type: AgentV1UpdatePromptType """ Message type identifier for prompt update request """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak.py b/src/deepgram/agent/v1/requests/agent_v1update_speak.py index c5820907..23ffd54d 100644 --- a/src/deepgram/agent/v1/requests/agent_v1update_speak.py +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak.py @@ -1,15 +1,14 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions -from ....requests.speak_settings_v1 import SpeakSettingsV1Params +from ..types.agent_v1update_speak_type import AgentV1UpdateSpeakType +from .agent_v1update_speak_speak import AgentV1UpdateSpeakSpeakParams class AgentV1UpdateSpeakParams(typing_extensions.TypedDict): - type: typing.Literal["UpdateSpeak"] + type: AgentV1UpdateSpeakType """ Message type identifier for updating the speak model """ - speak: SpeakSettingsV1Params + speak: AgentV1UpdateSpeakSpeakParams diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak.py new file mode 100644 index 00000000..c71e5376 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak.py @@ -0,0 +1,10 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from .agent_v1update_speak_speak_endpoint import AgentV1UpdateSpeakSpeakEndpointParams +from .agent_v1update_speak_speak_one_item import AgentV1UpdateSpeakSpeakOneItemParams + +AgentV1UpdateSpeakSpeakParams = typing.Union[ + AgentV1UpdateSpeakSpeakEndpointParams, typing.Sequence[AgentV1UpdateSpeakSpeakOneItemParams] +] diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint.py new file mode 100644 index 00000000..c8fba1a8 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint.py @@ -0,0 +1,14 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from .agent_v1update_speak_speak_endpoint_endpoint import AgentV1UpdateSpeakSpeakEndpointEndpointParams +from .agent_v1update_speak_speak_endpoint_provider import AgentV1UpdateSpeakSpeakEndpointProviderParams + + +class AgentV1UpdateSpeakSpeakEndpointParams(typing_extensions.TypedDict): + provider: AgentV1UpdateSpeakSpeakEndpointProviderParams + endpoint: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointEndpointParams] + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_endpoint.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_endpoint.py new file mode 100644 index 00000000..bfa152d1 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_endpoint.py @@ -0,0 +1,19 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import typing_extensions + + +class AgentV1UpdateSpeakSpeakEndpointEndpointParams(typing_extensions.TypedDict): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing_extensions.NotRequired[str] + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing_extensions.NotRequired[typing.Dict[str, str]] diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider.py new file mode 100644 index 00000000..a9f5ec70 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider.py @@ -0,0 +1,93 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams, +) + + +class AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams(typing_extensions.TypedDict): + type: typing.Literal["deepgram"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion] + model: AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel + + +class AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams(typing_extensions.TypedDict): + type: typing.Literal["eleven_labs"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion] + model_id: AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId + language: typing_extensions.NotRequired[str] + language_code: typing_extensions.NotRequired[str] + + +class AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams(typing_extensions.TypedDict): + type: typing.Literal["cartesia"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion] + model_id: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId + voice: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + + +class AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams(typing_extensions.TypedDict): + type: typing.Literal["open_ai"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion] + model: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel + voice: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice + + +class AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams(typing_extensions.TypedDict): + type: typing.Literal["aws_polly"] + voice: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice + language: str + language_code: typing_extensions.NotRequired[str] + engine: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams + + +AgentV1UpdateSpeakSpeakEndpointProviderParams = typing.Union[ + AgentV1UpdateSpeakSpeakEndpointProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakEndpointProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPollyParams, +] diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly.py new file mode 100644 index 00000000..ab5964ed --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly.py @@ -0,0 +1,32 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyParams(typing_extensions.TypedDict): + voice: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice + """ + AWS Polly voice name + """ + + language: str + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ + + engine: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py new file mode 100644 index 00000000..3553faa1 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py @@ -0,0 +1,17 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsParams(typing_extensions.TypedDict): + type: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing_extensions.NotRequired[str] + """ + Required for STS only + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia.py new file mode 100644 index 00000000..fa84c7e7 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderCartesiaParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion] + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId + """ + Cartesia model ID + """ + + voice: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + """ + Cartesia language code + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py new file mode 100644 index 00000000..15e27850 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py @@ -0,0 +1,15 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions + + +class AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoiceParams(typing_extensions.TypedDict): + mode: str + """ + Cartesia voice mode + """ + + id: str + """ + Cartesia voice ID + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_deepgram.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_deepgram.py new file mode 100644 index 00000000..cb1fdb92 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_deepgram.py @@ -0,0 +1,21 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderDeepgramParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion] + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel + """ + Deepgram TTS model + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py new file mode 100644 index 00000000..f6eccbdf --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py @@ -0,0 +1,31 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion] + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId + """ + Eleven Labs model ID + """ + + language: typing_extensions.NotRequired[str] + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_open_ai.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_open_ai.py new file mode 100644 index 00000000..edf3daec --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_endpoint_provider_open_ai.py @@ -0,0 +1,29 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, +) +from ..types.agent_v1update_speak_speak_endpoint_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderOpenAiParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion] + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel + """ + OpenAI TTS model + """ + + voice: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice + """ + OpenAI voice + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item.py new file mode 100644 index 00000000..416e766a --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item.py @@ -0,0 +1,14 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from .agent_v1update_speak_speak_one_item_endpoint import AgentV1UpdateSpeakSpeakOneItemEndpointParams +from .agent_v1update_speak_speak_one_item_provider import AgentV1UpdateSpeakSpeakOneItemProviderParams + + +class AgentV1UpdateSpeakSpeakOneItemParams(typing_extensions.TypedDict): + provider: AgentV1UpdateSpeakSpeakOneItemProviderParams + endpoint: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemEndpointParams] + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ diff --git a/src/deepgram/requests/speak_settings_v1endpoint.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_endpoint.py similarity index 87% rename from src/deepgram/requests/speak_settings_v1endpoint.py rename to src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_endpoint.py index 1961d66c..f051beef 100644 --- a/src/deepgram/requests/speak_settings_v1endpoint.py +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_endpoint.py @@ -5,7 +5,7 @@ import typing_extensions -class SpeakSettingsV1EndpointParams(typing_extensions.TypedDict): +class AgentV1UpdateSpeakSpeakOneItemEndpointParams(typing_extensions.TypedDict): """ Optional if provider is Deepgram. Required for non-Deepgram TTS providers. When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider.py new file mode 100644 index 00000000..bb8f3bc7 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider.py @@ -0,0 +1,93 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, +) +from ..types.agent_v1update_speak_speak_one_item_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, +) +from ..types.agent_v1update_speak_speak_one_item_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, +) +from ..types.agent_v1update_speak_speak_one_item_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, +) +from ..types.agent_v1update_speak_speak_one_item_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, +) +from ..types.agent_v1update_speak_speak_one_item_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, +) +from ..types.agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, +) +from ..types.agent_v1update_speak_speak_one_item_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, +) +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, +) +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, +) +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams, +) + + +class AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams(typing_extensions.TypedDict): + type: typing.Literal["deepgram"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion] + model: AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel + + +class AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams(typing_extensions.TypedDict): + type: typing.Literal["eleven_labs"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion] + model_id: AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId + language: typing_extensions.NotRequired[str] + language_code: typing_extensions.NotRequired[str] + + +class AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams(typing_extensions.TypedDict): + type: typing.Literal["cartesia"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion] + model_id: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId + voice: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + + +class AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams(typing_extensions.TypedDict): + type: typing.Literal["open_ai"] + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion] + model: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel + voice: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice + + +class AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams(typing_extensions.TypedDict): + type: typing.Literal["aws_polly"] + voice: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice + language: str + language_code: typing_extensions.NotRequired[str] + engine: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams + + +AgentV1UpdateSpeakSpeakOneItemProviderParams = typing.Union[ + AgentV1UpdateSpeakSpeakOneItemProvider_DeepgramParams, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabsParams, + AgentV1UpdateSpeakSpeakOneItemProvider_CartesiaParams, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAiParams, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPollyParams, +] diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly.py new file mode 100644 index 00000000..b79609a0 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly.py @@ -0,0 +1,32 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, +) +from ..types.agent_v1update_speak_speak_one_item_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyParams(typing_extensions.TypedDict): + voice: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice + """ + AWS Polly voice name + """ + + language: str + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ + + engine: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py new file mode 100644 index 00000000..f47f7a00 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py @@ -0,0 +1,17 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsParams(typing_extensions.TypedDict): + type: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing_extensions.NotRequired[str] + """ + Required for STS only + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia.py new file mode 100644 index 00000000..d07ba108 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, +) +from ..types.agent_v1update_speak_speak_one_item_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderCartesiaParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion] + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId + """ + Cartesia model ID + """ + + voice: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + """ + Cartesia language code + """ diff --git a/src/deepgram/requests/cartesia_speak_provider_voice.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py similarity index 67% rename from src/deepgram/requests/cartesia_speak_provider_voice.py rename to src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py index 89929e27..5702bdb7 100644 --- a/src/deepgram/requests/cartesia_speak_provider_voice.py +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py @@ -3,7 +3,7 @@ import typing_extensions -class CartesiaSpeakProviderVoiceParams(typing_extensions.TypedDict): +class AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoiceParams(typing_extensions.TypedDict): mode: str """ Cartesia voice mode diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_deepgram.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_deepgram.py new file mode 100644 index 00000000..73e08249 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_deepgram.py @@ -0,0 +1,21 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, +) +from ..types.agent_v1update_speak_speak_one_item_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderDeepgramParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion] + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel + """ + Deepgram TTS model + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_eleven_labs.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_eleven_labs.py new file mode 100644 index 00000000..1906bbe1 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_eleven_labs.py @@ -0,0 +1,31 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, +) +from ..types.agent_v1update_speak_speak_one_item_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion] + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId + """ + Eleven Labs model ID + """ + + language: typing_extensions.NotRequired[str] + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_open_ai.py b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_open_ai.py new file mode 100644 index 00000000..2dbd8cb9 --- /dev/null +++ b/src/deepgram/agent/v1/requests/agent_v1update_speak_speak_one_item_provider_open_ai.py @@ -0,0 +1,29 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, +) +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, +) +from ..types.agent_v1update_speak_speak_one_item_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderOpenAiParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion] + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel + """ + OpenAI TTS model + """ + + voice: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice + """ + OpenAI voice + """ diff --git a/src/deepgram/agent/v1/requests/agent_v1user_started_speaking.py b/src/deepgram/agent/v1/requests/agent_v1user_started_speaking.py index c883119c..77fb500e 100644 --- a/src/deepgram/agent/v1/requests/agent_v1user_started_speaking.py +++ b/src/deepgram/agent/v1/requests/agent_v1user_started_speaking.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1user_started_speaking_type import AgentV1UserStartedSpeakingType class AgentV1UserStartedSpeakingParams(typing_extensions.TypedDict): - type: typing.Literal["UserStartedSpeaking"] + type: AgentV1UserStartedSpeakingType """ Message type identifier indicating that the user has begun speaking """ diff --git a/src/deepgram/agent/v1/requests/agent_v1warning.py b/src/deepgram/agent/v1/requests/agent_v1warning.py index f1e75051..0afeaedb 100644 --- a/src/deepgram/agent/v1/requests/agent_v1warning.py +++ b/src/deepgram/agent/v1/requests/agent_v1warning.py @@ -1,8 +1,7 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1warning_type import AgentV1WarningType class AgentV1WarningParams(typing_extensions.TypedDict): @@ -10,7 +9,7 @@ class AgentV1WarningParams(typing_extensions.TypedDict): Notifies the client of non-fatal errors or warnings """ - type: typing.Literal["Warning"] + type: AgentV1WarningType """ Message type identifier for warnings """ diff --git a/src/deepgram/agent/v1/requests/agent_v1welcome.py b/src/deepgram/agent/v1/requests/agent_v1welcome.py index 5168a4f0..9804dc6f 100644 --- a/src/deepgram/agent/v1/requests/agent_v1welcome.py +++ b/src/deepgram/agent/v1/requests/agent_v1welcome.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_v1welcome_type import AgentV1WelcomeType class AgentV1WelcomeParams(typing_extensions.TypedDict): - type: typing.Literal["Welcome"] + type: AgentV1WelcomeType """ Message type identifier for welcome message """ diff --git a/src/deepgram/agent/v1/requests/cartesia.py b/src/deepgram/agent/v1/requests/cartesia.py new file mode 100644 index 00000000..9e1d3db2 --- /dev/null +++ b/src/deepgram/agent/v1/requests/cartesia.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams, +) + + +class CartesiaParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId + """ + Cartesia model ID + """ + + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoiceParams + language: typing_extensions.NotRequired[str] + """ + Cartesia language code + """ diff --git a/src/deepgram/agent/v1/requests/deepgram.py b/src/deepgram/agent/v1/requests/deepgram.py new file mode 100644 index 00000000..bafefe78 --- /dev/null +++ b/src/deepgram/agent/v1/requests/deepgram.py @@ -0,0 +1,21 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) + + +class DeepgramParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel + """ + Deepgram TTS model + """ diff --git a/src/deepgram/agent/v1/requests/eleven_labs.py b/src/deepgram/agent/v1/requests/eleven_labs.py new file mode 100644 index 00000000..083e2c33 --- /dev/null +++ b/src/deepgram/agent/v1/requests/eleven_labs.py @@ -0,0 +1,31 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing_extensions +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from ..types.agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) + + +class ElevenLabsParams(typing_extensions.TypedDict): + version: typing_extensions.NotRequired[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId + """ + Eleven Labs model ID + """ + + language: typing_extensions.NotRequired[str] + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing_extensions.NotRequired[str] + """ + Use the `language` field instead. + """ diff --git a/src/deepgram/agent/v1/socket_client.py b/src/deepgram/agent/v1/socket_client.py index e486dc67..1ef9014d 100644 --- a/src/deepgram/agent/v1/socket_client.py +++ b/src/deepgram/agent/v1/socket_client.py @@ -1,6 +1,7 @@ # This file was auto-generated by Fern from our API Definition. import json +import logging import typing from json.decoder import JSONDecodeError @@ -35,6 +36,9 @@ except ImportError: from websockets import WebSocketClientProtocol # type: ignore +_logger = logging.getLogger(__name__) + + def _sanitize_numeric_types(obj: typing.Any) -> typing.Any: """ Recursively convert float values that are whole numbers to int. @@ -53,8 +57,6 @@ def _sanitize_numeric_types(obj: typing.Any) -> typing.Any: elif isinstance(obj, float) and obj.is_integer(): return int(obj) return obj - - V1SocketClientResponse = typing.Union[ AgentV1ReceiveFunctionCallResponse, AgentV1PromptUpdated, @@ -84,7 +86,13 @@ async def __aiter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue async def start_listening(self): """ @@ -103,7 +111,13 @@ async def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue await self._emit_async(EventType.MESSAGE, parsed) except Exception as exc: await self._emit_async(EventType.ERROR, exc) @@ -145,12 +159,12 @@ async def send_function_call_response(self, message: AgentV1SendFunctionCallResp """ await self._send_model(message) - async def send_keep_alive(self, message: typing.Optional[AgentV1KeepAlive] = None) -> None: + async def send_keep_alive(self, message: AgentV1KeepAlive) -> None: """ Send a message to the websocket connection. The message will be sent as a AgentV1KeepAlive. """ - await self._send_model(message or AgentV1KeepAlive()) + await self._send_model(message) async def send_update_prompt(self, message: AgentV1UpdatePrompt) -> None: """ @@ -174,7 +188,11 @@ async def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore async def _send(self, data: typing.Any) -> None: """ @@ -201,7 +219,13 @@ def __iter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue def start_listening(self): """ @@ -220,7 +244,13 @@ def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue self._emit(EventType.MESSAGE, parsed) except Exception as exc: self._emit(EventType.ERROR, exc) @@ -262,12 +292,12 @@ def send_function_call_response(self, message: AgentV1SendFunctionCallResponse) """ self._send_model(message) - def send_keep_alive(self, message: typing.Optional[AgentV1KeepAlive] = None) -> None: + def send_keep_alive(self, message: AgentV1KeepAlive) -> None: """ Send a message to the websocket connection. The message will be sent as a AgentV1KeepAlive. """ - self._send_model(message or AgentV1KeepAlive()) + self._send_model(message) def send_update_prompt(self, message: AgentV1UpdatePrompt) -> None: """ @@ -291,7 +321,11 @@ def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore def _send(self, data: typing.Any) -> None: """ diff --git a/src/deepgram/agent/v1/types/__init__.py b/src/deepgram/agent/v1/types/__init__.py index 6cc1e513..58919d1e 100644 --- a/src/deepgram/agent/v1/types/__init__.py +++ b/src/deepgram/agent/v1/types/__init__.py @@ -7,20 +7,33 @@ if typing.TYPE_CHECKING: from .agent_v1agent_audio_done import AgentV1AgentAudioDone + from .agent_v1agent_audio_done_type import AgentV1AgentAudioDoneType from .agent_v1agent_started_speaking import AgentV1AgentStartedSpeaking + from .agent_v1agent_started_speaking_type import AgentV1AgentStartedSpeakingType from .agent_v1agent_thinking import AgentV1AgentThinking + from .agent_v1agent_thinking_type import AgentV1AgentThinkingType from .agent_v1conversation_text import AgentV1ConversationText from .agent_v1conversation_text_role import AgentV1ConversationTextRole + from .agent_v1conversation_text_type import AgentV1ConversationTextType from .agent_v1error import AgentV1Error + from .agent_v1error_type import AgentV1ErrorType from .agent_v1function_call_request import AgentV1FunctionCallRequest from .agent_v1function_call_request_functions_item import AgentV1FunctionCallRequestFunctionsItem + from .agent_v1function_call_request_type import AgentV1FunctionCallRequestType from .agent_v1inject_agent_message import AgentV1InjectAgentMessage + from .agent_v1inject_agent_message_type import AgentV1InjectAgentMessageType from .agent_v1inject_user_message import AgentV1InjectUserMessage + from .agent_v1inject_user_message_type import AgentV1InjectUserMessageType from .agent_v1injection_refused import AgentV1InjectionRefused + from .agent_v1injection_refused_type import AgentV1InjectionRefusedType from .agent_v1keep_alive import AgentV1KeepAlive + from .agent_v1keep_alive_type import AgentV1KeepAliveType from .agent_v1prompt_updated import AgentV1PromptUpdated + from .agent_v1prompt_updated_type import AgentV1PromptUpdatedType from .agent_v1receive_function_call_response import AgentV1ReceiveFunctionCallResponse + from .agent_v1receive_function_call_response_type import AgentV1ReceiveFunctionCallResponseType from .agent_v1send_function_call_response import AgentV1SendFunctionCallResponse + from .agent_v1send_function_call_response_type import AgentV1SendFunctionCallResponseType from .agent_v1settings import AgentV1Settings from .agent_v1settings_agent import AgentV1SettingsAgent from .agent_v1settings_agent_context import AgentV1SettingsAgentContext @@ -29,12 +42,18 @@ from .agent_v1settings_agent_context_messages_item_content_role import ( AgentV1SettingsAgentContextMessagesItemContentRole, ) + from .agent_v1settings_agent_context_messages_item_content_type import ( + AgentV1SettingsAgentContextMessagesItemContentType, + ) from .agent_v1settings_agent_context_messages_item_function_calls import ( AgentV1SettingsAgentContextMessagesItemFunctionCalls, ) from .agent_v1settings_agent_context_messages_item_function_calls_function_calls_item import ( AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem, ) + from .agent_v1settings_agent_context_messages_item_function_calls_type import ( + AgentV1SettingsAgentContextMessagesItemFunctionCallsType, + ) from .agent_v1settings_agent_listen import AgentV1SettingsAgentListen from .agent_v1settings_agent_listen_provider import ( AgentV1SettingsAgentListenProvider, @@ -42,67 +61,458 @@ AgentV1SettingsAgentListenProvider_V2, ) from .agent_v1settings_agent_listen_provider_v1 import AgentV1SettingsAgentListenProviderV1 + from .agent_v1settings_agent_listen_provider_v1type import AgentV1SettingsAgentListenProviderV1Type from .agent_v1settings_agent_listen_provider_v2 import AgentV1SettingsAgentListenProviderV2 + from .agent_v1settings_agent_listen_provider_v2type import AgentV1SettingsAgentListenProviderV2Type from .agent_v1settings_agent_speak import AgentV1SettingsAgentSpeak + from .agent_v1settings_agent_speak_endpoint import AgentV1SettingsAgentSpeakEndpoint + from .agent_v1settings_agent_speak_endpoint_endpoint import AgentV1SettingsAgentSpeakEndpointEndpoint + from .agent_v1settings_agent_speak_endpoint_provider import ( + AgentV1SettingsAgentSpeakEndpointProvider, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly, + AgentV1SettingsAgentSpeakEndpointProvider_Cartesia, + AgentV1SettingsAgentSpeakEndpointProvider_Deepgram, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAi, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPolly, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, + ) + from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, + ) + from .agent_v1settings_agent_speak_endpoint_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId, + ) + from .agent_v1settings_agent_speak_endpoint_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion, + ) + from .agent_v1settings_agent_speak_endpoint_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice, + ) + from .agent_v1settings_agent_speak_endpoint_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel, + ) + from .agent_v1settings_agent_speak_endpoint_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion, + ) + from .agent_v1settings_agent_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId, + ) + from .agent_v1settings_agent_speak_endpoint_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion, + ) + from .agent_v1settings_agent_speak_endpoint_provider_open_ai import AgentV1SettingsAgentSpeakEndpointProviderOpenAi + from .agent_v1settings_agent_speak_endpoint_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, + ) + from .agent_v1settings_agent_speak_endpoint_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, + ) + from .agent_v1settings_agent_speak_endpoint_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, + ) + from .agent_v1settings_agent_speak_one_item import AgentV1SettingsAgentSpeakOneItem + from .agent_v1settings_agent_speak_one_item_endpoint import AgentV1SettingsAgentSpeakOneItemEndpoint + from .agent_v1settings_agent_speak_one_item_provider import ( + AgentV1SettingsAgentSpeakOneItemProvider, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly, + AgentV1SettingsAgentSpeakOneItemProvider_Cartesia, + AgentV1SettingsAgentSpeakOneItemProvider_Deepgram, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAi, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPolly, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, + ) + from .agent_v1settings_agent_speak_one_item_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, + ) + from .agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, + ) + from .agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, + ) + from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, + ) + from .agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, + ) + from .agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, + ) + from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, + ) + from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, + ) + from .agent_v1settings_agent_speak_one_item_provider_open_ai import AgentV1SettingsAgentSpeakOneItemProviderOpenAi + from .agent_v1settings_agent_speak_one_item_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, + ) + from .agent_v1settings_agent_speak_one_item_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, + ) + from .agent_v1settings_agent_speak_one_item_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, + ) from .agent_v1settings_agent_think import AgentV1SettingsAgentThink + from .agent_v1settings_agent_think_one_item import AgentV1SettingsAgentThinkOneItem + from .agent_v1settings_agent_think_one_item_context_length import AgentV1SettingsAgentThinkOneItemContextLength + from .agent_v1settings_agent_think_one_item_endpoint import AgentV1SettingsAgentThinkOneItemEndpoint + from .agent_v1settings_agent_think_one_item_functions_item import AgentV1SettingsAgentThinkOneItemFunctionsItem + from .agent_v1settings_agent_think_one_item_functions_item_endpoint import ( + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint, + ) + from .agent_v1settings_agent_think_one_item_provider import AgentV1SettingsAgentThinkOneItemProvider from .agent_v1settings_applied import AgentV1SettingsApplied + from .agent_v1settings_applied_type import AgentV1SettingsAppliedType from .agent_v1settings_audio import AgentV1SettingsAudio from .agent_v1settings_audio_input import AgentV1SettingsAudioInput from .agent_v1settings_audio_input_encoding import AgentV1SettingsAudioInputEncoding from .agent_v1settings_audio_output import AgentV1SettingsAudioOutput from .agent_v1settings_audio_output_encoding import AgentV1SettingsAudioOutputEncoding from .agent_v1settings_flags import AgentV1SettingsFlags + from .agent_v1settings_type import AgentV1SettingsType from .agent_v1speak_updated import AgentV1SpeakUpdated + from .agent_v1speak_updated_type import AgentV1SpeakUpdatedType from .agent_v1update_prompt import AgentV1UpdatePrompt + from .agent_v1update_prompt_type import AgentV1UpdatePromptType from .agent_v1update_speak import AgentV1UpdateSpeak + from .agent_v1update_speak_speak import AgentV1UpdateSpeakSpeak + from .agent_v1update_speak_speak_endpoint import AgentV1UpdateSpeakSpeakEndpoint + from .agent_v1update_speak_speak_endpoint_endpoint import AgentV1UpdateSpeakSpeakEndpointEndpoint + from .agent_v1update_speak_speak_endpoint_provider import ( + AgentV1UpdateSpeakSpeakEndpointProvider, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly, + AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia, + AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly import AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly + from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, + ) + from .agent_v1update_speak_speak_endpoint_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, + ) + from .agent_v1update_speak_speak_endpoint_provider_cartesia import AgentV1UpdateSpeakSpeakEndpointProviderCartesia + from .agent_v1update_speak_speak_endpoint_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, + ) + from .agent_v1update_speak_speak_endpoint_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, + ) + from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice, + ) + from .agent_v1update_speak_speak_endpoint_provider_deepgram import AgentV1UpdateSpeakSpeakEndpointProviderDeepgram + from .agent_v1update_speak_speak_endpoint_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, + ) + from .agent_v1update_speak_speak_endpoint_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, + ) + from .agent_v1update_speak_speak_endpoint_provider_eleven_labs import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs, + ) + from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, + ) + from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, + ) + from .agent_v1update_speak_speak_endpoint_provider_open_ai import AgentV1UpdateSpeakSpeakEndpointProviderOpenAi + from .agent_v1update_speak_speak_endpoint_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, + ) + from .agent_v1update_speak_speak_endpoint_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, + ) + from .agent_v1update_speak_speak_endpoint_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, + ) + from .agent_v1update_speak_speak_one_item import AgentV1UpdateSpeakSpeakOneItem + from .agent_v1update_speak_speak_one_item_endpoint import AgentV1UpdateSpeakSpeakOneItemEndpoint + from .agent_v1update_speak_speak_one_item_provider import ( + AgentV1UpdateSpeakSpeakOneItemProvider, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly, + AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia, + AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly import AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly + from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, + ) + from .agent_v1update_speak_speak_one_item_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, + ) + from .agent_v1update_speak_speak_one_item_provider_cartesia import AgentV1UpdateSpeakSpeakOneItemProviderCartesia + from .agent_v1update_speak_speak_one_item_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, + ) + from .agent_v1update_speak_speak_one_item_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, + ) + from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice, + ) + from .agent_v1update_speak_speak_one_item_provider_deepgram import AgentV1UpdateSpeakSpeakOneItemProviderDeepgram + from .agent_v1update_speak_speak_one_item_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, + ) + from .agent_v1update_speak_speak_one_item_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, + ) + from .agent_v1update_speak_speak_one_item_provider_eleven_labs import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs, + ) + from .agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, + ) + from .agent_v1update_speak_speak_one_item_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, + ) + from .agent_v1update_speak_speak_one_item_provider_open_ai import AgentV1UpdateSpeakSpeakOneItemProviderOpenAi + from .agent_v1update_speak_speak_one_item_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, + ) + from .agent_v1update_speak_speak_one_item_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, + ) + from .agent_v1update_speak_speak_one_item_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, + ) + from .agent_v1update_speak_type import AgentV1UpdateSpeakType from .agent_v1user_started_speaking import AgentV1UserStartedSpeaking + from .agent_v1user_started_speaking_type import AgentV1UserStartedSpeakingType from .agent_v1warning import AgentV1Warning + from .agent_v1warning_type import AgentV1WarningType from .agent_v1welcome import AgentV1Welcome + from .agent_v1welcome_type import AgentV1WelcomeType + from .cartesia import Cartesia + from .deepgram import Deepgram + from .eleven_labs import ElevenLabs + from .max import Max _dynamic_imports: typing.Dict[str, str] = { "AgentV1AgentAudioDone": ".agent_v1agent_audio_done", + "AgentV1AgentAudioDoneType": ".agent_v1agent_audio_done_type", "AgentV1AgentStartedSpeaking": ".agent_v1agent_started_speaking", + "AgentV1AgentStartedSpeakingType": ".agent_v1agent_started_speaking_type", "AgentV1AgentThinking": ".agent_v1agent_thinking", + "AgentV1AgentThinkingType": ".agent_v1agent_thinking_type", "AgentV1ConversationText": ".agent_v1conversation_text", "AgentV1ConversationTextRole": ".agent_v1conversation_text_role", + "AgentV1ConversationTextType": ".agent_v1conversation_text_type", "AgentV1Error": ".agent_v1error", + "AgentV1ErrorType": ".agent_v1error_type", "AgentV1FunctionCallRequest": ".agent_v1function_call_request", "AgentV1FunctionCallRequestFunctionsItem": ".agent_v1function_call_request_functions_item", + "AgentV1FunctionCallRequestType": ".agent_v1function_call_request_type", "AgentV1InjectAgentMessage": ".agent_v1inject_agent_message", + "AgentV1InjectAgentMessageType": ".agent_v1inject_agent_message_type", "AgentV1InjectUserMessage": ".agent_v1inject_user_message", + "AgentV1InjectUserMessageType": ".agent_v1inject_user_message_type", "AgentV1InjectionRefused": ".agent_v1injection_refused", + "AgentV1InjectionRefusedType": ".agent_v1injection_refused_type", "AgentV1KeepAlive": ".agent_v1keep_alive", + "AgentV1KeepAliveType": ".agent_v1keep_alive_type", "AgentV1PromptUpdated": ".agent_v1prompt_updated", + "AgentV1PromptUpdatedType": ".agent_v1prompt_updated_type", "AgentV1ReceiveFunctionCallResponse": ".agent_v1receive_function_call_response", + "AgentV1ReceiveFunctionCallResponseType": ".agent_v1receive_function_call_response_type", "AgentV1SendFunctionCallResponse": ".agent_v1send_function_call_response", + "AgentV1SendFunctionCallResponseType": ".agent_v1send_function_call_response_type", "AgentV1Settings": ".agent_v1settings", "AgentV1SettingsAgent": ".agent_v1settings_agent", "AgentV1SettingsAgentContext": ".agent_v1settings_agent_context", "AgentV1SettingsAgentContextMessagesItem": ".agent_v1settings_agent_context_messages_item", "AgentV1SettingsAgentContextMessagesItemContent": ".agent_v1settings_agent_context_messages_item_content", "AgentV1SettingsAgentContextMessagesItemContentRole": ".agent_v1settings_agent_context_messages_item_content_role", + "AgentV1SettingsAgentContextMessagesItemContentType": ".agent_v1settings_agent_context_messages_item_content_type", "AgentV1SettingsAgentContextMessagesItemFunctionCalls": ".agent_v1settings_agent_context_messages_item_function_calls", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem": ".agent_v1settings_agent_context_messages_item_function_calls_function_calls_item", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType": ".agent_v1settings_agent_context_messages_item_function_calls_type", "AgentV1SettingsAgentListen": ".agent_v1settings_agent_listen", "AgentV1SettingsAgentListenProvider": ".agent_v1settings_agent_listen_provider", "AgentV1SettingsAgentListenProviderV1": ".agent_v1settings_agent_listen_provider_v1", + "AgentV1SettingsAgentListenProviderV1Type": ".agent_v1settings_agent_listen_provider_v1type", "AgentV1SettingsAgentListenProviderV2": ".agent_v1settings_agent_listen_provider_v2", + "AgentV1SettingsAgentListenProviderV2Type": ".agent_v1settings_agent_listen_provider_v2type", "AgentV1SettingsAgentListenProvider_V1": ".agent_v1settings_agent_listen_provider", "AgentV1SettingsAgentListenProvider_V2": ".agent_v1settings_agent_listen_provider", "AgentV1SettingsAgentSpeak": ".agent_v1settings_agent_speak", + "AgentV1SettingsAgentSpeakEndpoint": ".agent_v1settings_agent_speak_endpoint", + "AgentV1SettingsAgentSpeakEndpointEndpoint": ".agent_v1settings_agent_speak_endpoint_endpoint", + "AgentV1SettingsAgentSpeakEndpointProvider": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice": ".agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId": ".agent_v1settings_agent_speak_endpoint_provider_cartesia_model_id", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion": ".agent_v1settings_agent_speak_endpoint_provider_cartesia_version", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice": ".agent_v1settings_agent_speak_endpoint_provider_cartesia_voice", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel": ".agent_v1settings_agent_speak_endpoint_provider_deepgram_model", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion": ".agent_v1settings_agent_speak_endpoint_provider_deepgram_version", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId": ".agent_v1settings_agent_speak_endpoint_provider_eleven_labs_model_id", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion": ".agent_v1settings_agent_speak_endpoint_provider_eleven_labs_version", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi": ".agent_v1settings_agent_speak_endpoint_provider_open_ai", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel": ".agent_v1settings_agent_speak_endpoint_provider_open_ai_model", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion": ".agent_v1settings_agent_speak_endpoint_provider_open_ai_version", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice": ".agent_v1settings_agent_speak_endpoint_provider_open_ai_voice", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi": ".agent_v1settings_agent_speak_endpoint_provider", + "AgentV1SettingsAgentSpeakOneItem": ".agent_v1settings_agent_speak_one_item", + "AgentV1SettingsAgentSpeakOneItemEndpoint": ".agent_v1settings_agent_speak_one_item_endpoint", + "AgentV1SettingsAgentSpeakOneItemProvider": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly": ".agent_v1settings_agent_speak_one_item_provider_aws_polly", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials": ".agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType": ".agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine": ".agent_v1settings_agent_speak_one_item_provider_aws_polly_engine", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice": ".agent_v1settings_agent_speak_one_item_provider_aws_polly_voice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId": ".agent_v1settings_agent_speak_one_item_provider_cartesia_model_id", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion": ".agent_v1settings_agent_speak_one_item_provider_cartesia_version", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice": ".agent_v1settings_agent_speak_one_item_provider_cartesia_voice", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel": ".agent_v1settings_agent_speak_one_item_provider_deepgram_model", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion": ".agent_v1settings_agent_speak_one_item_provider_deepgram_version", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId": ".agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion": ".agent_v1settings_agent_speak_one_item_provider_eleven_labs_version", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi": ".agent_v1settings_agent_speak_one_item_provider_open_ai", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel": ".agent_v1settings_agent_speak_one_item_provider_open_ai_model", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion": ".agent_v1settings_agent_speak_one_item_provider_open_ai_version", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice": ".agent_v1settings_agent_speak_one_item_provider_open_ai_voice", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs": ".agent_v1settings_agent_speak_one_item_provider", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi": ".agent_v1settings_agent_speak_one_item_provider", "AgentV1SettingsAgentThink": ".agent_v1settings_agent_think", + "AgentV1SettingsAgentThinkOneItem": ".agent_v1settings_agent_think_one_item", + "AgentV1SettingsAgentThinkOneItemContextLength": ".agent_v1settings_agent_think_one_item_context_length", + "AgentV1SettingsAgentThinkOneItemEndpoint": ".agent_v1settings_agent_think_one_item_endpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItem": ".agent_v1settings_agent_think_one_item_functions_item", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint": ".agent_v1settings_agent_think_one_item_functions_item_endpoint", + "AgentV1SettingsAgentThinkOneItemProvider": ".agent_v1settings_agent_think_one_item_provider", "AgentV1SettingsApplied": ".agent_v1settings_applied", + "AgentV1SettingsAppliedType": ".agent_v1settings_applied_type", "AgentV1SettingsAudio": ".agent_v1settings_audio", "AgentV1SettingsAudioInput": ".agent_v1settings_audio_input", "AgentV1SettingsAudioInputEncoding": ".agent_v1settings_audio_input_encoding", "AgentV1SettingsAudioOutput": ".agent_v1settings_audio_output", "AgentV1SettingsAudioOutputEncoding": ".agent_v1settings_audio_output_encoding", "AgentV1SettingsFlags": ".agent_v1settings_flags", + "AgentV1SettingsType": ".agent_v1settings_type", "AgentV1SpeakUpdated": ".agent_v1speak_updated", + "AgentV1SpeakUpdatedType": ".agent_v1speak_updated_type", "AgentV1UpdatePrompt": ".agent_v1update_prompt", + "AgentV1UpdatePromptType": ".agent_v1update_prompt_type", "AgentV1UpdateSpeak": ".agent_v1update_speak", + "AgentV1UpdateSpeakSpeak": ".agent_v1update_speak_speak", + "AgentV1UpdateSpeakSpeakEndpoint": ".agent_v1update_speak_speak_endpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpoint": ".agent_v1update_speak_speak_endpoint_endpoint", + "AgentV1UpdateSpeakSpeakEndpointProvider": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly": ".agent_v1update_speak_speak_endpoint_provider_aws_polly", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials": ".agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType": ".agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine": ".agent_v1update_speak_speak_endpoint_provider_aws_polly_engine", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice": ".agent_v1update_speak_speak_endpoint_provider_aws_polly_voice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia": ".agent_v1update_speak_speak_endpoint_provider_cartesia", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId": ".agent_v1update_speak_speak_endpoint_provider_cartesia_model_id", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion": ".agent_v1update_speak_speak_endpoint_provider_cartesia_version", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice": ".agent_v1update_speak_speak_endpoint_provider_cartesia_voice", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram": ".agent_v1update_speak_speak_endpoint_provider_deepgram", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel": ".agent_v1update_speak_speak_endpoint_provider_deepgram_model", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion": ".agent_v1update_speak_speak_endpoint_provider_deepgram_version", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs": ".agent_v1update_speak_speak_endpoint_provider_eleven_labs", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId": ".agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion": ".agent_v1update_speak_speak_endpoint_provider_eleven_labs_version", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi": ".agent_v1update_speak_speak_endpoint_provider_open_ai", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel": ".agent_v1update_speak_speak_endpoint_provider_open_ai_model", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion": ".agent_v1update_speak_speak_endpoint_provider_open_ai_version", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice": ".agent_v1update_speak_speak_endpoint_provider_open_ai_voice", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi": ".agent_v1update_speak_speak_endpoint_provider", + "AgentV1UpdateSpeakSpeakOneItem": ".agent_v1update_speak_speak_one_item", + "AgentV1UpdateSpeakSpeakOneItemEndpoint": ".agent_v1update_speak_speak_one_item_endpoint", + "AgentV1UpdateSpeakSpeakOneItemProvider": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly": ".agent_v1update_speak_speak_one_item_provider_aws_polly", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials": ".agent_v1update_speak_speak_one_item_provider_aws_polly_credentials", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType": ".agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine": ".agent_v1update_speak_speak_one_item_provider_aws_polly_engine", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice": ".agent_v1update_speak_speak_one_item_provider_aws_polly_voice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia": ".agent_v1update_speak_speak_one_item_provider_cartesia", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId": ".agent_v1update_speak_speak_one_item_provider_cartesia_model_id", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion": ".agent_v1update_speak_speak_one_item_provider_cartesia_version", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice": ".agent_v1update_speak_speak_one_item_provider_cartesia_voice", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram": ".agent_v1update_speak_speak_one_item_provider_deepgram", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel": ".agent_v1update_speak_speak_one_item_provider_deepgram_model", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion": ".agent_v1update_speak_speak_one_item_provider_deepgram_version", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs": ".agent_v1update_speak_speak_one_item_provider_eleven_labs", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId": ".agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion": ".agent_v1update_speak_speak_one_item_provider_eleven_labs_version", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi": ".agent_v1update_speak_speak_one_item_provider_open_ai", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel": ".agent_v1update_speak_speak_one_item_provider_open_ai_model", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion": ".agent_v1update_speak_speak_one_item_provider_open_ai_version", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice": ".agent_v1update_speak_speak_one_item_provider_open_ai_voice", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi": ".agent_v1update_speak_speak_one_item_provider", + "AgentV1UpdateSpeakType": ".agent_v1update_speak_type", "AgentV1UserStartedSpeaking": ".agent_v1user_started_speaking", + "AgentV1UserStartedSpeakingType": ".agent_v1user_started_speaking_type", "AgentV1Warning": ".agent_v1warning", + "AgentV1WarningType": ".agent_v1warning_type", "AgentV1Welcome": ".agent_v1welcome", + "AgentV1WelcomeType": ".agent_v1welcome_type", + "Cartesia": ".cartesia", + "Deepgram": ".deepgram", + "ElevenLabs": ".eleven_labs", + "Max": ".max", } @@ -129,47 +539,185 @@ def __dir__(): __all__ = [ "AgentV1AgentAudioDone", + "AgentV1AgentAudioDoneType", "AgentV1AgentStartedSpeaking", + "AgentV1AgentStartedSpeakingType", "AgentV1AgentThinking", + "AgentV1AgentThinkingType", "AgentV1ConversationText", "AgentV1ConversationTextRole", + "AgentV1ConversationTextType", "AgentV1Error", + "AgentV1ErrorType", "AgentV1FunctionCallRequest", "AgentV1FunctionCallRequestFunctionsItem", + "AgentV1FunctionCallRequestType", "AgentV1InjectAgentMessage", + "AgentV1InjectAgentMessageType", "AgentV1InjectUserMessage", + "AgentV1InjectUserMessageType", "AgentV1InjectionRefused", + "AgentV1InjectionRefusedType", "AgentV1KeepAlive", + "AgentV1KeepAliveType", "AgentV1PromptUpdated", + "AgentV1PromptUpdatedType", "AgentV1ReceiveFunctionCallResponse", + "AgentV1ReceiveFunctionCallResponseType", "AgentV1SendFunctionCallResponse", + "AgentV1SendFunctionCallResponseType", "AgentV1Settings", "AgentV1SettingsAgent", "AgentV1SettingsAgentContext", "AgentV1SettingsAgentContextMessagesItem", "AgentV1SettingsAgentContextMessagesItemContent", "AgentV1SettingsAgentContextMessagesItemContentRole", + "AgentV1SettingsAgentContextMessagesItemContentType", "AgentV1SettingsAgentContextMessagesItemFunctionCalls", "AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem", + "AgentV1SettingsAgentContextMessagesItemFunctionCallsType", "AgentV1SettingsAgentListen", "AgentV1SettingsAgentListenProvider", "AgentV1SettingsAgentListenProviderV1", + "AgentV1SettingsAgentListenProviderV1Type", "AgentV1SettingsAgentListenProviderV2", + "AgentV1SettingsAgentListenProviderV2Type", "AgentV1SettingsAgentListenProvider_V1", "AgentV1SettingsAgentListenProvider_V2", "AgentV1SettingsAgentSpeak", + "AgentV1SettingsAgentSpeakEndpoint", + "AgentV1SettingsAgentSpeakEndpointEndpoint", + "AgentV1SettingsAgentSpeakEndpointProvider", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPolly", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel", + "AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAi", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly", + "AgentV1SettingsAgentSpeakEndpointProvider_Cartesia", + "AgentV1SettingsAgentSpeakEndpointProvider_Deepgram", + "AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakEndpointProvider_OpenAi", + "AgentV1SettingsAgentSpeakOneItem", + "AgentV1SettingsAgentSpeakOneItemEndpoint", + "AgentV1SettingsAgentSpeakOneItemProvider", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPolly", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine", + "AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion", + "AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel", + "AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId", + "AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAi", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion", + "AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice", + "AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly", + "AgentV1SettingsAgentSpeakOneItemProvider_Cartesia", + "AgentV1SettingsAgentSpeakOneItemProvider_Deepgram", + "AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs", + "AgentV1SettingsAgentSpeakOneItemProvider_OpenAi", "AgentV1SettingsAgentThink", + "AgentV1SettingsAgentThinkOneItem", + "AgentV1SettingsAgentThinkOneItemContextLength", + "AgentV1SettingsAgentThinkOneItemEndpoint", + "AgentV1SettingsAgentThinkOneItemFunctionsItem", + "AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint", + "AgentV1SettingsAgentThinkOneItemProvider", "AgentV1SettingsApplied", + "AgentV1SettingsAppliedType", "AgentV1SettingsAudio", "AgentV1SettingsAudioInput", "AgentV1SettingsAudioInputEncoding", "AgentV1SettingsAudioOutput", "AgentV1SettingsAudioOutputEncoding", "AgentV1SettingsFlags", + "AgentV1SettingsType", "AgentV1SpeakUpdated", + "AgentV1SpeakUpdatedType", "AgentV1UpdatePrompt", + "AgentV1UpdatePromptType", "AgentV1UpdateSpeak", + "AgentV1UpdateSpeakSpeak", + "AgentV1UpdateSpeakSpeakEndpoint", + "AgentV1UpdateSpeakSpeakEndpointEndpoint", + "AgentV1UpdateSpeakSpeakEndpointProvider", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesia", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgram", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAi", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia", + "AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram", + "AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi", + "AgentV1UpdateSpeakSpeakOneItem", + "AgentV1UpdateSpeakSpeakOneItemEndpoint", + "AgentV1UpdateSpeakSpeakOneItemProvider", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine", + "AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesia", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgram", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel", + "AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId", + "AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAi", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion", + "AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice", + "AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly", + "AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia", + "AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram", + "AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs", + "AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi", + "AgentV1UpdateSpeakType", "AgentV1UserStartedSpeaking", + "AgentV1UserStartedSpeakingType", "AgentV1Warning", + "AgentV1WarningType", "AgentV1Welcome", + "AgentV1WelcomeType", + "Cartesia", + "Deepgram", + "ElevenLabs", + "Max", ] diff --git a/src/deepgram/agent/v1/types/agent_v1agent_audio_done.py b/src/deepgram/agent/v1/types/agent_v1agent_audio_done.py index a8d9fddd..36b8a608 100644 --- a/src/deepgram/agent/v1/types/agent_v1agent_audio_done.py +++ b/src/deepgram/agent/v1/types/agent_v1agent_audio_done.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1agent_audio_done_type import AgentV1AgentAudioDoneType class AgentV1AgentAudioDone(UncheckedBaseModel): - type: typing.Literal["AgentAudioDone"] = pydantic.Field(default="AgentAudioDone") + type: AgentV1AgentAudioDoneType = pydantic.Field() """ Message type identifier indicating the agent has finished sending audio """ diff --git a/src/deepgram/agent/v1/types/agent_v1agent_audio_done_type.py b/src/deepgram/agent/v1/types/agent_v1agent_audio_done_type.py new file mode 100644 index 00000000..3fafb98d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1agent_audio_done_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1AgentAudioDoneType = typing.Union[typing.Literal["AgentAudioDone"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1agent_started_speaking.py b/src/deepgram/agent/v1/types/agent_v1agent_started_speaking.py index ba913bf5..a85ba7d2 100644 --- a/src/deepgram/agent/v1/types/agent_v1agent_started_speaking.py +++ b/src/deepgram/agent/v1/types/agent_v1agent_started_speaking.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1agent_started_speaking_type import AgentV1AgentStartedSpeakingType class AgentV1AgentStartedSpeaking(UncheckedBaseModel): - type: typing.Literal["AgentStartedSpeaking"] = pydantic.Field(default="AgentStartedSpeaking") + type: AgentV1AgentStartedSpeakingType = pydantic.Field() """ Message type identifier for agent started speaking """ diff --git a/src/deepgram/agent/v1/types/agent_v1agent_started_speaking_type.py b/src/deepgram/agent/v1/types/agent_v1agent_started_speaking_type.py new file mode 100644 index 00000000..e677505a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1agent_started_speaking_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1AgentStartedSpeakingType = typing.Union[typing.Literal["AgentStartedSpeaking"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1agent_thinking.py b/src/deepgram/agent/v1/types/agent_v1agent_thinking.py index 7c3e6531..77dad48b 100644 --- a/src/deepgram/agent/v1/types/agent_v1agent_thinking.py +++ b/src/deepgram/agent/v1/types/agent_v1agent_thinking.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1agent_thinking_type import AgentV1AgentThinkingType class AgentV1AgentThinking(UncheckedBaseModel): - type: typing.Literal["AgentThinking"] = pydantic.Field(default="AgentThinking") + type: AgentV1AgentThinkingType = pydantic.Field() """ Message type identifier for agent thinking """ diff --git a/src/deepgram/agent/v1/types/agent_v1agent_thinking_type.py b/src/deepgram/agent/v1/types/agent_v1agent_thinking_type.py new file mode 100644 index 00000000..34516f80 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1agent_thinking_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1AgentThinkingType = typing.Union[typing.Literal["AgentThinking"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1conversation_text.py b/src/deepgram/agent/v1/types/agent_v1conversation_text.py index 10148aec..30722582 100644 --- a/src/deepgram/agent/v1/types/agent_v1conversation_text.py +++ b/src/deepgram/agent/v1/types/agent_v1conversation_text.py @@ -6,10 +6,11 @@ from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel from .agent_v1conversation_text_role import AgentV1ConversationTextRole +from .agent_v1conversation_text_type import AgentV1ConversationTextType class AgentV1ConversationText(UncheckedBaseModel): - type: typing.Literal["ConversationText"] = pydantic.Field(default="ConversationText") + type: AgentV1ConversationTextType = pydantic.Field() """ Message type identifier for conversation text """ diff --git a/src/deepgram/agent/v1/types/agent_v1conversation_text_type.py b/src/deepgram/agent/v1/types/agent_v1conversation_text_type.py new file mode 100644 index 00000000..6bb3a7cb --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1conversation_text_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1ConversationTextType = typing.Union[typing.Literal["ConversationText"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1error.py b/src/deepgram/agent/v1/types/agent_v1error.py index aea1228b..ae292e40 100644 --- a/src/deepgram/agent/v1/types/agent_v1error.py +++ b/src/deepgram/agent/v1/types/agent_v1error.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1error_type import AgentV1ErrorType class AgentV1Error(UncheckedBaseModel): - type: typing.Literal["Error"] = pydantic.Field(default="Error") + type: AgentV1ErrorType = pydantic.Field() """ Message type identifier for error responses """ diff --git a/src/deepgram/types/think_settings_v1context_length.py b/src/deepgram/agent/v1/types/agent_v1error_type.py similarity index 51% rename from src/deepgram/types/think_settings_v1context_length.py rename to src/deepgram/agent/v1/types/agent_v1error_type.py index dc9a6555..3838a7b1 100644 --- a/src/deepgram/types/think_settings_v1context_length.py +++ b/src/deepgram/agent/v1/types/agent_v1error_type.py @@ -2,4 +2,4 @@ import typing -ThinkSettingsV1ContextLength = typing.Union[typing.Literal["max"], float] +AgentV1ErrorType = typing.Union[typing.Literal["Error"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1function_call_request.py b/src/deepgram/agent/v1/types/agent_v1function_call_request.py index c587938e..8470e291 100644 --- a/src/deepgram/agent/v1/types/agent_v1function_call_request.py +++ b/src/deepgram/agent/v1/types/agent_v1function_call_request.py @@ -6,10 +6,11 @@ from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel from .agent_v1function_call_request_functions_item import AgentV1FunctionCallRequestFunctionsItem +from .agent_v1function_call_request_type import AgentV1FunctionCallRequestType class AgentV1FunctionCallRequest(UncheckedBaseModel): - type: typing.Literal["FunctionCallRequest"] = pydantic.Field(default="FunctionCallRequest") + type: AgentV1FunctionCallRequestType = pydantic.Field() """ Message type identifier for function call requests """ diff --git a/src/deepgram/agent/v1/types/agent_v1function_call_request_type.py b/src/deepgram/agent/v1/types/agent_v1function_call_request_type.py new file mode 100644 index 00000000..8fbcd24d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1function_call_request_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1FunctionCallRequestType = typing.Union[typing.Literal["FunctionCallRequest"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1inject_agent_message.py b/src/deepgram/agent/v1/types/agent_v1inject_agent_message.py index a2ca5886..b7435ced 100644 --- a/src/deepgram/agent/v1/types/agent_v1inject_agent_message.py +++ b/src/deepgram/agent/v1/types/agent_v1inject_agent_message.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1inject_agent_message_type import AgentV1InjectAgentMessageType class AgentV1InjectAgentMessage(UncheckedBaseModel): - type: typing.Literal["InjectAgentMessage"] = pydantic.Field(default="InjectAgentMessage") + type: AgentV1InjectAgentMessageType = pydantic.Field() """ Message type identifier for injecting an agent message """ diff --git a/src/deepgram/agent/v1/types/agent_v1inject_agent_message_type.py b/src/deepgram/agent/v1/types/agent_v1inject_agent_message_type.py new file mode 100644 index 00000000..0bda4261 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1inject_agent_message_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1InjectAgentMessageType = typing.Union[typing.Literal["InjectAgentMessage"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1inject_user_message.py b/src/deepgram/agent/v1/types/agent_v1inject_user_message.py index f10cd987..bf0df4b2 100644 --- a/src/deepgram/agent/v1/types/agent_v1inject_user_message.py +++ b/src/deepgram/agent/v1/types/agent_v1inject_user_message.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1inject_user_message_type import AgentV1InjectUserMessageType class AgentV1InjectUserMessage(UncheckedBaseModel): - type: typing.Literal["InjectUserMessage"] = pydantic.Field(default="InjectUserMessage") + type: AgentV1InjectUserMessageType = pydantic.Field() """ Message type identifier for injecting a user message """ diff --git a/src/deepgram/agent/v1/types/agent_v1inject_user_message_type.py b/src/deepgram/agent/v1/types/agent_v1inject_user_message_type.py new file mode 100644 index 00000000..54f3d351 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1inject_user_message_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1InjectUserMessageType = typing.Union[typing.Literal["InjectUserMessage"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1injection_refused.py b/src/deepgram/agent/v1/types/agent_v1injection_refused.py index 619fb932..f557ce09 100644 --- a/src/deepgram/agent/v1/types/agent_v1injection_refused.py +++ b/src/deepgram/agent/v1/types/agent_v1injection_refused.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1injection_refused_type import AgentV1InjectionRefusedType class AgentV1InjectionRefused(UncheckedBaseModel): - type: typing.Literal["InjectionRefused"] = pydantic.Field(default="InjectionRefused") + type: AgentV1InjectionRefusedType = pydantic.Field() """ Message type identifier for injection refused """ diff --git a/src/deepgram/agent/v1/types/agent_v1injection_refused_type.py b/src/deepgram/agent/v1/types/agent_v1injection_refused_type.py new file mode 100644 index 00000000..d44019c0 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1injection_refused_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1InjectionRefusedType = typing.Union[typing.Literal["InjectionRefused"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1keep_alive.py b/src/deepgram/agent/v1/types/agent_v1keep_alive.py index 4dcfc4e0..1b69a573 100644 --- a/src/deepgram/agent/v1/types/agent_v1keep_alive.py +++ b/src/deepgram/agent/v1/types/agent_v1keep_alive.py @@ -5,6 +5,7 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1keep_alive_type import AgentV1KeepAliveType class AgentV1KeepAlive(UncheckedBaseModel): @@ -12,7 +13,7 @@ class AgentV1KeepAlive(UncheckedBaseModel): Send a control message to the agent """ - type: typing.Literal["KeepAlive"] = pydantic.Field(default="KeepAlive") + type: AgentV1KeepAliveType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/read/v1/text/types/text_analyze_request_summarize.py b/src/deepgram/agent/v1/types/agent_v1keep_alive_type.py similarity index 59% rename from src/deepgram/read/v1/text/types/text_analyze_request_summarize.py rename to src/deepgram/agent/v1/types/agent_v1keep_alive_type.py index 74aa7e3b..01adb138 100644 --- a/src/deepgram/read/v1/text/types/text_analyze_request_summarize.py +++ b/src/deepgram/agent/v1/types/agent_v1keep_alive_type.py @@ -2,4 +2,4 @@ import typing -TextAnalyzeRequestSummarize = typing.Union[typing.Literal["v2"], typing.Any] +AgentV1KeepAliveType = typing.Union[typing.Literal["KeepAlive"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1prompt_updated.py b/src/deepgram/agent/v1/types/agent_v1prompt_updated.py index f0840cd5..6f5c7d43 100644 --- a/src/deepgram/agent/v1/types/agent_v1prompt_updated.py +++ b/src/deepgram/agent/v1/types/agent_v1prompt_updated.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1prompt_updated_type import AgentV1PromptUpdatedType class AgentV1PromptUpdated(UncheckedBaseModel): - type: typing.Literal["PromptUpdated"] = pydantic.Field(default="PromptUpdated") + type: AgentV1PromptUpdatedType = pydantic.Field() """ Message type identifier for prompt update confirmation """ diff --git a/src/deepgram/agent/v1/types/agent_v1prompt_updated_type.py b/src/deepgram/agent/v1/types/agent_v1prompt_updated_type.py new file mode 100644 index 00000000..7b3cd88d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1prompt_updated_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1PromptUpdatedType = typing.Union[typing.Literal["PromptUpdated"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1receive_function_call_response.py b/src/deepgram/agent/v1/types/agent_v1receive_function_call_response.py index 3501450d..b7dcd120 100644 --- a/src/deepgram/agent/v1/types/agent_v1receive_function_call_response.py +++ b/src/deepgram/agent/v1/types/agent_v1receive_function_call_response.py @@ -5,6 +5,7 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1receive_function_call_response_type import AgentV1ReceiveFunctionCallResponseType class AgentV1ReceiveFunctionCallResponse(UncheckedBaseModel): @@ -20,7 +21,7 @@ class AgentV1ReceiveFunctionCallResponse(UncheckedBaseModel): interface for function call responses regardless of execution location. """ - type: typing.Literal["FunctionCallResponse"] = pydantic.Field(default="FunctionCallResponse") + type: AgentV1ReceiveFunctionCallResponseType = pydantic.Field() """ Message type identifier for function call responses """ diff --git a/src/deepgram/agent/v1/types/agent_v1receive_function_call_response_type.py b/src/deepgram/agent/v1/types/agent_v1receive_function_call_response_type.py new file mode 100644 index 00000000..8886b2fd --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1receive_function_call_response_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1ReceiveFunctionCallResponseType = typing.Union[typing.Literal["FunctionCallResponse"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1send_function_call_response.py b/src/deepgram/agent/v1/types/agent_v1send_function_call_response.py index 6cdef2ff..5d635197 100644 --- a/src/deepgram/agent/v1/types/agent_v1send_function_call_response.py +++ b/src/deepgram/agent/v1/types/agent_v1send_function_call_response.py @@ -5,6 +5,7 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1send_function_call_response_type import AgentV1SendFunctionCallResponseType class AgentV1SendFunctionCallResponse(UncheckedBaseModel): @@ -20,7 +21,7 @@ class AgentV1SendFunctionCallResponse(UncheckedBaseModel): interface for function call responses regardless of execution location. """ - type: typing.Literal["FunctionCallResponse"] = pydantic.Field(default="FunctionCallResponse") + type: AgentV1SendFunctionCallResponseType = pydantic.Field() """ Message type identifier for function call responses """ diff --git a/src/deepgram/agent/v1/types/agent_v1send_function_call_response_type.py b/src/deepgram/agent/v1/types/agent_v1send_function_call_response_type.py new file mode 100644 index 00000000..3712790e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1send_function_call_response_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SendFunctionCallResponseType = typing.Union[typing.Literal["FunctionCallResponse"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings.py b/src/deepgram/agent/v1/types/agent_v1settings.py index 11601dfc..82de5bbf 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings.py +++ b/src/deepgram/agent/v1/types/agent_v1settings.py @@ -8,10 +8,11 @@ from .agent_v1settings_agent import AgentV1SettingsAgent from .agent_v1settings_audio import AgentV1SettingsAudio from .agent_v1settings_flags import AgentV1SettingsFlags +from .agent_v1settings_type import AgentV1SettingsType class AgentV1Settings(UncheckedBaseModel): - type: typing.Literal["Settings"] = "Settings" + type: AgentV1SettingsType tags: typing.Optional[typing.List[str]] = pydantic.Field(default=None) """ Tags to associate with the request diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content.py index c1ad74ec..812518f9 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content.py @@ -8,6 +8,9 @@ from .agent_v1settings_agent_context_messages_item_content_role import ( AgentV1SettingsAgentContextMessagesItemContentRole, ) +from .agent_v1settings_agent_context_messages_item_content_type import ( + AgentV1SettingsAgentContextMessagesItemContentType, +) class AgentV1SettingsAgentContextMessagesItemContent(UncheckedBaseModel): @@ -15,7 +18,7 @@ class AgentV1SettingsAgentContextMessagesItemContent(UncheckedBaseModel): Conversation text as part of the conversation history """ - type: typing.Literal["History"] = pydantic.Field(default="History") + type: AgentV1SettingsAgentContextMessagesItemContentType = pydantic.Field() """ Message type identifier for conversation text """ diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content_type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content_type.py new file mode 100644 index 00000000..49b525f9 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_content_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentContextMessagesItemContentType = typing.Union[typing.Literal["History"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls.py index dd693535..61afaf26 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls.py @@ -8,6 +8,9 @@ from .agent_v1settings_agent_context_messages_item_function_calls_function_calls_item import ( AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem, ) +from .agent_v1settings_agent_context_messages_item_function_calls_type import ( + AgentV1SettingsAgentContextMessagesItemFunctionCallsType, +) class AgentV1SettingsAgentContextMessagesItemFunctionCalls(UncheckedBaseModel): @@ -15,7 +18,7 @@ class AgentV1SettingsAgentContextMessagesItemFunctionCalls(UncheckedBaseModel): Client-side or server-side function call request and response as part of the conversation history """ - type: typing.Literal["History"] = "History" + type: AgentV1SettingsAgentContextMessagesItemFunctionCallsType function_calls: typing.List[AgentV1SettingsAgentContextMessagesItemFunctionCallsFunctionCallsItem] = ( pydantic.Field() ) diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls_type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls_type.py new file mode 100644 index 00000000..18d6ace1 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_context_messages_item_function_calls_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentContextMessagesItemFunctionCallsType = typing.Union[typing.Literal["History"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider.py index 9800b27d..c87d09ab 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider.py @@ -8,11 +8,13 @@ import typing_extensions from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel, UnionMetadata +from .agent_v1settings_agent_listen_provider_v1type import AgentV1SettingsAgentListenProviderV1Type +from .agent_v1settings_agent_listen_provider_v2type import AgentV1SettingsAgentListenProviderV2Type class AgentV1SettingsAgentListenProvider_V1(UncheckedBaseModel): version: typing.Literal["v1"] = "v1" - type: typing.Literal["deepgram"] = "deepgram" + type: AgentV1SettingsAgentListenProviderV1Type model: typing.Optional[str] = None language: typing.Optional[str] = None keyterms: typing.Optional[typing.List[str]] = None @@ -30,7 +32,7 @@ class Config: class AgentV1SettingsAgentListenProvider_V2(UncheckedBaseModel): version: typing.Literal["v2"] = "v2" - type: typing.Literal["deepgram"] = "deepgram" + type: AgentV1SettingsAgentListenProviderV2Type model: str keyterms: typing.Optional[typing.List[str]] = None diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1.py index 3370c965..13bcf34f 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_listen_provider_v1type import AgentV1SettingsAgentListenProviderV1Type class AgentV1SettingsAgentListenProviderV1(UncheckedBaseModel): - type: typing.Literal["deepgram"] = pydantic.Field(default="deepgram") + type: AgentV1SettingsAgentListenProviderV1Type = pydantic.Field() """ Provider type for speech-to-text """ diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1type.py new file mode 100644 index 00000000..1baf7137 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v1type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentListenProviderV1Type = typing.Union[typing.Literal["deepgram"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2.py index 5975d23e..058ec6bd 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_listen_provider_v2type import AgentV1SettingsAgentListenProviderV2Type class AgentV1SettingsAgentListenProviderV2(UncheckedBaseModel): - type: typing.Literal["deepgram"] = pydantic.Field(default="deepgram") + type: AgentV1SettingsAgentListenProviderV2Type = pydantic.Field() """ Provider type for speech-to-text """ diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2type.py new file mode 100644 index 00000000..90180c18 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_listen_provider_v2type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentListenProviderV2Type = typing.Union[typing.Literal["deepgram"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak.py index 84a61a59..0ad6d1c5 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak.py @@ -2,6 +2,9 @@ import typing -from ....types.speak_settings_v1 import SpeakSettingsV1 +from .agent_v1settings_agent_speak_endpoint import AgentV1SettingsAgentSpeakEndpoint +from .agent_v1settings_agent_speak_one_item import AgentV1SettingsAgentSpeakOneItem -AgentV1SettingsAgentSpeak = typing.Union[SpeakSettingsV1, typing.List[SpeakSettingsV1]] +AgentV1SettingsAgentSpeak = typing.Union[ + AgentV1SettingsAgentSpeakEndpoint, typing.List[AgentV1SettingsAgentSpeakOneItem] +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint.py new file mode 100644 index 00000000..d0b3ec7b --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint.py @@ -0,0 +1,27 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_endpoint_endpoint import AgentV1SettingsAgentSpeakEndpointEndpoint +from .agent_v1settings_agent_speak_endpoint_provider import AgentV1SettingsAgentSpeakEndpointProvider + + +class AgentV1SettingsAgentSpeakEndpoint(UncheckedBaseModel): + provider: AgentV1SettingsAgentSpeakEndpointProvider + endpoint: typing.Optional[AgentV1SettingsAgentSpeakEndpointEndpoint] = pydantic.Field(default=None) + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_endpoint.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_endpoint.py new file mode 100644 index 00000000..2778435f --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_endpoint.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1SettingsAgentSpeakEndpointEndpoint(UncheckedBaseModel): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing.Optional[str] = pydantic.Field(default=None) + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing.Optional[typing.Dict[str, str]] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider.py new file mode 100644 index 00000000..c7d724b2 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider.py @@ -0,0 +1,144 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import pydantic +import typing_extensions +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel, UnionMetadata +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, +) +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, +) +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, +) +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from .agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) + + +class AgentV1SettingsAgentSpeakEndpointProvider_Deepgram(UncheckedBaseModel): + type: typing.Literal["deepgram"] = "deepgram" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] = None + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs(UncheckedBaseModel): + type: typing.Literal["eleven_labs"] = "eleven_labs" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] = None + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId + language: typing.Optional[str] = None + language_code: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakEndpointProvider_Cartesia(UncheckedBaseModel): + type: typing.Literal["cartesia"] = "cartesia" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] = None + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice + language: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakEndpointProvider_OpenAi(UncheckedBaseModel): + type: typing.Literal["open_ai"] = "open_ai" + version: typing.Optional[AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion] = None + model: AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel + voice: AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly(UncheckedBaseModel): + type: typing.Literal["aws_polly"] = "aws_polly" + voice: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice + language: str + language_code: typing.Optional[str] = None + engine: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +AgentV1SettingsAgentSpeakEndpointProvider = typing_extensions.Annotated[ + typing.Union[ + AgentV1SettingsAgentSpeakEndpointProvider_Deepgram, + AgentV1SettingsAgentSpeakEndpointProvider_ElevenLabs, + AgentV1SettingsAgentSpeakEndpointProvider_Cartesia, + AgentV1SettingsAgentSpeakEndpointProvider_OpenAi, + AgentV1SettingsAgentSpeakEndpointProvider_AwsPolly, + ], + UnionMetadata(discriminant="type"), +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py new file mode 100644 index 00000000..256cfba6 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly.py @@ -0,0 +1,45 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine, +) +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderAwsPolly(UncheckedBaseModel): + voice: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice = pydantic.Field() + """ + AWS Polly voice name + """ + + language: str = pydantic.Field() + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + engine: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py new file mode 100644 index 00000000..95284bf1 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentials(UncheckedBaseModel): + type: AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing.Optional[str] = pydantic.Field(default=None) + """ + Required for STS only + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type.py new file mode 100644 index 00000000..515f0617 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_credentials_type.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderAwsPollyCredentialsType = typing.Union[ + typing.Literal["sts", "iam"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine.py new file mode 100644 index 00000000..2f182419 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_engine.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderAwsPollyEngine = typing.Union[ + typing.Literal["generative", "long-form", "standard", "neural"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice.py new file mode 100644 index 00000000..0079e7b3 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_aws_polly_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderAwsPollyVoice = typing.Union[ + typing.Literal["Matthew", "Joanna", "Amy", "Emma", "Brian", "Arthur", "Aria", "Ayanda"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_model_id.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_model_id.py new file mode 100644 index 00000000..b81e30b3 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderCartesiaModelId = typing.Union[ + typing.Literal["sonic-2", "sonic-multilingual"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_version.py new file mode 100644 index 00000000..f29cec69 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderCartesiaVersion = typing.Union[typing.Literal["2025-03-17"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py new file mode 100644 index 00000000..4801692c --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_cartesia_voice.py @@ -0,0 +1,28 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1SettingsAgentSpeakEndpointProviderCartesiaVoice(UncheckedBaseModel): + mode: str = pydantic.Field() + """ + Cartesia voice mode + """ + + id: str = pydantic.Field() + """ + Cartesia voice ID + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_model.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_model.py new file mode 100644 index 00000000..161119d8 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_model.py @@ -0,0 +1,72 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderDeepgramModel = typing.Union[ + typing.Literal[ + "aura-asteria-en", + "aura-luna-en", + "aura-stella-en", + "aura-athena-en", + "aura-hera-en", + "aura-orion-en", + "aura-arcas-en", + "aura-perseus-en", + "aura-angus-en", + "aura-orpheus-en", + "aura-helios-en", + "aura-zeus-en", + "aura-2-amalthea-en", + "aura-2-andromeda-en", + "aura-2-apollo-en", + "aura-2-arcas-en", + "aura-2-aries-en", + "aura-2-asteria-en", + "aura-2-athena-en", + "aura-2-atlas-en", + "aura-2-aurora-en", + "aura-2-callista-en", + "aura-2-cora-en", + "aura-2-cordelia-en", + "aura-2-delia-en", + "aura-2-draco-en", + "aura-2-electra-en", + "aura-2-harmonia-en", + "aura-2-helena-en", + "aura-2-hera-en", + "aura-2-hermes-en", + "aura-2-hyperion-en", + "aura-2-iris-en", + "aura-2-janus-en", + "aura-2-juno-en", + "aura-2-jupiter-en", + "aura-2-luna-en", + "aura-2-mars-en", + "aura-2-minerva-en", + "aura-2-neptune-en", + "aura-2-odysseus-en", + "aura-2-ophelia-en", + "aura-2-orion-en", + "aura-2-orpheus-en", + "aura-2-pandora-en", + "aura-2-phoebe-en", + "aura-2-pluto-en", + "aura-2-saturn-en", + "aura-2-selene-en", + "aura-2-thalia-en", + "aura-2-theia-en", + "aura-2-vesta-en", + "aura-2-zeus-en", + "aura-2-sirio-es", + "aura-2-nestor-es", + "aura-2-carina-es", + "aura-2-celeste-es", + "aura-2-alvaro-es", + "aura-2-diana-es", + "aura-2-aquila-es", + "aura-2-selena-es", + "aura-2-estrella-es", + "aura-2-javier-es", + ], + typing.Any, +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_version.py new file mode 100644 index 00000000..e9cd4832 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_deepgram_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderDeepgramVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_model_id.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_model_id.py new file mode 100644 index 00000000..4ed8c7e8 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderElevenLabsModelId = typing.Union[ + typing.Literal["eleven_turbo_v2_5", "eleven_monolingual_v1", "eleven_multilingual_v2"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_version.py new file mode 100644 index 00000000..2a21cb4c --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_eleven_labs_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderElevenLabsVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai.py new file mode 100644 index 00000000..506f5245 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai.py @@ -0,0 +1,42 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel, +) +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion, +) +from .agent_v1settings_agent_speak_endpoint_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice, +) + + +class AgentV1SettingsAgentSpeakEndpointProviderOpenAi(UncheckedBaseModel): + version: typing.Optional[AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion] = pydantic.Field(default=None) + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel = pydantic.Field() + """ + OpenAI TTS model + """ + + voice: AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice = pydantic.Field() + """ + OpenAI voice + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_model.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_model.py new file mode 100644 index 00000000..f83a1943 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_model.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderOpenAiModel = typing.Union[typing.Literal["tts-1", "tts-1-hd"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_version.py new file mode 100644 index 00000000..8dfcd1d7 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderOpenAiVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_voice.py new file mode 100644 index 00000000..0e8a10eb --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_endpoint_provider_open_ai_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakEndpointProviderOpenAiVoice = typing.Union[ + typing.Literal["alloy", "echo", "fable", "onyx", "nova", "shimmer"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item.py new file mode 100644 index 00000000..b95fc91a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item.py @@ -0,0 +1,27 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_endpoint import AgentV1SettingsAgentSpeakOneItemEndpoint +from .agent_v1settings_agent_speak_one_item_provider import AgentV1SettingsAgentSpeakOneItemProvider + + +class AgentV1SettingsAgentSpeakOneItem(UncheckedBaseModel): + provider: AgentV1SettingsAgentSpeakOneItemProvider + endpoint: typing.Optional[AgentV1SettingsAgentSpeakOneItemEndpoint] = pydantic.Field(default=None) + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_endpoint.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_endpoint.py new file mode 100644 index 00000000..50e760d6 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_endpoint.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1SettingsAgentSpeakOneItemEndpoint(UncheckedBaseModel): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing.Optional[str] = pydantic.Field(default=None) + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing.Optional[typing.Dict[str, str]] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider.py new file mode 100644 index 00000000..5c7c7c82 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider.py @@ -0,0 +1,144 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import pydantic +import typing_extensions +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel, UnionMetadata +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, +) +from .agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from .agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, +) +from .agent_v1settings_agent_speak_one_item_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1SettingsAgentSpeakOneItemProvider_Deepgram(UncheckedBaseModel): + type: typing.Literal["deepgram"] = "deepgram" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] = None + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs(UncheckedBaseModel): + type: typing.Literal["eleven_labs"] = "eleven_labs" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] = None + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId + language: typing.Optional[str] = None + language_code: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakOneItemProvider_Cartesia(UncheckedBaseModel): + type: typing.Literal["cartesia"] = "cartesia" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] = None + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice + language: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakOneItemProvider_OpenAi(UncheckedBaseModel): + type: typing.Literal["open_ai"] = "open_ai" + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion] = None + model: AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel + voice: AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly(UncheckedBaseModel): + type: typing.Literal["aws_polly"] = "aws_polly" + voice: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice + language: str + language_code: typing.Optional[str] = None + engine: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +AgentV1SettingsAgentSpeakOneItemProvider = typing_extensions.Annotated[ + typing.Union[ + AgentV1SettingsAgentSpeakOneItemProvider_Deepgram, + AgentV1SettingsAgentSpeakOneItemProvider_ElevenLabs, + AgentV1SettingsAgentSpeakOneItemProvider_Cartesia, + AgentV1SettingsAgentSpeakOneItemProvider_OpenAi, + AgentV1SettingsAgentSpeakOneItemProvider_AwsPolly, + ], + UnionMetadata(discriminant="type"), +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly.py new file mode 100644 index 00000000..11c3ab25 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly.py @@ -0,0 +1,45 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_engine import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine, +) +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderAwsPolly(UncheckedBaseModel): + voice: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice = pydantic.Field() + """ + AWS Polly voice name + """ + + language: str = pydantic.Field() + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + engine: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py new file mode 100644 index 00000000..d8dcb07e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentials(UncheckedBaseModel): + type: AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing.Optional[str] = pydantic.Field(default=None) + """ + Required for STS only + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type.py new file mode 100644 index 00000000..3cd1c64f --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_credentials_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderAwsPollyCredentialsType = typing.Union[typing.Literal["sts", "iam"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_engine.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_engine.py new file mode 100644 index 00000000..7313bd82 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_engine.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderAwsPollyEngine = typing.Union[ + typing.Literal["generative", "long-form", "standard", "neural"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_voice.py new file mode 100644 index 00000000..ad77ee0f --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_aws_polly_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderAwsPollyVoice = typing.Union[ + typing.Literal["Matthew", "Joanna", "Amy", "Emma", "Brian", "Arthur", "Aria", "Ayanda"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_model_id.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_model_id.py new file mode 100644 index 00000000..8d062938 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId = typing.Union[ + typing.Literal["sonic-2", "sonic-multilingual"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_version.py new file mode 100644 index 00000000..8a7e04a5 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion = typing.Union[typing.Literal["2025-03-17"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py new file mode 100644 index 00000000..783363d3 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_cartesia_voice.py @@ -0,0 +1,28 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice(UncheckedBaseModel): + mode: str = pydantic.Field() + """ + Cartesia voice mode + """ + + id: str = pydantic.Field() + """ + Cartesia voice ID + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_model.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_model.py new file mode 100644 index 00000000..2c896d8d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_model.py @@ -0,0 +1,72 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel = typing.Union[ + typing.Literal[ + "aura-asteria-en", + "aura-luna-en", + "aura-stella-en", + "aura-athena-en", + "aura-hera-en", + "aura-orion-en", + "aura-arcas-en", + "aura-perseus-en", + "aura-angus-en", + "aura-orpheus-en", + "aura-helios-en", + "aura-zeus-en", + "aura-2-amalthea-en", + "aura-2-andromeda-en", + "aura-2-apollo-en", + "aura-2-arcas-en", + "aura-2-aries-en", + "aura-2-asteria-en", + "aura-2-athena-en", + "aura-2-atlas-en", + "aura-2-aurora-en", + "aura-2-callista-en", + "aura-2-cora-en", + "aura-2-cordelia-en", + "aura-2-delia-en", + "aura-2-draco-en", + "aura-2-electra-en", + "aura-2-harmonia-en", + "aura-2-helena-en", + "aura-2-hera-en", + "aura-2-hermes-en", + "aura-2-hyperion-en", + "aura-2-iris-en", + "aura-2-janus-en", + "aura-2-juno-en", + "aura-2-jupiter-en", + "aura-2-luna-en", + "aura-2-mars-en", + "aura-2-minerva-en", + "aura-2-neptune-en", + "aura-2-odysseus-en", + "aura-2-ophelia-en", + "aura-2-orion-en", + "aura-2-orpheus-en", + "aura-2-pandora-en", + "aura-2-phoebe-en", + "aura-2-pluto-en", + "aura-2-saturn-en", + "aura-2-selene-en", + "aura-2-thalia-en", + "aura-2-theia-en", + "aura-2-vesta-en", + "aura-2-zeus-en", + "aura-2-sirio-es", + "aura-2-nestor-es", + "aura-2-carina-es", + "aura-2-celeste-es", + "aura-2-alvaro-es", + "aura-2-diana-es", + "aura-2-aquila-es", + "aura-2-selena-es", + "aura-2-estrella-es", + "aura-2-javier-es", + ], + typing.Any, +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_version.py new file mode 100644 index 00000000..db5b1714 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_deepgram_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id.py new file mode 100644 index 00000000..f2bbee02 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId = typing.Union[ + typing.Literal["eleven_turbo_v2_5", "eleven_monolingual_v1", "eleven_multilingual_v2"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_version.py new file mode 100644 index 00000000..5498f864 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_eleven_labs_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai.py new file mode 100644 index 00000000..c21f1b63 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai.py @@ -0,0 +1,42 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_open_ai_model import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel, +) +from .agent_v1settings_agent_speak_one_item_provider_open_ai_version import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_open_ai_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1SettingsAgentSpeakOneItemProviderOpenAi(UncheckedBaseModel): + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion] = pydantic.Field(default=None) + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel = pydantic.Field() + """ + OpenAI TTS model + """ + + voice: AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice = pydantic.Field() + """ + OpenAI voice + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_model.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_model.py new file mode 100644 index 00000000..21ea5697 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_model.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderOpenAiModel = typing.Union[typing.Literal["tts-1", "tts-1-hd"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_version.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_version.py new file mode 100644 index 00000000..f5201c98 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderOpenAiVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_voice.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_voice.py new file mode 100644 index 00000000..7e4bc122 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_speak_one_item_provider_open_ai_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAgentSpeakOneItemProviderOpenAiVoice = typing.Union[ + typing.Literal["alloy", "echo", "fable", "onyx", "nova", "shimmer"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_think.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think.py index 2521acf2..f4a89bf9 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_agent_think.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think.py @@ -3,5 +3,6 @@ import typing from ....types.think_settings_v1 import ThinkSettingsV1 +from .agent_v1settings_agent_think_one_item import AgentV1SettingsAgentThinkOneItem -AgentV1SettingsAgentThink = typing.Union[ThinkSettingsV1, typing.List[ThinkSettingsV1]] +AgentV1SettingsAgentThink = typing.Union[ThinkSettingsV1, typing.List[AgentV1SettingsAgentThinkOneItem]] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item.py new file mode 100644 index 00000000..d0fa3ebb --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item.py @@ -0,0 +1,35 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_think_one_item_context_length import AgentV1SettingsAgentThinkOneItemContextLength +from .agent_v1settings_agent_think_one_item_endpoint import AgentV1SettingsAgentThinkOneItemEndpoint +from .agent_v1settings_agent_think_one_item_functions_item import AgentV1SettingsAgentThinkOneItemFunctionsItem +from .agent_v1settings_agent_think_one_item_provider import AgentV1SettingsAgentThinkOneItemProvider + + +class AgentV1SettingsAgentThinkOneItem(UncheckedBaseModel): + provider: AgentV1SettingsAgentThinkOneItemProvider + endpoint: typing.Optional[AgentV1SettingsAgentThinkOneItemEndpoint] = pydantic.Field(default=None) + """ + Optional for non-Deepgram LLM providers. When present, must include url field and headers object + """ + + functions: typing.Optional[typing.List[AgentV1SettingsAgentThinkOneItemFunctionsItem]] = None + prompt: typing.Optional[str] = None + context_length: typing.Optional[AgentV1SettingsAgentThinkOneItemContextLength] = pydantic.Field(default=None) + """ + Specifies the number of characters retained in context between user messages, agent responses, and function calls. This setting is only configurable when a custom think endpoint is used + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_context_length.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_context_length.py new file mode 100644 index 00000000..e2c64663 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_context_length.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from .max import Max + +AgentV1SettingsAgentThinkOneItemContextLength = typing.Union[Max, float] diff --git a/src/deepgram/types/think_settings_v1endpoint.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_endpoint.py similarity index 80% rename from src/deepgram/types/think_settings_v1endpoint.py rename to src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_endpoint.py index a98942ca..09e0a99d 100644 --- a/src/deepgram/types/think_settings_v1endpoint.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_endpoint.py @@ -3,11 +3,11 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel -class ThinkSettingsV1Endpoint(UncheckedBaseModel): +class AgentV1SettingsAgentThinkOneItemEndpoint(UncheckedBaseModel): """ Optional for non-Deepgram LLM providers. When present, must include url field and headers object """ diff --git a/src/deepgram/types/think_settings_v1functions_item.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item.py similarity index 65% rename from src/deepgram/types/think_settings_v1functions_item.py rename to src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item.py index 6a142ae5..0bdf2a29 100644 --- a/src/deepgram/types/think_settings_v1functions_item.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item.py @@ -3,12 +3,14 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .think_settings_v1functions_item_endpoint import ThinkSettingsV1FunctionsItemEndpoint +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_think_one_item_functions_item_endpoint import ( + AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint, +) -class ThinkSettingsV1FunctionsItem(UncheckedBaseModel): +class AgentV1SettingsAgentThinkOneItemFunctionsItem(UncheckedBaseModel): name: typing.Optional[str] = pydantic.Field(default=None) """ Function name @@ -24,7 +26,7 @@ class ThinkSettingsV1FunctionsItem(UncheckedBaseModel): Function parameters """ - endpoint: typing.Optional[ThinkSettingsV1FunctionsItemEndpoint] = pydantic.Field(default=None) + endpoint: typing.Optional[AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint] = pydantic.Field(default=None) """ The Function endpoint to call. if not passed, function is called client-side """ diff --git a/src/deepgram/types/think_settings_v1functions_item_endpoint.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item_endpoint.py similarity index 78% rename from src/deepgram/types/think_settings_v1functions_item_endpoint.py rename to src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item_endpoint.py index b3e0148e..74886529 100644 --- a/src/deepgram/types/think_settings_v1functions_item_endpoint.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_functions_item_endpoint.py @@ -3,11 +3,11 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel -class ThinkSettingsV1FunctionsItemEndpoint(UncheckedBaseModel): +class AgentV1SettingsAgentThinkOneItemFunctionsItemEndpoint(UncheckedBaseModel): """ The Function endpoint to call. if not passed, function is called client-side """ diff --git a/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_provider.py b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_provider.py new file mode 100644 index 00000000..239fc6c9 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_agent_think_one_item_provider.py @@ -0,0 +1,13 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from ....types.anthropic import Anthropic +from ....types.aws_bedrock_think_provider import AwsBedrockThinkProvider +from ....types.google import Google +from ....types.groq import Groq +from ....types.open_ai_think_provider import OpenAiThinkProvider + +AgentV1SettingsAgentThinkOneItemProvider = typing.Union[ + OpenAiThinkProvider, AwsBedrockThinkProvider, Anthropic, Google, Groq +] diff --git a/src/deepgram/agent/v1/types/agent_v1settings_applied.py b/src/deepgram/agent/v1/types/agent_v1settings_applied.py index 065643b6..de27cee7 100644 --- a/src/deepgram/agent/v1/types/agent_v1settings_applied.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_applied.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_applied_type import AgentV1SettingsAppliedType class AgentV1SettingsApplied(UncheckedBaseModel): - type: typing.Literal["SettingsApplied"] = pydantic.Field(default="SettingsApplied") + type: AgentV1SettingsAppliedType = pydantic.Field() """ Message type identifier for settings applied confirmation """ diff --git a/src/deepgram/agent/v1/types/agent_v1settings_applied_type.py b/src/deepgram/agent/v1/types/agent_v1settings_applied_type.py new file mode 100644 index 00000000..51a5cf6b --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1settings_applied_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SettingsAppliedType = typing.Union[typing.Literal["SettingsApplied"], typing.Any] diff --git a/src/deepgram/requests/think_settings_v1context_length.py b/src/deepgram/agent/v1/types/agent_v1settings_type.py similarity index 50% rename from src/deepgram/requests/think_settings_v1context_length.py rename to src/deepgram/agent/v1/types/agent_v1settings_type.py index e1955c18..0018c56b 100644 --- a/src/deepgram/requests/think_settings_v1context_length.py +++ b/src/deepgram/agent/v1/types/agent_v1settings_type.py @@ -2,4 +2,4 @@ import typing -ThinkSettingsV1ContextLengthParams = typing.Union[typing.Literal["max"], float] +AgentV1SettingsType = typing.Union[typing.Literal["Settings"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1speak_updated.py b/src/deepgram/agent/v1/types/agent_v1speak_updated.py index 956dcfd7..41a6e4ed 100644 --- a/src/deepgram/agent/v1/types/agent_v1speak_updated.py +++ b/src/deepgram/agent/v1/types/agent_v1speak_updated.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1speak_updated_type import AgentV1SpeakUpdatedType class AgentV1SpeakUpdated(UncheckedBaseModel): - type: typing.Literal["SpeakUpdated"] = pydantic.Field(default="SpeakUpdated") + type: AgentV1SpeakUpdatedType = pydantic.Field() """ Message type identifier for speak update confirmation """ diff --git a/src/deepgram/agent/v1/types/agent_v1speak_updated_type.py b/src/deepgram/agent/v1/types/agent_v1speak_updated_type.py new file mode 100644 index 00000000..ccab8cfc --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1speak_updated_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1SpeakUpdatedType = typing.Union[typing.Literal["SpeakUpdated"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_prompt.py b/src/deepgram/agent/v1/types/agent_v1update_prompt.py index 8cbada41..90162802 100644 --- a/src/deepgram/agent/v1/types/agent_v1update_prompt.py +++ b/src/deepgram/agent/v1/types/agent_v1update_prompt.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_prompt_type import AgentV1UpdatePromptType class AgentV1UpdatePrompt(UncheckedBaseModel): - type: typing.Literal["UpdatePrompt"] = pydantic.Field(default="UpdatePrompt") + type: AgentV1UpdatePromptType = pydantic.Field() """ Message type identifier for prompt update request """ diff --git a/src/deepgram/agent/v1/types/agent_v1update_prompt_type.py b/src/deepgram/agent/v1/types/agent_v1update_prompt_type.py new file mode 100644 index 00000000..ca8122be --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_prompt_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdatePromptType = typing.Union[typing.Literal["UpdatePrompt"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak.py b/src/deepgram/agent/v1/types/agent_v1update_speak.py index 4afc8d37..5d7a7e41 100644 --- a/src/deepgram/agent/v1/types/agent_v1update_speak.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak.py @@ -5,16 +5,17 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel -from ....types.speak_settings_v1 import SpeakSettingsV1 +from .agent_v1update_speak_speak import AgentV1UpdateSpeakSpeak +from .agent_v1update_speak_type import AgentV1UpdateSpeakType class AgentV1UpdateSpeak(UncheckedBaseModel): - type: typing.Literal["UpdateSpeak"] = pydantic.Field(default="UpdateSpeak") + type: AgentV1UpdateSpeakType = pydantic.Field() """ Message type identifier for updating the speak model """ - speak: SpeakSettingsV1 + speak: AgentV1UpdateSpeakSpeak if IS_PYDANTIC_V2: model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak.py new file mode 100644 index 00000000..a5a48320 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak.py @@ -0,0 +1,8 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +from .agent_v1update_speak_speak_endpoint import AgentV1UpdateSpeakSpeakEndpoint +from .agent_v1update_speak_speak_one_item import AgentV1UpdateSpeakSpeakOneItem + +AgentV1UpdateSpeakSpeak = typing.Union[AgentV1UpdateSpeakSpeakEndpoint, typing.List[AgentV1UpdateSpeakSpeakOneItem]] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint.py new file mode 100644 index 00000000..86a7280b --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint.py @@ -0,0 +1,27 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_endpoint import AgentV1UpdateSpeakSpeakEndpointEndpoint +from .agent_v1update_speak_speak_endpoint_provider import AgentV1UpdateSpeakSpeakEndpointProvider + + +class AgentV1UpdateSpeakSpeakEndpoint(UncheckedBaseModel): + provider: AgentV1UpdateSpeakSpeakEndpointProvider + endpoint: typing.Optional[AgentV1UpdateSpeakSpeakEndpointEndpoint] = pydantic.Field(default=None) + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_endpoint.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_endpoint.py new file mode 100644 index 00000000..e8095d3e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_endpoint.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1UpdateSpeakSpeakEndpointEndpoint(UncheckedBaseModel): + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + url: typing.Optional[str] = pydantic.Field(default=None) + """ + Custom TTS endpoint URL. Cannot contain `output_format` or `model_id` query parameters when the provider is Eleven Labs. + """ + + headers: typing.Optional[typing.Dict[str, str]] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider.py new file mode 100644 index 00000000..f2446d53 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider.py @@ -0,0 +1,144 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import pydantic +import typing_extensions +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel, UnionMetadata +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice, +) +from .agent_v1update_speak_speak_endpoint_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, +) +from .agent_v1update_speak_speak_endpoint_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, +) +from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, +) +from .agent_v1update_speak_speak_endpoint_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram(UncheckedBaseModel): + type: typing.Literal["deepgram"] = "deepgram" + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion] = None + model: AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs(UncheckedBaseModel): + type: typing.Literal["eleven_labs"] = "eleven_labs" + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion] = None + model_id: AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId + language: typing.Optional[str] = None + language_code: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia(UncheckedBaseModel): + type: typing.Literal["cartesia"] = "cartesia" + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion] = None + model_id: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId + voice: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice + language: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi(UncheckedBaseModel): + type: typing.Literal["open_ai"] = "open_ai" + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion] = None + model: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel + voice: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly(UncheckedBaseModel): + type: typing.Literal["aws_polly"] = "aws_polly" + voice: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice + language: str + language_code: typing.Optional[str] = None + engine: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +AgentV1UpdateSpeakSpeakEndpointProvider = typing_extensions.Annotated[ + typing.Union[ + AgentV1UpdateSpeakSpeakEndpointProvider_Deepgram, + AgentV1UpdateSpeakSpeakEndpointProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakEndpointProvider_Cartesia, + AgentV1UpdateSpeakSpeakEndpointProvider_OpenAi, + AgentV1UpdateSpeakSpeakEndpointProvider_AwsPolly, + ], + UnionMetadata(discriminant="type"), +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly.py new file mode 100644 index 00000000..11d2fe03 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly.py @@ -0,0 +1,45 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine, +) +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderAwsPolly(UncheckedBaseModel): + voice: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice = pydantic.Field() + """ + AWS Polly voice name + """ + + language: str = pydantic.Field() + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + engine: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py new file mode 100644 index 00000000..f042ae94 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials.py @@ -0,0 +1,30 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentials(UncheckedBaseModel): + type: AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType + region: str + access_key_id: str + secret_access_key: str + session_token: typing.Optional[str] = pydantic.Field(default=None) + """ + Required for STS only + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type.py new file mode 100644 index 00000000..87063f91 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_credentials_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyCredentialsType = typing.Union[typing.Literal["sts", "iam"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_engine.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_engine.py new file mode 100644 index 00000000..5d9f0a82 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_engine.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyEngine = typing.Union[ + typing.Literal["generative", "long-form", "standard", "neural"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_voice.py new file mode 100644 index 00000000..e8a6592d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_aws_polly_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderAwsPollyVoice = typing.Union[ + typing.Literal["Matthew", "Joanna", "Amy", "Emma", "Brian", "Arthur", "Aria", "Ayanda"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia.py new file mode 100644 index 00000000..c8e1d77a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia.py @@ -0,0 +1,43 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderCartesia(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion] = pydantic.Field(default=None) + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId = pydantic.Field() + """ + Cartesia model ID + """ + + voice: AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Cartesia language code + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_model_id.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_model_id.py new file mode 100644 index 00000000..d9032180 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderCartesiaModelId = typing.Union[ + typing.Literal["sonic-2", "sonic-multilingual"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_version.py new file mode 100644 index 00000000..464d35d4 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVersion = typing.Union[typing.Literal["2025-03-17"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py new file mode 100644 index 00000000..c44e1eac --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_cartesia_voice.py @@ -0,0 +1,28 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel + + +class AgentV1UpdateSpeakSpeakEndpointProviderCartesiaVoice(UncheckedBaseModel): + mode: str = pydantic.Field() + """ + Cartesia voice mode + """ + + id: str = pydantic.Field() + """ + Cartesia voice ID + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram.py new file mode 100644 index 00000000..36b8ba77 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram.py @@ -0,0 +1,34 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel, +) +from .agent_v1update_speak_speak_endpoint_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderDeepgram(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion] = pydantic.Field(default=None) + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel = pydantic.Field() + """ + Deepgram TTS model + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_model.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_model.py new file mode 100644 index 00000000..59ca4fd1 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_model.py @@ -0,0 +1,72 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderDeepgramModel = typing.Union[ + typing.Literal[ + "aura-asteria-en", + "aura-luna-en", + "aura-stella-en", + "aura-athena-en", + "aura-hera-en", + "aura-orion-en", + "aura-arcas-en", + "aura-perseus-en", + "aura-angus-en", + "aura-orpheus-en", + "aura-helios-en", + "aura-zeus-en", + "aura-2-amalthea-en", + "aura-2-andromeda-en", + "aura-2-apollo-en", + "aura-2-arcas-en", + "aura-2-aries-en", + "aura-2-asteria-en", + "aura-2-athena-en", + "aura-2-atlas-en", + "aura-2-aurora-en", + "aura-2-callista-en", + "aura-2-cora-en", + "aura-2-cordelia-en", + "aura-2-delia-en", + "aura-2-draco-en", + "aura-2-electra-en", + "aura-2-harmonia-en", + "aura-2-helena-en", + "aura-2-hera-en", + "aura-2-hermes-en", + "aura-2-hyperion-en", + "aura-2-iris-en", + "aura-2-janus-en", + "aura-2-juno-en", + "aura-2-jupiter-en", + "aura-2-luna-en", + "aura-2-mars-en", + "aura-2-minerva-en", + "aura-2-neptune-en", + "aura-2-odysseus-en", + "aura-2-ophelia-en", + "aura-2-orion-en", + "aura-2-orpheus-en", + "aura-2-pandora-en", + "aura-2-phoebe-en", + "aura-2-pluto-en", + "aura-2-saturn-en", + "aura-2-selene-en", + "aura-2-thalia-en", + "aura-2-theia-en", + "aura-2-vesta-en", + "aura-2-zeus-en", + "aura-2-sirio-es", + "aura-2-nestor-es", + "aura-2-carina-es", + "aura-2-celeste-es", + "aura-2-alvaro-es", + "aura-2-diana-es", + "aura-2-aquila-es", + "aura-2-selena-es", + "aura-2-estrella-es", + "aura-2-javier-es", + ], + typing.Any, +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_version.py new file mode 100644 index 00000000..367d022a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_deepgram_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderDeepgramVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py new file mode 100644 index 00000000..6a8d1e7e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs.py @@ -0,0 +1,44 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId, +) +from .agent_v1update_speak_speak_endpoint_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderElevenLabs(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion] = pydantic.Field(default=None) + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId = pydantic.Field() + """ + Eleven Labs model ID + """ + + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id.py new file mode 100644 index 00000000..f8b807ca --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsModelId = typing.Union[ + typing.Literal["eleven_turbo_v2_5", "eleven_monolingual_v1", "eleven_multilingual_v2"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_version.py new file mode 100644 index 00000000..127403d2 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_eleven_labs_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderElevenLabsVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai.py new file mode 100644 index 00000000..29cac1b3 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai.py @@ -0,0 +1,42 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_endpoint_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel, +) +from .agent_v1update_speak_speak_endpoint_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion, +) +from .agent_v1update_speak_speak_endpoint_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakEndpointProviderOpenAi(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion] = pydantic.Field(default=None) + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel = pydantic.Field() + """ + OpenAI TTS model + """ + + voice: AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice = pydantic.Field() + """ + OpenAI voice + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_model.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_model.py new file mode 100644 index 00000000..76522c3e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_model.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderOpenAiModel = typing.Union[typing.Literal["tts-1", "tts-1-hd"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_version.py new file mode 100644 index 00000000..e25be55a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_voice.py new file mode 100644 index 00000000..d360b702 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_endpoint_provider_open_ai_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakEndpointProviderOpenAiVoice = typing.Union[ + typing.Literal["alloy", "echo", "fable", "onyx", "nova", "shimmer"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item.py new file mode 100644 index 00000000..6310beb5 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item.py @@ -0,0 +1,27 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_endpoint import AgentV1UpdateSpeakSpeakOneItemEndpoint +from .agent_v1update_speak_speak_one_item_provider import AgentV1UpdateSpeakSpeakOneItemProvider + + +class AgentV1UpdateSpeakSpeakOneItem(UncheckedBaseModel): + provider: AgentV1UpdateSpeakSpeakOneItemProvider + endpoint: typing.Optional[AgentV1UpdateSpeakSpeakOneItemEndpoint] = pydantic.Field(default=None) + """ + Optional if provider is Deepgram. Required for non-Deepgram TTS providers. + When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/types/speak_settings_v1endpoint.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_endpoint.py similarity index 82% rename from src/deepgram/types/speak_settings_v1endpoint.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_endpoint.py index ac806e29..45694cc7 100644 --- a/src/deepgram/types/speak_settings_v1endpoint.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_endpoint.py @@ -3,11 +3,11 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel -class SpeakSettingsV1Endpoint(UncheckedBaseModel): +class AgentV1UpdateSpeakSpeakOneItemEndpoint(UncheckedBaseModel): """ Optional if provider is Deepgram. Required for non-Deepgram TTS providers. When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider.py new file mode 100644 index 00000000..56702f29 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider.py @@ -0,0 +1,144 @@ +# This file was auto-generated by Fern from our API Definition. + +from __future__ import annotations + +import typing + +import pydantic +import typing_extensions +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel, UnionMetadata +from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice, +) +from .agent_v1update_speak_speak_one_item_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, +) +from .agent_v1update_speak_speak_one_item_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, +) +from .agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, +) +from .agent_v1update_speak_speak_one_item_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, +) +from .agent_v1update_speak_speak_one_item_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, +) +from .agent_v1update_speak_speak_one_item_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, +) +from .agent_v1update_speak_speak_one_item_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram(UncheckedBaseModel): + type: typing.Literal["deepgram"] = "deepgram" + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion] = None + model: AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs(UncheckedBaseModel): + type: typing.Literal["eleven_labs"] = "eleven_labs" + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion] = None + model_id: AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId + language: typing.Optional[str] = None + language_code: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia(UncheckedBaseModel): + type: typing.Literal["cartesia"] = "cartesia" + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion] = None + model_id: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId + voice: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice + language: typing.Optional[str] = None + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi(UncheckedBaseModel): + type: typing.Literal["open_ai"] = "open_ai" + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion] = None + model: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel + voice: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +class AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly(UncheckedBaseModel): + type: typing.Literal["aws_polly"] = "aws_polly" + voice: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice + language: str + language_code: typing.Optional[str] = None + engine: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow + + +AgentV1UpdateSpeakSpeakOneItemProvider = typing_extensions.Annotated[ + typing.Union[ + AgentV1UpdateSpeakSpeakOneItemProvider_Deepgram, + AgentV1UpdateSpeakSpeakOneItemProvider_ElevenLabs, + AgentV1UpdateSpeakSpeakOneItemProvider_Cartesia, + AgentV1UpdateSpeakSpeakOneItemProvider_OpenAi, + AgentV1UpdateSpeakSpeakOneItemProvider_AwsPolly, + ], + UnionMetadata(discriminant="type"), +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly.py new file mode 100644 index 00000000..3ef8938c --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly.py @@ -0,0 +1,45 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_engine import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine, +) +from .agent_v1update_speak_speak_one_item_provider_aws_polly_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderAwsPolly(UncheckedBaseModel): + voice: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice = pydantic.Field() + """ + AWS Polly voice name + """ + + language: str = pydantic.Field() + """ + Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + engine: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine + credentials: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/types/aws_polly_speak_provider_credentials.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py similarity index 57% rename from src/deepgram/types/aws_polly_speak_provider_credentials.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py index eca5defc..daf92957 100644 --- a/src/deepgram/types/aws_polly_speak_provider_credentials.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials.py @@ -3,13 +3,15 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .aws_polly_speak_provider_credentials_type import AwsPollySpeakProviderCredentialsType +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type import ( + AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType, +) -class AwsPollySpeakProviderCredentials(UncheckedBaseModel): - type: AwsPollySpeakProviderCredentialsType +class AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentials(UncheckedBaseModel): + type: AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType region: str access_key_id: str secret_access_key: str diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type.py new file mode 100644 index 00000000..a10f9924 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_credentials_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyCredentialsType = typing.Union[typing.Literal["sts", "iam"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_engine.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_engine.py new file mode 100644 index 00000000..1ba214f0 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_engine.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyEngine = typing.Union[ + typing.Literal["generative", "long-form", "standard", "neural"], typing.Any +] diff --git a/src/deepgram/types/aws_polly_speak_provider_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_voice.py similarity index 73% rename from src/deepgram/types/aws_polly_speak_provider_voice.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_voice.py index c186c9b0..3f077a0e 100644 --- a/src/deepgram/types/aws_polly_speak_provider_voice.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_aws_polly_voice.py @@ -2,6 +2,6 @@ import typing -AwsPollySpeakProviderVoice = typing.Union[ +AgentV1UpdateSpeakSpeakOneItemProviderAwsPollyVoice = typing.Union[ typing.Literal["Matthew", "Joanna", "Amy", "Emma", "Brian", "Arthur", "Aria", "Ayanda"], typing.Any ] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia.py new file mode 100644 index 00000000..4ebf676c --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia.py @@ -0,0 +1,43 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_cartesia_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1update_speak_speak_one_item_provider_cartesia_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderCartesia(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion] = pydantic.Field(default=None) + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId = pydantic.Field() + """ + Cartesia model ID + """ + + voice: AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Cartesia language code + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_model_id.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_model_id.py new file mode 100644 index 00000000..9749b8a5 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_model_id.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderCartesiaModelId = typing.Union[ + typing.Literal["sonic-2", "sonic-multilingual"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_version.py new file mode 100644 index 00000000..9a6f6c37 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVersion = typing.Union[typing.Literal["2025-03-17"], typing.Any] diff --git a/src/deepgram/types/cartesia_speak_provider_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py similarity index 73% rename from src/deepgram/types/cartesia_speak_provider_voice.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py index cfc897e7..78756f0b 100644 --- a/src/deepgram/types/cartesia_speak_provider_voice.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_cartesia_voice.py @@ -3,11 +3,11 @@ import typing import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel -class CartesiaSpeakProviderVoice(UncheckedBaseModel): +class AgentV1UpdateSpeakSpeakOneItemProviderCartesiaVoice(UncheckedBaseModel): mode: str = pydantic.Field() """ Cartesia voice mode diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram.py new file mode 100644 index 00000000..5ac39d7d --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram.py @@ -0,0 +1,34 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_deepgram_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel, +) +from .agent_v1update_speak_speak_one_item_provider_deepgram_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderDeepgram(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion] = pydantic.Field(default=None) + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel = pydantic.Field() + """ + Deepgram TTS model + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/types/deepgram_speak_provider_model.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_model.py similarity index 96% rename from src/deepgram/types/deepgram_speak_provider_model.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_model.py index 72bcae5c..7600a539 100644 --- a/src/deepgram/types/deepgram_speak_provider_model.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_model.py @@ -2,7 +2,7 @@ import typing -DeepgramSpeakProviderModel = typing.Union[ +AgentV1UpdateSpeakSpeakOneItemProviderDeepgramModel = typing.Union[ typing.Literal[ "aura-asteria-en", "aura-luna-en", diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_version.py new file mode 100644 index 00000000..757270a9 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_deepgram_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderDeepgramVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs.py new file mode 100644 index 00000000..5e74832b --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs.py @@ -0,0 +1,44 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId, +) +from .agent_v1update_speak_speak_one_item_provider_eleven_labs_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderElevenLabs(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion] = pydantic.Field(default=None) + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId = pydantic.Field() + """ + Eleven Labs model ID + """ + + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/types/eleven_labs_speak_provider_model_id.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id.py similarity index 71% rename from src/deepgram/types/eleven_labs_speak_provider_model_id.py rename to src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id.py index ce0427e2..33b709fe 100644 --- a/src/deepgram/types/eleven_labs_speak_provider_model_id.py +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_model_id.py @@ -2,6 +2,6 @@ import typing -ElevenLabsSpeakProviderModelId = typing.Union[ +AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsModelId = typing.Union[ typing.Literal["eleven_turbo_v2_5", "eleven_monolingual_v1", "eleven_multilingual_v2"], typing.Any ] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_version.py new file mode 100644 index 00000000..5440d576 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_eleven_labs_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderElevenLabsVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai.py new file mode 100644 index 00000000..025f5cc4 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai.py @@ -0,0 +1,42 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1update_speak_speak_one_item_provider_open_ai_model import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel, +) +from .agent_v1update_speak_speak_one_item_provider_open_ai_version import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion, +) +from .agent_v1update_speak_speak_one_item_provider_open_ai_voice import ( + AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice, +) + + +class AgentV1UpdateSpeakSpeakOneItemProviderOpenAi(UncheckedBaseModel): + version: typing.Optional[AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion] = pydantic.Field(default=None) + """ + The REST API version for the OpenAI text-to-speech API + """ + + model: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel = pydantic.Field() + """ + OpenAI TTS model + """ + + voice: AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice = pydantic.Field() + """ + OpenAI voice + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_model.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_model.py new file mode 100644 index 00000000..f2e71577 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_model.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderOpenAiModel = typing.Union[typing.Literal["tts-1", "tts-1-hd"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_version.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_version.py new file mode 100644 index 00000000..8d06f519 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_version.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVersion = typing.Union[typing.Literal["v1"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_voice.py b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_voice.py new file mode 100644 index 00000000..742d729e --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_speak_one_item_provider_open_ai_voice.py @@ -0,0 +1,7 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakSpeakOneItemProviderOpenAiVoice = typing.Union[ + typing.Literal["alloy", "echo", "fable", "onyx", "nova", "shimmer"], typing.Any +] diff --git a/src/deepgram/agent/v1/types/agent_v1update_speak_type.py b/src/deepgram/agent/v1/types/agent_v1update_speak_type.py new file mode 100644 index 00000000..1c1f127a --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1update_speak_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UpdateSpeakType = typing.Union[typing.Literal["UpdateSpeak"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1user_started_speaking.py b/src/deepgram/agent/v1/types/agent_v1user_started_speaking.py index afc96fed..7f991033 100644 --- a/src/deepgram/agent/v1/types/agent_v1user_started_speaking.py +++ b/src/deepgram/agent/v1/types/agent_v1user_started_speaking.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1user_started_speaking_type import AgentV1UserStartedSpeakingType class AgentV1UserStartedSpeaking(UncheckedBaseModel): - type: typing.Literal["UserStartedSpeaking"] = pydantic.Field(default="UserStartedSpeaking") + type: AgentV1UserStartedSpeakingType = pydantic.Field() """ Message type identifier indicating that the user has begun speaking """ diff --git a/src/deepgram/agent/v1/types/agent_v1user_started_speaking_type.py b/src/deepgram/agent/v1/types/agent_v1user_started_speaking_type.py new file mode 100644 index 00000000..b23ebccb --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1user_started_speaking_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1UserStartedSpeakingType = typing.Union[typing.Literal["UserStartedSpeaking"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1warning.py b/src/deepgram/agent/v1/types/agent_v1warning.py index c4741050..4ac416c5 100644 --- a/src/deepgram/agent/v1/types/agent_v1warning.py +++ b/src/deepgram/agent/v1/types/agent_v1warning.py @@ -5,6 +5,7 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1warning_type import AgentV1WarningType class AgentV1Warning(UncheckedBaseModel): @@ -12,7 +13,7 @@ class AgentV1Warning(UncheckedBaseModel): Notifies the client of non-fatal errors or warnings """ - type: typing.Literal["Warning"] = pydantic.Field(default="Warning") + type: AgentV1WarningType = pydantic.Field() """ Message type identifier for warnings """ diff --git a/src/deepgram/agent/v1/types/agent_v1warning_type.py b/src/deepgram/agent/v1/types/agent_v1warning_type.py new file mode 100644 index 00000000..af9abcdb --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1warning_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1WarningType = typing.Union[typing.Literal["Warning"], typing.Any] diff --git a/src/deepgram/agent/v1/types/agent_v1welcome.py b/src/deepgram/agent/v1/types/agent_v1welcome.py index 128700c0..c9c04406 100644 --- a/src/deepgram/agent/v1/types/agent_v1welcome.py +++ b/src/deepgram/agent/v1/types/agent_v1welcome.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1welcome_type import AgentV1WelcomeType class AgentV1Welcome(UncheckedBaseModel): - type: typing.Literal["Welcome"] = pydantic.Field(default="Welcome") + type: AgentV1WelcomeType = pydantic.Field() """ Message type identifier for welcome message """ diff --git a/src/deepgram/agent/v1/types/agent_v1welcome_type.py b/src/deepgram/agent/v1/types/agent_v1welcome_type.py new file mode 100644 index 00000000..d2902f24 --- /dev/null +++ b/src/deepgram/agent/v1/types/agent_v1welcome_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentV1WelcomeType = typing.Union[typing.Literal["Welcome"], typing.Any] diff --git a/src/deepgram/agent/v1/types/cartesia.py b/src/deepgram/agent/v1/types/cartesia.py new file mode 100644 index 00000000..d5fa86be --- /dev/null +++ b/src/deepgram/agent/v1/types/cartesia.py @@ -0,0 +1,43 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_cartesia_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_version import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion, +) +from .agent_v1settings_agent_speak_one_item_provider_cartesia_voice import ( + AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice, +) + + +class Cartesia(UncheckedBaseModel): + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderCartesiaVersion] = pydantic.Field(default=None) + """ + The API version header for the Cartesia text-to-speech API + """ + + model_id: AgentV1SettingsAgentSpeakOneItemProviderCartesiaModelId = pydantic.Field() + """ + Cartesia model ID + """ + + voice: AgentV1SettingsAgentSpeakOneItemProviderCartesiaVoice + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Cartesia language code + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/deepgram.py b/src/deepgram/agent/v1/types/deepgram.py new file mode 100644 index 00000000..78fec994 --- /dev/null +++ b/src/deepgram/agent/v1/types/deepgram.py @@ -0,0 +1,34 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_deepgram_model import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel, +) +from .agent_v1settings_agent_speak_one_item_provider_deepgram_version import ( + AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion, +) + + +class Deepgram(UncheckedBaseModel): + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderDeepgramVersion] = pydantic.Field(default=None) + """ + The REST API version for the Deepgram text-to-speech API + """ + + model: AgentV1SettingsAgentSpeakOneItemProviderDeepgramModel = pydantic.Field() + """ + Deepgram TTS model + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/eleven_labs.py b/src/deepgram/agent/v1/types/eleven_labs.py new file mode 100644 index 00000000..ef3ea856 --- /dev/null +++ b/src/deepgram/agent/v1/types/eleven_labs.py @@ -0,0 +1,44 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +import pydantic +from ....core.pydantic_utilities import IS_PYDANTIC_V2 +from ....core.unchecked_base_model import UncheckedBaseModel +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_model_id import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId, +) +from .agent_v1settings_agent_speak_one_item_provider_eleven_labs_version import ( + AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion, +) + + +class ElevenLabs(UncheckedBaseModel): + version: typing.Optional[AgentV1SettingsAgentSpeakOneItemProviderElevenLabsVersion] = pydantic.Field(default=None) + """ + The REST API version for the ElevenLabs text-to-speech API + """ + + model_id: AgentV1SettingsAgentSpeakOneItemProviderElevenLabsModelId = pydantic.Field() + """ + Eleven Labs model ID + """ + + language: typing.Optional[str] = pydantic.Field(default=None) + """ + Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API + """ + + language_code: typing.Optional[str] = pydantic.Field(default=None) + """ + Use the `language` field instead. + """ + + if IS_PYDANTIC_V2: + model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 + else: + + class Config: + frozen = True + smart_union = True + extra = pydantic.Extra.allow diff --git a/src/deepgram/agent/v1/types/max.py b/src/deepgram/agent/v1/types/max.py new file mode 100644 index 00000000..71bcad70 --- /dev/null +++ b/src/deepgram/agent/v1/types/max.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +Max = typing.Union[typing.Literal["max"], typing.Any] diff --git a/src/deepgram/base_client.py b/src/deepgram/base_client.py index 34261bc6..7ee628e3 100644 --- a/src/deepgram/base_client.py +++ b/src/deepgram/base_client.py @@ -8,6 +8,7 @@ import httpx from .core.api_error import ApiError from .core.client_wrapper import AsyncClientWrapper, SyncClientWrapper +from .core.logging import LogConfig, Logger from .environment import DeepgramClientEnvironment if typing.TYPE_CHECKING: @@ -48,6 +49,9 @@ class BaseClient: httpx_client : typing.Optional[httpx.Client] The httpx client to use for making requests, a preconfigured client is used by default, however this is useful should you want to pass in any custom httpx configuration. + logging : typing.Optional[typing.Union[LogConfig, Logger]] + Configure logging for the SDK. Accepts a LogConfig dict with 'level' (debug/info/warn/error), 'logger' (custom logger implementation), and 'silent' (boolean, defaults to True) fields. You can also pass a pre-configured Logger instance. + Examples -------- from deepgram import DeepgramClient @@ -66,6 +70,7 @@ def __init__( timeout: typing.Optional[float] = None, follow_redirects: typing.Optional[bool] = True, httpx_client: typing.Optional[httpx.Client] = None, + logging: typing.Optional[typing.Union[LogConfig, Logger]] = None, ): _defaulted_timeout = ( timeout if timeout is not None else 60 if httpx_client is None else httpx_client.timeout.read @@ -84,6 +89,7 @@ def __init__( if follow_redirects is not None else httpx.Client(timeout=_defaulted_timeout), timeout=_defaulted_timeout, + logging=logging, ) self._agent: typing.Optional[AgentClient] = None self._auth: typing.Optional[AuthClient] = None @@ -178,6 +184,9 @@ class AsyncBaseClient: httpx_client : typing.Optional[httpx.AsyncClient] The httpx client to use for making requests, a preconfigured client is used by default, however this is useful should you want to pass in any custom httpx configuration. + logging : typing.Optional[typing.Union[LogConfig, Logger]] + Configure logging for the SDK. Accepts a LogConfig dict with 'level' (debug/info/warn/error), 'logger' (custom logger implementation), and 'silent' (boolean, defaults to True) fields. You can also pass a pre-configured Logger instance. + Examples -------- from deepgram import AsyncDeepgramClient @@ -196,6 +205,7 @@ def __init__( timeout: typing.Optional[float] = None, follow_redirects: typing.Optional[bool] = True, httpx_client: typing.Optional[httpx.AsyncClient] = None, + logging: typing.Optional[typing.Union[LogConfig, Logger]] = None, ): _defaulted_timeout = ( timeout if timeout is not None else 60 if httpx_client is None else httpx_client.timeout.read @@ -214,6 +224,7 @@ def __init__( if follow_redirects is not None else httpx.AsyncClient(timeout=_defaulted_timeout), timeout=_defaulted_timeout, + logging=logging, ) self._agent: typing.Optional[AsyncAgentClient] = None self._auth: typing.Optional[AsyncAuthClient] = None diff --git a/src/deepgram/core/__init__.py b/src/deepgram/core/__init__.py index 8547a17c..d247a71c 100644 --- a/src/deepgram/core/__init__.py +++ b/src/deepgram/core/__init__.py @@ -8,12 +8,13 @@ if typing.TYPE_CHECKING: from .api_error import ApiError from .client_wrapper import AsyncClientWrapper, BaseClientWrapper, SyncClientWrapper - from .datetime_utils import serialize_datetime + from .datetime_utils import Rfc2822DateTime, parse_rfc2822_datetime, serialize_datetime from .events import EventEmitterMixin, EventType from .file import File, convert_file_dict_to_httpx_tuples, with_content_type from .http_client import AsyncHttpClient, HttpClient from .http_response import AsyncHttpResponse, HttpResponse from .jsonable_encoder import jsonable_encoder + from .logging import ConsoleLogger, ILogger, LogConfig, LogLevel, Logger, create_logger from .pydantic_utilities import ( IS_PYDANTIC_V2, UniversalBaseModel, @@ -35,15 +36,21 @@ "AsyncHttpClient": ".http_client", "AsyncHttpResponse": ".http_response", "BaseClientWrapper": ".client_wrapper", + "ConsoleLogger": ".logging", "EventEmitterMixin": ".events", "EventType": ".events", "FieldMetadata": ".serialization", "File": ".file", "HttpClient": ".http_client", "HttpResponse": ".http_response", + "ILogger": ".logging", "IS_PYDANTIC_V2": ".pydantic_utilities", "InvalidWebSocketStatus": ".websocket_compat", + "LogConfig": ".logging", + "LogLevel": ".logging", + "Logger": ".logging", "RequestOptions": ".request_options", + "Rfc2822DateTime": ".datetime_utils", "SyncClientWrapper": ".client_wrapper", "UncheckedBaseModel": ".unchecked_base_model", "UnionMetadata": ".unchecked_base_model", @@ -52,10 +59,12 @@ "construct_type": ".unchecked_base_model", "convert_and_respect_annotation_metadata": ".serialization", "convert_file_dict_to_httpx_tuples": ".file", + "create_logger": ".logging", "encode_query": ".query_encoder", "get_status_code": ".websocket_compat", "jsonable_encoder": ".jsonable_encoder", "parse_obj_as": ".pydantic_utilities", + "parse_rfc2822_datetime": ".datetime_utils", "remove_none_from_dict": ".remove_none_from_dict", "serialize_datetime": ".datetime_utils", "universal_field_validator": ".pydantic_utilities", @@ -92,15 +101,21 @@ def __dir__(): "AsyncHttpClient", "AsyncHttpResponse", "BaseClientWrapper", + "ConsoleLogger", "EventEmitterMixin", "EventType", "FieldMetadata", "File", "HttpClient", "HttpResponse", + "ILogger", "IS_PYDANTIC_V2", "InvalidWebSocketStatus", + "LogConfig", + "LogLevel", + "Logger", "RequestOptions", + "Rfc2822DateTime", "SyncClientWrapper", "UncheckedBaseModel", "UnionMetadata", @@ -109,10 +124,12 @@ def __dir__(): "construct_type", "convert_and_respect_annotation_metadata", "convert_file_dict_to_httpx_tuples", + "create_logger", "encode_query", "get_status_code", "jsonable_encoder", "parse_obj_as", + "parse_rfc2822_datetime", "remove_none_from_dict", "serialize_datetime", "universal_field_validator", diff --git a/src/deepgram/core/client_wrapper.py b/src/deepgram/core/client_wrapper.py index 33b53ce1..51402d11 100644 --- a/src/deepgram/core/client_wrapper.py +++ b/src/deepgram/core/client_wrapper.py @@ -5,6 +5,7 @@ import httpx from ..environment import DeepgramClientEnvironment from .http_client import AsyncHttpClient, HttpClient +from .logging import LogConfig, Logger class BaseClientWrapper: @@ -15,22 +16,24 @@ def __init__( headers: typing.Optional[typing.Dict[str, str]] = None, environment: DeepgramClientEnvironment, timeout: typing.Optional[float] = None, + logging: typing.Optional[typing.Union[LogConfig, Logger]] = None, ): self.api_key = api_key self._headers = headers self._environment = environment self._timeout = timeout + self._logging = logging def get_headers(self) -> typing.Dict[str, str]: import platform headers: typing.Dict[str, str] = { - "User-Agent": "deepgram-sdk/6.0.1", + "User-Agent": "deepgram-sdk/6.0.2", "X-Fern-Language": "Python", "X-Fern-Runtime": f"python/{platform.python_version()}", "X-Fern-Platform": f"{platform.system().lower()}/{platform.release()}", "X-Fern-SDK-Name": "deepgram-sdk", - "X-Fern-SDK-Version": "6.0.1", + "X-Fern-SDK-Version": "6.0.2", **(self.get_custom_headers() or {}), } headers["Authorization"] = f"Token {self.api_key}" @@ -54,11 +57,15 @@ def __init__( headers: typing.Optional[typing.Dict[str, str]] = None, environment: DeepgramClientEnvironment, timeout: typing.Optional[float] = None, + logging: typing.Optional[typing.Union[LogConfig, Logger]] = None, httpx_client: httpx.Client, ): - super().__init__(api_key=api_key, headers=headers, environment=environment, timeout=timeout) + super().__init__(api_key=api_key, headers=headers, environment=environment, timeout=timeout, logging=logging) self.httpx_client = HttpClient( - httpx_client=httpx_client, base_headers=self.get_headers, base_timeout=self.get_timeout + httpx_client=httpx_client, + base_headers=self.get_headers, + base_timeout=self.get_timeout, + logging_config=self._logging, ) @@ -70,16 +77,18 @@ def __init__( headers: typing.Optional[typing.Dict[str, str]] = None, environment: DeepgramClientEnvironment, timeout: typing.Optional[float] = None, + logging: typing.Optional[typing.Union[LogConfig, Logger]] = None, async_token: typing.Optional[typing.Callable[[], typing.Awaitable[str]]] = None, httpx_client: httpx.AsyncClient, ): - super().__init__(api_key=api_key, headers=headers, environment=environment, timeout=timeout) + super().__init__(api_key=api_key, headers=headers, environment=environment, timeout=timeout, logging=logging) self._async_token = async_token self.httpx_client = AsyncHttpClient( httpx_client=httpx_client, base_headers=self.get_headers, base_timeout=self.get_timeout, async_base_headers=self.async_get_headers, + logging_config=self._logging, ) async def async_get_headers(self) -> typing.Dict[str, str]: diff --git a/src/deepgram/core/datetime_utils.py b/src/deepgram/core/datetime_utils.py index 7c9864a9..a12b2ad0 100644 --- a/src/deepgram/core/datetime_utils.py +++ b/src/deepgram/core/datetime_utils.py @@ -1,6 +1,48 @@ # This file was auto-generated by Fern from our API Definition. import datetime as dt +from email.utils import parsedate_to_datetime +from typing import Any + +import pydantic + +IS_PYDANTIC_V2 = pydantic.VERSION.startswith("2.") + + +def parse_rfc2822_datetime(v: Any) -> dt.datetime: + """ + Parse an RFC 2822 datetime string (e.g., "Wed, 02 Oct 2002 13:00:00 GMT") + into a datetime object. If the value is already a datetime, return it as-is. + Falls back to ISO 8601 parsing if RFC 2822 parsing fails. + """ + if isinstance(v, dt.datetime): + return v + if isinstance(v, str): + try: + return parsedate_to_datetime(v) + except Exception: + pass + # Fallback to ISO 8601 parsing + return dt.datetime.fromisoformat(v.replace("Z", "+00:00")) + raise ValueError(f"Expected str or datetime, got {type(v)}") + + +class Rfc2822DateTime(dt.datetime): + """A datetime subclass that parses RFC 2822 date strings. + + On Pydantic V1, uses __get_validators__ for pre-validation. + On Pydantic V2, uses __get_pydantic_core_schema__ for BeforeValidator-style parsing. + """ + + @classmethod + def __get_validators__(cls): # type: ignore[no-untyped-def] + yield parse_rfc2822_datetime + + @classmethod + def __get_pydantic_core_schema__(cls, _source_type: Any, _handler: Any) -> Any: # type: ignore[override] + from pydantic_core import core_schema + + return core_schema.no_info_before_validator_function(parse_rfc2822_datetime, core_schema.datetime_schema()) def serialize_datetime(v: dt.datetime) -> str: diff --git a/src/deepgram/core/http_client.py b/src/deepgram/core/http_client.py index 7c6c936f..ee937589 100644 --- a/src/deepgram/core/http_client.py +++ b/src/deepgram/core/http_client.py @@ -12,6 +12,7 @@ from .file import File, convert_file_dict_to_httpx_tuples from .force_multipart import FORCE_MULTIPART from .jsonable_encoder import jsonable_encoder +from .logging import LogConfig, Logger, create_logger from .query_encoder import encode_query from .remove_none_from_dict import remove_none_from_dict as remove_none_from_dict from .request_options import RequestOptions @@ -122,6 +123,32 @@ def _should_retry(response: httpx.Response) -> bool: return response.status_code >= 500 or response.status_code in retryable_400s +_SENSITIVE_HEADERS = frozenset( + { + "authorization", + "www-authenticate", + "x-api-key", + "api-key", + "apikey", + "x-api-token", + "x-auth-token", + "auth-token", + "cookie", + "set-cookie", + "proxy-authorization", + "proxy-authenticate", + "x-csrf-token", + "x-xsrf-token", + "x-session-token", + "x-access-token", + } +) + + +def _redact_headers(headers: typing.Dict[str, str]) -> typing.Dict[str, str]: + return {k: ("[REDACTED]" if k.lower() in _SENSITIVE_HEADERS else v) for k, v in headers.items()} + + def _build_url(base_url: str, path: typing.Optional[str]) -> str: """ Build a full URL by joining a base URL with a path. @@ -238,11 +265,13 @@ def __init__( base_timeout: typing.Callable[[], typing.Optional[float]], base_headers: typing.Callable[[], typing.Dict[str, str]], base_url: typing.Optional[typing.Callable[[], str]] = None, + logging_config: typing.Optional[typing.Union[LogConfig, Logger]] = None, ): self.base_url = base_url self.base_timeout = base_timeout self.base_headers = base_headers self.httpx_client = httpx_client + self.logger = create_logger(logging_config) def get_base_url(self, maybe_base_url: typing.Optional[str]) -> str: base_url = maybe_base_url @@ -315,18 +344,30 @@ def request( ) ) + _request_url = _build_url(base_url, path) + _request_headers = jsonable_encoder( + remove_none_from_dict( + { + **self.base_headers(), + **(headers if headers is not None else {}), + **(request_options.get("additional_headers", {}) or {} if request_options is not None else {}), + } + ) + ) + + if self.logger.is_debug(): + self.logger.debug( + "Making HTTP request", + method=method, + url=_request_url, + headers=_redact_headers(_request_headers), + has_body=json_body is not None or data_body is not None, + ) + response = self.httpx_client.request( method=method, - url=_build_url(base_url, path), - headers=jsonable_encoder( - remove_none_from_dict( - { - **self.base_headers(), - **(headers if headers is not None else {}), - **(request_options.get("additional_headers", {}) or {} if request_options is not None else {}), - } - ) - ), + url=_request_url, + headers=_request_headers, params=_encoded_params if _encoded_params else None, json=json_body, data=data_body, @@ -353,6 +394,24 @@ def request( omit=omit, ) + if self.logger.is_debug(): + if 200 <= response.status_code < 400: + self.logger.debug( + "HTTP request succeeded", + method=method, + url=_request_url, + status_code=response.status_code, + ) + + if self.logger.is_error(): + if response.status_code >= 400: + self.logger.error( + "HTTP request failed with error status", + method=method, + url=_request_url, + status_code=response.status_code, + ) + return response @contextmanager @@ -418,18 +477,29 @@ def stream( ) ) + _request_url = _build_url(base_url, path) + _request_headers = jsonable_encoder( + remove_none_from_dict( + { + **self.base_headers(), + **(headers if headers is not None else {}), + **(request_options.get("additional_headers", {}) if request_options is not None else {}), + } + ) + ) + + if self.logger.is_debug(): + self.logger.debug( + "Making streaming HTTP request", + method=method, + url=_request_url, + headers=_redact_headers(_request_headers), + ) + with self.httpx_client.stream( method=method, - url=_build_url(base_url, path), - headers=jsonable_encoder( - remove_none_from_dict( - { - **self.base_headers(), - **(headers if headers is not None else {}), - **(request_options.get("additional_headers", {}) if request_options is not None else {}), - } - ) - ), + url=_request_url, + headers=_request_headers, params=_encoded_params if _encoded_params else None, json=json_body, data=data_body, @@ -449,12 +519,14 @@ def __init__( base_headers: typing.Callable[[], typing.Dict[str, str]], base_url: typing.Optional[typing.Callable[[], str]] = None, async_base_headers: typing.Optional[typing.Callable[[], typing.Awaitable[typing.Dict[str, str]]]] = None, + logging_config: typing.Optional[typing.Union[LogConfig, Logger]] = None, ): self.base_url = base_url self.base_timeout = base_timeout self.base_headers = base_headers self.async_base_headers = async_base_headers self.httpx_client = httpx_client + self.logger = create_logger(logging_config) async def _get_headers(self) -> typing.Dict[str, str]: if self.async_base_headers is not None: @@ -535,19 +607,30 @@ async def request( ) ) - # Add the input to each of these and do None-safety checks + _request_url = _build_url(base_url, path) + _request_headers = jsonable_encoder( + remove_none_from_dict( + { + **_headers, + **(headers if headers is not None else {}), + **(request_options.get("additional_headers", {}) or {} if request_options is not None else {}), + } + ) + ) + + if self.logger.is_debug(): + self.logger.debug( + "Making HTTP request", + method=method, + url=_request_url, + headers=_redact_headers(_request_headers), + has_body=json_body is not None or data_body is not None, + ) + response = await self.httpx_client.request( method=method, - url=_build_url(base_url, path), - headers=jsonable_encoder( - remove_none_from_dict( - { - **_headers, - **(headers if headers is not None else {}), - **(request_options.get("additional_headers", {}) or {} if request_options is not None else {}), - } - ) - ), + url=_request_url, + headers=_request_headers, params=_encoded_params if _encoded_params else None, json=json_body, data=data_body, @@ -573,6 +656,25 @@ async def request( retries=retries + 1, omit=omit, ) + + if self.logger.is_debug(): + if 200 <= response.status_code < 400: + self.logger.debug( + "HTTP request succeeded", + method=method, + url=_request_url, + status_code=response.status_code, + ) + + if self.logger.is_error(): + if response.status_code >= 400: + self.logger.error( + "HTTP request failed with error status", + method=method, + url=_request_url, + status_code=response.status_code, + ) + return response @asynccontextmanager @@ -641,18 +743,29 @@ async def stream( ) ) + _request_url = _build_url(base_url, path) + _request_headers = jsonable_encoder( + remove_none_from_dict( + { + **_headers, + **(headers if headers is not None else {}), + **(request_options.get("additional_headers", {}) if request_options is not None else {}), + } + ) + ) + + if self.logger.is_debug(): + self.logger.debug( + "Making streaming HTTP request", + method=method, + url=_request_url, + headers=_redact_headers(_request_headers), + ) + async with self.httpx_client.stream( method=method, - url=_build_url(base_url, path), - headers=jsonable_encoder( - remove_none_from_dict( - { - **_headers, - **(headers if headers is not None else {}), - **(request_options.get("additional_headers", {}) if request_options is not None else {}), - } - ) - ), + url=_request_url, + headers=_request_headers, params=_encoded_params if _encoded_params else None, json=json_body, data=data_body, diff --git a/src/deepgram/core/logging.py b/src/deepgram/core/logging.py new file mode 100644 index 00000000..e5e57245 --- /dev/null +++ b/src/deepgram/core/logging.py @@ -0,0 +1,107 @@ +# This file was auto-generated by Fern from our API Definition. + +import logging +import typing + +LogLevel = typing.Literal["debug", "info", "warn", "error"] + +_LOG_LEVEL_MAP: typing.Dict[LogLevel, int] = { + "debug": 1, + "info": 2, + "warn": 3, + "error": 4, +} + + +class ILogger(typing.Protocol): + def debug(self, message: str, **kwargs: typing.Any) -> None: ... + def info(self, message: str, **kwargs: typing.Any) -> None: ... + def warn(self, message: str, **kwargs: typing.Any) -> None: ... + def error(self, message: str, **kwargs: typing.Any) -> None: ... + + +class ConsoleLogger: + _logger: logging.Logger + + def __init__(self) -> None: + self._logger = logging.getLogger("fern") + if not self._logger.handlers: + handler = logging.StreamHandler() + handler.setFormatter(logging.Formatter("%(levelname)s - %(message)s")) + self._logger.addHandler(handler) + self._logger.setLevel(logging.DEBUG) + + def debug(self, message: str, **kwargs: typing.Any) -> None: + self._logger.debug(message, extra=kwargs) + + def info(self, message: str, **kwargs: typing.Any) -> None: + self._logger.info(message, extra=kwargs) + + def warn(self, message: str, **kwargs: typing.Any) -> None: + self._logger.warning(message, extra=kwargs) + + def error(self, message: str, **kwargs: typing.Any) -> None: + self._logger.error(message, extra=kwargs) + + +class LogConfig(typing.TypedDict, total=False): + level: LogLevel + logger: ILogger + silent: bool + + +class Logger: + _level: int + _logger: ILogger + _silent: bool + + def __init__(self, *, level: LogLevel, logger: ILogger, silent: bool) -> None: + self._level = _LOG_LEVEL_MAP[level] + self._logger = logger + self._silent = silent + + def _should_log(self, level: LogLevel) -> bool: + return not self._silent and self._level <= _LOG_LEVEL_MAP[level] + + def is_debug(self) -> bool: + return self._should_log("debug") + + def is_info(self) -> bool: + return self._should_log("info") + + def is_warn(self) -> bool: + return self._should_log("warn") + + def is_error(self) -> bool: + return self._should_log("error") + + def debug(self, message: str, **kwargs: typing.Any) -> None: + if self.is_debug(): + self._logger.debug(message, **kwargs) + + def info(self, message: str, **kwargs: typing.Any) -> None: + if self.is_info(): + self._logger.info(message, **kwargs) + + def warn(self, message: str, **kwargs: typing.Any) -> None: + if self.is_warn(): + self._logger.warn(message, **kwargs) + + def error(self, message: str, **kwargs: typing.Any) -> None: + if self.is_error(): + self._logger.error(message, **kwargs) + + +_default_logger: Logger = Logger(level="info", logger=ConsoleLogger(), silent=True) + + +def create_logger(config: typing.Optional[typing.Union[LogConfig, Logger]] = None) -> Logger: + if config is None: + return _default_logger + if isinstance(config, Logger): + return config + return Logger( + level=config.get("level", "info"), + logger=config.get("logger", ConsoleLogger()), + silent=config.get("silent", True), + ) diff --git a/src/deepgram/environment.py b/src/deepgram/environment.py index 240d7ca4..e514ef69 100644 --- a/src/deepgram/environment.py +++ b/src/deepgram/environment.py @@ -7,15 +7,15 @@ class DeepgramClientEnvironment: PRODUCTION: DeepgramClientEnvironment AGENT: DeepgramClientEnvironment - def __init__(self, *, base: str, production: str, agent: str): + def __init__(self, *, base: str, agent: str, production: str): self.base = base - self.production = production self.agent = agent + self.production = production DeepgramClientEnvironment.PRODUCTION = DeepgramClientEnvironment( - base="https://api.deepgram.com", production="wss://api.deepgram.com", agent="wss://agent.deepgram.com" + base="https://api.deepgram.com", agent="wss://agent.deepgram.com", production="wss://api.deepgram.com" ) DeepgramClientEnvironment.AGENT = DeepgramClientEnvironment( - base="https://agent.deepgram.com", production="wss://api.deepgram.com", agent="wss://agent.deepgram.com" + base="https://agent.deepgram.com", agent="wss://agent.deepgram.com", production="wss://api.deepgram.com" ) diff --git a/src/deepgram/listen/__init__.py b/src/deepgram/listen/__init__.py index 42a8267f..236cdc61 100644 --- a/src/deepgram/listen/__init__.py +++ b/src/deepgram/listen/__init__.py @@ -19,6 +19,7 @@ ListenV1KeepAliveType, ListenV1Metadata, ListenV1MetadataParams, + ListenV1MetadataType, ListenV1Results, ListenV1ResultsChannel, ListenV1ResultsChannelAlternativesItem, @@ -33,10 +34,13 @@ ListenV1ResultsMetadataModelInfoParams, ListenV1ResultsMetadataParams, ListenV1ResultsParams, + ListenV1ResultsType, ListenV1SpeechStarted, ListenV1SpeechStartedParams, + ListenV1SpeechStartedType, ListenV1UtteranceEnd, ListenV1UtteranceEndParams, + ListenV1UtteranceEndType, ) from .v2 import ( ListenV2CloseStream, @@ -44,11 +48,14 @@ ListenV2CloseStreamType, ListenV2Connected, ListenV2ConnectedParams, + ListenV2ConnectedType, ListenV2FatalError, ListenV2FatalErrorParams, + ListenV2FatalErrorType, ListenV2TurnInfo, ListenV2TurnInfoEvent, ListenV2TurnInfoParams, + ListenV2TurnInfoType, ListenV2TurnInfoWordsItem, ListenV2TurnInfoWordsItemParams, ) @@ -64,6 +71,7 @@ "ListenV1KeepAliveType": ".v1", "ListenV1Metadata": ".v1", "ListenV1MetadataParams": ".v1", + "ListenV1MetadataType": ".v1", "ListenV1Results": ".v1", "ListenV1ResultsChannel": ".v1", "ListenV1ResultsChannelAlternativesItem": ".v1", @@ -78,20 +86,26 @@ "ListenV1ResultsMetadataModelInfoParams": ".v1", "ListenV1ResultsMetadataParams": ".v1", "ListenV1ResultsParams": ".v1", + "ListenV1ResultsType": ".v1", "ListenV1SpeechStarted": ".v1", "ListenV1SpeechStartedParams": ".v1", + "ListenV1SpeechStartedType": ".v1", "ListenV1UtteranceEnd": ".v1", "ListenV1UtteranceEndParams": ".v1", + "ListenV1UtteranceEndType": ".v1", "ListenV2CloseStream": ".v2", "ListenV2CloseStreamParams": ".v2", "ListenV2CloseStreamType": ".v2", "ListenV2Connected": ".v2", "ListenV2ConnectedParams": ".v2", + "ListenV2ConnectedType": ".v2", "ListenV2FatalError": ".v2", "ListenV2FatalErrorParams": ".v2", + "ListenV2FatalErrorType": ".v2", "ListenV2TurnInfo": ".v2", "ListenV2TurnInfoEvent": ".v2", "ListenV2TurnInfoParams": ".v2", + "ListenV2TurnInfoType": ".v2", "ListenV2TurnInfoWordsItem": ".v2", "ListenV2TurnInfoWordsItemParams": ".v2", "v1": ".v1", @@ -132,6 +146,7 @@ def __dir__(): "ListenV1KeepAliveType", "ListenV1Metadata", "ListenV1MetadataParams", + "ListenV1MetadataType", "ListenV1Results", "ListenV1ResultsChannel", "ListenV1ResultsChannelAlternativesItem", @@ -146,20 +161,26 @@ def __dir__(): "ListenV1ResultsMetadataModelInfoParams", "ListenV1ResultsMetadataParams", "ListenV1ResultsParams", + "ListenV1ResultsType", "ListenV1SpeechStarted", "ListenV1SpeechStartedParams", + "ListenV1SpeechStartedType", "ListenV1UtteranceEnd", "ListenV1UtteranceEndParams", + "ListenV1UtteranceEndType", "ListenV2CloseStream", "ListenV2CloseStreamParams", "ListenV2CloseStreamType", "ListenV2Connected", "ListenV2ConnectedParams", + "ListenV2ConnectedType", "ListenV2FatalError", "ListenV2FatalErrorParams", + "ListenV2FatalErrorType", "ListenV2TurnInfo", "ListenV2TurnInfoEvent", "ListenV2TurnInfoParams", + "ListenV2TurnInfoType", "ListenV2TurnInfoWordsItem", "ListenV2TurnInfoWordsItemParams", "v1", diff --git a/src/deepgram/listen/v1/__init__.py b/src/deepgram/listen/v1/__init__.py index 8b43c4ff..a8115d75 100644 --- a/src/deepgram/listen/v1/__init__.py +++ b/src/deepgram/listen/v1/__init__.py @@ -14,6 +14,7 @@ ListenV1KeepAlive, ListenV1KeepAliveType, ListenV1Metadata, + ListenV1MetadataType, ListenV1Results, ListenV1ResultsChannel, ListenV1ResultsChannelAlternativesItem, @@ -21,8 +22,11 @@ ListenV1ResultsEntitiesItem, ListenV1ResultsMetadata, ListenV1ResultsMetadataModelInfo, + ListenV1ResultsType, ListenV1SpeechStarted, + ListenV1SpeechStartedType, ListenV1UtteranceEnd, + ListenV1UtteranceEndType, ) from . import media from .media import ( @@ -30,9 +34,9 @@ MediaTranscribeRequestCustomIntentMode, MediaTranscribeRequestCustomTopicMode, MediaTranscribeRequestEncoding, - MediaTranscribeRequestModel, - MediaTranscribeRequestSummarize, - MediaTranscribeRequestVersion, + MediaTranscribeRequestModelZero, + MediaTranscribeRequestSummarizeZero, + MediaTranscribeRequestVersionZero, MediaTranscribeResponse, MediaTranscribeResponseParams, ) @@ -63,6 +67,7 @@ "ListenV1KeepAliveType": ".types", "ListenV1Metadata": ".types", "ListenV1MetadataParams": ".requests", + "ListenV1MetadataType": ".types", "ListenV1Results": ".types", "ListenV1ResultsChannel": ".types", "ListenV1ResultsChannelAlternativesItem": ".types", @@ -77,17 +82,20 @@ "ListenV1ResultsMetadataModelInfoParams": ".requests", "ListenV1ResultsMetadataParams": ".requests", "ListenV1ResultsParams": ".requests", + "ListenV1ResultsType": ".types", "ListenV1SpeechStarted": ".types", "ListenV1SpeechStartedParams": ".requests", + "ListenV1SpeechStartedType": ".types", "ListenV1UtteranceEnd": ".types", "ListenV1UtteranceEndParams": ".requests", + "ListenV1UtteranceEndType": ".types", "MediaTranscribeRequestCallbackMethod": ".media", "MediaTranscribeRequestCustomIntentMode": ".media", "MediaTranscribeRequestCustomTopicMode": ".media", "MediaTranscribeRequestEncoding": ".media", - "MediaTranscribeRequestModel": ".media", - "MediaTranscribeRequestSummarize": ".media", - "MediaTranscribeRequestVersion": ".media", + "MediaTranscribeRequestModelZero": ".media", + "MediaTranscribeRequestSummarizeZero": ".media", + "MediaTranscribeRequestVersionZero": ".media", "MediaTranscribeResponse": ".media", "MediaTranscribeResponseParams": ".media", "media": ".media", @@ -127,6 +135,7 @@ def __dir__(): "ListenV1KeepAliveType", "ListenV1Metadata", "ListenV1MetadataParams", + "ListenV1MetadataType", "ListenV1Results", "ListenV1ResultsChannel", "ListenV1ResultsChannelAlternativesItem", @@ -141,17 +150,20 @@ def __dir__(): "ListenV1ResultsMetadataModelInfoParams", "ListenV1ResultsMetadataParams", "ListenV1ResultsParams", + "ListenV1ResultsType", "ListenV1SpeechStarted", "ListenV1SpeechStartedParams", + "ListenV1SpeechStartedType", "ListenV1UtteranceEnd", "ListenV1UtteranceEndParams", + "ListenV1UtteranceEndType", "MediaTranscribeRequestCallbackMethod", "MediaTranscribeRequestCustomIntentMode", "MediaTranscribeRequestCustomTopicMode", "MediaTranscribeRequestEncoding", - "MediaTranscribeRequestModel", - "MediaTranscribeRequestSummarize", - "MediaTranscribeRequestVersion", + "MediaTranscribeRequestModelZero", + "MediaTranscribeRequestSummarizeZero", + "MediaTranscribeRequestVersionZero", "MediaTranscribeResponse", "MediaTranscribeResponseParams", "media", diff --git a/src/deepgram/listen/v1/client.py b/src/deepgram/listen/v1/client.py index 7dd8baf3..0a4ac15e 100644 --- a/src/deepgram/listen/v1/client.py +++ b/src/deepgram/listen/v1/client.py @@ -14,6 +14,34 @@ from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...types.listen_v1callback import ListenV1Callback +from ...types.listen_v1callback_method import ListenV1CallbackMethod +from ...types.listen_v1channels import ListenV1Channels +from ...types.listen_v1detect_entities import ListenV1DetectEntities +from ...types.listen_v1diarize import ListenV1Diarize +from ...types.listen_v1dictation import ListenV1Dictation +from ...types.listen_v1encoding import ListenV1Encoding +from ...types.listen_v1endpointing import ListenV1Endpointing +from ...types.listen_v1extra import ListenV1Extra +from ...types.listen_v1interim_results import ListenV1InterimResults +from ...types.listen_v1keyterm import ListenV1Keyterm +from ...types.listen_v1keywords import ListenV1Keywords +from ...types.listen_v1language import ListenV1Language +from ...types.listen_v1mip_opt_out import ListenV1MipOptOut +from ...types.listen_v1model import ListenV1Model +from ...types.listen_v1multichannel import ListenV1Multichannel +from ...types.listen_v1numerals import ListenV1Numerals +from ...types.listen_v1profanity_filter import ListenV1ProfanityFilter +from ...types.listen_v1punctuate import ListenV1Punctuate +from ...types.listen_v1redact import ListenV1Redact +from ...types.listen_v1replace import ListenV1Replace +from ...types.listen_v1sample_rate import ListenV1SampleRate +from ...types.listen_v1search import ListenV1Search +from ...types.listen_v1smart_format import ListenV1SmartFormat +from ...types.listen_v1tag import ListenV1Tag +from ...types.listen_v1utterance_end_ms import ListenV1UtteranceEndMs +from ...types.listen_v1vad_events import ListenV1VadEvents +from ...types.listen_v1version import ListenV1Version from .raw_client import AsyncRawV1Client, RawV1Client from .socket_client import AsyncV1SocketClient, V1SocketClient @@ -47,34 +75,34 @@ def with_raw_response(self) -> RawV1Client: def connect( self, *, - callback: typing.Optional[str] = None, - callback_method: typing.Optional[str] = None, - channels: typing.Optional[str] = None, - detect_entities: typing.Optional[str] = None, - diarize: typing.Optional[str] = None, - dictation: typing.Optional[str] = None, - encoding: typing.Optional[str] = None, - endpointing: typing.Optional[str] = None, - extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - interim_results: typing.Optional[str] = None, - keyterm: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - language: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: str, - multichannel: typing.Optional[str] = None, - numerals: typing.Optional[str] = None, - profanity_filter: typing.Optional[str] = None, - punctuate: typing.Optional[str] = None, - redact: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - replace: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - sample_rate: typing.Optional[str] = None, - search: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - smart_format: typing.Optional[str] = None, - tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - utterance_end_ms: typing.Optional[str] = None, - vad_events: typing.Optional[str] = None, - version: typing.Optional[str] = None, + callback: typing.Optional[ListenV1Callback] = None, + callback_method: typing.Optional[ListenV1CallbackMethod] = None, + channels: typing.Optional[ListenV1Channels] = None, + detect_entities: typing.Optional[ListenV1DetectEntities] = None, + diarize: typing.Optional[ListenV1Diarize] = None, + dictation: typing.Optional[ListenV1Dictation] = None, + encoding: typing.Optional[ListenV1Encoding] = None, + endpointing: typing.Optional[ListenV1Endpointing] = None, + extra: typing.Optional[ListenV1Extra] = None, + interim_results: typing.Optional[ListenV1InterimResults] = None, + keyterm: typing.Optional[ListenV1Keyterm] = None, + keywords: typing.Optional[ListenV1Keywords] = None, + language: typing.Optional[ListenV1Language] = None, + mip_opt_out: typing.Optional[ListenV1MipOptOut] = None, + model: ListenV1Model, + multichannel: typing.Optional[ListenV1Multichannel] = None, + numerals: typing.Optional[ListenV1Numerals] = None, + profanity_filter: typing.Optional[ListenV1ProfanityFilter] = None, + punctuate: typing.Optional[ListenV1Punctuate] = None, + redact: typing.Optional[ListenV1Redact] = None, + replace: typing.Optional[ListenV1Replace] = None, + sample_rate: typing.Optional[ListenV1SampleRate] = None, + search: typing.Optional[ListenV1Search] = None, + smart_format: typing.Optional[ListenV1SmartFormat] = None, + tag: typing.Optional[ListenV1Tag] = None, + utterance_end_ms: typing.Optional[ListenV1UtteranceEndMs] = None, + vad_events: typing.Optional[ListenV1VadEvents] = None, + version: typing.Optional[ListenV1Version] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V1SocketClient]: @@ -83,67 +111,67 @@ def connect( Parameters ---------- - callback : typing.Optional[str] + callback : typing.Optional[ListenV1Callback] - callback_method : typing.Optional[str] + callback_method : typing.Optional[ListenV1CallbackMethod] - channels : typing.Optional[str] + channels : typing.Optional[ListenV1Channels] - detect_entities : typing.Optional[str] + detect_entities : typing.Optional[ListenV1DetectEntities] - diarize : typing.Optional[str] + diarize : typing.Optional[ListenV1Diarize] - dictation : typing.Optional[str] + dictation : typing.Optional[ListenV1Dictation] - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV1Encoding] - endpointing : typing.Optional[str] + endpointing : typing.Optional[ListenV1Endpointing] - extra : typing.Optional[typing.Union[str, typing.Sequence[str]]] + extra : typing.Optional[ListenV1Extra] - interim_results : typing.Optional[str] + interim_results : typing.Optional[ListenV1InterimResults] - keyterm : typing.Optional[typing.Union[str, typing.Sequence[str]]] + keyterm : typing.Optional[ListenV1Keyterm] - keywords : typing.Optional[typing.Union[str, typing.Sequence[str]]] + keywords : typing.Optional[ListenV1Keywords] - language : typing.Optional[str] + language : typing.Optional[ListenV1Language] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV1MipOptOut] - model : str + model : ListenV1Model AI model to use for the transcription - multichannel : typing.Optional[str] + multichannel : typing.Optional[ListenV1Multichannel] - numerals : typing.Optional[str] + numerals : typing.Optional[ListenV1Numerals] - profanity_filter : typing.Optional[str] + profanity_filter : typing.Optional[ListenV1ProfanityFilter] - punctuate : typing.Optional[str] + punctuate : typing.Optional[ListenV1Punctuate] - redact : typing.Optional[typing.Union[str, typing.Sequence[str]]] + redact : typing.Optional[ListenV1Redact] - replace : typing.Optional[typing.Union[str, typing.Sequence[str]]] + replace : typing.Optional[ListenV1Replace] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV1SampleRate] - search : typing.Optional[typing.Union[str, typing.Sequence[str]]] + search : typing.Optional[ListenV1Search] - smart_format : typing.Optional[str] + smart_format : typing.Optional[ListenV1SmartFormat] - tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] + tag : typing.Optional[ListenV1Tag] - utterance_end_ms : typing.Optional[str] + utterance_end_ms : typing.Optional[ListenV1UtteranceEndMs] - vad_events : typing.Optional[str] + vad_events : typing.Optional[ListenV1VadEvents] - version : typing.Optional[str] + version : typing.Optional[ListenV1Version] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -248,34 +276,34 @@ def with_raw_response(self) -> AsyncRawV1Client: async def connect( self, *, - callback: typing.Optional[str] = None, - callback_method: typing.Optional[str] = None, - channels: typing.Optional[str] = None, - detect_entities: typing.Optional[str] = None, - diarize: typing.Optional[str] = None, - dictation: typing.Optional[str] = None, - encoding: typing.Optional[str] = None, - endpointing: typing.Optional[str] = None, - extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - interim_results: typing.Optional[str] = None, - keyterm: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - language: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: str, - multichannel: typing.Optional[str] = None, - numerals: typing.Optional[str] = None, - profanity_filter: typing.Optional[str] = None, - punctuate: typing.Optional[str] = None, - redact: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - replace: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - sample_rate: typing.Optional[str] = None, - search: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - smart_format: typing.Optional[str] = None, - tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - utterance_end_ms: typing.Optional[str] = None, - vad_events: typing.Optional[str] = None, - version: typing.Optional[str] = None, + callback: typing.Optional[ListenV1Callback] = None, + callback_method: typing.Optional[ListenV1CallbackMethod] = None, + channels: typing.Optional[ListenV1Channels] = None, + detect_entities: typing.Optional[ListenV1DetectEntities] = None, + diarize: typing.Optional[ListenV1Diarize] = None, + dictation: typing.Optional[ListenV1Dictation] = None, + encoding: typing.Optional[ListenV1Encoding] = None, + endpointing: typing.Optional[ListenV1Endpointing] = None, + extra: typing.Optional[ListenV1Extra] = None, + interim_results: typing.Optional[ListenV1InterimResults] = None, + keyterm: typing.Optional[ListenV1Keyterm] = None, + keywords: typing.Optional[ListenV1Keywords] = None, + language: typing.Optional[ListenV1Language] = None, + mip_opt_out: typing.Optional[ListenV1MipOptOut] = None, + model: ListenV1Model, + multichannel: typing.Optional[ListenV1Multichannel] = None, + numerals: typing.Optional[ListenV1Numerals] = None, + profanity_filter: typing.Optional[ListenV1ProfanityFilter] = None, + punctuate: typing.Optional[ListenV1Punctuate] = None, + redact: typing.Optional[ListenV1Redact] = None, + replace: typing.Optional[ListenV1Replace] = None, + sample_rate: typing.Optional[ListenV1SampleRate] = None, + search: typing.Optional[ListenV1Search] = None, + smart_format: typing.Optional[ListenV1SmartFormat] = None, + tag: typing.Optional[ListenV1Tag] = None, + utterance_end_ms: typing.Optional[ListenV1UtteranceEndMs] = None, + vad_events: typing.Optional[ListenV1VadEvents] = None, + version: typing.Optional[ListenV1Version] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV1SocketClient]: @@ -284,67 +312,67 @@ async def connect( Parameters ---------- - callback : typing.Optional[str] + callback : typing.Optional[ListenV1Callback] - callback_method : typing.Optional[str] + callback_method : typing.Optional[ListenV1CallbackMethod] - channels : typing.Optional[str] + channels : typing.Optional[ListenV1Channels] - detect_entities : typing.Optional[str] + detect_entities : typing.Optional[ListenV1DetectEntities] - diarize : typing.Optional[str] + diarize : typing.Optional[ListenV1Diarize] - dictation : typing.Optional[str] + dictation : typing.Optional[ListenV1Dictation] - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV1Encoding] - endpointing : typing.Optional[str] + endpointing : typing.Optional[ListenV1Endpointing] - extra : typing.Optional[typing.Union[str, typing.Sequence[str]]] + extra : typing.Optional[ListenV1Extra] - interim_results : typing.Optional[str] + interim_results : typing.Optional[ListenV1InterimResults] - keyterm : typing.Optional[typing.Union[str, typing.Sequence[str]]] + keyterm : typing.Optional[ListenV1Keyterm] - keywords : typing.Optional[typing.Union[str, typing.Sequence[str]]] + keywords : typing.Optional[ListenV1Keywords] - language : typing.Optional[str] + language : typing.Optional[ListenV1Language] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV1MipOptOut] - model : str + model : ListenV1Model AI model to use for the transcription - multichannel : typing.Optional[str] + multichannel : typing.Optional[ListenV1Multichannel] - numerals : typing.Optional[str] + numerals : typing.Optional[ListenV1Numerals] - profanity_filter : typing.Optional[str] + profanity_filter : typing.Optional[ListenV1ProfanityFilter] - punctuate : typing.Optional[str] + punctuate : typing.Optional[ListenV1Punctuate] - redact : typing.Optional[typing.Union[str, typing.Sequence[str]]] + redact : typing.Optional[ListenV1Redact] - replace : typing.Optional[typing.Union[str, typing.Sequence[str]]] + replace : typing.Optional[ListenV1Replace] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV1SampleRate] - search : typing.Optional[typing.Union[str, typing.Sequence[str]]] + search : typing.Optional[ListenV1Search] - smart_format : typing.Optional[str] + smart_format : typing.Optional[ListenV1SmartFormat] - tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] + tag : typing.Optional[ListenV1Tag] - utterance_end_ms : typing.Optional[str] + utterance_end_ms : typing.Optional[ListenV1UtteranceEndMs] - vad_events : typing.Optional[str] + vad_events : typing.Optional[ListenV1VadEvents] - version : typing.Optional[str] + version : typing.Optional[ListenV1Version] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/listen/v1/media/__init__.py b/src/deepgram/listen/v1/media/__init__.py index 495ed32d..6cadb71a 100644 --- a/src/deepgram/listen/v1/media/__init__.py +++ b/src/deepgram/listen/v1/media/__init__.py @@ -11,9 +11,9 @@ MediaTranscribeRequestCustomIntentMode, MediaTranscribeRequestCustomTopicMode, MediaTranscribeRequestEncoding, - MediaTranscribeRequestModel, - MediaTranscribeRequestSummarize, - MediaTranscribeRequestVersion, + MediaTranscribeRequestModelZero, + MediaTranscribeRequestSummarizeZero, + MediaTranscribeRequestVersionZero, MediaTranscribeResponse, ) from .requests import MediaTranscribeResponseParams @@ -22,9 +22,9 @@ "MediaTranscribeRequestCustomIntentMode": ".types", "MediaTranscribeRequestCustomTopicMode": ".types", "MediaTranscribeRequestEncoding": ".types", - "MediaTranscribeRequestModel": ".types", - "MediaTranscribeRequestSummarize": ".types", - "MediaTranscribeRequestVersion": ".types", + "MediaTranscribeRequestModelZero": ".types", + "MediaTranscribeRequestSummarizeZero": ".types", + "MediaTranscribeRequestVersionZero": ".types", "MediaTranscribeResponse": ".types", "MediaTranscribeResponseParams": ".requests", } @@ -56,9 +56,9 @@ def __dir__(): "MediaTranscribeRequestCustomIntentMode", "MediaTranscribeRequestCustomTopicMode", "MediaTranscribeRequestEncoding", - "MediaTranscribeRequestModel", - "MediaTranscribeRequestSummarize", - "MediaTranscribeRequestVersion", + "MediaTranscribeRequestModelZero", + "MediaTranscribeRequestSummarizeZero", + "MediaTranscribeRequestVersionZero", "MediaTranscribeResponse", "MediaTranscribeResponseParams", ] diff --git a/src/deepgram/listen/v1/media/client.py b/src/deepgram/listen/v1/media/client.py index ab81ae3b..bbc8b98f 100644 --- a/src/deepgram/listen/v1/media/client.py +++ b/src/deepgram/listen/v1/media/client.py @@ -9,9 +9,9 @@ from .types.media_transcribe_request_custom_intent_mode import MediaTranscribeRequestCustomIntentMode from .types.media_transcribe_request_custom_topic_mode import MediaTranscribeRequestCustomTopicMode from .types.media_transcribe_request_encoding import MediaTranscribeRequestEncoding -from .types.media_transcribe_request_model import MediaTranscribeRequestModel -from .types.media_transcribe_request_summarize import MediaTranscribeRequestSummarize -from .types.media_transcribe_request_version import MediaTranscribeRequestVersion +from .types.media_transcribe_request_model_zero import MediaTranscribeRequestModelZero +from .types.media_transcribe_request_summarize_zero import MediaTranscribeRequestSummarizeZero +from .types.media_transcribe_request_version_zero import MediaTranscribeRequestVersionZero from .types.media_transcribe_response import MediaTranscribeResponse # this is used as the default value for optional parameters @@ -41,7 +41,7 @@ def transcribe_url( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -59,7 +59,7 @@ def transcribe_url( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -71,7 +71,7 @@ def transcribe_url( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> MediaTranscribeResponse: @@ -94,7 +94,7 @@ def transcribe_url( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -148,7 +148,7 @@ def transcribe_url( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -184,7 +184,7 @@ def transcribe_url( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -294,7 +294,7 @@ def transcribe_file( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -312,7 +312,7 @@ def transcribe_file( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -324,7 +324,7 @@ def transcribe_file( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> MediaTranscribeResponse: @@ -347,7 +347,7 @@ def transcribe_file( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -401,7 +401,7 @@ def transcribe_file( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -437,7 +437,7 @@ def transcribe_file( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -526,7 +526,7 @@ async def transcribe_url( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -544,7 +544,7 @@ async def transcribe_url( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -556,7 +556,7 @@ async def transcribe_url( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> MediaTranscribeResponse: @@ -579,7 +579,7 @@ async def transcribe_url( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -633,7 +633,7 @@ async def transcribe_url( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -669,7 +669,7 @@ async def transcribe_url( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -787,7 +787,7 @@ async def transcribe_file( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -805,7 +805,7 @@ async def transcribe_file( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -817,7 +817,7 @@ async def transcribe_file( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> MediaTranscribeResponse: @@ -840,7 +840,7 @@ async def transcribe_file( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -894,7 +894,7 @@ async def transcribe_file( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -930,7 +930,7 @@ async def transcribe_file( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] diff --git a/src/deepgram/listen/v1/media/raw_client.py b/src/deepgram/listen/v1/media/raw_client.py index 830d4259..88d1c5e9 100644 --- a/src/deepgram/listen/v1/media/raw_client.py +++ b/src/deepgram/listen/v1/media/raw_client.py @@ -13,9 +13,9 @@ from .types.media_transcribe_request_custom_intent_mode import MediaTranscribeRequestCustomIntentMode from .types.media_transcribe_request_custom_topic_mode import MediaTranscribeRequestCustomTopicMode from .types.media_transcribe_request_encoding import MediaTranscribeRequestEncoding -from .types.media_transcribe_request_model import MediaTranscribeRequestModel -from .types.media_transcribe_request_summarize import MediaTranscribeRequestSummarize -from .types.media_transcribe_request_version import MediaTranscribeRequestVersion +from .types.media_transcribe_request_model_zero import MediaTranscribeRequestModelZero +from .types.media_transcribe_request_summarize_zero import MediaTranscribeRequestSummarizeZero +from .types.media_transcribe_request_version_zero import MediaTranscribeRequestVersionZero from .types.media_transcribe_response import MediaTranscribeResponse # this is used as the default value for optional parameters @@ -34,7 +34,7 @@ def transcribe_url( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -52,7 +52,7 @@ def transcribe_url( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -64,7 +64,7 @@ def transcribe_url( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> HttpResponse[MediaTranscribeResponse]: @@ -87,7 +87,7 @@ def transcribe_url( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -141,7 +141,7 @@ def transcribe_url( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -177,7 +177,7 @@ def transcribe_url( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -276,7 +276,7 @@ def transcribe_file( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -294,7 +294,7 @@ def transcribe_file( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -306,7 +306,7 @@ def transcribe_file( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> HttpResponse[MediaTranscribeResponse]: @@ -329,7 +329,7 @@ def transcribe_file( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -383,7 +383,7 @@ def transcribe_file( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -419,7 +419,7 @@ def transcribe_file( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -521,7 +521,7 @@ async def transcribe_url( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -539,7 +539,7 @@ async def transcribe_url( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -551,7 +551,7 @@ async def transcribe_url( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> AsyncHttpResponse[MediaTranscribeResponse]: @@ -574,7 +574,7 @@ async def transcribe_url( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -628,7 +628,7 @@ async def transcribe_url( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -664,7 +664,7 @@ async def transcribe_url( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] @@ -763,7 +763,7 @@ async def transcribe_file( callback_method: typing.Optional[MediaTranscribeRequestCallbackMethod] = None, extra: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[MediaTranscribeRequestSummarize] = None, + summarize: typing.Optional[MediaTranscribeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -781,7 +781,7 @@ async def transcribe_file( keywords: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, language: typing.Optional[str] = None, measurements: typing.Optional[bool] = None, - model: typing.Optional[MediaTranscribeRequestModel] = None, + model: typing.Optional[MediaTranscribeRequestModelZero] = None, multichannel: typing.Optional[bool] = None, numerals: typing.Optional[bool] = None, paragraphs: typing.Optional[bool] = None, @@ -793,7 +793,7 @@ async def transcribe_file( smart_format: typing.Optional[bool] = None, utterances: typing.Optional[bool] = None, utt_split: typing.Optional[float] = None, - version: typing.Optional[MediaTranscribeRequestVersion] = None, + version: typing.Optional[MediaTranscribeRequestVersionZero] = None, mip_opt_out: typing.Optional[bool] = None, request_options: typing.Optional[RequestOptions] = None, ) -> AsyncHttpResponse[MediaTranscribeResponse]: @@ -816,7 +816,7 @@ async def transcribe_file( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[MediaTranscribeRequestSummarize] + summarize : typing.Optional[MediaTranscribeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -870,7 +870,7 @@ async def transcribe_file( measurements : typing.Optional[bool] Spoken measurements will be converted to their corresponding abbreviations - model : typing.Optional[MediaTranscribeRequestModel] + model : typing.Optional[MediaTranscribeRequestModelZero] AI model used to process submitted audio multichannel : typing.Optional[bool] @@ -906,7 +906,7 @@ async def transcribe_file( utt_split : typing.Optional[float] Seconds to wait before detecting a pause between words in submitted audio - version : typing.Optional[MediaTranscribeRequestVersion] + version : typing.Optional[MediaTranscribeRequestVersionZero] Version of an AI model to use mip_opt_out : typing.Optional[bool] diff --git a/src/deepgram/listen/v1/media/types/__init__.py b/src/deepgram/listen/v1/media/types/__init__.py index d0ff0fb9..81d1d91b 100644 --- a/src/deepgram/listen/v1/media/types/__init__.py +++ b/src/deepgram/listen/v1/media/types/__init__.py @@ -10,18 +10,18 @@ from .media_transcribe_request_custom_intent_mode import MediaTranscribeRequestCustomIntentMode from .media_transcribe_request_custom_topic_mode import MediaTranscribeRequestCustomTopicMode from .media_transcribe_request_encoding import MediaTranscribeRequestEncoding - from .media_transcribe_request_model import MediaTranscribeRequestModel - from .media_transcribe_request_summarize import MediaTranscribeRequestSummarize - from .media_transcribe_request_version import MediaTranscribeRequestVersion + from .media_transcribe_request_model_zero import MediaTranscribeRequestModelZero + from .media_transcribe_request_summarize_zero import MediaTranscribeRequestSummarizeZero + from .media_transcribe_request_version_zero import MediaTranscribeRequestVersionZero from .media_transcribe_response import MediaTranscribeResponse _dynamic_imports: typing.Dict[str, str] = { "MediaTranscribeRequestCallbackMethod": ".media_transcribe_request_callback_method", "MediaTranscribeRequestCustomIntentMode": ".media_transcribe_request_custom_intent_mode", "MediaTranscribeRequestCustomTopicMode": ".media_transcribe_request_custom_topic_mode", "MediaTranscribeRequestEncoding": ".media_transcribe_request_encoding", - "MediaTranscribeRequestModel": ".media_transcribe_request_model", - "MediaTranscribeRequestSummarize": ".media_transcribe_request_summarize", - "MediaTranscribeRequestVersion": ".media_transcribe_request_version", + "MediaTranscribeRequestModelZero": ".media_transcribe_request_model_zero", + "MediaTranscribeRequestSummarizeZero": ".media_transcribe_request_summarize_zero", + "MediaTranscribeRequestVersionZero": ".media_transcribe_request_version_zero", "MediaTranscribeResponse": ".media_transcribe_response", } @@ -52,8 +52,8 @@ def __dir__(): "MediaTranscribeRequestCustomIntentMode", "MediaTranscribeRequestCustomTopicMode", "MediaTranscribeRequestEncoding", - "MediaTranscribeRequestModel", - "MediaTranscribeRequestSummarize", - "MediaTranscribeRequestVersion", + "MediaTranscribeRequestModelZero", + "MediaTranscribeRequestSummarizeZero", + "MediaTranscribeRequestVersionZero", "MediaTranscribeResponse", ] diff --git a/src/deepgram/listen/v1/media/types/media_transcribe_request_model.py b/src/deepgram/listen/v1/media/types/media_transcribe_request_model_zero.py similarity index 94% rename from src/deepgram/listen/v1/media/types/media_transcribe_request_model.py rename to src/deepgram/listen/v1/media/types/media_transcribe_request_model_zero.py index eaebbc7a..6608514c 100644 --- a/src/deepgram/listen/v1/media/types/media_transcribe_request_model.py +++ b/src/deepgram/listen/v1/media/types/media_transcribe_request_model_zero.py @@ -2,7 +2,7 @@ import typing -MediaTranscribeRequestModel = typing.Union[ +MediaTranscribeRequestModelZero = typing.Union[ typing.Literal[ "nova-3", "nova-3-general", diff --git a/src/deepgram/listen/v1/media/types/media_transcribe_request_summarize_zero.py b/src/deepgram/listen/v1/media/types/media_transcribe_request_summarize_zero.py new file mode 100644 index 00000000..a91b33f1 --- /dev/null +++ b/src/deepgram/listen/v1/media/types/media_transcribe_request_summarize_zero.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +MediaTranscribeRequestSummarizeZero = typing.Union[typing.Literal["v2"], typing.Any] diff --git a/src/deepgram/listen/v1/media/types/media_transcribe_request_version.py b/src/deepgram/listen/v1/media/types/media_transcribe_request_version.py deleted file mode 100644 index e6ae1cd5..00000000 --- a/src/deepgram/listen/v1/media/types/media_transcribe_request_version.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -MediaTranscribeRequestVersion = typing.Union[typing.Literal["latest"], typing.Any] diff --git a/src/deepgram/listen/v1/media/types/media_transcribe_request_version_zero.py b/src/deepgram/listen/v1/media/types/media_transcribe_request_version_zero.py new file mode 100644 index 00000000..032f5c31 --- /dev/null +++ b/src/deepgram/listen/v1/media/types/media_transcribe_request_version_zero.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +MediaTranscribeRequestVersionZero = typing.Union[typing.Literal["latest"], typing.Any] diff --git a/src/deepgram/listen/v1/raw_client.py b/src/deepgram/listen/v1/raw_client.py index 7b2d6bd6..f85fe623 100644 --- a/src/deepgram/listen/v1/raw_client.py +++ b/src/deepgram/listen/v1/raw_client.py @@ -12,6 +12,34 @@ from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...types.listen_v1callback import ListenV1Callback +from ...types.listen_v1callback_method import ListenV1CallbackMethod +from ...types.listen_v1channels import ListenV1Channels +from ...types.listen_v1detect_entities import ListenV1DetectEntities +from ...types.listen_v1diarize import ListenV1Diarize +from ...types.listen_v1dictation import ListenV1Dictation +from ...types.listen_v1encoding import ListenV1Encoding +from ...types.listen_v1endpointing import ListenV1Endpointing +from ...types.listen_v1extra import ListenV1Extra +from ...types.listen_v1interim_results import ListenV1InterimResults +from ...types.listen_v1keyterm import ListenV1Keyterm +from ...types.listen_v1keywords import ListenV1Keywords +from ...types.listen_v1language import ListenV1Language +from ...types.listen_v1mip_opt_out import ListenV1MipOptOut +from ...types.listen_v1model import ListenV1Model +from ...types.listen_v1multichannel import ListenV1Multichannel +from ...types.listen_v1numerals import ListenV1Numerals +from ...types.listen_v1profanity_filter import ListenV1ProfanityFilter +from ...types.listen_v1punctuate import ListenV1Punctuate +from ...types.listen_v1redact import ListenV1Redact +from ...types.listen_v1replace import ListenV1Replace +from ...types.listen_v1sample_rate import ListenV1SampleRate +from ...types.listen_v1search import ListenV1Search +from ...types.listen_v1smart_format import ListenV1SmartFormat +from ...types.listen_v1tag import ListenV1Tag +from ...types.listen_v1utterance_end_ms import ListenV1UtteranceEndMs +from ...types.listen_v1vad_events import ListenV1VadEvents +from ...types.listen_v1version import ListenV1Version from .socket_client import AsyncV1SocketClient, V1SocketClient try: @@ -28,34 +56,34 @@ def __init__(self, *, client_wrapper: SyncClientWrapper): def connect( self, *, - callback: typing.Optional[str] = None, - callback_method: typing.Optional[str] = None, - channels: typing.Optional[str] = None, - detect_entities: typing.Optional[str] = None, - diarize: typing.Optional[str] = None, - dictation: typing.Optional[str] = None, - encoding: typing.Optional[str] = None, - endpointing: typing.Optional[str] = None, - extra: typing.Optional[str] = None, - interim_results: typing.Optional[str] = None, - keyterm: typing.Optional[str] = None, - keywords: typing.Optional[str] = None, - language: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: str, - multichannel: typing.Optional[str] = None, - numerals: typing.Optional[str] = None, - profanity_filter: typing.Optional[str] = None, - punctuate: typing.Optional[str] = None, - redact: typing.Optional[str] = None, - replace: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - search: typing.Optional[str] = None, - smart_format: typing.Optional[str] = None, - tag: typing.Optional[str] = None, - utterance_end_ms: typing.Optional[str] = None, - vad_events: typing.Optional[str] = None, - version: typing.Optional[str] = None, + callback: typing.Optional[ListenV1Callback] = None, + callback_method: typing.Optional[ListenV1CallbackMethod] = None, + channels: typing.Optional[ListenV1Channels] = None, + detect_entities: typing.Optional[ListenV1DetectEntities] = None, + diarize: typing.Optional[ListenV1Diarize] = None, + dictation: typing.Optional[ListenV1Dictation] = None, + encoding: typing.Optional[ListenV1Encoding] = None, + endpointing: typing.Optional[ListenV1Endpointing] = None, + extra: typing.Optional[ListenV1Extra] = None, + interim_results: typing.Optional[ListenV1InterimResults] = None, + keyterm: typing.Optional[ListenV1Keyterm] = None, + keywords: typing.Optional[ListenV1Keywords] = None, + language: typing.Optional[ListenV1Language] = None, + mip_opt_out: typing.Optional[ListenV1MipOptOut] = None, + model: ListenV1Model, + multichannel: typing.Optional[ListenV1Multichannel] = None, + numerals: typing.Optional[ListenV1Numerals] = None, + profanity_filter: typing.Optional[ListenV1ProfanityFilter] = None, + punctuate: typing.Optional[ListenV1Punctuate] = None, + redact: typing.Optional[ListenV1Redact] = None, + replace: typing.Optional[ListenV1Replace] = None, + sample_rate: typing.Optional[ListenV1SampleRate] = None, + search: typing.Optional[ListenV1Search] = None, + smart_format: typing.Optional[ListenV1SmartFormat] = None, + tag: typing.Optional[ListenV1Tag] = None, + utterance_end_ms: typing.Optional[ListenV1UtteranceEndMs] = None, + vad_events: typing.Optional[ListenV1VadEvents] = None, + version: typing.Optional[ListenV1Version] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V1SocketClient]: @@ -64,67 +92,67 @@ def connect( Parameters ---------- - callback : typing.Optional[str] + callback : typing.Optional[ListenV1Callback] - callback_method : typing.Optional[str] + callback_method : typing.Optional[ListenV1CallbackMethod] - channels : typing.Optional[str] + channels : typing.Optional[ListenV1Channels] - detect_entities : typing.Optional[str] + detect_entities : typing.Optional[ListenV1DetectEntities] - diarize : typing.Optional[str] + diarize : typing.Optional[ListenV1Diarize] - dictation : typing.Optional[str] + dictation : typing.Optional[ListenV1Dictation] - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV1Encoding] - endpointing : typing.Optional[str] + endpointing : typing.Optional[ListenV1Endpointing] - extra : typing.Optional[str] + extra : typing.Optional[ListenV1Extra] - interim_results : typing.Optional[str] + interim_results : typing.Optional[ListenV1InterimResults] - keyterm : typing.Optional[str] + keyterm : typing.Optional[ListenV1Keyterm] - keywords : typing.Optional[str] + keywords : typing.Optional[ListenV1Keywords] - language : typing.Optional[str] + language : typing.Optional[ListenV1Language] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV1MipOptOut] - model : str + model : ListenV1Model AI model to use for the transcription - multichannel : typing.Optional[str] + multichannel : typing.Optional[ListenV1Multichannel] - numerals : typing.Optional[str] + numerals : typing.Optional[ListenV1Numerals] - profanity_filter : typing.Optional[str] + profanity_filter : typing.Optional[ListenV1ProfanityFilter] - punctuate : typing.Optional[str] + punctuate : typing.Optional[ListenV1Punctuate] - redact : typing.Optional[str] + redact : typing.Optional[ListenV1Redact] - replace : typing.Optional[str] + replace : typing.Optional[ListenV1Replace] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV1SampleRate] - search : typing.Optional[str] + search : typing.Optional[ListenV1Search] - smart_format : typing.Optional[str] + smart_format : typing.Optional[ListenV1SmartFormat] - tag : typing.Optional[str] + tag : typing.Optional[ListenV1Tag] - utterance_end_ms : typing.Optional[str] + utterance_end_ms : typing.Optional[ListenV1UtteranceEndMs] - vad_events : typing.Optional[str] + vad_events : typing.Optional[ListenV1VadEvents] - version : typing.Optional[str] + version : typing.Optional[ListenV1Version] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -208,34 +236,34 @@ def __init__(self, *, client_wrapper: AsyncClientWrapper): async def connect( self, *, - callback: typing.Optional[str] = None, - callback_method: typing.Optional[str] = None, - channels: typing.Optional[str] = None, - detect_entities: typing.Optional[str] = None, - diarize: typing.Optional[str] = None, - dictation: typing.Optional[str] = None, - encoding: typing.Optional[str] = None, - endpointing: typing.Optional[str] = None, - extra: typing.Optional[str] = None, - interim_results: typing.Optional[str] = None, - keyterm: typing.Optional[str] = None, - keywords: typing.Optional[str] = None, - language: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: str, - multichannel: typing.Optional[str] = None, - numerals: typing.Optional[str] = None, - profanity_filter: typing.Optional[str] = None, - punctuate: typing.Optional[str] = None, - redact: typing.Optional[str] = None, - replace: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - search: typing.Optional[str] = None, - smart_format: typing.Optional[str] = None, - tag: typing.Optional[str] = None, - utterance_end_ms: typing.Optional[str] = None, - vad_events: typing.Optional[str] = None, - version: typing.Optional[str] = None, + callback: typing.Optional[ListenV1Callback] = None, + callback_method: typing.Optional[ListenV1CallbackMethod] = None, + channels: typing.Optional[ListenV1Channels] = None, + detect_entities: typing.Optional[ListenV1DetectEntities] = None, + diarize: typing.Optional[ListenV1Diarize] = None, + dictation: typing.Optional[ListenV1Dictation] = None, + encoding: typing.Optional[ListenV1Encoding] = None, + endpointing: typing.Optional[ListenV1Endpointing] = None, + extra: typing.Optional[ListenV1Extra] = None, + interim_results: typing.Optional[ListenV1InterimResults] = None, + keyterm: typing.Optional[ListenV1Keyterm] = None, + keywords: typing.Optional[ListenV1Keywords] = None, + language: typing.Optional[ListenV1Language] = None, + mip_opt_out: typing.Optional[ListenV1MipOptOut] = None, + model: ListenV1Model, + multichannel: typing.Optional[ListenV1Multichannel] = None, + numerals: typing.Optional[ListenV1Numerals] = None, + profanity_filter: typing.Optional[ListenV1ProfanityFilter] = None, + punctuate: typing.Optional[ListenV1Punctuate] = None, + redact: typing.Optional[ListenV1Redact] = None, + replace: typing.Optional[ListenV1Replace] = None, + sample_rate: typing.Optional[ListenV1SampleRate] = None, + search: typing.Optional[ListenV1Search] = None, + smart_format: typing.Optional[ListenV1SmartFormat] = None, + tag: typing.Optional[ListenV1Tag] = None, + utterance_end_ms: typing.Optional[ListenV1UtteranceEndMs] = None, + vad_events: typing.Optional[ListenV1VadEvents] = None, + version: typing.Optional[ListenV1Version] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV1SocketClient]: @@ -244,67 +272,67 @@ async def connect( Parameters ---------- - callback : typing.Optional[str] + callback : typing.Optional[ListenV1Callback] - callback_method : typing.Optional[str] + callback_method : typing.Optional[ListenV1CallbackMethod] - channels : typing.Optional[str] + channels : typing.Optional[ListenV1Channels] - detect_entities : typing.Optional[str] + detect_entities : typing.Optional[ListenV1DetectEntities] - diarize : typing.Optional[str] + diarize : typing.Optional[ListenV1Diarize] - dictation : typing.Optional[str] + dictation : typing.Optional[ListenV1Dictation] - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV1Encoding] - endpointing : typing.Optional[str] + endpointing : typing.Optional[ListenV1Endpointing] - extra : typing.Optional[str] + extra : typing.Optional[ListenV1Extra] - interim_results : typing.Optional[str] + interim_results : typing.Optional[ListenV1InterimResults] - keyterm : typing.Optional[str] + keyterm : typing.Optional[ListenV1Keyterm] - keywords : typing.Optional[str] + keywords : typing.Optional[ListenV1Keywords] - language : typing.Optional[str] + language : typing.Optional[ListenV1Language] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV1MipOptOut] - model : str + model : ListenV1Model AI model to use for the transcription - multichannel : typing.Optional[str] + multichannel : typing.Optional[ListenV1Multichannel] - numerals : typing.Optional[str] + numerals : typing.Optional[ListenV1Numerals] - profanity_filter : typing.Optional[str] + profanity_filter : typing.Optional[ListenV1ProfanityFilter] - punctuate : typing.Optional[str] + punctuate : typing.Optional[ListenV1Punctuate] - redact : typing.Optional[str] + redact : typing.Optional[ListenV1Redact] - replace : typing.Optional[str] + replace : typing.Optional[ListenV1Replace] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV1SampleRate] - search : typing.Optional[str] + search : typing.Optional[ListenV1Search] - smart_format : typing.Optional[str] + smart_format : typing.Optional[ListenV1SmartFormat] - tag : typing.Optional[str] + tag : typing.Optional[ListenV1Tag] - utterance_end_ms : typing.Optional[str] + utterance_end_ms : typing.Optional[ListenV1UtteranceEndMs] - vad_events : typing.Optional[str] + vad_events : typing.Optional[ListenV1VadEvents] - version : typing.Optional[str] + version : typing.Optional[ListenV1Version] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/listen/v1/requests/listen_v1metadata.py b/src/deepgram/listen/v1/requests/listen_v1metadata.py index 2b648b48..c6564659 100644 --- a/src/deepgram/listen/v1/requests/listen_v1metadata.py +++ b/src/deepgram/listen/v1/requests/listen_v1metadata.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.listen_v1metadata_type import ListenV1MetadataType class ListenV1MetadataParams(typing_extensions.TypedDict): - type: typing.Literal["Metadata"] + type: ListenV1MetadataType """ Message type identifier """ diff --git a/src/deepgram/listen/v1/requests/listen_v1results.py b/src/deepgram/listen/v1/requests/listen_v1results.py index 96e152f0..31577afa 100644 --- a/src/deepgram/listen/v1/requests/listen_v1results.py +++ b/src/deepgram/listen/v1/requests/listen_v1results.py @@ -3,13 +3,14 @@ import typing import typing_extensions +from ..types.listen_v1results_type import ListenV1ResultsType from .listen_v1results_channel import ListenV1ResultsChannelParams from .listen_v1results_entities_item import ListenV1ResultsEntitiesItemParams from .listen_v1results_metadata import ListenV1ResultsMetadataParams class ListenV1ResultsParams(typing_extensions.TypedDict): - type: typing.Literal["Results"] + type: ListenV1ResultsType """ Message type identifier """ diff --git a/src/deepgram/listen/v1/requests/listen_v1speech_started.py b/src/deepgram/listen/v1/requests/listen_v1speech_started.py index 1cc1dcfa..5741ca46 100644 --- a/src/deepgram/listen/v1/requests/listen_v1speech_started.py +++ b/src/deepgram/listen/v1/requests/listen_v1speech_started.py @@ -3,10 +3,11 @@ import typing import typing_extensions +from ..types.listen_v1speech_started_type import ListenV1SpeechStartedType class ListenV1SpeechStartedParams(typing_extensions.TypedDict): - type: typing.Literal["SpeechStarted"] + type: ListenV1SpeechStartedType """ Message type identifier """ diff --git a/src/deepgram/listen/v1/requests/listen_v1utterance_end.py b/src/deepgram/listen/v1/requests/listen_v1utterance_end.py index 37ae57b8..cac51946 100644 --- a/src/deepgram/listen/v1/requests/listen_v1utterance_end.py +++ b/src/deepgram/listen/v1/requests/listen_v1utterance_end.py @@ -3,10 +3,11 @@ import typing import typing_extensions +from ..types.listen_v1utterance_end_type import ListenV1UtteranceEndType class ListenV1UtteranceEndParams(typing_extensions.TypedDict): - type: typing.Literal["UtteranceEnd"] + type: ListenV1UtteranceEndType """ Message type identifier """ diff --git a/src/deepgram/listen/v1/socket_client.py b/src/deepgram/listen/v1/socket_client.py index c16839d3..68feddb6 100644 --- a/src/deepgram/listen/v1/socket_client.py +++ b/src/deepgram/listen/v1/socket_client.py @@ -1,6 +1,7 @@ # This file was auto-generated by Fern from our API Definition. import json +import logging import typing from json.decoder import JSONDecodeError @@ -21,6 +22,7 @@ except ImportError: from websockets import WebSocketClientProtocol # type: ignore +_logger = logging.getLogger(__name__) V1SocketClientResponse = typing.Union[ListenV1Results, ListenV1Metadata, ListenV1UtteranceEnd, ListenV1SpeechStarted] @@ -34,7 +36,13 @@ async def __aiter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue async def start_listening(self): """ @@ -53,7 +61,13 @@ async def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue await self._emit_async(EventType.MESSAGE, parsed) except Exception as exc: await self._emit_async(EventType.ERROR, exc) @@ -67,26 +81,26 @@ async def send_media(self, message: bytes) -> None: """ await self._send(message) - async def send_finalize(self, message: typing.Optional[ListenV1Finalize] = None) -> None: + async def send_finalize(self, message: ListenV1Finalize) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1Finalize. """ - await self._send_model(message or ListenV1Finalize(type="Finalize")) + await self._send_model(message) - async def send_close_stream(self, message: typing.Optional[ListenV1CloseStream] = None) -> None: + async def send_close_stream(self, message: ListenV1CloseStream) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1CloseStream. """ - await self._send_model(message or ListenV1CloseStream(type="CloseStream")) + await self._send_model(message) - async def send_keep_alive(self, message: typing.Optional[ListenV1KeepAlive] = None) -> None: + async def send_keep_alive(self, message: ListenV1KeepAlive) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1KeepAlive. """ - await self._send_model(message or ListenV1KeepAlive(type="KeepAlive")) + await self._send_model(message) async def recv(self) -> V1SocketClientResponse: """ @@ -96,7 +110,11 @@ async def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore async def _send(self, data: typing.Any) -> None: """ @@ -123,7 +141,13 @@ def __iter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue def start_listening(self): """ @@ -142,7 +166,13 @@ def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue self._emit(EventType.MESSAGE, parsed) except Exception as exc: self._emit(EventType.ERROR, exc) @@ -156,26 +186,26 @@ def send_media(self, message: bytes) -> None: """ self._send(message) - def send_finalize(self, message: typing.Optional[ListenV1Finalize] = None) -> None: + def send_finalize(self, message: ListenV1Finalize) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1Finalize. """ - self._send_model(message or ListenV1Finalize(type="Finalize")) + self._send_model(message) - def send_close_stream(self, message: typing.Optional[ListenV1CloseStream] = None) -> None: + def send_close_stream(self, message: ListenV1CloseStream) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1CloseStream. """ - self._send_model(message or ListenV1CloseStream(type="CloseStream")) + self._send_model(message) - def send_keep_alive(self, message: typing.Optional[ListenV1KeepAlive] = None) -> None: + def send_keep_alive(self, message: ListenV1KeepAlive) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV1KeepAlive. """ - self._send_model(message or ListenV1KeepAlive(type="KeepAlive")) + self._send_model(message) def recv(self) -> V1SocketClientResponse: """ @@ -185,7 +215,11 @@ def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore def _send(self, data: typing.Any) -> None: """ diff --git a/src/deepgram/listen/v1/types/__init__.py b/src/deepgram/listen/v1/types/__init__.py index 30c6849a..4eab0761 100644 --- a/src/deepgram/listen/v1/types/__init__.py +++ b/src/deepgram/listen/v1/types/__init__.py @@ -13,6 +13,7 @@ from .listen_v1keep_alive import ListenV1KeepAlive from .listen_v1keep_alive_type import ListenV1KeepAliveType from .listen_v1metadata import ListenV1Metadata + from .listen_v1metadata_type import ListenV1MetadataType from .listen_v1results import ListenV1Results from .listen_v1results_channel import ListenV1ResultsChannel from .listen_v1results_channel_alternatives_item import ListenV1ResultsChannelAlternativesItem @@ -20,8 +21,11 @@ from .listen_v1results_entities_item import ListenV1ResultsEntitiesItem from .listen_v1results_metadata import ListenV1ResultsMetadata from .listen_v1results_metadata_model_info import ListenV1ResultsMetadataModelInfo + from .listen_v1results_type import ListenV1ResultsType from .listen_v1speech_started import ListenV1SpeechStarted + from .listen_v1speech_started_type import ListenV1SpeechStartedType from .listen_v1utterance_end import ListenV1UtteranceEnd + from .listen_v1utterance_end_type import ListenV1UtteranceEndType _dynamic_imports: typing.Dict[str, str] = { "ListenV1CloseStream": ".listen_v1close_stream", "ListenV1CloseStreamType": ".listen_v1close_stream_type", @@ -30,6 +34,7 @@ "ListenV1KeepAlive": ".listen_v1keep_alive", "ListenV1KeepAliveType": ".listen_v1keep_alive_type", "ListenV1Metadata": ".listen_v1metadata", + "ListenV1MetadataType": ".listen_v1metadata_type", "ListenV1Results": ".listen_v1results", "ListenV1ResultsChannel": ".listen_v1results_channel", "ListenV1ResultsChannelAlternativesItem": ".listen_v1results_channel_alternatives_item", @@ -37,8 +42,11 @@ "ListenV1ResultsEntitiesItem": ".listen_v1results_entities_item", "ListenV1ResultsMetadata": ".listen_v1results_metadata", "ListenV1ResultsMetadataModelInfo": ".listen_v1results_metadata_model_info", + "ListenV1ResultsType": ".listen_v1results_type", "ListenV1SpeechStarted": ".listen_v1speech_started", + "ListenV1SpeechStartedType": ".listen_v1speech_started_type", "ListenV1UtteranceEnd": ".listen_v1utterance_end", + "ListenV1UtteranceEndType": ".listen_v1utterance_end_type", } @@ -71,6 +79,7 @@ def __dir__(): "ListenV1KeepAlive", "ListenV1KeepAliveType", "ListenV1Metadata", + "ListenV1MetadataType", "ListenV1Results", "ListenV1ResultsChannel", "ListenV1ResultsChannelAlternativesItem", @@ -78,6 +87,9 @@ def __dir__(): "ListenV1ResultsEntitiesItem", "ListenV1ResultsMetadata", "ListenV1ResultsMetadataModelInfo", + "ListenV1ResultsType", "ListenV1SpeechStarted", + "ListenV1SpeechStartedType", "ListenV1UtteranceEnd", + "ListenV1UtteranceEndType", ] diff --git a/src/deepgram/listen/v1/types/listen_v1metadata.py b/src/deepgram/listen/v1/types/listen_v1metadata.py index 5a048330..975bc63f 100644 --- a/src/deepgram/listen/v1/types/listen_v1metadata.py +++ b/src/deepgram/listen/v1/types/listen_v1metadata.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .listen_v1metadata_type import ListenV1MetadataType class ListenV1Metadata(UncheckedBaseModel): - type: typing.Literal["Metadata"] = pydantic.Field(default="Metadata") + type: ListenV1MetadataType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v1/types/listen_v1metadata_type.py b/src/deepgram/listen/v1/types/listen_v1metadata_type.py new file mode 100644 index 00000000..c6e15ad0 --- /dev/null +++ b/src/deepgram/listen/v1/types/listen_v1metadata_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV1MetadataType = typing.Union[typing.Literal["Metadata"], typing.Any] diff --git a/src/deepgram/listen/v1/types/listen_v1results.py b/src/deepgram/listen/v1/types/listen_v1results.py index 2ac0f05c..db98dec5 100644 --- a/src/deepgram/listen/v1/types/listen_v1results.py +++ b/src/deepgram/listen/v1/types/listen_v1results.py @@ -8,10 +8,11 @@ from .listen_v1results_channel import ListenV1ResultsChannel from .listen_v1results_entities_item import ListenV1ResultsEntitiesItem from .listen_v1results_metadata import ListenV1ResultsMetadata +from .listen_v1results_type import ListenV1ResultsType class ListenV1Results(UncheckedBaseModel): - type: typing.Literal["Results"] = pydantic.Field(default="Results") + type: ListenV1ResultsType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v1/types/listen_v1results_type.py b/src/deepgram/listen/v1/types/listen_v1results_type.py new file mode 100644 index 00000000..7d93db8b --- /dev/null +++ b/src/deepgram/listen/v1/types/listen_v1results_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV1ResultsType = typing.Union[typing.Literal["Results"], typing.Any] diff --git a/src/deepgram/listen/v1/types/listen_v1speech_started.py b/src/deepgram/listen/v1/types/listen_v1speech_started.py index ff986c28..14115993 100644 --- a/src/deepgram/listen/v1/types/listen_v1speech_started.py +++ b/src/deepgram/listen/v1/types/listen_v1speech_started.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .listen_v1speech_started_type import ListenV1SpeechStartedType class ListenV1SpeechStarted(UncheckedBaseModel): - type: typing.Literal["SpeechStarted"] = pydantic.Field(default="SpeechStarted") + type: ListenV1SpeechStartedType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v1/types/listen_v1speech_started_type.py b/src/deepgram/listen/v1/types/listen_v1speech_started_type.py new file mode 100644 index 00000000..1fc14b5a --- /dev/null +++ b/src/deepgram/listen/v1/types/listen_v1speech_started_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV1SpeechStartedType = typing.Union[typing.Literal["SpeechStarted"], typing.Any] diff --git a/src/deepgram/listen/v1/types/listen_v1utterance_end.py b/src/deepgram/listen/v1/types/listen_v1utterance_end.py index 893468fc..e6488578 100644 --- a/src/deepgram/listen/v1/types/listen_v1utterance_end.py +++ b/src/deepgram/listen/v1/types/listen_v1utterance_end.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .listen_v1utterance_end_type import ListenV1UtteranceEndType class ListenV1UtteranceEnd(UncheckedBaseModel): - type: typing.Literal["UtteranceEnd"] = pydantic.Field(default="UtteranceEnd") + type: ListenV1UtteranceEndType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v1/types/listen_v1utterance_end_type.py b/src/deepgram/listen/v1/types/listen_v1utterance_end_type.py new file mode 100644 index 00000000..a8348503 --- /dev/null +++ b/src/deepgram/listen/v1/types/listen_v1utterance_end_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV1UtteranceEndType = typing.Union[typing.Literal["UtteranceEnd"], typing.Any] diff --git a/src/deepgram/listen/v2/__init__.py b/src/deepgram/listen/v2/__init__.py index db7c724c..8568cbd8 100644 --- a/src/deepgram/listen/v2/__init__.py +++ b/src/deepgram/listen/v2/__init__.py @@ -10,9 +10,12 @@ ListenV2CloseStream, ListenV2CloseStreamType, ListenV2Connected, + ListenV2ConnectedType, ListenV2FatalError, + ListenV2FatalErrorType, ListenV2TurnInfo, ListenV2TurnInfoEvent, + ListenV2TurnInfoType, ListenV2TurnInfoWordsItem, ) from .requests import ( @@ -28,11 +31,14 @@ "ListenV2CloseStreamType": ".types", "ListenV2Connected": ".types", "ListenV2ConnectedParams": ".requests", + "ListenV2ConnectedType": ".types", "ListenV2FatalError": ".types", "ListenV2FatalErrorParams": ".requests", + "ListenV2FatalErrorType": ".types", "ListenV2TurnInfo": ".types", "ListenV2TurnInfoEvent": ".types", "ListenV2TurnInfoParams": ".requests", + "ListenV2TurnInfoType": ".types", "ListenV2TurnInfoWordsItem": ".types", "ListenV2TurnInfoWordsItemParams": ".requests", } @@ -65,11 +71,14 @@ def __dir__(): "ListenV2CloseStreamType", "ListenV2Connected", "ListenV2ConnectedParams", + "ListenV2ConnectedType", "ListenV2FatalError", "ListenV2FatalErrorParams", + "ListenV2FatalErrorType", "ListenV2TurnInfo", "ListenV2TurnInfoEvent", "ListenV2TurnInfoParams", + "ListenV2TurnInfoType", "ListenV2TurnInfoWordsItem", "ListenV2TurnInfoWordsItemParams", ] diff --git a/src/deepgram/listen/v2/client.py b/src/deepgram/listen/v2/client.py index 949f0c75..6b7d03b9 100644 --- a/src/deepgram/listen/v2/client.py +++ b/src/deepgram/listen/v2/client.py @@ -11,7 +11,17 @@ from ...core.query_encoder import encode_query from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions +from ...core.serialization import convert_and_respect_annotation_metadata from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...requests.listen_v2keyterm import ListenV2KeytermParams +from ...types.listen_v2eager_eot_threshold import ListenV2EagerEotThreshold +from ...types.listen_v2encoding import ListenV2Encoding +from ...types.listen_v2eot_threshold import ListenV2EotThreshold +from ...types.listen_v2eot_timeout_ms import ListenV2EotTimeoutMs +from ...types.listen_v2mip_opt_out import ListenV2MipOptOut +from ...types.listen_v2model import ListenV2Model +from ...types.listen_v2sample_rate import ListenV2SampleRate +from ...types.listen_v2tag import ListenV2Tag from .raw_client import AsyncRawV2Client, RawV2Client from .socket_client import AsyncV2SocketClient, V2SocketClient @@ -40,15 +50,15 @@ def with_raw_response(self) -> RawV2Client: def connect( self, *, - model: str, - encoding: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - eager_eot_threshold: typing.Optional[str] = None, - eot_threshold: typing.Optional[str] = None, - eot_timeout_ms: typing.Optional[str] = None, - keyterm: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - mip_opt_out: typing.Optional[str] = None, - tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, + model: ListenV2Model, + encoding: typing.Optional[ListenV2Encoding] = None, + sample_rate: typing.Optional[ListenV2SampleRate] = None, + eager_eot_threshold: typing.Optional[ListenV2EagerEotThreshold] = None, + eot_threshold: typing.Optional[ListenV2EotThreshold] = None, + eot_timeout_ms: typing.Optional[ListenV2EotTimeoutMs] = None, + keyterm: typing.Optional[ListenV2KeytermParams] = None, + mip_opt_out: typing.Optional[ListenV2MipOptOut] = None, + tag: typing.Optional[ListenV2Tag] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V2SocketClient]: @@ -58,30 +68,28 @@ def connect( Parameters ---------- - model : str + model : ListenV2Model - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV2Encoding] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV2SampleRate] - eager_eot_threshold : typing.Optional[str] + eager_eot_threshold : typing.Optional[ListenV2EagerEotThreshold] - eot_threshold : typing.Optional[str] + eot_threshold : typing.Optional[ListenV2EotThreshold] - eot_timeout_ms : typing.Optional[str] + eot_timeout_ms : typing.Optional[ListenV2EotTimeoutMs] - keyterm : typing.Optional[typing.Union[str, typing.Sequence[str]]] - Keyterm prompting can improve recognition of specialized terminology. Pass a single string or a list of strings. + keyterm : typing.Optional[ListenV2KeytermParams] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV2MipOptOut] - tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] - Label your requests for the purpose of identification during usage reporting. Pass a single string or a list of strings. + tag : typing.Optional[ListenV2Tag] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -101,7 +109,13 @@ def connect( "eager_eot_threshold": eager_eot_threshold, "eot_threshold": eot_threshold, "eot_timeout_ms": eot_timeout_ms, - "keyterm": keyterm, + "keyterm": convert_and_respect_annotation_metadata( + object_=convert_and_respect_annotation_metadata( + object_=keyterm, annotation=ListenV2KeytermParams, direction="write" + ), + annotation=ListenV2KeytermParams, + direction="write", + ), "mip_opt_out": mip_opt_out, "tag": tag, **( @@ -157,15 +171,15 @@ def with_raw_response(self) -> AsyncRawV2Client: async def connect( self, *, - model: str, - encoding: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - eager_eot_threshold: typing.Optional[str] = None, - eot_threshold: typing.Optional[str] = None, - eot_timeout_ms: typing.Optional[str] = None, - keyterm: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, - mip_opt_out: typing.Optional[str] = None, - tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, + model: ListenV2Model, + encoding: typing.Optional[ListenV2Encoding] = None, + sample_rate: typing.Optional[ListenV2SampleRate] = None, + eager_eot_threshold: typing.Optional[ListenV2EagerEotThreshold] = None, + eot_threshold: typing.Optional[ListenV2EotThreshold] = None, + eot_timeout_ms: typing.Optional[ListenV2EotTimeoutMs] = None, + keyterm: typing.Optional[ListenV2KeytermParams] = None, + mip_opt_out: typing.Optional[ListenV2MipOptOut] = None, + tag: typing.Optional[ListenV2Tag] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV2SocketClient]: @@ -175,30 +189,28 @@ async def connect( Parameters ---------- - model : str + model : ListenV2Model - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV2Encoding] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV2SampleRate] - eager_eot_threshold : typing.Optional[str] + eager_eot_threshold : typing.Optional[ListenV2EagerEotThreshold] - eot_threshold : typing.Optional[str] + eot_threshold : typing.Optional[ListenV2EotThreshold] - eot_timeout_ms : typing.Optional[str] + eot_timeout_ms : typing.Optional[ListenV2EotTimeoutMs] - keyterm : typing.Optional[typing.Union[str, typing.Sequence[str]]] - Keyterm prompting can improve recognition of specialized terminology. Pass a single string or a list of strings. + keyterm : typing.Optional[ListenV2KeytermParams] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV2MipOptOut] - tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] - Label your requests for the purpose of identification during usage reporting. Pass a single string or a list of strings. + tag : typing.Optional[ListenV2Tag] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -218,7 +230,13 @@ async def connect( "eager_eot_threshold": eager_eot_threshold, "eot_threshold": eot_threshold, "eot_timeout_ms": eot_timeout_ms, - "keyterm": keyterm, + "keyterm": convert_and_respect_annotation_metadata( + object_=convert_and_respect_annotation_metadata( + object_=keyterm, annotation=ListenV2KeytermParams, direction="write" + ), + annotation=ListenV2KeytermParams, + direction="write", + ), "mip_opt_out": mip_opt_out, "tag": tag, **( diff --git a/src/deepgram/listen/v2/raw_client.py b/src/deepgram/listen/v2/raw_client.py index 5b3819c9..53d2a13a 100644 --- a/src/deepgram/listen/v2/raw_client.py +++ b/src/deepgram/listen/v2/raw_client.py @@ -11,7 +11,17 @@ from ...core.query_encoder import encode_query from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions +from ...core.serialization import convert_and_respect_annotation_metadata from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...requests.listen_v2keyterm import ListenV2KeytermParams +from ...types.listen_v2eager_eot_threshold import ListenV2EagerEotThreshold +from ...types.listen_v2encoding import ListenV2Encoding +from ...types.listen_v2eot_threshold import ListenV2EotThreshold +from ...types.listen_v2eot_timeout_ms import ListenV2EotTimeoutMs +from ...types.listen_v2mip_opt_out import ListenV2MipOptOut +from ...types.listen_v2model import ListenV2Model +from ...types.listen_v2sample_rate import ListenV2SampleRate +from ...types.listen_v2tag import ListenV2Tag from .socket_client import AsyncV2SocketClient, V2SocketClient try: @@ -28,15 +38,15 @@ def __init__(self, *, client_wrapper: SyncClientWrapper): def connect( self, *, - model: str, - encoding: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - eager_eot_threshold: typing.Optional[str] = None, - eot_threshold: typing.Optional[str] = None, - eot_timeout_ms: typing.Optional[str] = None, - keyterm: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - tag: typing.Optional[str] = None, + model: ListenV2Model, + encoding: typing.Optional[ListenV2Encoding] = None, + sample_rate: typing.Optional[ListenV2SampleRate] = None, + eager_eot_threshold: typing.Optional[ListenV2EagerEotThreshold] = None, + eot_threshold: typing.Optional[ListenV2EotThreshold] = None, + eot_timeout_ms: typing.Optional[ListenV2EotTimeoutMs] = None, + keyterm: typing.Optional[ListenV2KeytermParams] = None, + mip_opt_out: typing.Optional[ListenV2MipOptOut] = None, + tag: typing.Optional[ListenV2Tag] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V2SocketClient]: @@ -46,28 +56,28 @@ def connect( Parameters ---------- - model : str + model : ListenV2Model - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV2Encoding] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV2SampleRate] - eager_eot_threshold : typing.Optional[str] + eager_eot_threshold : typing.Optional[ListenV2EagerEotThreshold] - eot_threshold : typing.Optional[str] + eot_threshold : typing.Optional[ListenV2EotThreshold] - eot_timeout_ms : typing.Optional[str] + eot_timeout_ms : typing.Optional[ListenV2EotTimeoutMs] - keyterm : typing.Optional[str] + keyterm : typing.Optional[ListenV2KeytermParams] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV2MipOptOut] - tag : typing.Optional[str] + tag : typing.Optional[ListenV2Tag] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -87,7 +97,13 @@ def connect( "eager_eot_threshold": eager_eot_threshold, "eot_threshold": eot_threshold, "eot_timeout_ms": eot_timeout_ms, - "keyterm": keyterm, + "keyterm": convert_and_respect_annotation_metadata( + object_=convert_and_respect_annotation_metadata( + object_=keyterm, annotation=ListenV2KeytermParams, direction="write" + ), + annotation=ListenV2KeytermParams, + direction="write", + ), "mip_opt_out": mip_opt_out, "tag": tag, **( @@ -132,15 +148,15 @@ def __init__(self, *, client_wrapper: AsyncClientWrapper): async def connect( self, *, - model: str, - encoding: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, - eager_eot_threshold: typing.Optional[str] = None, - eot_threshold: typing.Optional[str] = None, - eot_timeout_ms: typing.Optional[str] = None, - keyterm: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - tag: typing.Optional[str] = None, + model: ListenV2Model, + encoding: typing.Optional[ListenV2Encoding] = None, + sample_rate: typing.Optional[ListenV2SampleRate] = None, + eager_eot_threshold: typing.Optional[ListenV2EagerEotThreshold] = None, + eot_threshold: typing.Optional[ListenV2EotThreshold] = None, + eot_timeout_ms: typing.Optional[ListenV2EotTimeoutMs] = None, + keyterm: typing.Optional[ListenV2KeytermParams] = None, + mip_opt_out: typing.Optional[ListenV2MipOptOut] = None, + tag: typing.Optional[ListenV2Tag] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV2SocketClient]: @@ -150,28 +166,28 @@ async def connect( Parameters ---------- - model : str + model : ListenV2Model - encoding : typing.Optional[str] + encoding : typing.Optional[ListenV2Encoding] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[ListenV2SampleRate] - eager_eot_threshold : typing.Optional[str] + eager_eot_threshold : typing.Optional[ListenV2EagerEotThreshold] - eot_threshold : typing.Optional[str] + eot_threshold : typing.Optional[ListenV2EotThreshold] - eot_timeout_ms : typing.Optional[str] + eot_timeout_ms : typing.Optional[ListenV2EotTimeoutMs] - keyterm : typing.Optional[str] + keyterm : typing.Optional[ListenV2KeytermParams] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[ListenV2MipOptOut] - tag : typing.Optional[str] + tag : typing.Optional[ListenV2Tag] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -191,7 +207,13 @@ async def connect( "eager_eot_threshold": eager_eot_threshold, "eot_threshold": eot_threshold, "eot_timeout_ms": eot_timeout_ms, - "keyterm": keyterm, + "keyterm": convert_and_respect_annotation_metadata( + object_=convert_and_respect_annotation_metadata( + object_=keyterm, annotation=ListenV2KeytermParams, direction="write" + ), + annotation=ListenV2KeytermParams, + direction="write", + ), "mip_opt_out": mip_opt_out, "tag": tag, **( diff --git a/src/deepgram/listen/v2/requests/listen_v2connected.py b/src/deepgram/listen/v2/requests/listen_v2connected.py index c931eec2..97955889 100644 --- a/src/deepgram/listen/v2/requests/listen_v2connected.py +++ b/src/deepgram/listen/v2/requests/listen_v2connected.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.listen_v2connected_type import ListenV2ConnectedType class ListenV2ConnectedParams(typing_extensions.TypedDict): - type: typing.Literal["Connected"] + type: ListenV2ConnectedType """ Message type identifier """ diff --git a/src/deepgram/listen/v2/requests/listen_v2fatal_error.py b/src/deepgram/listen/v2/requests/listen_v2fatal_error.py index 05cb3041..c124f959 100644 --- a/src/deepgram/listen/v2/requests/listen_v2fatal_error.py +++ b/src/deepgram/listen/v2/requests/listen_v2fatal_error.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.listen_v2fatal_error_type import ListenV2FatalErrorType class ListenV2FatalErrorParams(typing_extensions.TypedDict): - type: typing.Literal["Error"] + type: ListenV2FatalErrorType """ Message type identifier """ diff --git a/src/deepgram/listen/v2/requests/listen_v2turn_info.py b/src/deepgram/listen/v2/requests/listen_v2turn_info.py index d1a15fec..97cd3eda 100644 --- a/src/deepgram/listen/v2/requests/listen_v2turn_info.py +++ b/src/deepgram/listen/v2/requests/listen_v2turn_info.py @@ -4,6 +4,7 @@ import typing_extensions from ..types.listen_v2turn_info_event import ListenV2TurnInfoEvent +from ..types.listen_v2turn_info_type import ListenV2TurnInfoType from .listen_v2turn_info_words_item import ListenV2TurnInfoWordsItemParams @@ -12,7 +13,7 @@ class ListenV2TurnInfoParams(typing_extensions.TypedDict): Describes the current turn and latest state of the turn """ - type: typing.Literal["TurnInfo"] + type: ListenV2TurnInfoType request_id: str """ The unique identifier of the request diff --git a/src/deepgram/listen/v2/socket_client.py b/src/deepgram/listen/v2/socket_client.py index 136f7e74..0811393c 100644 --- a/src/deepgram/listen/v2/socket_client.py +++ b/src/deepgram/listen/v2/socket_client.py @@ -1,6 +1,7 @@ # This file was auto-generated by Fern from our API Definition. import json +import logging import typing from json.decoder import JSONDecodeError @@ -18,6 +19,7 @@ except ImportError: from websockets import WebSocketClientProtocol # type: ignore +_logger = logging.getLogger(__name__) V2SocketClientResponse = typing.Union[ListenV2Connected, ListenV2TurnInfo, ListenV2FatalError] @@ -31,7 +33,13 @@ async def __aiter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V2SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V2SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue async def start_listening(self): """ @@ -50,7 +58,13 @@ async def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue await self._emit_async(EventType.MESSAGE, parsed) except Exception as exc: await self._emit_async(EventType.ERROR, exc) @@ -64,12 +78,12 @@ async def send_media(self, message: bytes) -> None: """ await self._send(message) - async def send_close_stream(self, message: typing.Optional[ListenV2CloseStream] = None) -> None: + async def send_close_stream(self, message: ListenV2CloseStream) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV2CloseStream. """ - await self._send_model(message or ListenV2CloseStream(type="CloseStream")) + await self._send_model(message) async def recv(self) -> V2SocketClientResponse: """ @@ -79,7 +93,11 @@ async def recv(self) -> V2SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore async def _send(self, data: typing.Any) -> None: """ @@ -106,7 +124,13 @@ def __iter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V2SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V2SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue def start_listening(self): """ @@ -125,7 +149,13 @@ def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue self._emit(EventType.MESSAGE, parsed) except Exception as exc: self._emit(EventType.ERROR, exc) @@ -139,12 +169,12 @@ def send_media(self, message: bytes) -> None: """ self._send(message) - def send_close_stream(self, message: typing.Optional[ListenV2CloseStream] = None) -> None: + def send_close_stream(self, message: ListenV2CloseStream) -> None: """ Send a message to the websocket connection. The message will be sent as a ListenV2CloseStream. """ - self._send_model(message or ListenV2CloseStream(type="CloseStream")) + self._send_model(message) def recv(self) -> V2SocketClientResponse: """ @@ -154,7 +184,11 @@ def recv(self) -> V2SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V2SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore def _send(self, data: typing.Any) -> None: """ diff --git a/src/deepgram/listen/v2/types/__init__.py b/src/deepgram/listen/v2/types/__init__.py index 229417bf..e30397cc 100644 --- a/src/deepgram/listen/v2/types/__init__.py +++ b/src/deepgram/listen/v2/types/__init__.py @@ -9,17 +9,23 @@ from .listen_v2close_stream import ListenV2CloseStream from .listen_v2close_stream_type import ListenV2CloseStreamType from .listen_v2connected import ListenV2Connected + from .listen_v2connected_type import ListenV2ConnectedType from .listen_v2fatal_error import ListenV2FatalError + from .listen_v2fatal_error_type import ListenV2FatalErrorType from .listen_v2turn_info import ListenV2TurnInfo from .listen_v2turn_info_event import ListenV2TurnInfoEvent + from .listen_v2turn_info_type import ListenV2TurnInfoType from .listen_v2turn_info_words_item import ListenV2TurnInfoWordsItem _dynamic_imports: typing.Dict[str, str] = { "ListenV2CloseStream": ".listen_v2close_stream", "ListenV2CloseStreamType": ".listen_v2close_stream_type", "ListenV2Connected": ".listen_v2connected", + "ListenV2ConnectedType": ".listen_v2connected_type", "ListenV2FatalError": ".listen_v2fatal_error", + "ListenV2FatalErrorType": ".listen_v2fatal_error_type", "ListenV2TurnInfo": ".listen_v2turn_info", "ListenV2TurnInfoEvent": ".listen_v2turn_info_event", + "ListenV2TurnInfoType": ".listen_v2turn_info_type", "ListenV2TurnInfoWordsItem": ".listen_v2turn_info_words_item", } @@ -49,8 +55,11 @@ def __dir__(): "ListenV2CloseStream", "ListenV2CloseStreamType", "ListenV2Connected", + "ListenV2ConnectedType", "ListenV2FatalError", + "ListenV2FatalErrorType", "ListenV2TurnInfo", "ListenV2TurnInfoEvent", + "ListenV2TurnInfoType", "ListenV2TurnInfoWordsItem", ] diff --git a/src/deepgram/listen/v2/types/listen_v2connected.py b/src/deepgram/listen/v2/types/listen_v2connected.py index b174be68..dc60c365 100644 --- a/src/deepgram/listen/v2/types/listen_v2connected.py +++ b/src/deepgram/listen/v2/types/listen_v2connected.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .listen_v2connected_type import ListenV2ConnectedType class ListenV2Connected(UncheckedBaseModel): - type: typing.Literal["Connected"] = pydantic.Field(default="Connected") + type: ListenV2ConnectedType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v2/types/listen_v2connected_type.py b/src/deepgram/listen/v2/types/listen_v2connected_type.py new file mode 100644 index 00000000..f2c23e1f --- /dev/null +++ b/src/deepgram/listen/v2/types/listen_v2connected_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV2ConnectedType = typing.Union[typing.Literal["Connected"], typing.Any] diff --git a/src/deepgram/listen/v2/types/listen_v2fatal_error.py b/src/deepgram/listen/v2/types/listen_v2fatal_error.py index 0e65c00e..92b58581 100644 --- a/src/deepgram/listen/v2/types/listen_v2fatal_error.py +++ b/src/deepgram/listen/v2/types/listen_v2fatal_error.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .listen_v2fatal_error_type import ListenV2FatalErrorType class ListenV2FatalError(UncheckedBaseModel): - type: typing.Literal["Error"] = pydantic.Field(default="Error") + type: ListenV2FatalErrorType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/listen/v2/types/listen_v2fatal_error_type.py b/src/deepgram/listen/v2/types/listen_v2fatal_error_type.py new file mode 100644 index 00000000..bb3395f9 --- /dev/null +++ b/src/deepgram/listen/v2/types/listen_v2fatal_error_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV2FatalErrorType = typing.Union[typing.Literal["Error"], typing.Any] diff --git a/src/deepgram/listen/v2/types/listen_v2turn_info.py b/src/deepgram/listen/v2/types/listen_v2turn_info.py index 7ba2770c..c14590a3 100644 --- a/src/deepgram/listen/v2/types/listen_v2turn_info.py +++ b/src/deepgram/listen/v2/types/listen_v2turn_info.py @@ -6,6 +6,7 @@ from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel from .listen_v2turn_info_event import ListenV2TurnInfoEvent +from .listen_v2turn_info_type import ListenV2TurnInfoType from .listen_v2turn_info_words_item import ListenV2TurnInfoWordsItem @@ -14,7 +15,7 @@ class ListenV2TurnInfo(UncheckedBaseModel): Describes the current turn and latest state of the turn """ - type: typing.Literal["TurnInfo"] = "TurnInfo" + type: ListenV2TurnInfoType request_id: str = pydantic.Field() """ The unique identifier of the request diff --git a/src/deepgram/listen/v2/types/listen_v2turn_info_type.py b/src/deepgram/listen/v2/types/listen_v2turn_info_type.py new file mode 100644 index 00000000..5ec2c86a --- /dev/null +++ b/src/deepgram/listen/v2/types/listen_v2turn_info_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +ListenV2TurnInfoType = typing.Union[typing.Literal["TurnInfo"], typing.Any] diff --git a/src/deepgram/read/v1/__init__.py b/src/deepgram/read/v1/__init__.py index 6e2939a4..56569a29 100644 --- a/src/deepgram/read/v1/__init__.py +++ b/src/deepgram/read/v1/__init__.py @@ -11,13 +11,13 @@ TextAnalyzeRequestCallbackMethod, TextAnalyzeRequestCustomIntentMode, TextAnalyzeRequestCustomTopicMode, - TextAnalyzeRequestSummarize, + TextAnalyzeRequestSummarizeZero, ) _dynamic_imports: typing.Dict[str, str] = { "TextAnalyzeRequestCallbackMethod": ".text", "TextAnalyzeRequestCustomIntentMode": ".text", "TextAnalyzeRequestCustomTopicMode": ".text", - "TextAnalyzeRequestSummarize": ".text", + "TextAnalyzeRequestSummarizeZero": ".text", "text": ".text", } @@ -47,6 +47,6 @@ def __dir__(): "TextAnalyzeRequestCallbackMethod", "TextAnalyzeRequestCustomIntentMode", "TextAnalyzeRequestCustomTopicMode", - "TextAnalyzeRequestSummarize", + "TextAnalyzeRequestSummarizeZero", "text", ] diff --git a/src/deepgram/read/v1/text/__init__.py b/src/deepgram/read/v1/text/__init__.py index 069d04c2..839a83c9 100644 --- a/src/deepgram/read/v1/text/__init__.py +++ b/src/deepgram/read/v1/text/__init__.py @@ -10,13 +10,13 @@ TextAnalyzeRequestCallbackMethod, TextAnalyzeRequestCustomIntentMode, TextAnalyzeRequestCustomTopicMode, - TextAnalyzeRequestSummarize, + TextAnalyzeRequestSummarizeZero, ) _dynamic_imports: typing.Dict[str, str] = { "TextAnalyzeRequestCallbackMethod": ".types", "TextAnalyzeRequestCustomIntentMode": ".types", "TextAnalyzeRequestCustomTopicMode": ".types", - "TextAnalyzeRequestSummarize": ".types", + "TextAnalyzeRequestSummarizeZero": ".types", } @@ -45,5 +45,5 @@ def __dir__(): "TextAnalyzeRequestCallbackMethod", "TextAnalyzeRequestCustomIntentMode", "TextAnalyzeRequestCustomTopicMode", - "TextAnalyzeRequestSummarize", + "TextAnalyzeRequestSummarizeZero", ] diff --git a/src/deepgram/read/v1/text/client.py b/src/deepgram/read/v1/text/client.py index d04b8a7c..44f607f4 100644 --- a/src/deepgram/read/v1/text/client.py +++ b/src/deepgram/read/v1/text/client.py @@ -10,7 +10,7 @@ from .types.text_analyze_request_callback_method import TextAnalyzeRequestCallbackMethod from .types.text_analyze_request_custom_intent_mode import TextAnalyzeRequestCustomIntentMode from .types.text_analyze_request_custom_topic_mode import TextAnalyzeRequestCustomTopicMode -from .types.text_analyze_request_summarize import TextAnalyzeRequestSummarize +from .types.text_analyze_request_summarize_zero import TextAnalyzeRequestSummarizeZero # this is used as the default value for optional parameters OMIT = typing.cast(typing.Any, ...) @@ -38,7 +38,7 @@ def analyze( callback: typing.Optional[str] = None, callback_method: typing.Optional[TextAnalyzeRequestCallbackMethod] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[TextAnalyzeRequestSummarize] = None, + summarize: typing.Optional[TextAnalyzeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -65,7 +65,7 @@ def analyze( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[TextAnalyzeRequestSummarize] + summarize : typing.Optional[TextAnalyzeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -164,7 +164,7 @@ async def analyze( callback: typing.Optional[str] = None, callback_method: typing.Optional[TextAnalyzeRequestCallbackMethod] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[TextAnalyzeRequestSummarize] = None, + summarize: typing.Optional[TextAnalyzeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -191,7 +191,7 @@ async def analyze( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[TextAnalyzeRequestSummarize] + summarize : typing.Optional[TextAnalyzeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] diff --git a/src/deepgram/read/v1/text/raw_client.py b/src/deepgram/read/v1/text/raw_client.py index 65b7ebca..66defd3c 100644 --- a/src/deepgram/read/v1/text/raw_client.py +++ b/src/deepgram/read/v1/text/raw_client.py @@ -15,7 +15,7 @@ from .types.text_analyze_request_callback_method import TextAnalyzeRequestCallbackMethod from .types.text_analyze_request_custom_intent_mode import TextAnalyzeRequestCustomIntentMode from .types.text_analyze_request_custom_topic_mode import TextAnalyzeRequestCustomTopicMode -from .types.text_analyze_request_summarize import TextAnalyzeRequestSummarize +from .types.text_analyze_request_summarize_zero import TextAnalyzeRequestSummarizeZero # this is used as the default value for optional parameters OMIT = typing.cast(typing.Any, ...) @@ -32,7 +32,7 @@ def analyze( callback: typing.Optional[str] = None, callback_method: typing.Optional[TextAnalyzeRequestCallbackMethod] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[TextAnalyzeRequestSummarize] = None, + summarize: typing.Optional[TextAnalyzeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -59,7 +59,7 @@ def analyze( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[TextAnalyzeRequestSummarize] + summarize : typing.Optional[TextAnalyzeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] @@ -159,7 +159,7 @@ async def analyze( callback: typing.Optional[str] = None, callback_method: typing.Optional[TextAnalyzeRequestCallbackMethod] = None, sentiment: typing.Optional[bool] = None, - summarize: typing.Optional[TextAnalyzeRequestSummarize] = None, + summarize: typing.Optional[TextAnalyzeRequestSummarizeZero] = None, tag: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, topics: typing.Optional[bool] = None, custom_topic: typing.Optional[typing.Union[str, typing.Sequence[str]]] = None, @@ -186,7 +186,7 @@ async def analyze( sentiment : typing.Optional[bool] Recognizes the sentiment throughout a transcript or text - summarize : typing.Optional[TextAnalyzeRequestSummarize] + summarize : typing.Optional[TextAnalyzeRequestSummarizeZero] Summarize content. For Listen API, supports string version option. For Read API, accepts boolean only. tag : typing.Optional[typing.Union[str, typing.Sequence[str]]] diff --git a/src/deepgram/read/v1/text/types/__init__.py b/src/deepgram/read/v1/text/types/__init__.py index cb955d49..f4086661 100644 --- a/src/deepgram/read/v1/text/types/__init__.py +++ b/src/deepgram/read/v1/text/types/__init__.py @@ -9,12 +9,12 @@ from .text_analyze_request_callback_method import TextAnalyzeRequestCallbackMethod from .text_analyze_request_custom_intent_mode import TextAnalyzeRequestCustomIntentMode from .text_analyze_request_custom_topic_mode import TextAnalyzeRequestCustomTopicMode - from .text_analyze_request_summarize import TextAnalyzeRequestSummarize + from .text_analyze_request_summarize_zero import TextAnalyzeRequestSummarizeZero _dynamic_imports: typing.Dict[str, str] = { "TextAnalyzeRequestCallbackMethod": ".text_analyze_request_callback_method", "TextAnalyzeRequestCustomIntentMode": ".text_analyze_request_custom_intent_mode", "TextAnalyzeRequestCustomTopicMode": ".text_analyze_request_custom_topic_mode", - "TextAnalyzeRequestSummarize": ".text_analyze_request_summarize", + "TextAnalyzeRequestSummarizeZero": ".text_analyze_request_summarize_zero", } @@ -43,5 +43,5 @@ def __dir__(): "TextAnalyzeRequestCallbackMethod", "TextAnalyzeRequestCustomIntentMode", "TextAnalyzeRequestCustomTopicMode", - "TextAnalyzeRequestSummarize", + "TextAnalyzeRequestSummarizeZero", ] diff --git a/src/deepgram/listen/v1/media/types/media_transcribe_request_summarize.py b/src/deepgram/read/v1/text/types/text_analyze_request_summarize_zero.py similarity index 60% rename from src/deepgram/listen/v1/media/types/media_transcribe_request_summarize.py rename to src/deepgram/read/v1/text/types/text_analyze_request_summarize_zero.py index cc44a279..4552004e 100644 --- a/src/deepgram/listen/v1/media/types/media_transcribe_request_summarize.py +++ b/src/deepgram/read/v1/text/types/text_analyze_request_summarize_zero.py @@ -2,4 +2,4 @@ import typing -MediaTranscribeRequestSummarize = typing.Union[typing.Literal["v2"], typing.Any] +TextAnalyzeRequestSummarizeZero = typing.Union[typing.Literal["v2"], typing.Any] diff --git a/src/deepgram/requests/__init__.py b/src/deepgram/requests/__init__.py index 5a3a81b7..8406dfe5 100644 --- a/src/deepgram/requests/__init__.py +++ b/src/deepgram/requests/__init__.py @@ -13,17 +13,10 @@ from .agent_think_models_v1response_models_item_three import AgentThinkModelsV1ResponseModelsItemThreeParams from .agent_think_models_v1response_models_item_two import AgentThinkModelsV1ResponseModelsItemTwoParams from .agent_think_models_v1response_models_item_zero import AgentThinkModelsV1ResponseModelsItemZeroParams - from .anthropic import AnthropicParams - from .aws_bedrock_think_provider import AwsBedrockThinkProviderParams - from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentialsParams - from .aws_polly_speak_provider import AwsPollySpeakProviderParams - from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentialsParams from .billing_breakdown_v1response import BillingBreakdownV1ResponseParams from .billing_breakdown_v1response_resolution import BillingBreakdownV1ResponseResolutionParams from .billing_breakdown_v1response_results_item import BillingBreakdownV1ResponseResultsItemParams from .billing_breakdown_v1response_results_item_grouping import BillingBreakdownV1ResponseResultsItemGroupingParams - from .cartesia import CartesiaParams - from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoiceParams from .create_key_v1response import CreateKeyV1ResponseParams from .create_project_distribution_credentials_v1response import CreateProjectDistributionCredentialsV1ResponseParams from .create_project_distribution_credentials_v1response_distribution_credentials import ( @@ -33,12 +26,10 @@ CreateProjectDistributionCredentialsV1ResponseMemberParams, ) from .create_project_invite_v1response import CreateProjectInviteV1ResponseParams - from .deepgram import DeepgramParams from .delete_project_invite_v1response import DeleteProjectInviteV1ResponseParams from .delete_project_key_v1response import DeleteProjectKeyV1ResponseParams from .delete_project_member_v1response import DeleteProjectMemberV1ResponseParams from .delete_project_v1response import DeleteProjectV1ResponseParams - from .eleven_labs_speak_provider import ElevenLabsSpeakProviderParams from .error_response import ErrorResponseParams from .error_response_legacy_error import ErrorResponseLegacyErrorParams from .error_response_modern_error import ErrorResponseModernErrorParams @@ -60,9 +51,7 @@ from .get_project_key_v1response_item_member_api_key import GetProjectKeyV1ResponseItemMemberApiKeyParams from .get_project_request_v1response import GetProjectRequestV1ResponseParams from .get_project_v1response import GetProjectV1ResponseParams - from .google import GoogleParams from .grant_v1response import GrantV1ResponseParams - from .groq import GroqParams from .leave_project_v1response import LeaveProjectV1ResponseParams from .list_billing_fields_v1response import ListBillingFieldsV1ResponseParams from .list_models_v1response import ListModelsV1ResponseParams @@ -140,8 +129,6 @@ ListenV1ResponseResultsUtterancesItemWordsItemParams, ) from .listen_v2keyterm import ListenV2KeytermParams - from .open_ai_speak_provider import OpenAiSpeakProviderParams - from .open_ai_think_provider import OpenAiThinkProviderParams from .project_request_response import ProjectRequestResponseParams from .read_v1request import ReadV1RequestParams from .read_v1request_text import ReadV1RequestTextParams @@ -174,29 +161,7 @@ from .shared_topics_results_topics_segments_item_topics_item import ( SharedTopicsResultsTopicsSegmentsItemTopicsItemParams, ) - from .speak_settings_v1 import SpeakSettingsV1Params - from .speak_settings_v1endpoint import SpeakSettingsV1EndpointParams - from .speak_settings_v1provider import ( - SpeakSettingsV1ProviderParams, - SpeakSettingsV1Provider_AwsPollyParams, - SpeakSettingsV1Provider_CartesiaParams, - SpeakSettingsV1Provider_DeepgramParams, - SpeakSettingsV1Provider_ElevenLabsParams, - SpeakSettingsV1Provider_OpenAiParams, - ) from .think_settings_v1 import ThinkSettingsV1Params - from .think_settings_v1context_length import ThinkSettingsV1ContextLengthParams - from .think_settings_v1endpoint import ThinkSettingsV1EndpointParams - from .think_settings_v1functions_item import ThinkSettingsV1FunctionsItemParams - from .think_settings_v1functions_item_endpoint import ThinkSettingsV1FunctionsItemEndpointParams - from .think_settings_v1provider import ( - ThinkSettingsV1ProviderParams, - ThinkSettingsV1Provider_AnthropicParams, - ThinkSettingsV1Provider_AwsBedrockParams, - ThinkSettingsV1Provider_GoogleParams, - ThinkSettingsV1Provider_GroqParams, - ThinkSettingsV1Provider_OpenAiParams, - ) from .update_project_member_scopes_v1response import UpdateProjectMemberScopesV1ResponseParams from .update_project_v1response import UpdateProjectV1ResponseParams from .usage_breakdown_v1response import UsageBreakdownV1ResponseParams @@ -215,28 +180,19 @@ "AgentThinkModelsV1ResponseModelsItemTwoParams": ".agent_think_models_v1response_models_item_two", "AgentThinkModelsV1ResponseModelsItemZeroParams": ".agent_think_models_v1response_models_item_zero", "AgentThinkModelsV1ResponseParams": ".agent_think_models_v1response", - "AnthropicParams": ".anthropic", - "AwsBedrockThinkProviderCredentialsParams": ".aws_bedrock_think_provider_credentials", - "AwsBedrockThinkProviderParams": ".aws_bedrock_think_provider", - "AwsPollySpeakProviderCredentialsParams": ".aws_polly_speak_provider_credentials", - "AwsPollySpeakProviderParams": ".aws_polly_speak_provider", "BillingBreakdownV1ResponseParams": ".billing_breakdown_v1response", "BillingBreakdownV1ResponseResolutionParams": ".billing_breakdown_v1response_resolution", "BillingBreakdownV1ResponseResultsItemGroupingParams": ".billing_breakdown_v1response_results_item_grouping", "BillingBreakdownV1ResponseResultsItemParams": ".billing_breakdown_v1response_results_item", - "CartesiaParams": ".cartesia", - "CartesiaSpeakProviderVoiceParams": ".cartesia_speak_provider_voice", "CreateKeyV1ResponseParams": ".create_key_v1response", "CreateProjectDistributionCredentialsV1ResponseDistributionCredentialsParams": ".create_project_distribution_credentials_v1response_distribution_credentials", "CreateProjectDistributionCredentialsV1ResponseMemberParams": ".create_project_distribution_credentials_v1response_member", "CreateProjectDistributionCredentialsV1ResponseParams": ".create_project_distribution_credentials_v1response", "CreateProjectInviteV1ResponseParams": ".create_project_invite_v1response", - "DeepgramParams": ".deepgram", "DeleteProjectInviteV1ResponseParams": ".delete_project_invite_v1response", "DeleteProjectKeyV1ResponseParams": ".delete_project_key_v1response", "DeleteProjectMemberV1ResponseParams": ".delete_project_member_v1response", "DeleteProjectV1ResponseParams": ".delete_project_v1response", - "ElevenLabsSpeakProviderParams": ".eleven_labs_speak_provider", "ErrorResponseLegacyErrorParams": ".error_response_legacy_error", "ErrorResponseModernErrorParams": ".error_response_modern_error", "ErrorResponseParams": ".error_response", @@ -254,9 +210,7 @@ "GetProjectKeyV1ResponseParams": ".get_project_key_v1response", "GetProjectRequestV1ResponseParams": ".get_project_request_v1response", "GetProjectV1ResponseParams": ".get_project_v1response", - "GoogleParams": ".google", "GrantV1ResponseParams": ".grant_v1response", - "GroqParams": ".groq", "LeaveProjectV1ResponseParams": ".leave_project_v1response", "ListBillingFieldsV1ResponseParams": ".list_billing_fields_v1response", "ListModelsV1ResponseParams": ".list_models_v1response", @@ -308,8 +262,6 @@ "ListenV1ResponseResultsUtterancesItemWordsItemParams": ".listen_v1response_results_utterances_item_words_item", "ListenV1ResponseResultsUtterancesParams": ".listen_v1response_results_utterances", "ListenV2KeytermParams": ".listen_v2keyterm", - "OpenAiSpeakProviderParams": ".open_ai_speak_provider", - "OpenAiThinkProviderParams": ".open_ai_think_provider", "ProjectRequestResponseParams": ".project_request_response", "ReadV1RequestParams": ".read_v1request", "ReadV1RequestTextParams": ".read_v1request_text", @@ -338,25 +290,7 @@ "SharedTopicsResultsTopicsParams": ".shared_topics_results_topics", "SharedTopicsResultsTopicsSegmentsItemParams": ".shared_topics_results_topics_segments_item", "SharedTopicsResultsTopicsSegmentsItemTopicsItemParams": ".shared_topics_results_topics_segments_item_topics_item", - "SpeakSettingsV1EndpointParams": ".speak_settings_v1endpoint", - "SpeakSettingsV1Params": ".speak_settings_v1", - "SpeakSettingsV1ProviderParams": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_AwsPollyParams": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_CartesiaParams": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_DeepgramParams": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_ElevenLabsParams": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_OpenAiParams": ".speak_settings_v1provider", - "ThinkSettingsV1ContextLengthParams": ".think_settings_v1context_length", - "ThinkSettingsV1EndpointParams": ".think_settings_v1endpoint", - "ThinkSettingsV1FunctionsItemEndpointParams": ".think_settings_v1functions_item_endpoint", - "ThinkSettingsV1FunctionsItemParams": ".think_settings_v1functions_item", "ThinkSettingsV1Params": ".think_settings_v1", - "ThinkSettingsV1ProviderParams": ".think_settings_v1provider", - "ThinkSettingsV1Provider_AnthropicParams": ".think_settings_v1provider", - "ThinkSettingsV1Provider_AwsBedrockParams": ".think_settings_v1provider", - "ThinkSettingsV1Provider_GoogleParams": ".think_settings_v1provider", - "ThinkSettingsV1Provider_GroqParams": ".think_settings_v1provider", - "ThinkSettingsV1Provider_OpenAiParams": ".think_settings_v1provider", "UpdateProjectMemberScopesV1ResponseParams": ".update_project_member_scopes_v1response", "UpdateProjectV1ResponseParams": ".update_project_v1response", "UsageBreakdownV1ResponseParams": ".usage_breakdown_v1response", @@ -399,28 +333,19 @@ def __dir__(): "AgentThinkModelsV1ResponseModelsItemTwoParams", "AgentThinkModelsV1ResponseModelsItemZeroParams", "AgentThinkModelsV1ResponseParams", - "AnthropicParams", - "AwsBedrockThinkProviderCredentialsParams", - "AwsBedrockThinkProviderParams", - "AwsPollySpeakProviderCredentialsParams", - "AwsPollySpeakProviderParams", "BillingBreakdownV1ResponseParams", "BillingBreakdownV1ResponseResolutionParams", "BillingBreakdownV1ResponseResultsItemGroupingParams", "BillingBreakdownV1ResponseResultsItemParams", - "CartesiaParams", - "CartesiaSpeakProviderVoiceParams", "CreateKeyV1ResponseParams", "CreateProjectDistributionCredentialsV1ResponseDistributionCredentialsParams", "CreateProjectDistributionCredentialsV1ResponseMemberParams", "CreateProjectDistributionCredentialsV1ResponseParams", "CreateProjectInviteV1ResponseParams", - "DeepgramParams", "DeleteProjectInviteV1ResponseParams", "DeleteProjectKeyV1ResponseParams", "DeleteProjectMemberV1ResponseParams", "DeleteProjectV1ResponseParams", - "ElevenLabsSpeakProviderParams", "ErrorResponseLegacyErrorParams", "ErrorResponseModernErrorParams", "ErrorResponseParams", @@ -438,9 +363,7 @@ def __dir__(): "GetProjectKeyV1ResponseParams", "GetProjectRequestV1ResponseParams", "GetProjectV1ResponseParams", - "GoogleParams", "GrantV1ResponseParams", - "GroqParams", "LeaveProjectV1ResponseParams", "ListBillingFieldsV1ResponseParams", "ListModelsV1ResponseParams", @@ -492,8 +415,6 @@ def __dir__(): "ListenV1ResponseResultsUtterancesItemWordsItemParams", "ListenV1ResponseResultsUtterancesParams", "ListenV2KeytermParams", - "OpenAiSpeakProviderParams", - "OpenAiThinkProviderParams", "ProjectRequestResponseParams", "ReadV1RequestParams", "ReadV1RequestTextParams", @@ -522,25 +443,7 @@ def __dir__(): "SharedTopicsResultsTopicsParams", "SharedTopicsResultsTopicsSegmentsItemParams", "SharedTopicsResultsTopicsSegmentsItemTopicsItemParams", - "SpeakSettingsV1EndpointParams", - "SpeakSettingsV1Params", - "SpeakSettingsV1ProviderParams", - "SpeakSettingsV1Provider_AwsPollyParams", - "SpeakSettingsV1Provider_CartesiaParams", - "SpeakSettingsV1Provider_DeepgramParams", - "SpeakSettingsV1Provider_ElevenLabsParams", - "SpeakSettingsV1Provider_OpenAiParams", - "ThinkSettingsV1ContextLengthParams", - "ThinkSettingsV1EndpointParams", - "ThinkSettingsV1FunctionsItemEndpointParams", - "ThinkSettingsV1FunctionsItemParams", "ThinkSettingsV1Params", - "ThinkSettingsV1ProviderParams", - "ThinkSettingsV1Provider_AnthropicParams", - "ThinkSettingsV1Provider_AwsBedrockParams", - "ThinkSettingsV1Provider_GoogleParams", - "ThinkSettingsV1Provider_GroqParams", - "ThinkSettingsV1Provider_OpenAiParams", "UpdateProjectMemberScopesV1ResponseParams", "UpdateProjectV1ResponseParams", "UsageBreakdownV1ResponseParams", diff --git a/src/deepgram/requests/agent_think_models_v1response_models_item_id.py b/src/deepgram/requests/agent_think_models_v1response_models_item_id.py index 4bfeab9b..b1b4c7da 100644 --- a/src/deepgram/requests/agent_think_models_v1response_models_item_id.py +++ b/src/deepgram/requests/agent_think_models_v1response_models_item_id.py @@ -1,8 +1,7 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_think_models_v1response_models_item_id_provider import AgentThinkModelsV1ResponseModelsItemIdProvider class AgentThinkModelsV1ResponseModelsItemIdParams(typing_extensions.TypedDict): @@ -20,7 +19,7 @@ class AgentThinkModelsV1ResponseModelsItemIdParams(typing_extensions.TypedDict): The display name of the model """ - provider: typing.Literal["aws_bedrock"] + provider: AgentThinkModelsV1ResponseModelsItemIdProvider """ The provider of the model """ diff --git a/src/deepgram/requests/agent_think_models_v1response_models_item_one.py b/src/deepgram/requests/agent_think_models_v1response_models_item_one.py index 88f45618..5246711a 100644 --- a/src/deepgram/requests/agent_think_models_v1response_models_item_one.py +++ b/src/deepgram/requests/agent_think_models_v1response_models_item_one.py @@ -1,9 +1,10 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions from ..types.agent_think_models_v1response_models_item_one_id import AgentThinkModelsV1ResponseModelsItemOneId +from ..types.agent_think_models_v1response_models_item_one_provider import ( + AgentThinkModelsV1ResponseModelsItemOneProvider, +) class AgentThinkModelsV1ResponseModelsItemOneParams(typing_extensions.TypedDict): @@ -21,7 +22,7 @@ class AgentThinkModelsV1ResponseModelsItemOneParams(typing_extensions.TypedDict) The display name of the model """ - provider: typing.Literal["anthropic"] + provider: AgentThinkModelsV1ResponseModelsItemOneProvider """ The provider of the model """ diff --git a/src/deepgram/requests/agent_think_models_v1response_models_item_three.py b/src/deepgram/requests/agent_think_models_v1response_models_item_three.py index cabb968b..4db87165 100644 --- a/src/deepgram/requests/agent_think_models_v1response_models_item_three.py +++ b/src/deepgram/requests/agent_think_models_v1response_models_item_three.py @@ -1,8 +1,10 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.agent_think_models_v1response_models_item_three_id import AgentThinkModelsV1ResponseModelsItemThreeId +from ..types.agent_think_models_v1response_models_item_three_provider import ( + AgentThinkModelsV1ResponseModelsItemThreeProvider, +) class AgentThinkModelsV1ResponseModelsItemThreeParams(typing_extensions.TypedDict): @@ -10,7 +12,7 @@ class AgentThinkModelsV1ResponseModelsItemThreeParams(typing_extensions.TypedDic Groq models """ - id: typing.Literal["openai/gpt-oss-20b"] + id: AgentThinkModelsV1ResponseModelsItemThreeId """ The unique identifier of the Groq model """ @@ -20,7 +22,7 @@ class AgentThinkModelsV1ResponseModelsItemThreeParams(typing_extensions.TypedDic The display name of the model """ - provider: typing.Literal["groq"] + provider: AgentThinkModelsV1ResponseModelsItemThreeProvider """ The provider of the model """ diff --git a/src/deepgram/requests/agent_think_models_v1response_models_item_two.py b/src/deepgram/requests/agent_think_models_v1response_models_item_two.py index 5466ad3d..1ec41562 100644 --- a/src/deepgram/requests/agent_think_models_v1response_models_item_two.py +++ b/src/deepgram/requests/agent_think_models_v1response_models_item_two.py @@ -1,9 +1,10 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions from ..types.agent_think_models_v1response_models_item_two_id import AgentThinkModelsV1ResponseModelsItemTwoId +from ..types.agent_think_models_v1response_models_item_two_provider import ( + AgentThinkModelsV1ResponseModelsItemTwoProvider, +) class AgentThinkModelsV1ResponseModelsItemTwoParams(typing_extensions.TypedDict): @@ -21,7 +22,7 @@ class AgentThinkModelsV1ResponseModelsItemTwoParams(typing_extensions.TypedDict) The display name of the model """ - provider: typing.Literal["google"] + provider: AgentThinkModelsV1ResponseModelsItemTwoProvider """ The provider of the model """ diff --git a/src/deepgram/requests/agent_think_models_v1response_models_item_zero.py b/src/deepgram/requests/agent_think_models_v1response_models_item_zero.py index 5b027b7c..e0a6c466 100644 --- a/src/deepgram/requests/agent_think_models_v1response_models_item_zero.py +++ b/src/deepgram/requests/agent_think_models_v1response_models_item_zero.py @@ -1,9 +1,10 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions from ..types.agent_think_models_v1response_models_item_zero_id import AgentThinkModelsV1ResponseModelsItemZeroId +from ..types.agent_think_models_v1response_models_item_zero_provider import ( + AgentThinkModelsV1ResponseModelsItemZeroProvider, +) class AgentThinkModelsV1ResponseModelsItemZeroParams(typing_extensions.TypedDict): @@ -21,7 +22,7 @@ class AgentThinkModelsV1ResponseModelsItemZeroParams(typing_extensions.TypedDict The display name of the model """ - provider: typing.Literal["open_ai"] + provider: AgentThinkModelsV1ResponseModelsItemZeroProvider """ The provider of the model """ diff --git a/src/deepgram/requests/anthropic.py b/src/deepgram/requests/anthropic.py deleted file mode 100644 index 3b60c72d..00000000 --- a/src/deepgram/requests/anthropic.py +++ /dev/null @@ -1,24 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.anthropic_think_provider_model import AnthropicThinkProviderModel - - -class AnthropicParams(typing_extensions.TypedDict): - type: typing.Literal["anthropic"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the Anthropic Messages API - """ - - model: AnthropicThinkProviderModel - """ - Anthropic model to use - """ - - temperature: typing_extensions.NotRequired[float] - """ - Anthropic temperature (0-1) - """ diff --git a/src/deepgram/requests/aws_bedrock_think_provider.py b/src/deepgram/requests/aws_bedrock_think_provider.py deleted file mode 100644 index 1bdfd6f4..00000000 --- a/src/deepgram/requests/aws_bedrock_think_provider.py +++ /dev/null @@ -1,25 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.aws_bedrock_think_provider_model import AwsBedrockThinkProviderModel -from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentialsParams - - -class AwsBedrockThinkProviderParams(typing_extensions.TypedDict): - type: typing.Literal["aws_bedrock"] - model: AwsBedrockThinkProviderModel - """ - AWS Bedrock model to use - """ - - temperature: typing_extensions.NotRequired[float] - """ - AWS Bedrock temperature (0-2) - """ - - credentials: typing_extensions.NotRequired[AwsBedrockThinkProviderCredentialsParams] - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ diff --git a/src/deepgram/requests/aws_bedrock_think_provider_credentials.py b/src/deepgram/requests/aws_bedrock_think_provider_credentials.py deleted file mode 100644 index ff2a0cb4..00000000 --- a/src/deepgram/requests/aws_bedrock_think_provider_credentials.py +++ /dev/null @@ -1,35 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing_extensions -from ..types.aws_bedrock_think_provider_credentials_type import AwsBedrockThinkProviderCredentialsType - - -class AwsBedrockThinkProviderCredentialsParams(typing_extensions.TypedDict): - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ - - type: typing_extensions.NotRequired[AwsBedrockThinkProviderCredentialsType] - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ - - region: typing_extensions.NotRequired[str] - """ - AWS region - """ - - access_key_id: typing_extensions.NotRequired[str] - """ - AWS access key - """ - - secret_access_key: typing_extensions.NotRequired[str] - """ - AWS secret access key - """ - - session_token: typing_extensions.NotRequired[str] - """ - AWS session token (required for STS only) - """ diff --git a/src/deepgram/requests/aws_polly_speak_provider.py b/src/deepgram/requests/aws_polly_speak_provider.py deleted file mode 100644 index d5cbb726..00000000 --- a/src/deepgram/requests/aws_polly_speak_provider.py +++ /dev/null @@ -1,29 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.aws_polly_speak_provider_engine import AwsPollySpeakProviderEngine -from ..types.aws_polly_speak_provider_voice import AwsPollySpeakProviderVoice -from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentialsParams - - -class AwsPollySpeakProviderParams(typing_extensions.TypedDict): - type: typing.Literal["aws_polly"] - voice: AwsPollySpeakProviderVoice - """ - AWS Polly voice name - """ - - language: str - """ - Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API - """ - - language_code: typing_extensions.NotRequired[str] - """ - Use the `language` field instead. - """ - - engine: AwsPollySpeakProviderEngine - credentials: AwsPollySpeakProviderCredentialsParams diff --git a/src/deepgram/requests/aws_polly_speak_provider_credentials.py b/src/deepgram/requests/aws_polly_speak_provider_credentials.py deleted file mode 100644 index fd279a88..00000000 --- a/src/deepgram/requests/aws_polly_speak_provider_credentials.py +++ /dev/null @@ -1,15 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing_extensions -from ..types.aws_polly_speak_provider_credentials_type import AwsPollySpeakProviderCredentialsType - - -class AwsPollySpeakProviderCredentialsParams(typing_extensions.TypedDict): - type: AwsPollySpeakProviderCredentialsType - region: str - access_key_id: str - secret_access_key: str - session_token: typing_extensions.NotRequired[str] - """ - Required for STS only - """ diff --git a/src/deepgram/requests/cartesia.py b/src/deepgram/requests/cartesia.py deleted file mode 100644 index f51b6c58..00000000 --- a/src/deepgram/requests/cartesia.py +++ /dev/null @@ -1,26 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.cartesia_speak_provider_model_id import CartesiaSpeakProviderModelId -from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoiceParams - - -class CartesiaParams(typing_extensions.TypedDict): - type: typing.Literal["cartesia"] - version: typing_extensions.NotRequired[typing.Literal["2025-03-17"]] - """ - The API version header for the Cartesia text-to-speech API - """ - - model_id: CartesiaSpeakProviderModelId - """ - Cartesia model ID - """ - - voice: CartesiaSpeakProviderVoiceParams - language: typing_extensions.NotRequired[str] - """ - Cartesia language code - """ diff --git a/src/deepgram/requests/deepgram.py b/src/deepgram/requests/deepgram.py deleted file mode 100644 index 57769bf9..00000000 --- a/src/deepgram/requests/deepgram.py +++ /dev/null @@ -1,19 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.deepgram_speak_provider_model import DeepgramSpeakProviderModel - - -class DeepgramParams(typing_extensions.TypedDict): - type: typing.Literal["deepgram"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the Deepgram text-to-speech API - """ - - model: DeepgramSpeakProviderModel - """ - Deepgram TTS model - """ diff --git a/src/deepgram/requests/eleven_labs_speak_provider.py b/src/deepgram/requests/eleven_labs_speak_provider.py deleted file mode 100644 index e0f24bf4..00000000 --- a/src/deepgram/requests/eleven_labs_speak_provider.py +++ /dev/null @@ -1,29 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.eleven_labs_speak_provider_model_id import ElevenLabsSpeakProviderModelId - - -class ElevenLabsSpeakProviderParams(typing_extensions.TypedDict): - type: typing.Literal["eleven_labs"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the ElevenLabs text-to-speech API - """ - - model_id: ElevenLabsSpeakProviderModelId - """ - Eleven Labs model ID - """ - - language: typing_extensions.NotRequired[str] - """ - Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API - """ - - language_code: typing_extensions.NotRequired[str] - """ - Use the `language` field instead. - """ diff --git a/src/deepgram/requests/google.py b/src/deepgram/requests/google.py deleted file mode 100644 index 7c183372..00000000 --- a/src/deepgram/requests/google.py +++ /dev/null @@ -1,24 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.google_think_provider_model import GoogleThinkProviderModel - - -class GoogleParams(typing_extensions.TypedDict): - type: typing.Literal["google"] - version: typing_extensions.NotRequired[typing.Literal["v1beta"]] - """ - The REST API version for the Google generative language API - """ - - model: GoogleThinkProviderModel - """ - Google model to use - """ - - temperature: typing_extensions.NotRequired[float] - """ - Google temperature (0-2) - """ diff --git a/src/deepgram/requests/groq.py b/src/deepgram/requests/groq.py deleted file mode 100644 index 526dbff3..00000000 --- a/src/deepgram/requests/groq.py +++ /dev/null @@ -1,23 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions - - -class GroqParams(typing_extensions.TypedDict): - type: typing.Literal["groq"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the Groq's chat completions API (mostly OpenAI-compatible) - """ - - model: typing.Literal["openai/gpt-oss-20b"] - """ - Groq model to use - """ - - temperature: typing_extensions.NotRequired[float] - """ - Groq temperature (0-2) - """ diff --git a/src/deepgram/requests/open_ai_speak_provider.py b/src/deepgram/requests/open_ai_speak_provider.py deleted file mode 100644 index dd341584..00000000 --- a/src/deepgram/requests/open_ai_speak_provider.py +++ /dev/null @@ -1,25 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.open_ai_speak_provider_model import OpenAiSpeakProviderModel -from ..types.open_ai_speak_provider_voice import OpenAiSpeakProviderVoice - - -class OpenAiSpeakProviderParams(typing_extensions.TypedDict): - type: typing.Literal["open_ai"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the OpenAI text-to-speech API - """ - - model: OpenAiSpeakProviderModel - """ - OpenAI TTS model - """ - - voice: OpenAiSpeakProviderVoice - """ - OpenAI voice - """ diff --git a/src/deepgram/requests/open_ai_think_provider.py b/src/deepgram/requests/open_ai_think_provider.py deleted file mode 100644 index 447fd842..00000000 --- a/src/deepgram/requests/open_ai_think_provider.py +++ /dev/null @@ -1,24 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import typing_extensions -from ..types.open_ai_think_provider_model import OpenAiThinkProviderModel - - -class OpenAiThinkProviderParams(typing_extensions.TypedDict): - type: typing.Literal["open_ai"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - """ - The REST API version for the OpenAI chat completions API - """ - - model: OpenAiThinkProviderModel - """ - OpenAI model to use - """ - - temperature: typing_extensions.NotRequired[float] - """ - OpenAI temperature (0-2) - """ diff --git a/src/deepgram/requests/speak_settings_v1.py b/src/deepgram/requests/speak_settings_v1.py deleted file mode 100644 index c11bc1c7..00000000 --- a/src/deepgram/requests/speak_settings_v1.py +++ /dev/null @@ -1,14 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing_extensions -from .speak_settings_v1endpoint import SpeakSettingsV1EndpointParams -from .speak_settings_v1provider import SpeakSettingsV1ProviderParams - - -class SpeakSettingsV1Params(typing_extensions.TypedDict): - provider: SpeakSettingsV1ProviderParams - endpoint: typing_extensions.NotRequired[SpeakSettingsV1EndpointParams] - """ - Optional if provider is Deepgram. Required for non-Deepgram TTS providers. - When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. - """ diff --git a/src/deepgram/requests/speak_settings_v1provider.py b/src/deepgram/requests/speak_settings_v1provider.py deleted file mode 100644 index 3db9f563..00000000 --- a/src/deepgram/requests/speak_settings_v1provider.py +++ /dev/null @@ -1,63 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -from __future__ import annotations - -import typing - -import typing_extensions -from ..types.aws_polly_speak_provider_engine import AwsPollySpeakProviderEngine -from ..types.aws_polly_speak_provider_voice import AwsPollySpeakProviderVoice -from ..types.cartesia_speak_provider_model_id import CartesiaSpeakProviderModelId -from ..types.deepgram_speak_provider_model import DeepgramSpeakProviderModel -from ..types.eleven_labs_speak_provider_model_id import ElevenLabsSpeakProviderModelId -from ..types.open_ai_speak_provider_model import OpenAiSpeakProviderModel -from ..types.open_ai_speak_provider_voice import OpenAiSpeakProviderVoice -from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentialsParams -from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoiceParams - - -class SpeakSettingsV1Provider_DeepgramParams(typing_extensions.TypedDict): - type: typing.Literal["deepgram"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model: DeepgramSpeakProviderModel - - -class SpeakSettingsV1Provider_ElevenLabsParams(typing_extensions.TypedDict): - type: typing.Literal["eleven_labs"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model_id: ElevenLabsSpeakProviderModelId - language: typing_extensions.NotRequired[str] - language_code: typing_extensions.NotRequired[str] - - -class SpeakSettingsV1Provider_CartesiaParams(typing_extensions.TypedDict): - type: typing.Literal["cartesia"] - version: typing_extensions.NotRequired[typing.Literal["2025-03-17"]] - model_id: CartesiaSpeakProviderModelId - voice: CartesiaSpeakProviderVoiceParams - language: typing_extensions.NotRequired[str] - - -class SpeakSettingsV1Provider_OpenAiParams(typing_extensions.TypedDict): - type: typing.Literal["open_ai"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model: OpenAiSpeakProviderModel - voice: OpenAiSpeakProviderVoice - - -class SpeakSettingsV1Provider_AwsPollyParams(typing_extensions.TypedDict): - type: typing.Literal["aws_polly"] - voice: AwsPollySpeakProviderVoice - language: str - language_code: typing_extensions.NotRequired[str] - engine: AwsPollySpeakProviderEngine - credentials: AwsPollySpeakProviderCredentialsParams - - -SpeakSettingsV1ProviderParams = typing.Union[ - SpeakSettingsV1Provider_DeepgramParams, - SpeakSettingsV1Provider_ElevenLabsParams, - SpeakSettingsV1Provider_CartesiaParams, - SpeakSettingsV1Provider_OpenAiParams, - SpeakSettingsV1Provider_AwsPollyParams, -] diff --git a/src/deepgram/requests/think_settings_v1.py b/src/deepgram/requests/think_settings_v1.py index 260b488d..5a04e8f7 100644 --- a/src/deepgram/requests/think_settings_v1.py +++ b/src/deepgram/requests/think_settings_v1.py @@ -3,22 +3,7 @@ import typing import typing_extensions -from .think_settings_v1context_length import ThinkSettingsV1ContextLengthParams -from .think_settings_v1endpoint import ThinkSettingsV1EndpointParams -from .think_settings_v1functions_item import ThinkSettingsV1FunctionsItemParams -from .think_settings_v1provider import ThinkSettingsV1ProviderParams class ThinkSettingsV1Params(typing_extensions.TypedDict): - provider: ThinkSettingsV1ProviderParams - endpoint: typing_extensions.NotRequired[ThinkSettingsV1EndpointParams] - """ - Optional for non-Deepgram LLM providers. When present, must include url field and headers object - """ - - functions: typing_extensions.NotRequired[typing.Sequence[ThinkSettingsV1FunctionsItemParams]] - prompt: typing_extensions.NotRequired[str] - context_length: typing_extensions.NotRequired[ThinkSettingsV1ContextLengthParams] - """ - Specifies the number of characters retained in context between user messages, agent responses, and function calls. This setting is only configurable when a custom think endpoint is used - """ + context_length: typing_extensions.NotRequired[typing.Any] diff --git a/src/deepgram/requests/think_settings_v1provider.py b/src/deepgram/requests/think_settings_v1provider.py deleted file mode 100644 index 6de775b2..00000000 --- a/src/deepgram/requests/think_settings_v1provider.py +++ /dev/null @@ -1,56 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -from __future__ import annotations - -import typing - -import typing_extensions -from ..types.anthropic_think_provider_model import AnthropicThinkProviderModel -from ..types.aws_bedrock_think_provider_model import AwsBedrockThinkProviderModel -from ..types.google_think_provider_model import GoogleThinkProviderModel -from ..types.open_ai_think_provider_model import OpenAiThinkProviderModel -from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentialsParams - - -class ThinkSettingsV1Provider_OpenAiParams(typing_extensions.TypedDict): - type: typing.Literal["open_ai"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model: OpenAiThinkProviderModel - temperature: typing_extensions.NotRequired[float] - - -class ThinkSettingsV1Provider_AwsBedrockParams(typing_extensions.TypedDict): - type: typing.Literal["aws_bedrock"] - model: AwsBedrockThinkProviderModel - temperature: typing_extensions.NotRequired[float] - credentials: typing_extensions.NotRequired[AwsBedrockThinkProviderCredentialsParams] - - -class ThinkSettingsV1Provider_AnthropicParams(typing_extensions.TypedDict): - type: typing.Literal["anthropic"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model: AnthropicThinkProviderModel - temperature: typing_extensions.NotRequired[float] - - -class ThinkSettingsV1Provider_GoogleParams(typing_extensions.TypedDict): - type: typing.Literal["google"] - version: typing_extensions.NotRequired[typing.Literal["v1beta"]] - model: GoogleThinkProviderModel - temperature: typing_extensions.NotRequired[float] - - -class ThinkSettingsV1Provider_GroqParams(typing_extensions.TypedDict): - type: typing.Literal["groq"] - version: typing_extensions.NotRequired[typing.Literal["v1"]] - model: typing.Literal["openai/gpt-oss-20b"] - temperature: typing_extensions.NotRequired[float] - - -ThinkSettingsV1ProviderParams = typing.Union[ - ThinkSettingsV1Provider_OpenAiParams, - ThinkSettingsV1Provider_AwsBedrockParams, - ThinkSettingsV1Provider_AnthropicParams, - ThinkSettingsV1Provider_GoogleParams, - ThinkSettingsV1Provider_GroqParams, -] diff --git a/src/deepgram/requests/usage_breakdown_v1response_results_item_grouping.py b/src/deepgram/requests/usage_breakdown_v1response_results_item_grouping.py index 9529479b..087a86bb 100644 --- a/src/deepgram/requests/usage_breakdown_v1response_results_item_grouping.py +++ b/src/deepgram/requests/usage_breakdown_v1response_results_item_grouping.py @@ -1,5 +1,7 @@ # This file was auto-generated by Fern from our API Definition. +import typing + import typing_extensions @@ -29,11 +31,7 @@ class UsageBreakdownV1ResponseResultsItemGroupingParams(typing_extensions.TypedD Optional feature set identifier """ - models: typing_extensions.NotRequired[str] - """ - Optional models identifier - """ - + models: typing_extensions.NotRequired[typing.Sequence[typing.Optional[str]]] method: typing_extensions.NotRequired[str] """ Optional method identifier diff --git a/src/deepgram/self_hosted/v1/__init__.py b/src/deepgram/self_hosted/v1/__init__.py index 7d71b379..c0b97e44 100644 --- a/src/deepgram/self_hosted/v1/__init__.py +++ b/src/deepgram/self_hosted/v1/__init__.py @@ -7,8 +7,12 @@ if typing.TYPE_CHECKING: from . import distribution_credentials - from .distribution_credentials import DistributionCredentialsCreateRequestScopesItem + from .distribution_credentials import ( + DistributionCredentialsCreateRequestProvider, + DistributionCredentialsCreateRequestScopesItem, + ) _dynamic_imports: typing.Dict[str, str] = { + "DistributionCredentialsCreateRequestProvider": ".distribution_credentials", "DistributionCredentialsCreateRequestScopesItem": ".distribution_credentials", "distribution_credentials": ".distribution_credentials", } @@ -35,4 +39,8 @@ def __dir__(): return sorted(lazy_attrs) -__all__ = ["DistributionCredentialsCreateRequestScopesItem", "distribution_credentials"] +__all__ = [ + "DistributionCredentialsCreateRequestProvider", + "DistributionCredentialsCreateRequestScopesItem", + "distribution_credentials", +] diff --git a/src/deepgram/self_hosted/v1/distribution_credentials/__init__.py b/src/deepgram/self_hosted/v1/distribution_credentials/__init__.py index 020a49b5..cb4dba07 100644 --- a/src/deepgram/self_hosted/v1/distribution_credentials/__init__.py +++ b/src/deepgram/self_hosted/v1/distribution_credentials/__init__.py @@ -6,8 +6,11 @@ from importlib import import_module if typing.TYPE_CHECKING: - from .types import DistributionCredentialsCreateRequestScopesItem -_dynamic_imports: typing.Dict[str, str] = {"DistributionCredentialsCreateRequestScopesItem": ".types"} + from .types import DistributionCredentialsCreateRequestProvider, DistributionCredentialsCreateRequestScopesItem +_dynamic_imports: typing.Dict[str, str] = { + "DistributionCredentialsCreateRequestProvider": ".types", + "DistributionCredentialsCreateRequestScopesItem": ".types", +} def __getattr__(attr_name: str) -> typing.Any: @@ -31,4 +34,4 @@ def __dir__(): return sorted(lazy_attrs) -__all__ = ["DistributionCredentialsCreateRequestScopesItem"] +__all__ = ["DistributionCredentialsCreateRequestProvider", "DistributionCredentialsCreateRequestScopesItem"] diff --git a/src/deepgram/self_hosted/v1/distribution_credentials/client.py b/src/deepgram/self_hosted/v1/distribution_credentials/client.py index 63371d29..e60f4af2 100644 --- a/src/deepgram/self_hosted/v1/distribution_credentials/client.py +++ b/src/deepgram/self_hosted/v1/distribution_credentials/client.py @@ -8,6 +8,7 @@ from ....types.get_project_distribution_credentials_v1response import GetProjectDistributionCredentialsV1Response from ....types.list_project_distribution_credentials_v1response import ListProjectDistributionCredentialsV1Response from .raw_client import AsyncRawDistributionCredentialsClient, RawDistributionCredentialsClient +from .types.distribution_credentials_create_request_provider import DistributionCredentialsCreateRequestProvider from .types.distribution_credentials_create_request_scopes_item import DistributionCredentialsCreateRequestScopesItem # this is used as the default value for optional parameters @@ -72,7 +73,7 @@ def create( typing.Sequence[DistributionCredentialsCreateRequestScopesItem], ] ] = None, - provider: typing.Optional[typing.Literal["quay"]] = None, + provider: typing.Optional[DistributionCredentialsCreateRequestProvider] = None, comment: typing.Optional[str] = OMIT, request_options: typing.Optional[RequestOptions] = None, ) -> CreateProjectDistributionCredentialsV1Response: @@ -87,7 +88,7 @@ def create( scopes : typing.Optional[typing.Union[DistributionCredentialsCreateRequestScopesItem, typing.Sequence[DistributionCredentialsCreateRequestScopesItem]]] List of permission scopes for the credentials - provider : typing.Optional[typing.Literal["quay"]] + provider : typing.Optional[DistributionCredentialsCreateRequestProvider] The provider of the distribution service comment : typing.Optional[str] @@ -110,6 +111,7 @@ def create( ) client.self_hosted.v1.distribution_credentials.create( project_id="123456-7890-1234-5678-901234", + provider="quay", ) """ _response = self._raw_client.create( @@ -266,7 +268,7 @@ async def create( typing.Sequence[DistributionCredentialsCreateRequestScopesItem], ] ] = None, - provider: typing.Optional[typing.Literal["quay"]] = None, + provider: typing.Optional[DistributionCredentialsCreateRequestProvider] = None, comment: typing.Optional[str] = OMIT, request_options: typing.Optional[RequestOptions] = None, ) -> CreateProjectDistributionCredentialsV1Response: @@ -281,7 +283,7 @@ async def create( scopes : typing.Optional[typing.Union[DistributionCredentialsCreateRequestScopesItem, typing.Sequence[DistributionCredentialsCreateRequestScopesItem]]] List of permission scopes for the credentials - provider : typing.Optional[typing.Literal["quay"]] + provider : typing.Optional[DistributionCredentialsCreateRequestProvider] The provider of the distribution service comment : typing.Optional[str] @@ -309,6 +311,7 @@ async def create( async def main() -> None: await client.self_hosted.v1.distribution_credentials.create( project_id="123456-7890-1234-5678-901234", + provider="quay", ) diff --git a/src/deepgram/self_hosted/v1/distribution_credentials/raw_client.py b/src/deepgram/self_hosted/v1/distribution_credentials/raw_client.py index f7e60ef9..5861d960 100644 --- a/src/deepgram/self_hosted/v1/distribution_credentials/raw_client.py +++ b/src/deepgram/self_hosted/v1/distribution_credentials/raw_client.py @@ -13,6 +13,7 @@ from ....types.create_project_distribution_credentials_v1response import CreateProjectDistributionCredentialsV1Response from ....types.get_project_distribution_credentials_v1response import GetProjectDistributionCredentialsV1Response from ....types.list_project_distribution_credentials_v1response import ListProjectDistributionCredentialsV1Response +from .types.distribution_credentials_create_request_provider import DistributionCredentialsCreateRequestProvider from .types.distribution_credentials_create_request_scopes_item import DistributionCredentialsCreateRequestScopesItem # this is used as the default value for optional parameters @@ -84,7 +85,7 @@ def create( typing.Sequence[DistributionCredentialsCreateRequestScopesItem], ] ] = None, - provider: typing.Optional[typing.Literal["quay"]] = None, + provider: typing.Optional[DistributionCredentialsCreateRequestProvider] = None, comment: typing.Optional[str] = OMIT, request_options: typing.Optional[RequestOptions] = None, ) -> HttpResponse[CreateProjectDistributionCredentialsV1Response]: @@ -99,7 +100,7 @@ def create( scopes : typing.Optional[typing.Union[DistributionCredentialsCreateRequestScopesItem, typing.Sequence[DistributionCredentialsCreateRequestScopesItem]]] List of permission scopes for the credentials - provider : typing.Optional[typing.Literal["quay"]] + provider : typing.Optional[DistributionCredentialsCreateRequestProvider] The provider of the distribution service comment : typing.Optional[str] @@ -338,7 +339,7 @@ async def create( typing.Sequence[DistributionCredentialsCreateRequestScopesItem], ] ] = None, - provider: typing.Optional[typing.Literal["quay"]] = None, + provider: typing.Optional[DistributionCredentialsCreateRequestProvider] = None, comment: typing.Optional[str] = OMIT, request_options: typing.Optional[RequestOptions] = None, ) -> AsyncHttpResponse[CreateProjectDistributionCredentialsV1Response]: @@ -353,7 +354,7 @@ async def create( scopes : typing.Optional[typing.Union[DistributionCredentialsCreateRequestScopesItem, typing.Sequence[DistributionCredentialsCreateRequestScopesItem]]] List of permission scopes for the credentials - provider : typing.Optional[typing.Literal["quay"]] + provider : typing.Optional[DistributionCredentialsCreateRequestProvider] The provider of the distribution service comment : typing.Optional[str] diff --git a/src/deepgram/self_hosted/v1/distribution_credentials/types/__init__.py b/src/deepgram/self_hosted/v1/distribution_credentials/types/__init__.py index 05c1485a..6f834f97 100644 --- a/src/deepgram/self_hosted/v1/distribution_credentials/types/__init__.py +++ b/src/deepgram/self_hosted/v1/distribution_credentials/types/__init__.py @@ -6,9 +6,11 @@ from importlib import import_module if typing.TYPE_CHECKING: + from .distribution_credentials_create_request_provider import DistributionCredentialsCreateRequestProvider from .distribution_credentials_create_request_scopes_item import DistributionCredentialsCreateRequestScopesItem _dynamic_imports: typing.Dict[str, str] = { - "DistributionCredentialsCreateRequestScopesItem": ".distribution_credentials_create_request_scopes_item" + "DistributionCredentialsCreateRequestProvider": ".distribution_credentials_create_request_provider", + "DistributionCredentialsCreateRequestScopesItem": ".distribution_credentials_create_request_scopes_item", } @@ -33,4 +35,4 @@ def __dir__(): return sorted(lazy_attrs) -__all__ = ["DistributionCredentialsCreateRequestScopesItem"] +__all__ = ["DistributionCredentialsCreateRequestProvider", "DistributionCredentialsCreateRequestScopesItem"] diff --git a/src/deepgram/self_hosted/v1/distribution_credentials/types/distribution_credentials_create_request_provider.py b/src/deepgram/self_hosted/v1/distribution_credentials/types/distribution_credentials_create_request_provider.py new file mode 100644 index 00000000..20487087 --- /dev/null +++ b/src/deepgram/self_hosted/v1/distribution_credentials/types/distribution_credentials_create_request_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +DistributionCredentialsCreateRequestProvider = typing.Union[typing.Literal["quay"], typing.Any] diff --git a/src/deepgram/speak/__init__.py b/src/deepgram/speak/__init__.py index 73eda24c..3adcdd07 100644 --- a/src/deepgram/speak/__init__.py +++ b/src/deepgram/speak/__init__.py @@ -25,10 +25,13 @@ SpeakV1FlushedType, SpeakV1Metadata, SpeakV1MetadataParams, + SpeakV1MetadataType, SpeakV1Text, SpeakV1TextParams, + SpeakV1TextType, SpeakV1Warning, SpeakV1WarningParams, + SpeakV1WarningType, ) _dynamic_imports: typing.Dict[str, str] = { "SpeakV1Clear": ".v1", @@ -48,10 +51,13 @@ "SpeakV1FlushedType": ".v1", "SpeakV1Metadata": ".v1", "SpeakV1MetadataParams": ".v1", + "SpeakV1MetadataType": ".v1", "SpeakV1Text": ".v1", "SpeakV1TextParams": ".v1", + "SpeakV1TextType": ".v1", "SpeakV1Warning": ".v1", "SpeakV1WarningParams": ".v1", + "SpeakV1WarningType": ".v1", "v1": ".v1", } @@ -95,9 +101,12 @@ def __dir__(): "SpeakV1FlushedType", "SpeakV1Metadata", "SpeakV1MetadataParams", + "SpeakV1MetadataType", "SpeakV1Text", "SpeakV1TextParams", + "SpeakV1TextType", "SpeakV1Warning", "SpeakV1WarningParams", + "SpeakV1WarningType", "v1", ] diff --git a/src/deepgram/speak/v1/__init__.py b/src/deepgram/speak/v1/__init__.py index c874ef09..54c2fbfc 100644 --- a/src/deepgram/speak/v1/__init__.py +++ b/src/deepgram/speak/v1/__init__.py @@ -18,8 +18,11 @@ SpeakV1Flushed, SpeakV1FlushedType, SpeakV1Metadata, + SpeakV1MetadataType, SpeakV1Text, + SpeakV1TextType, SpeakV1Warning, + SpeakV1WarningType, ) from . import audio from .audio import ( @@ -60,10 +63,13 @@ "SpeakV1FlushedType": ".types", "SpeakV1Metadata": ".types", "SpeakV1MetadataParams": ".requests", + "SpeakV1MetadataType": ".types", "SpeakV1Text": ".types", "SpeakV1TextParams": ".requests", + "SpeakV1TextType": ".types", "SpeakV1Warning": ".types", "SpeakV1WarningParams": ".requests", + "SpeakV1WarningType": ".types", "audio": ".audio", } @@ -111,9 +117,12 @@ def __dir__(): "SpeakV1FlushedType", "SpeakV1Metadata", "SpeakV1MetadataParams", + "SpeakV1MetadataType", "SpeakV1Text", "SpeakV1TextParams", + "SpeakV1TextType", "SpeakV1Warning", "SpeakV1WarningParams", + "SpeakV1WarningType", "audio", ] diff --git a/src/deepgram/speak/v1/client.py b/src/deepgram/speak/v1/client.py index 1780af66..92785112 100644 --- a/src/deepgram/speak/v1/client.py +++ b/src/deepgram/speak/v1/client.py @@ -14,6 +14,10 @@ from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...types.speak_v1encoding import SpeakV1Encoding +from ...types.speak_v1mip_opt_out import SpeakV1MipOptOut +from ...types.speak_v1model import SpeakV1Model +from ...types.speak_v1sample_rate import SpeakV1SampleRate from .raw_client import AsyncRawV1Client, RawV1Client from .socket_client import AsyncV1SocketClient, V1SocketClient @@ -47,10 +51,10 @@ def with_raw_response(self) -> RawV1Client: def connect( self, *, - encoding: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, + encoding: typing.Optional[SpeakV1Encoding] = None, + mip_opt_out: typing.Optional[SpeakV1MipOptOut] = None, + model: typing.Optional[SpeakV1Model] = None, + sample_rate: typing.Optional[SpeakV1SampleRate] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V1SocketClient]: @@ -59,18 +63,18 @@ def connect( Parameters ---------- - encoding : typing.Optional[str] + encoding : typing.Optional[SpeakV1Encoding] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[SpeakV1MipOptOut] - model : typing.Optional[str] + model : typing.Optional[SpeakV1Model] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[SpeakV1SampleRate] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -151,10 +155,10 @@ def with_raw_response(self) -> AsyncRawV1Client: async def connect( self, *, - encoding: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, + encoding: typing.Optional[SpeakV1Encoding] = None, + mip_opt_out: typing.Optional[SpeakV1MipOptOut] = None, + model: typing.Optional[SpeakV1Model] = None, + sample_rate: typing.Optional[SpeakV1SampleRate] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV1SocketClient]: @@ -163,18 +167,18 @@ async def connect( Parameters ---------- - encoding : typing.Optional[str] + encoding : typing.Optional[SpeakV1Encoding] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[SpeakV1MipOptOut] - model : typing.Optional[str] + model : typing.Optional[SpeakV1Model] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[SpeakV1SampleRate] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/speak/v1/raw_client.py b/src/deepgram/speak/v1/raw_client.py index 0ade1093..fd739c1d 100644 --- a/src/deepgram/speak/v1/raw_client.py +++ b/src/deepgram/speak/v1/raw_client.py @@ -12,6 +12,10 @@ from ...core.remove_none_from_dict import remove_none_from_dict from ...core.request_options import RequestOptions from ...core.websocket_compat import InvalidWebSocketStatus, get_status_code +from ...types.speak_v1encoding import SpeakV1Encoding +from ...types.speak_v1mip_opt_out import SpeakV1MipOptOut +from ...types.speak_v1model import SpeakV1Model +from ...types.speak_v1sample_rate import SpeakV1SampleRate from .socket_client import AsyncV1SocketClient, V1SocketClient try: @@ -28,10 +32,10 @@ def __init__(self, *, client_wrapper: SyncClientWrapper): def connect( self, *, - encoding: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, + encoding: typing.Optional[SpeakV1Encoding] = None, + mip_opt_out: typing.Optional[SpeakV1MipOptOut] = None, + model: typing.Optional[SpeakV1Model] = None, + sample_rate: typing.Optional[SpeakV1SampleRate] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.Iterator[V1SocketClient]: @@ -40,18 +44,18 @@ def connect( Parameters ---------- - encoding : typing.Optional[str] + encoding : typing.Optional[SpeakV1Encoding] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[SpeakV1MipOptOut] - model : typing.Optional[str] + model : typing.Optional[SpeakV1Model] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[SpeakV1SampleRate] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. @@ -111,10 +115,10 @@ def __init__(self, *, client_wrapper: AsyncClientWrapper): async def connect( self, *, - encoding: typing.Optional[str] = None, - mip_opt_out: typing.Optional[str] = None, - model: typing.Optional[str] = None, - sample_rate: typing.Optional[str] = None, + encoding: typing.Optional[SpeakV1Encoding] = None, + mip_opt_out: typing.Optional[SpeakV1MipOptOut] = None, + model: typing.Optional[SpeakV1Model] = None, + sample_rate: typing.Optional[SpeakV1SampleRate] = None, authorization: typing.Optional[str] = None, request_options: typing.Optional[RequestOptions] = None, ) -> typing.AsyncIterator[AsyncV1SocketClient]: @@ -123,18 +127,18 @@ async def connect( Parameters ---------- - encoding : typing.Optional[str] + encoding : typing.Optional[SpeakV1Encoding] - mip_opt_out : typing.Optional[str] + mip_opt_out : typing.Optional[SpeakV1MipOptOut] - model : typing.Optional[str] + model : typing.Optional[SpeakV1Model] - sample_rate : typing.Optional[str] + sample_rate : typing.Optional[SpeakV1SampleRate] authorization : typing.Optional[str] - Use your API key for authentication, or alternatively generate a [temporary token](/guides/fundamentals/token-based-authentication) and pass it via the `token` query parameter. + Use your API key or a [temporary token](/guides/fundamentals/token-based-authentication) for authentication via the `Authorization` header. In client-side environments where custom headers are not supported, use the [`Sec-WebSocket-Protocol`](/guides/deep-dives/using-the-sec-websocket-protocol) header instead. - **Example:** `token %DEEPGRAM_API_KEY%` or `bearer %DEEPGRAM_TOKEN%` + **Example:** `Authorization: Token %DEEPGRAM_API_KEY%` or `Authorization: Bearer %DEEPGRAM_TOKEN%` request_options : typing.Optional[RequestOptions] Request-specific configuration. diff --git a/src/deepgram/speak/v1/requests/speak_v1metadata.py b/src/deepgram/speak/v1/requests/speak_v1metadata.py index 89fb6809..9c618ab3 100644 --- a/src/deepgram/speak/v1/requests/speak_v1metadata.py +++ b/src/deepgram/speak/v1/requests/speak_v1metadata.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.speak_v1metadata_type import SpeakV1MetadataType class SpeakV1MetadataParams(typing_extensions.TypedDict): - type: typing.Literal["Metadata"] + type: SpeakV1MetadataType """ Message type identifier """ diff --git a/src/deepgram/speak/v1/requests/speak_v1text.py b/src/deepgram/speak/v1/requests/speak_v1text.py index 78873194..1f66d1c9 100644 --- a/src/deepgram/speak/v1/requests/speak_v1text.py +++ b/src/deepgram/speak/v1/requests/speak_v1text.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.speak_v1text_type import SpeakV1TextType class SpeakV1TextParams(typing_extensions.TypedDict): - type: typing.Literal["Speak"] + type: SpeakV1TextType """ Message type identifier """ diff --git a/src/deepgram/speak/v1/requests/speak_v1warning.py b/src/deepgram/speak/v1/requests/speak_v1warning.py index ca6c78f8..69969e93 100644 --- a/src/deepgram/speak/v1/requests/speak_v1warning.py +++ b/src/deepgram/speak/v1/requests/speak_v1warning.py @@ -1,12 +1,11 @@ # This file was auto-generated by Fern from our API Definition. -import typing - import typing_extensions +from ..types.speak_v1warning_type import SpeakV1WarningType class SpeakV1WarningParams(typing_extensions.TypedDict): - type: typing.Literal["Warning"] + type: SpeakV1WarningType """ Message type identifier """ diff --git a/src/deepgram/speak/v1/socket_client.py b/src/deepgram/speak/v1/socket_client.py index 9f640575..e71ba038 100644 --- a/src/deepgram/speak/v1/socket_client.py +++ b/src/deepgram/speak/v1/socket_client.py @@ -1,6 +1,7 @@ # This file was auto-generated by Fern from our API Definition. import json +import logging import typing from json.decoder import JSONDecodeError @@ -22,6 +23,7 @@ except ImportError: from websockets import WebSocketClientProtocol # type: ignore +_logger = logging.getLogger(__name__) V1SocketClientResponse = typing.Union[bytes, SpeakV1Metadata, SpeakV1Flushed, SpeakV1Cleared, SpeakV1Warning] @@ -35,7 +37,13 @@ async def __aiter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue async def start_listening(self): """ @@ -54,7 +62,13 @@ async def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue await self._emit_async(EventType.MESSAGE, parsed) except Exception as exc: await self._emit_async(EventType.ERROR, exc) @@ -68,26 +82,26 @@ async def send_text(self, message: SpeakV1Text) -> None: """ await self._send_model(message) - async def send_flush(self, message: typing.Optional[SpeakV1Flush] = None) -> None: + async def send_flush(self, message: SpeakV1Flush) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Flush. """ - await self._send_model(message or SpeakV1Flush(type="Flush")) + await self._send_model(message) - async def send_clear(self, message: typing.Optional[SpeakV1Clear] = None) -> None: + async def send_clear(self, message: SpeakV1Clear) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Clear. """ - await self._send_model(message or SpeakV1Clear(type="Clear")) + await self._send_model(message) - async def send_close(self, message: typing.Optional[SpeakV1Close] = None) -> None: + async def send_close(self, message: SpeakV1Close) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Close. """ - await self._send_model(message or SpeakV1Close(type="Close")) + await self._send_model(message) async def recv(self) -> V1SocketClientResponse: """ @@ -97,7 +111,11 @@ async def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore async def _send(self, data: typing.Any) -> None: """ @@ -124,7 +142,13 @@ def __iter__(self): if isinstance(message, bytes): yield message else: - yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + try: + yield construct_type(type_=V1SocketClientResponse, object_=json.loads(message)) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue def start_listening(self): """ @@ -143,7 +167,13 @@ def start_listening(self): parsed = raw_message else: json_data = json.loads(raw_message) - parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + parsed = construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning( + "Skipping unknown WebSocket message; update your SDK version to support new message types." + ) + continue self._emit(EventType.MESSAGE, parsed) except Exception as exc: self._emit(EventType.ERROR, exc) @@ -157,26 +187,26 @@ def send_text(self, message: SpeakV1Text) -> None: """ self._send_model(message) - def send_flush(self, message: typing.Optional[SpeakV1Flush] = None) -> None: + def send_flush(self, message: SpeakV1Flush) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Flush. """ - self._send_model(message or SpeakV1Flush(type="Flush")) + self._send_model(message) - def send_clear(self, message: typing.Optional[SpeakV1Clear] = None) -> None: + def send_clear(self, message: SpeakV1Clear) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Clear. """ - self._send_model(message or SpeakV1Clear(type="Clear")) + self._send_model(message) - def send_close(self, message: typing.Optional[SpeakV1Close] = None) -> None: + def send_close(self, message: SpeakV1Close) -> None: """ Send a message to the websocket connection. The message will be sent as a SpeakV1Close. """ - self._send_model(message or SpeakV1Close(type="Close")) + self._send_model(message) def recv(self) -> V1SocketClientResponse: """ @@ -186,7 +216,11 @@ def recv(self) -> V1SocketClientResponse: if isinstance(data, bytes): return data # type: ignore json_data = json.loads(data) - return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + try: + return construct_type(type_=V1SocketClientResponse, object_=json_data) # type: ignore + except Exception: + _logger.warning("Skipping unknown WebSocket message; update your SDK version to support new message types.") + return json_data # type: ignore def _send(self, data: typing.Any) -> None: """ diff --git a/src/deepgram/speak/v1/types/__init__.py b/src/deepgram/speak/v1/types/__init__.py index 72a25d1b..cc5ee645 100644 --- a/src/deepgram/speak/v1/types/__init__.py +++ b/src/deepgram/speak/v1/types/__init__.py @@ -17,8 +17,11 @@ from .speak_v1flushed import SpeakV1Flushed from .speak_v1flushed_type import SpeakV1FlushedType from .speak_v1metadata import SpeakV1Metadata + from .speak_v1metadata_type import SpeakV1MetadataType from .speak_v1text import SpeakV1Text + from .speak_v1text_type import SpeakV1TextType from .speak_v1warning import SpeakV1Warning + from .speak_v1warning_type import SpeakV1WarningType _dynamic_imports: typing.Dict[str, str] = { "SpeakV1Clear": ".speak_v1clear", "SpeakV1ClearType": ".speak_v1clear_type", @@ -31,8 +34,11 @@ "SpeakV1Flushed": ".speak_v1flushed", "SpeakV1FlushedType": ".speak_v1flushed_type", "SpeakV1Metadata": ".speak_v1metadata", + "SpeakV1MetadataType": ".speak_v1metadata_type", "SpeakV1Text": ".speak_v1text", + "SpeakV1TextType": ".speak_v1text_type", "SpeakV1Warning": ".speak_v1warning", + "SpeakV1WarningType": ".speak_v1warning_type", } @@ -69,6 +75,9 @@ def __dir__(): "SpeakV1Flushed", "SpeakV1FlushedType", "SpeakV1Metadata", + "SpeakV1MetadataType", "SpeakV1Text", + "SpeakV1TextType", "SpeakV1Warning", + "SpeakV1WarningType", ] diff --git a/src/deepgram/speak/v1/types/speak_v1metadata.py b/src/deepgram/speak/v1/types/speak_v1metadata.py index e009a764..37ad50a8 100644 --- a/src/deepgram/speak/v1/types/speak_v1metadata.py +++ b/src/deepgram/speak/v1/types/speak_v1metadata.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .speak_v1metadata_type import SpeakV1MetadataType class SpeakV1Metadata(UncheckedBaseModel): - type: typing.Literal["Metadata"] = pydantic.Field(default="Metadata") + type: SpeakV1MetadataType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/speak/v1/types/speak_v1metadata_type.py b/src/deepgram/speak/v1/types/speak_v1metadata_type.py new file mode 100644 index 00000000..70b2b693 --- /dev/null +++ b/src/deepgram/speak/v1/types/speak_v1metadata_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +SpeakV1MetadataType = typing.Union[typing.Literal["Metadata"], typing.Any] diff --git a/src/deepgram/speak/v1/types/speak_v1text.py b/src/deepgram/speak/v1/types/speak_v1text.py index 3a566770..89f1a445 100644 --- a/src/deepgram/speak/v1/types/speak_v1text.py +++ b/src/deepgram/speak/v1/types/speak_v1text.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .speak_v1text_type import SpeakV1TextType class SpeakV1Text(UncheckedBaseModel): - type: typing.Literal["Speak"] = pydantic.Field(default="Speak") + type: SpeakV1TextType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/speak/v1/types/speak_v1text_type.py b/src/deepgram/speak/v1/types/speak_v1text_type.py new file mode 100644 index 00000000..f6b0345f --- /dev/null +++ b/src/deepgram/speak/v1/types/speak_v1text_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +SpeakV1TextType = typing.Union[typing.Literal["Speak"], typing.Any] diff --git a/src/deepgram/speak/v1/types/speak_v1warning.py b/src/deepgram/speak/v1/types/speak_v1warning.py index af374191..2219f352 100644 --- a/src/deepgram/speak/v1/types/speak_v1warning.py +++ b/src/deepgram/speak/v1/types/speak_v1warning.py @@ -5,10 +5,11 @@ import pydantic from ....core.pydantic_utilities import IS_PYDANTIC_V2 from ....core.unchecked_base_model import UncheckedBaseModel +from .speak_v1warning_type import SpeakV1WarningType class SpeakV1Warning(UncheckedBaseModel): - type: typing.Literal["Warning"] = pydantic.Field(default="Warning") + type: SpeakV1WarningType = pydantic.Field() """ Message type identifier """ diff --git a/src/deepgram/speak/v1/types/speak_v1warning_type.py b/src/deepgram/speak/v1/types/speak_v1warning_type.py new file mode 100644 index 00000000..7d0ede31 --- /dev/null +++ b/src/deepgram/speak/v1/types/speak_v1warning_type.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +SpeakV1WarningType = typing.Union[typing.Literal["Warning"], typing.Any] diff --git a/src/deepgram/types/__init__.py b/src/deepgram/types/__init__.py index da41b0d4..291d136f 100644 --- a/src/deepgram/types/__init__.py +++ b/src/deepgram/types/__init__.py @@ -9,31 +9,31 @@ from .agent_think_models_v1response import AgentThinkModelsV1Response from .agent_think_models_v1response_models_item import AgentThinkModelsV1ResponseModelsItem from .agent_think_models_v1response_models_item_id import AgentThinkModelsV1ResponseModelsItemId + from .agent_think_models_v1response_models_item_id_provider import AgentThinkModelsV1ResponseModelsItemIdProvider from .agent_think_models_v1response_models_item_one import AgentThinkModelsV1ResponseModelsItemOne from .agent_think_models_v1response_models_item_one_id import AgentThinkModelsV1ResponseModelsItemOneId + from .agent_think_models_v1response_models_item_one_provider import AgentThinkModelsV1ResponseModelsItemOneProvider from .agent_think_models_v1response_models_item_three import AgentThinkModelsV1ResponseModelsItemThree + from .agent_think_models_v1response_models_item_three_id import AgentThinkModelsV1ResponseModelsItemThreeId + from .agent_think_models_v1response_models_item_three_provider import ( + AgentThinkModelsV1ResponseModelsItemThreeProvider, + ) from .agent_think_models_v1response_models_item_two import AgentThinkModelsV1ResponseModelsItemTwo from .agent_think_models_v1response_models_item_two_id import AgentThinkModelsV1ResponseModelsItemTwoId + from .agent_think_models_v1response_models_item_two_provider import AgentThinkModelsV1ResponseModelsItemTwoProvider from .agent_think_models_v1response_models_item_zero import AgentThinkModelsV1ResponseModelsItemZero from .agent_think_models_v1response_models_item_zero_id import AgentThinkModelsV1ResponseModelsItemZeroId + from .agent_think_models_v1response_models_item_zero_provider import ( + AgentThinkModelsV1ResponseModelsItemZeroProvider, + ) from .anthropic import Anthropic - from .anthropic_think_provider_model import AnthropicThinkProviderModel from .aws_bedrock_think_provider import AwsBedrockThinkProvider - from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentials - from .aws_bedrock_think_provider_credentials_type import AwsBedrockThinkProviderCredentialsType - from .aws_bedrock_think_provider_model import AwsBedrockThinkProviderModel from .aws_polly_speak_provider import AwsPollySpeakProvider - from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentials - from .aws_polly_speak_provider_credentials_type import AwsPollySpeakProviderCredentialsType - from .aws_polly_speak_provider_engine import AwsPollySpeakProviderEngine - from .aws_polly_speak_provider_voice import AwsPollySpeakProviderVoice from .billing_breakdown_v1response import BillingBreakdownV1Response from .billing_breakdown_v1response_resolution import BillingBreakdownV1ResponseResolution from .billing_breakdown_v1response_results_item import BillingBreakdownV1ResponseResultsItem from .billing_breakdown_v1response_results_item_grouping import BillingBreakdownV1ResponseResultsItemGrouping from .cartesia import Cartesia - from .cartesia_speak_provider_model_id import CartesiaSpeakProviderModelId - from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoice from .create_key_v1request_one import CreateKeyV1RequestOne from .create_key_v1response import CreateKeyV1Response from .create_project_distribution_credentials_v1response import CreateProjectDistributionCredentialsV1Response @@ -45,13 +45,11 @@ ) from .create_project_invite_v1response import CreateProjectInviteV1Response from .deepgram import Deepgram - from .deepgram_speak_provider_model import DeepgramSpeakProviderModel from .delete_project_invite_v1response import DeleteProjectInviteV1Response from .delete_project_key_v1response import DeleteProjectKeyV1Response from .delete_project_member_v1response import DeleteProjectMemberV1Response from .delete_project_v1response import DeleteProjectV1Response from .eleven_labs_speak_provider import ElevenLabsSpeakProvider - from .eleven_labs_speak_provider_model_id import ElevenLabsSpeakProviderModelId from .error_response import ErrorResponse from .error_response_legacy_error import ErrorResponseLegacyError from .error_response_modern_error import ErrorResponseModernError @@ -75,7 +73,6 @@ from .get_project_request_v1response import GetProjectRequestV1Response from .get_project_v1response import GetProjectV1Response from .google import Google - from .google_think_provider_model import GoogleThinkProviderModel from .grant_v1response import GrantV1Response from .groq import Groq from .leave_project_v1response import LeaveProjectV1Response @@ -192,10 +189,7 @@ from .listen_v2sample_rate import ListenV2SampleRate from .listen_v2tag import ListenV2Tag from .open_ai_speak_provider import OpenAiSpeakProvider - from .open_ai_speak_provider_model import OpenAiSpeakProviderModel - from .open_ai_speak_provider_voice import OpenAiSpeakProviderVoice from .open_ai_think_provider import OpenAiThinkProvider - from .open_ai_think_provider_model import OpenAiThinkProviderModel from .project_request_response import ProjectRequestResponse from .read_v1request import ReadV1Request from .read_v1request_text import ReadV1RequestText @@ -227,33 +221,12 @@ from .shared_topics_results_topics_segments_item import SharedTopicsResultsTopicsSegmentsItem from .shared_topics_results_topics_segments_item_topics_item import SharedTopicsResultsTopicsSegmentsItemTopicsItem from .speak_settings_v1 import SpeakSettingsV1 - from .speak_settings_v1endpoint import SpeakSettingsV1Endpoint - from .speak_settings_v1provider import ( - SpeakSettingsV1Provider, - SpeakSettingsV1Provider_AwsPolly, - SpeakSettingsV1Provider_Cartesia, - SpeakSettingsV1Provider_Deepgram, - SpeakSettingsV1Provider_ElevenLabs, - SpeakSettingsV1Provider_OpenAi, - ) from .speak_v1encoding import SpeakV1Encoding from .speak_v1mip_opt_out import SpeakV1MipOptOut from .speak_v1model import SpeakV1Model from .speak_v1response import SpeakV1Response from .speak_v1sample_rate import SpeakV1SampleRate from .think_settings_v1 import ThinkSettingsV1 - from .think_settings_v1context_length import ThinkSettingsV1ContextLength - from .think_settings_v1endpoint import ThinkSettingsV1Endpoint - from .think_settings_v1functions_item import ThinkSettingsV1FunctionsItem - from .think_settings_v1functions_item_endpoint import ThinkSettingsV1FunctionsItemEndpoint - from .think_settings_v1provider import ( - ThinkSettingsV1Provider, - ThinkSettingsV1Provider_Anthropic, - ThinkSettingsV1Provider_AwsBedrock, - ThinkSettingsV1Provider_Google, - ThinkSettingsV1Provider_Groq, - ThinkSettingsV1Provider_OpenAi, - ) from .update_project_member_scopes_v1response import UpdateProjectMemberScopesV1Response from .update_project_v1response import UpdateProjectV1Response from .usage_breakdown_v1response import UsageBreakdownV1Response @@ -268,31 +241,27 @@ "AgentThinkModelsV1Response": ".agent_think_models_v1response", "AgentThinkModelsV1ResponseModelsItem": ".agent_think_models_v1response_models_item", "AgentThinkModelsV1ResponseModelsItemId": ".agent_think_models_v1response_models_item_id", + "AgentThinkModelsV1ResponseModelsItemIdProvider": ".agent_think_models_v1response_models_item_id_provider", "AgentThinkModelsV1ResponseModelsItemOne": ".agent_think_models_v1response_models_item_one", "AgentThinkModelsV1ResponseModelsItemOneId": ".agent_think_models_v1response_models_item_one_id", + "AgentThinkModelsV1ResponseModelsItemOneProvider": ".agent_think_models_v1response_models_item_one_provider", "AgentThinkModelsV1ResponseModelsItemThree": ".agent_think_models_v1response_models_item_three", + "AgentThinkModelsV1ResponseModelsItemThreeId": ".agent_think_models_v1response_models_item_three_id", + "AgentThinkModelsV1ResponseModelsItemThreeProvider": ".agent_think_models_v1response_models_item_three_provider", "AgentThinkModelsV1ResponseModelsItemTwo": ".agent_think_models_v1response_models_item_two", "AgentThinkModelsV1ResponseModelsItemTwoId": ".agent_think_models_v1response_models_item_two_id", + "AgentThinkModelsV1ResponseModelsItemTwoProvider": ".agent_think_models_v1response_models_item_two_provider", "AgentThinkModelsV1ResponseModelsItemZero": ".agent_think_models_v1response_models_item_zero", "AgentThinkModelsV1ResponseModelsItemZeroId": ".agent_think_models_v1response_models_item_zero_id", + "AgentThinkModelsV1ResponseModelsItemZeroProvider": ".agent_think_models_v1response_models_item_zero_provider", "Anthropic": ".anthropic", - "AnthropicThinkProviderModel": ".anthropic_think_provider_model", "AwsBedrockThinkProvider": ".aws_bedrock_think_provider", - "AwsBedrockThinkProviderCredentials": ".aws_bedrock_think_provider_credentials", - "AwsBedrockThinkProviderCredentialsType": ".aws_bedrock_think_provider_credentials_type", - "AwsBedrockThinkProviderModel": ".aws_bedrock_think_provider_model", "AwsPollySpeakProvider": ".aws_polly_speak_provider", - "AwsPollySpeakProviderCredentials": ".aws_polly_speak_provider_credentials", - "AwsPollySpeakProviderCredentialsType": ".aws_polly_speak_provider_credentials_type", - "AwsPollySpeakProviderEngine": ".aws_polly_speak_provider_engine", - "AwsPollySpeakProviderVoice": ".aws_polly_speak_provider_voice", "BillingBreakdownV1Response": ".billing_breakdown_v1response", "BillingBreakdownV1ResponseResolution": ".billing_breakdown_v1response_resolution", "BillingBreakdownV1ResponseResultsItem": ".billing_breakdown_v1response_results_item", "BillingBreakdownV1ResponseResultsItemGrouping": ".billing_breakdown_v1response_results_item_grouping", "Cartesia": ".cartesia", - "CartesiaSpeakProviderModelId": ".cartesia_speak_provider_model_id", - "CartesiaSpeakProviderVoice": ".cartesia_speak_provider_voice", "CreateKeyV1RequestOne": ".create_key_v1request_one", "CreateKeyV1Response": ".create_key_v1response", "CreateProjectDistributionCredentialsV1Response": ".create_project_distribution_credentials_v1response", @@ -300,13 +269,11 @@ "CreateProjectDistributionCredentialsV1ResponseMember": ".create_project_distribution_credentials_v1response_member", "CreateProjectInviteV1Response": ".create_project_invite_v1response", "Deepgram": ".deepgram", - "DeepgramSpeakProviderModel": ".deepgram_speak_provider_model", "DeleteProjectInviteV1Response": ".delete_project_invite_v1response", "DeleteProjectKeyV1Response": ".delete_project_key_v1response", "DeleteProjectMemberV1Response": ".delete_project_member_v1response", "DeleteProjectV1Response": ".delete_project_v1response", "ElevenLabsSpeakProvider": ".eleven_labs_speak_provider", - "ElevenLabsSpeakProviderModelId": ".eleven_labs_speak_provider_model_id", "ErrorResponse": ".error_response", "ErrorResponseLegacyError": ".error_response_legacy_error", "ErrorResponseModernError": ".error_response_modern_error", @@ -326,7 +293,6 @@ "GetProjectRequestV1Response": ".get_project_request_v1response", "GetProjectV1Response": ".get_project_v1response", "Google": ".google", - "GoogleThinkProviderModel": ".google_think_provider_model", "GrantV1Response": ".grant_v1response", "Groq": ".groq", "LeaveProjectV1Response": ".leave_project_v1response", @@ -419,10 +385,7 @@ "ListenV2SampleRate": ".listen_v2sample_rate", "ListenV2Tag": ".listen_v2tag", "OpenAiSpeakProvider": ".open_ai_speak_provider", - "OpenAiSpeakProviderModel": ".open_ai_speak_provider_model", - "OpenAiSpeakProviderVoice": ".open_ai_speak_provider_voice", "OpenAiThinkProvider": ".open_ai_think_provider", - "OpenAiThinkProviderModel": ".open_ai_think_provider_model", "ProjectRequestResponse": ".project_request_response", "ReadV1Request": ".read_v1request", "ReadV1RequestText": ".read_v1request_text", @@ -452,29 +415,12 @@ "SharedTopicsResultsTopicsSegmentsItem": ".shared_topics_results_topics_segments_item", "SharedTopicsResultsTopicsSegmentsItemTopicsItem": ".shared_topics_results_topics_segments_item_topics_item", "SpeakSettingsV1": ".speak_settings_v1", - "SpeakSettingsV1Endpoint": ".speak_settings_v1endpoint", - "SpeakSettingsV1Provider": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_AwsPolly": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_Cartesia": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_Deepgram": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_ElevenLabs": ".speak_settings_v1provider", - "SpeakSettingsV1Provider_OpenAi": ".speak_settings_v1provider", "SpeakV1Encoding": ".speak_v1encoding", "SpeakV1MipOptOut": ".speak_v1mip_opt_out", "SpeakV1Model": ".speak_v1model", "SpeakV1Response": ".speak_v1response", "SpeakV1SampleRate": ".speak_v1sample_rate", "ThinkSettingsV1": ".think_settings_v1", - "ThinkSettingsV1ContextLength": ".think_settings_v1context_length", - "ThinkSettingsV1Endpoint": ".think_settings_v1endpoint", - "ThinkSettingsV1FunctionsItem": ".think_settings_v1functions_item", - "ThinkSettingsV1FunctionsItemEndpoint": ".think_settings_v1functions_item_endpoint", - "ThinkSettingsV1Provider": ".think_settings_v1provider", - "ThinkSettingsV1Provider_Anthropic": ".think_settings_v1provider", - "ThinkSettingsV1Provider_AwsBedrock": ".think_settings_v1provider", - "ThinkSettingsV1Provider_Google": ".think_settings_v1provider", - "ThinkSettingsV1Provider_Groq": ".think_settings_v1provider", - "ThinkSettingsV1Provider_OpenAi": ".think_settings_v1provider", "UpdateProjectMemberScopesV1Response": ".update_project_member_scopes_v1response", "UpdateProjectV1Response": ".update_project_v1response", "UsageBreakdownV1Response": ".usage_breakdown_v1response", @@ -513,31 +459,27 @@ def __dir__(): "AgentThinkModelsV1Response", "AgentThinkModelsV1ResponseModelsItem", "AgentThinkModelsV1ResponseModelsItemId", + "AgentThinkModelsV1ResponseModelsItemIdProvider", "AgentThinkModelsV1ResponseModelsItemOne", "AgentThinkModelsV1ResponseModelsItemOneId", + "AgentThinkModelsV1ResponseModelsItemOneProvider", "AgentThinkModelsV1ResponseModelsItemThree", + "AgentThinkModelsV1ResponseModelsItemThreeId", + "AgentThinkModelsV1ResponseModelsItemThreeProvider", "AgentThinkModelsV1ResponseModelsItemTwo", "AgentThinkModelsV1ResponseModelsItemTwoId", + "AgentThinkModelsV1ResponseModelsItemTwoProvider", "AgentThinkModelsV1ResponseModelsItemZero", "AgentThinkModelsV1ResponseModelsItemZeroId", + "AgentThinkModelsV1ResponseModelsItemZeroProvider", "Anthropic", - "AnthropicThinkProviderModel", "AwsBedrockThinkProvider", - "AwsBedrockThinkProviderCredentials", - "AwsBedrockThinkProviderCredentialsType", - "AwsBedrockThinkProviderModel", "AwsPollySpeakProvider", - "AwsPollySpeakProviderCredentials", - "AwsPollySpeakProviderCredentialsType", - "AwsPollySpeakProviderEngine", - "AwsPollySpeakProviderVoice", "BillingBreakdownV1Response", "BillingBreakdownV1ResponseResolution", "BillingBreakdownV1ResponseResultsItem", "BillingBreakdownV1ResponseResultsItemGrouping", "Cartesia", - "CartesiaSpeakProviderModelId", - "CartesiaSpeakProviderVoice", "CreateKeyV1RequestOne", "CreateKeyV1Response", "CreateProjectDistributionCredentialsV1Response", @@ -545,13 +487,11 @@ def __dir__(): "CreateProjectDistributionCredentialsV1ResponseMember", "CreateProjectInviteV1Response", "Deepgram", - "DeepgramSpeakProviderModel", "DeleteProjectInviteV1Response", "DeleteProjectKeyV1Response", "DeleteProjectMemberV1Response", "DeleteProjectV1Response", "ElevenLabsSpeakProvider", - "ElevenLabsSpeakProviderModelId", "ErrorResponse", "ErrorResponseLegacyError", "ErrorResponseModernError", @@ -571,7 +511,6 @@ def __dir__(): "GetProjectRequestV1Response", "GetProjectV1Response", "Google", - "GoogleThinkProviderModel", "GrantV1Response", "Groq", "LeaveProjectV1Response", @@ -664,10 +603,7 @@ def __dir__(): "ListenV2SampleRate", "ListenV2Tag", "OpenAiSpeakProvider", - "OpenAiSpeakProviderModel", - "OpenAiSpeakProviderVoice", "OpenAiThinkProvider", - "OpenAiThinkProviderModel", "ProjectRequestResponse", "ReadV1Request", "ReadV1RequestText", @@ -697,29 +633,12 @@ def __dir__(): "SharedTopicsResultsTopicsSegmentsItem", "SharedTopicsResultsTopicsSegmentsItemTopicsItem", "SpeakSettingsV1", - "SpeakSettingsV1Endpoint", - "SpeakSettingsV1Provider", - "SpeakSettingsV1Provider_AwsPolly", - "SpeakSettingsV1Provider_Cartesia", - "SpeakSettingsV1Provider_Deepgram", - "SpeakSettingsV1Provider_ElevenLabs", - "SpeakSettingsV1Provider_OpenAi", "SpeakV1Encoding", "SpeakV1MipOptOut", "SpeakV1Model", "SpeakV1Response", "SpeakV1SampleRate", "ThinkSettingsV1", - "ThinkSettingsV1ContextLength", - "ThinkSettingsV1Endpoint", - "ThinkSettingsV1FunctionsItem", - "ThinkSettingsV1FunctionsItemEndpoint", - "ThinkSettingsV1Provider", - "ThinkSettingsV1Provider_Anthropic", - "ThinkSettingsV1Provider_AwsBedrock", - "ThinkSettingsV1Provider_Google", - "ThinkSettingsV1Provider_Groq", - "ThinkSettingsV1Provider_OpenAi", "UpdateProjectMemberScopesV1Response", "UpdateProjectV1Response", "UsageBreakdownV1Response", diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_id.py b/src/deepgram/types/agent_think_models_v1response_models_item_id.py index e203c954..786ef2be 100644 --- a/src/deepgram/types/agent_think_models_v1response_models_item_id.py +++ b/src/deepgram/types/agent_think_models_v1response_models_item_id.py @@ -5,6 +5,7 @@ import pydantic from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel +from .agent_think_models_v1response_models_item_id_provider import AgentThinkModelsV1ResponseModelsItemIdProvider class AgentThinkModelsV1ResponseModelsItemId(UncheckedBaseModel): @@ -22,7 +23,7 @@ class AgentThinkModelsV1ResponseModelsItemId(UncheckedBaseModel): The display name of the model """ - provider: typing.Literal["aws_bedrock"] = pydantic.Field(default="aws_bedrock") + provider: AgentThinkModelsV1ResponseModelsItemIdProvider = pydantic.Field() """ The provider of the model """ diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_id_provider.py b/src/deepgram/types/agent_think_models_v1response_models_item_id_provider.py new file mode 100644 index 00000000..51e56a22 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_id_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemIdProvider = typing.Union[typing.Literal["aws_bedrock"], typing.Any] diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_one.py b/src/deepgram/types/agent_think_models_v1response_models_item_one.py index 63981a6d..0aac30e8 100644 --- a/src/deepgram/types/agent_think_models_v1response_models_item_one.py +++ b/src/deepgram/types/agent_think_models_v1response_models_item_one.py @@ -6,6 +6,7 @@ from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel from .agent_think_models_v1response_models_item_one_id import AgentThinkModelsV1ResponseModelsItemOneId +from .agent_think_models_v1response_models_item_one_provider import AgentThinkModelsV1ResponseModelsItemOneProvider class AgentThinkModelsV1ResponseModelsItemOne(UncheckedBaseModel): @@ -23,7 +24,7 @@ class AgentThinkModelsV1ResponseModelsItemOne(UncheckedBaseModel): The display name of the model """ - provider: typing.Literal["anthropic"] = pydantic.Field(default="anthropic") + provider: AgentThinkModelsV1ResponseModelsItemOneProvider = pydantic.Field() """ The provider of the model """ diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_one_provider.py b/src/deepgram/types/agent_think_models_v1response_models_item_one_provider.py new file mode 100644 index 00000000..f00753d1 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_one_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemOneProvider = typing.Union[typing.Literal["anthropic"], typing.Any] diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_three.py b/src/deepgram/types/agent_think_models_v1response_models_item_three.py index 2e8118f9..391eca35 100644 --- a/src/deepgram/types/agent_think_models_v1response_models_item_three.py +++ b/src/deepgram/types/agent_think_models_v1response_models_item_three.py @@ -5,6 +5,8 @@ import pydantic from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel +from .agent_think_models_v1response_models_item_three_id import AgentThinkModelsV1ResponseModelsItemThreeId +from .agent_think_models_v1response_models_item_three_provider import AgentThinkModelsV1ResponseModelsItemThreeProvider class AgentThinkModelsV1ResponseModelsItemThree(UncheckedBaseModel): @@ -12,7 +14,7 @@ class AgentThinkModelsV1ResponseModelsItemThree(UncheckedBaseModel): Groq models """ - id: typing.Literal["openai/gpt-oss-20b"] = pydantic.Field(default="openai/gpt-oss-20b") + id: AgentThinkModelsV1ResponseModelsItemThreeId = pydantic.Field() """ The unique identifier of the Groq model """ @@ -22,7 +24,7 @@ class AgentThinkModelsV1ResponseModelsItemThree(UncheckedBaseModel): The display name of the model """ - provider: typing.Literal["groq"] = pydantic.Field(default="groq") + provider: AgentThinkModelsV1ResponseModelsItemThreeProvider = pydantic.Field() """ The provider of the model """ diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_three_id.py b/src/deepgram/types/agent_think_models_v1response_models_item_three_id.py new file mode 100644 index 00000000..c318c675 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_three_id.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemThreeId = typing.Union[typing.Literal["openai/gpt-oss-20b"], typing.Any] diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_three_provider.py b/src/deepgram/types/agent_think_models_v1response_models_item_three_provider.py new file mode 100644 index 00000000..0bf80206 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_three_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemThreeProvider = typing.Union[typing.Literal["groq"], typing.Any] diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_two.py b/src/deepgram/types/agent_think_models_v1response_models_item_two.py index 97e3f604..c38a46dc 100644 --- a/src/deepgram/types/agent_think_models_v1response_models_item_two.py +++ b/src/deepgram/types/agent_think_models_v1response_models_item_two.py @@ -6,6 +6,7 @@ from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel from .agent_think_models_v1response_models_item_two_id import AgentThinkModelsV1ResponseModelsItemTwoId +from .agent_think_models_v1response_models_item_two_provider import AgentThinkModelsV1ResponseModelsItemTwoProvider class AgentThinkModelsV1ResponseModelsItemTwo(UncheckedBaseModel): @@ -23,7 +24,7 @@ class AgentThinkModelsV1ResponseModelsItemTwo(UncheckedBaseModel): The display name of the model """ - provider: typing.Literal["google"] = pydantic.Field(default="google") + provider: AgentThinkModelsV1ResponseModelsItemTwoProvider = pydantic.Field() """ The provider of the model """ diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_two_provider.py b/src/deepgram/types/agent_think_models_v1response_models_item_two_provider.py new file mode 100644 index 00000000..4c660732 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_two_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemTwoProvider = typing.Union[typing.Literal["google"], typing.Any] diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_zero.py b/src/deepgram/types/agent_think_models_v1response_models_item_zero.py index 28850db2..543f6a93 100644 --- a/src/deepgram/types/agent_think_models_v1response_models_item_zero.py +++ b/src/deepgram/types/agent_think_models_v1response_models_item_zero.py @@ -6,6 +6,7 @@ from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel from .agent_think_models_v1response_models_item_zero_id import AgentThinkModelsV1ResponseModelsItemZeroId +from .agent_think_models_v1response_models_item_zero_provider import AgentThinkModelsV1ResponseModelsItemZeroProvider class AgentThinkModelsV1ResponseModelsItemZero(UncheckedBaseModel): @@ -23,7 +24,7 @@ class AgentThinkModelsV1ResponseModelsItemZero(UncheckedBaseModel): The display name of the model """ - provider: typing.Literal["open_ai"] = pydantic.Field(default="open_ai") + provider: AgentThinkModelsV1ResponseModelsItemZeroProvider = pydantic.Field() """ The provider of the model """ diff --git a/src/deepgram/types/agent_think_models_v1response_models_item_zero_provider.py b/src/deepgram/types/agent_think_models_v1response_models_item_zero_provider.py new file mode 100644 index 00000000..0b4e2034 --- /dev/null +++ b/src/deepgram/types/agent_think_models_v1response_models_item_zero_provider.py @@ -0,0 +1,5 @@ +# This file was auto-generated by Fern from our API Definition. + +import typing + +AgentThinkModelsV1ResponseModelsItemZeroProvider = typing.Union[typing.Literal["open_ai"], typing.Any] diff --git a/src/deepgram/types/anthropic.py b/src/deepgram/types/anthropic.py index c972254e..659b8cf8 100644 --- a/src/deepgram/types/anthropic.py +++ b/src/deepgram/types/anthropic.py @@ -2,34 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .anthropic_think_provider_model import AnthropicThinkProviderModel - - -class Anthropic(UncheckedBaseModel): - type: typing.Literal["anthropic"] = "anthropic" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the Anthropic Messages API - """ - - model: AnthropicThinkProviderModel = pydantic.Field() - """ - Anthropic model to use - """ - - temperature: typing.Optional[float] = pydantic.Field(default=None) - """ - Anthropic temperature (0-1) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +Anthropic = typing.Any diff --git a/src/deepgram/types/anthropic_think_provider_model.py b/src/deepgram/types/anthropic_think_provider_model.py deleted file mode 100644 index 7c78238f..00000000 --- a/src/deepgram/types/anthropic_think_provider_model.py +++ /dev/null @@ -1,7 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -AnthropicThinkProviderModel = typing.Union[ - typing.Literal["claude-3-5-haiku-latest", "claude-sonnet-4-20250514"], typing.Any -] diff --git a/src/deepgram/types/aws_bedrock_think_provider.py b/src/deepgram/types/aws_bedrock_think_provider.py index a601a278..4cd834f4 100644 --- a/src/deepgram/types/aws_bedrock_think_provider.py +++ b/src/deepgram/types/aws_bedrock_think_provider.py @@ -2,35 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentials -from .aws_bedrock_think_provider_model import AwsBedrockThinkProviderModel - - -class AwsBedrockThinkProvider(UncheckedBaseModel): - type: typing.Literal["aws_bedrock"] = "aws_bedrock" - model: AwsBedrockThinkProviderModel = pydantic.Field() - """ - AWS Bedrock model to use - """ - - temperature: typing.Optional[float] = pydantic.Field(default=None) - """ - AWS Bedrock temperature (0-2) - """ - - credentials: typing.Optional[AwsBedrockThinkProviderCredentials] = pydantic.Field(default=None) - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +AwsBedrockThinkProvider = typing.Any diff --git a/src/deepgram/types/aws_bedrock_think_provider_credentials.py b/src/deepgram/types/aws_bedrock_think_provider_credentials.py deleted file mode 100644 index c014d379..00000000 --- a/src/deepgram/types/aws_bedrock_think_provider_credentials.py +++ /dev/null @@ -1,48 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .aws_bedrock_think_provider_credentials_type import AwsBedrockThinkProviderCredentialsType - - -class AwsBedrockThinkProviderCredentials(UncheckedBaseModel): - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ - - type: typing.Optional[AwsBedrockThinkProviderCredentialsType] = pydantic.Field(default=None) - """ - AWS credentials type (STS short-lived or IAM long-lived) - """ - - region: typing.Optional[str] = pydantic.Field(default=None) - """ - AWS region - """ - - access_key_id: typing.Optional[str] = pydantic.Field(default=None) - """ - AWS access key - """ - - secret_access_key: typing.Optional[str] = pydantic.Field(default=None) - """ - AWS secret access key - """ - - session_token: typing.Optional[str] = pydantic.Field(default=None) - """ - AWS session token (required for STS only) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow diff --git a/src/deepgram/types/aws_bedrock_think_provider_credentials_type.py b/src/deepgram/types/aws_bedrock_think_provider_credentials_type.py deleted file mode 100644 index df30845b..00000000 --- a/src/deepgram/types/aws_bedrock_think_provider_credentials_type.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -AwsBedrockThinkProviderCredentialsType = typing.Union[typing.Literal["sts", "iam"], typing.Any] diff --git a/src/deepgram/types/aws_bedrock_think_provider_model.py b/src/deepgram/types/aws_bedrock_think_provider_model.py deleted file mode 100644 index 1b816d93..00000000 --- a/src/deepgram/types/aws_bedrock_think_provider_model.py +++ /dev/null @@ -1,7 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -AwsBedrockThinkProviderModel = typing.Union[ - typing.Literal["anthropic/claude-3-5-sonnet-20240620-v1:0", "anthropic/claude-3-5-haiku-20240307-v1:0"], typing.Any -] diff --git a/src/deepgram/types/aws_polly_speak_provider.py b/src/deepgram/types/aws_polly_speak_provider.py index 9dfa46bd..4c9020dc 100644 --- a/src/deepgram/types/aws_polly_speak_provider.py +++ b/src/deepgram/types/aws_polly_speak_provider.py @@ -2,39 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentials -from .aws_polly_speak_provider_engine import AwsPollySpeakProviderEngine -from .aws_polly_speak_provider_voice import AwsPollySpeakProviderVoice - - -class AwsPollySpeakProvider(UncheckedBaseModel): - type: typing.Literal["aws_polly"] = "aws_polly" - voice: AwsPollySpeakProviderVoice = pydantic.Field() - """ - AWS Polly voice name - """ - - language: str = pydantic.Field() - """ - Language code to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the AWS Polly API - """ - - language_code: typing.Optional[str] = pydantic.Field(default=None) - """ - Use the `language` field instead. - """ - - engine: AwsPollySpeakProviderEngine - credentials: AwsPollySpeakProviderCredentials - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +AwsPollySpeakProvider = typing.Any diff --git a/src/deepgram/types/aws_polly_speak_provider_credentials_type.py b/src/deepgram/types/aws_polly_speak_provider_credentials_type.py deleted file mode 100644 index 0f1785cf..00000000 --- a/src/deepgram/types/aws_polly_speak_provider_credentials_type.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -AwsPollySpeakProviderCredentialsType = typing.Union[typing.Literal["sts", "iam"], typing.Any] diff --git a/src/deepgram/types/aws_polly_speak_provider_engine.py b/src/deepgram/types/aws_polly_speak_provider_engine.py deleted file mode 100644 index 4f42ea6d..00000000 --- a/src/deepgram/types/aws_polly_speak_provider_engine.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -AwsPollySpeakProviderEngine = typing.Union[typing.Literal["generative", "long-form", "standard", "neural"], typing.Any] diff --git a/src/deepgram/types/cartesia.py b/src/deepgram/types/cartesia.py index bb6778bd..483f6949 100644 --- a/src/deepgram/types/cartesia.py +++ b/src/deepgram/types/cartesia.py @@ -2,36 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .cartesia_speak_provider_model_id import CartesiaSpeakProviderModelId -from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoice - - -class Cartesia(UncheckedBaseModel): - type: typing.Literal["cartesia"] = "cartesia" - version: typing.Optional[typing.Literal["2025-03-17"]] = pydantic.Field(default=None) - """ - The API version header for the Cartesia text-to-speech API - """ - - model_id: CartesiaSpeakProviderModelId = pydantic.Field() - """ - Cartesia model ID - """ - - voice: CartesiaSpeakProviderVoice - language: typing.Optional[str] = pydantic.Field(default=None) - """ - Cartesia language code - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +Cartesia = typing.Any diff --git a/src/deepgram/types/cartesia_speak_provider_model_id.py b/src/deepgram/types/cartesia_speak_provider_model_id.py deleted file mode 100644 index f8a9d7e9..00000000 --- a/src/deepgram/types/cartesia_speak_provider_model_id.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -CartesiaSpeakProviderModelId = typing.Union[typing.Literal["sonic-2", "sonic-multilingual"], typing.Any] diff --git a/src/deepgram/types/deepgram.py b/src/deepgram/types/deepgram.py index e3cd5f74..b18e81bd 100644 --- a/src/deepgram/types/deepgram.py +++ b/src/deepgram/types/deepgram.py @@ -2,29 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .deepgram_speak_provider_model import DeepgramSpeakProviderModel - - -class Deepgram(UncheckedBaseModel): - type: typing.Literal["deepgram"] = "deepgram" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the Deepgram text-to-speech API - """ - - model: DeepgramSpeakProviderModel = pydantic.Field() - """ - Deepgram TTS model - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +Deepgram = typing.Any diff --git a/src/deepgram/types/eleven_labs_speak_provider.py b/src/deepgram/types/eleven_labs_speak_provider.py index 070d7423..606028cc 100644 --- a/src/deepgram/types/eleven_labs_speak_provider.py +++ b/src/deepgram/types/eleven_labs_speak_provider.py @@ -2,39 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .eleven_labs_speak_provider_model_id import ElevenLabsSpeakProviderModelId - - -class ElevenLabsSpeakProvider(UncheckedBaseModel): - type: typing.Literal["eleven_labs"] = "eleven_labs" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the ElevenLabs text-to-speech API - """ - - model_id: ElevenLabsSpeakProviderModelId = pydantic.Field() - """ - Eleven Labs model ID - """ - - language: typing.Optional[str] = pydantic.Field(default=None) - """ - Optional language to use, e.g. 'en-US'. Corresponds to the `language_code` parameter in the ElevenLabs API - """ - - language_code: typing.Optional[str] = pydantic.Field(default=None) - """ - Use the `language` field instead. - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +ElevenLabsSpeakProvider = typing.Any diff --git a/src/deepgram/types/google.py b/src/deepgram/types/google.py index 56e86997..463900ac 100644 --- a/src/deepgram/types/google.py +++ b/src/deepgram/types/google.py @@ -2,34 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .google_think_provider_model import GoogleThinkProviderModel - - -class Google(UncheckedBaseModel): - type: typing.Literal["google"] = "google" - version: typing.Optional[typing.Literal["v1beta"]] = pydantic.Field(default=None) - """ - The REST API version for the Google generative language API - """ - - model: GoogleThinkProviderModel = pydantic.Field() - """ - Google model to use - """ - - temperature: typing.Optional[float] = pydantic.Field(default=None) - """ - Google temperature (0-2) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +Google = typing.Any diff --git a/src/deepgram/types/google_think_provider_model.py b/src/deepgram/types/google_think_provider_model.py deleted file mode 100644 index a6061bac..00000000 --- a/src/deepgram/types/google_think_provider_model.py +++ /dev/null @@ -1,7 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -GoogleThinkProviderModel = typing.Union[ - typing.Literal["gemini-2.0-flash", "gemini-2.0-flash-lite", "gemini-2.5-flash"], typing.Any -] diff --git a/src/deepgram/types/groq.py b/src/deepgram/types/groq.py index b04ae473..0df2d3c2 100644 --- a/src/deepgram/types/groq.py +++ b/src/deepgram/types/groq.py @@ -2,33 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel - - -class Groq(UncheckedBaseModel): - type: typing.Literal["groq"] = "groq" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the Groq's chat completions API (mostly OpenAI-compatible) - """ - - model: typing.Literal["openai/gpt-oss-20b"] = pydantic.Field(default="openai/gpt-oss-20b") - """ - Groq model to use - """ - - temperature: typing.Optional[float] = pydantic.Field(default=None) - """ - Groq temperature (0-2) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +Groq = typing.Any diff --git a/src/deepgram/types/listen_v1redact.py b/src/deepgram/types/listen_v1redact.py index 9187906c..7bf572f3 100644 --- a/src/deepgram/types/listen_v1redact.py +++ b/src/deepgram/types/listen_v1redact.py @@ -2,10 +2,6 @@ import typing -ListenV1Redact = typing.Optional[ - typing.Union[ - typing.Literal["true", "false", "pci", "numbers", "aggressive_numbers", "ssn"], - str, - typing.Sequence[str], - ] +ListenV1Redact = typing.Union[ + typing.Literal["true", "false", "pci", "numbers", "aggressive_numbers", "ssn"], typing.Any ] diff --git a/src/deepgram/types/listen_v2model.py b/src/deepgram/types/listen_v2model.py index a631e705..ad86649f 100644 --- a/src/deepgram/types/listen_v2model.py +++ b/src/deepgram/types/listen_v2model.py @@ -2,4 +2,4 @@ import typing -ListenV2Model = typing.Literal["flux-general-en"] +ListenV2Model = typing.Union[typing.Literal["flux-general-en"], typing.Any] diff --git a/src/deepgram/types/open_ai_speak_provider.py b/src/deepgram/types/open_ai_speak_provider.py index 204c2434..e21ee8f2 100644 --- a/src/deepgram/types/open_ai_speak_provider.py +++ b/src/deepgram/types/open_ai_speak_provider.py @@ -2,35 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .open_ai_speak_provider_model import OpenAiSpeakProviderModel -from .open_ai_speak_provider_voice import OpenAiSpeakProviderVoice - - -class OpenAiSpeakProvider(UncheckedBaseModel): - type: typing.Literal["open_ai"] = "open_ai" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the OpenAI text-to-speech API - """ - - model: OpenAiSpeakProviderModel = pydantic.Field() - """ - OpenAI TTS model - """ - - voice: OpenAiSpeakProviderVoice = pydantic.Field() - """ - OpenAI voice - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +OpenAiSpeakProvider = typing.Any diff --git a/src/deepgram/types/open_ai_speak_provider_model.py b/src/deepgram/types/open_ai_speak_provider_model.py deleted file mode 100644 index ed04c40f..00000000 --- a/src/deepgram/types/open_ai_speak_provider_model.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -OpenAiSpeakProviderModel = typing.Union[typing.Literal["tts-1", "tts-1-hd"], typing.Any] diff --git a/src/deepgram/types/open_ai_speak_provider_voice.py b/src/deepgram/types/open_ai_speak_provider_voice.py deleted file mode 100644 index 3d7608a5..00000000 --- a/src/deepgram/types/open_ai_speak_provider_voice.py +++ /dev/null @@ -1,5 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -OpenAiSpeakProviderVoice = typing.Union[typing.Literal["alloy", "echo", "fable", "onyx", "nova", "shimmer"], typing.Any] diff --git a/src/deepgram/types/open_ai_think_provider.py b/src/deepgram/types/open_ai_think_provider.py index fd3e2601..fed929a1 100644 --- a/src/deepgram/types/open_ai_think_provider.py +++ b/src/deepgram/types/open_ai_think_provider.py @@ -2,34 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .open_ai_think_provider_model import OpenAiThinkProviderModel - - -class OpenAiThinkProvider(UncheckedBaseModel): - type: typing.Literal["open_ai"] = "open_ai" - version: typing.Optional[typing.Literal["v1"]] = pydantic.Field(default=None) - """ - The REST API version for the OpenAI chat completions API - """ - - model: OpenAiThinkProviderModel = pydantic.Field() - """ - OpenAI model to use - """ - - temperature: typing.Optional[float] = pydantic.Field(default=None) - """ - OpenAI temperature (0-2) - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +OpenAiThinkProvider = typing.Any diff --git a/src/deepgram/types/open_ai_think_provider_model.py b/src/deepgram/types/open_ai_think_provider_model.py deleted file mode 100644 index 80b24e48..00000000 --- a/src/deepgram/types/open_ai_think_provider_model.py +++ /dev/null @@ -1,10 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -import typing - -OpenAiThinkProviderModel = typing.Union[ - typing.Literal[ - "gpt-5", "gpt-5-mini", "gpt-5-nano", "gpt-4.1", "gpt-4.1-mini", "gpt-4.1-nano", "gpt-4o", "gpt-4o-mini" - ], - typing.Any, -] diff --git a/src/deepgram/types/speak_settings_v1.py b/src/deepgram/types/speak_settings_v1.py index e4c41a60..20ab1f5a 100644 --- a/src/deepgram/types/speak_settings_v1.py +++ b/src/deepgram/types/speak_settings_v1.py @@ -2,26 +2,4 @@ import typing -import pydantic -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel -from .speak_settings_v1endpoint import SpeakSettingsV1Endpoint -from .speak_settings_v1provider import SpeakSettingsV1Provider - - -class SpeakSettingsV1(UncheckedBaseModel): - provider: SpeakSettingsV1Provider - endpoint: typing.Optional[SpeakSettingsV1Endpoint] = pydantic.Field(default=None) - """ - Optional if provider is Deepgram. Required for non-Deepgram TTS providers. - When present, must include url field and headers object. Valid schemes are https and wss with wss only supported for Eleven Labs. - """ - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow +SpeakSettingsV1 = typing.Any diff --git a/src/deepgram/types/speak_settings_v1provider.py b/src/deepgram/types/speak_settings_v1provider.py deleted file mode 100644 index 5fe6606d..00000000 --- a/src/deepgram/types/speak_settings_v1provider.py +++ /dev/null @@ -1,114 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -from __future__ import annotations - -import typing - -import pydantic -import typing_extensions -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel, UnionMetadata -from .aws_polly_speak_provider_credentials import AwsPollySpeakProviderCredentials -from .aws_polly_speak_provider_engine import AwsPollySpeakProviderEngine -from .aws_polly_speak_provider_voice import AwsPollySpeakProviderVoice -from .cartesia_speak_provider_model_id import CartesiaSpeakProviderModelId -from .cartesia_speak_provider_voice import CartesiaSpeakProviderVoice -from .deepgram_speak_provider_model import DeepgramSpeakProviderModel -from .eleven_labs_speak_provider_model_id import ElevenLabsSpeakProviderModelId -from .open_ai_speak_provider_model import OpenAiSpeakProviderModel -from .open_ai_speak_provider_voice import OpenAiSpeakProviderVoice - - -class SpeakSettingsV1Provider_Deepgram(UncheckedBaseModel): - type: typing.Literal["deepgram"] = "deepgram" - version: typing.Optional[typing.Literal["v1"]] = None - model: DeepgramSpeakProviderModel - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class SpeakSettingsV1Provider_ElevenLabs(UncheckedBaseModel): - type: typing.Literal["eleven_labs"] = "eleven_labs" - version: typing.Optional[typing.Literal["v1"]] = None - model_id: ElevenLabsSpeakProviderModelId - language: typing.Optional[str] = None - language_code: typing.Optional[str] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class SpeakSettingsV1Provider_Cartesia(UncheckedBaseModel): - type: typing.Literal["cartesia"] = "cartesia" - version: typing.Optional[typing.Literal["2025-03-17"]] = None - model_id: CartesiaSpeakProviderModelId - voice: CartesiaSpeakProviderVoice - language: typing.Optional[str] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class SpeakSettingsV1Provider_OpenAi(UncheckedBaseModel): - type: typing.Literal["open_ai"] = "open_ai" - version: typing.Optional[typing.Literal["v1"]] = None - model: OpenAiSpeakProviderModel - voice: OpenAiSpeakProviderVoice - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class SpeakSettingsV1Provider_AwsPolly(UncheckedBaseModel): - type: typing.Literal["aws_polly"] = "aws_polly" - voice: AwsPollySpeakProviderVoice - language: str - language_code: typing.Optional[str] = None - engine: AwsPollySpeakProviderEngine - credentials: AwsPollySpeakProviderCredentials - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -SpeakSettingsV1Provider = typing_extensions.Annotated[ - typing.Union[ - SpeakSettingsV1Provider_Deepgram, - SpeakSettingsV1Provider_ElevenLabs, - SpeakSettingsV1Provider_Cartesia, - SpeakSettingsV1Provider_OpenAi, - SpeakSettingsV1Provider_AwsPolly, - ], - UnionMetadata(discriminant="type"), -] diff --git a/src/deepgram/types/think_settings_v1.py b/src/deepgram/types/think_settings_v1.py index f74c63d1..1de56d5b 100644 --- a/src/deepgram/types/think_settings_v1.py +++ b/src/deepgram/types/think_settings_v1.py @@ -5,25 +5,10 @@ import pydantic from ..core.pydantic_utilities import IS_PYDANTIC_V2 from ..core.unchecked_base_model import UncheckedBaseModel -from .think_settings_v1context_length import ThinkSettingsV1ContextLength -from .think_settings_v1endpoint import ThinkSettingsV1Endpoint -from .think_settings_v1functions_item import ThinkSettingsV1FunctionsItem -from .think_settings_v1provider import ThinkSettingsV1Provider class ThinkSettingsV1(UncheckedBaseModel): - provider: ThinkSettingsV1Provider - endpoint: typing.Optional[ThinkSettingsV1Endpoint] = pydantic.Field(default=None) - """ - Optional for non-Deepgram LLM providers. When present, must include url field and headers object - """ - - functions: typing.Optional[typing.List[ThinkSettingsV1FunctionsItem]] = None - prompt: typing.Optional[str] = None - context_length: typing.Optional[ThinkSettingsV1ContextLength] = pydantic.Field(default=None) - """ - Specifies the number of characters retained in context between user messages, agent responses, and function calls. This setting is only configurable when a custom think endpoint is used - """ + context_length: typing.Optional[typing.Any] = None if IS_PYDANTIC_V2: model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 diff --git a/src/deepgram/types/think_settings_v1provider.py b/src/deepgram/types/think_settings_v1provider.py deleted file mode 100644 index 799ec07d..00000000 --- a/src/deepgram/types/think_settings_v1provider.py +++ /dev/null @@ -1,107 +0,0 @@ -# This file was auto-generated by Fern from our API Definition. - -from __future__ import annotations - -import typing - -import pydantic -import typing_extensions -from ..core.pydantic_utilities import IS_PYDANTIC_V2 -from ..core.unchecked_base_model import UncheckedBaseModel, UnionMetadata -from .anthropic_think_provider_model import AnthropicThinkProviderModel -from .aws_bedrock_think_provider_credentials import AwsBedrockThinkProviderCredentials -from .aws_bedrock_think_provider_model import AwsBedrockThinkProviderModel -from .google_think_provider_model import GoogleThinkProviderModel -from .open_ai_think_provider_model import OpenAiThinkProviderModel - - -class ThinkSettingsV1Provider_OpenAi(UncheckedBaseModel): - type: typing.Literal["open_ai"] = "open_ai" - version: typing.Optional[typing.Literal["v1"]] = None - model: OpenAiThinkProviderModel - temperature: typing.Optional[float] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class ThinkSettingsV1Provider_AwsBedrock(UncheckedBaseModel): - type: typing.Literal["aws_bedrock"] = "aws_bedrock" - model: AwsBedrockThinkProviderModel - temperature: typing.Optional[float] = None - credentials: typing.Optional[AwsBedrockThinkProviderCredentials] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class ThinkSettingsV1Provider_Anthropic(UncheckedBaseModel): - type: typing.Literal["anthropic"] = "anthropic" - version: typing.Optional[typing.Literal["v1"]] = None - model: AnthropicThinkProviderModel - temperature: typing.Optional[float] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class ThinkSettingsV1Provider_Google(UncheckedBaseModel): - type: typing.Literal["google"] = "google" - version: typing.Optional[typing.Literal["v1beta"]] = None - model: GoogleThinkProviderModel - temperature: typing.Optional[float] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -class ThinkSettingsV1Provider_Groq(UncheckedBaseModel): - type: typing.Literal["groq"] = "groq" - version: typing.Optional[typing.Literal["v1"]] = None - model: typing.Literal["openai/gpt-oss-20b"] = "openai/gpt-oss-20b" - temperature: typing.Optional[float] = None - - if IS_PYDANTIC_V2: - model_config: typing.ClassVar[pydantic.ConfigDict] = pydantic.ConfigDict(extra="allow", frozen=True) # type: ignore # Pydantic v2 - else: - - class Config: - frozen = True - smart_union = True - extra = pydantic.Extra.allow - - -ThinkSettingsV1Provider = typing_extensions.Annotated[ - typing.Union[ - ThinkSettingsV1Provider_OpenAi, - ThinkSettingsV1Provider_AwsBedrock, - ThinkSettingsV1Provider_Anthropic, - ThinkSettingsV1Provider_Google, - ThinkSettingsV1Provider_Groq, - ], - UnionMetadata(discriminant="type"), -] diff --git a/src/deepgram/types/usage_breakdown_v1response_results_item_grouping.py b/src/deepgram/types/usage_breakdown_v1response_results_item_grouping.py index c1966d9f..d71b368c 100644 --- a/src/deepgram/types/usage_breakdown_v1response_results_item_grouping.py +++ b/src/deepgram/types/usage_breakdown_v1response_results_item_grouping.py @@ -33,11 +33,7 @@ class UsageBreakdownV1ResponseResultsItemGrouping(UncheckedBaseModel): Optional feature set identifier """ - models: typing.Optional[str] = pydantic.Field(default=None) - """ - Optional models identifier - """ - + models: typing.Optional[typing.List[typing.Optional[str]]] = None method: typing.Optional[str] = pydantic.Field(default=None) """ Optional method identifier diff --git a/tests/conftest.py b/tests/conftest.py index ebaf0125..36a95eaa 100644 --- a/tests/conftest.py +++ b/tests/conftest.py @@ -15,7 +15,7 @@ import pytest _STARTED: bool = False -_WIREMOCK_PORT: str = "8080" # Default, will be updated after container starts +_WIREMOCK_URL: str = "http://localhost:8080" # Default, will be updated after container starts _PROJECT_NAME: str = "deepgram-api" # This file lives at tests/conftest.py, so the project root is one level up. @@ -41,7 +41,7 @@ def _get_wiremock_port() -> str: def _start_wiremock() -> None: """Starts the WireMock container using docker-compose.""" - global _STARTED, _WIREMOCK_PORT + global _STARTED, _WIREMOCK_URL if _STARTED: return @@ -54,8 +54,9 @@ def _start_wiremock() -> None: text=True, ) _WIREMOCK_PORT = _get_wiremock_port() - os.environ["WIREMOCK_PORT"] = _WIREMOCK_PORT - print(f"WireMock container is ready on port {_WIREMOCK_PORT}") + _WIREMOCK_URL = f"http://localhost:{_WIREMOCK_PORT}" + os.environ["WIREMOCK_URL"] = _WIREMOCK_URL + print(f"WireMock container is ready at {_WIREMOCK_URL}") _STARTED = True except subprocess.CalledProcessError as e: print(f"Failed to start WireMock: {e.stderr}") diff --git a/tests/wire/conftest.py b/tests/wire/conftest.py index 9d7c5e1d..d138641f 100644 --- a/tests/wire/conftest.py +++ b/tests/wire/conftest.py @@ -26,9 +26,8 @@ def _get_wiremock_base_url() -> str: - """Returns the WireMock base URL using the dynamically assigned port.""" - port = os.environ.get("WIREMOCK_PORT", "8080") - return f"http://localhost:{port}" + """Returns the WireMock base URL from the WIREMOCK_URL environment variable.""" + return os.environ.get("WIREMOCK_URL", "http://localhost:8080") def get_client(test_id: str) -> DeepgramClient: @@ -46,13 +45,13 @@ def get_client(test_id: str) -> DeepgramClient: if _CLIENT_SUPPORTS_HEADERS: return DeepgramClient( - environment=DeepgramClientEnvironment(base=base_url, production=base_url, agent=base_url), + environment=DeepgramClientEnvironment(base=base_url, agent=base_url, production=base_url), headers=test_headers, api_key="test_api_key", ) return DeepgramClient( - environment=DeepgramClientEnvironment(base=base_url, production=base_url, agent=base_url), + environment=DeepgramClientEnvironment(base=base_url, agent=base_url, production=base_url), httpx_client=httpx.Client(headers=test_headers), api_key="test_api_key", ) diff --git a/tests/wire/test_listen_v1_media.py b/tests/wire/test_listen_v1_media.py index d7943355..8c08f1ab 100644 --- a/tests/wire/test_listen_v1_media.py +++ b/tests/wire/test_listen_v1_media.py @@ -86,11 +86,3 @@ def test_listen_v1_media_transcribe_url() -> None: }, 1, ) - - -def test_listen_v1_media_transcribe_file() -> None: - """Test transcribeFile endpoint with WireMock""" - test_id = "listen.v1.media.transcribe_file.0" - client = get_client(test_id) - client.listen.v1.media.transcribe_file(request=b"test audio data") - verify_request_count(test_id, "POST", "/v1/listen", None, 1) diff --git a/tests/wire/test_manage_v1_models.py b/tests/wire/test_manage_v1_models.py index 9adc12a3..7bf62146 100644 --- a/tests/wire/test_manage_v1_models.py +++ b/tests/wire/test_manage_v1_models.py @@ -5,7 +5,9 @@ def test_manage_v1_models_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.models.list_.0" client = get_client(test_id) - client.manage.v1.models.list(include_outdated=True) + client.manage.v1.models.list( + include_outdated=True, + ) verify_request_count(test_id, "GET", "/v1/models", {"include_outdated": "true"}, 1) @@ -13,5 +15,7 @@ def test_manage_v1_models_get() -> None: """Test get endpoint with WireMock""" test_id = "manage.v1.models.get.0" client = get_client(test_id) - client.manage.v1.models.get(model_id="af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291") + client.manage.v1.models.get( + model_id="af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291", + ) verify_request_count(test_id, "GET", "/v1/models/af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291", None, 1) diff --git a/tests/wire/test_manage_v1_projects.py b/tests/wire/test_manage_v1_projects.py index 09ff5aaf..e3e02c6e 100644 --- a/tests/wire/test_manage_v1_projects.py +++ b/tests/wire/test_manage_v1_projects.py @@ -13,7 +13,11 @@ def test_manage_v1_projects_get() -> None: """Test get endpoint with WireMock""" test_id = "manage.v1.projects.get.0" client = get_client(test_id) - client.manage.v1.projects.get(project_id="123456-7890-1234-5678-901234", limit=1.1, page=1.1) + client.manage.v1.projects.get( + project_id="123456-7890-1234-5678-901234", + limit=1.1, + page=1.1, + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234", {"limit": "1.1", "page": "1.1"}, 1 ) @@ -23,7 +27,9 @@ def test_manage_v1_projects_delete() -> None: """Test delete endpoint with WireMock""" test_id = "manage.v1.projects.delete.0" client = get_client(test_id) - client.manage.v1.projects.delete(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.delete( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "DELETE", "/v1/projects/123456-7890-1234-5678-901234", None, 1) @@ -31,7 +37,9 @@ def test_manage_v1_projects_update() -> None: """Test update endpoint with WireMock""" test_id = "manage.v1.projects.update.0" client = get_client(test_id) - client.manage.v1.projects.update(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.update( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "PATCH", "/v1/projects/123456-7890-1234-5678-901234", None, 1) @@ -39,5 +47,7 @@ def test_manage_v1_projects_leave() -> None: """Test leave endpoint with WireMock""" test_id = "manage.v1.projects.leave.0" client = get_client(test_id) - client.manage.v1.projects.leave(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.leave( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "DELETE", "/v1/projects/123456-7890-1234-5678-901234/leave", None, 1) diff --git a/tests/wire/test_manage_v1_projects_billing_balances.py b/tests/wire/test_manage_v1_projects_billing_balances.py index b7619076..03d321d9 100644 --- a/tests/wire/test_manage_v1_projects_billing_balances.py +++ b/tests/wire/test_manage_v1_projects_billing_balances.py @@ -5,7 +5,9 @@ def test_manage_v1_projects_billing_balances_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.billing.balances.list_.0" client = get_client(test_id) - client.manage.v1.projects.billing.balances.list(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.billing.balances.list( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/balances", None, 1) @@ -14,7 +16,8 @@ def test_manage_v1_projects_billing_balances_get() -> None: test_id = "manage.v1.projects.billing.balances.get.0" client = get_client(test_id) client.manage.v1.projects.billing.balances.get( - project_id="123456-7890-1234-5678-901234", balance_id="123456-7890-1234-5678-901234" + project_id="123456-7890-1234-5678-901234", + balance_id="123456-7890-1234-5678-901234", ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/balances/123456-7890-1234-5678-901234", None, 1 diff --git a/tests/wire/test_manage_v1_projects_billing_fields.py b/tests/wire/test_manage_v1_projects_billing_fields.py index a8230bb5..2840b293 100644 --- a/tests/wire/test_manage_v1_projects_billing_fields.py +++ b/tests/wire/test_manage_v1_projects_billing_fields.py @@ -5,7 +5,11 @@ def test_manage_v1_projects_billing_fields_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.billing.fields.list_.0" client = get_client(test_id) - client.manage.v1.projects.billing.fields.list(project_id="123456-7890-1234-5678-901234", start="start", end="end") + client.manage.v1.projects.billing.fields.list( + project_id="123456-7890-1234-5678-901234", + start="start", + end="end", + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/billing/fields", {"start": "start", "end": "end"}, 1 ) diff --git a/tests/wire/test_manage_v1_projects_billing_purchases.py b/tests/wire/test_manage_v1_projects_billing_purchases.py index 3d1dd17f..f3937a21 100644 --- a/tests/wire/test_manage_v1_projects_billing_purchases.py +++ b/tests/wire/test_manage_v1_projects_billing_purchases.py @@ -5,5 +5,8 @@ def test_manage_v1_projects_billing_purchases_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.billing.purchases.list_.0" client = get_client(test_id) - client.manage.v1.projects.billing.purchases.list(project_id="123456-7890-1234-5678-901234", limit=1.1) + client.manage.v1.projects.billing.purchases.list( + project_id="123456-7890-1234-5678-901234", + limit=1.1, + ) verify_request_count(test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/purchases", {"limit": "1.1"}, 1) diff --git a/tests/wire/test_manage_v1_projects_keys.py b/tests/wire/test_manage_v1_projects_keys.py index 2640ff0d..7976bf7d 100644 --- a/tests/wire/test_manage_v1_projects_keys.py +++ b/tests/wire/test_manage_v1_projects_keys.py @@ -5,7 +5,10 @@ def test_manage_v1_projects_keys_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.keys.list_.0" client = get_client(test_id) - client.manage.v1.projects.keys.list(project_id="123456-7890-1234-5678-901234", status="active") + client.manage.v1.projects.keys.list( + project_id="123456-7890-1234-5678-901234", + status="active", + ) verify_request_count(test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/keys", {"status": "active"}, 1) @@ -15,9 +18,7 @@ def test_manage_v1_projects_keys_create() -> None: client = get_client(test_id) client.manage.v1.projects.keys.create( project_id="project_id", - request={ - "key": "value", - }, + request={"key": "value"}, ) verify_request_count(test_id, "POST", "/v1/projects/project_id/keys", None, 1) @@ -26,7 +27,10 @@ def test_manage_v1_projects_keys_get() -> None: """Test get endpoint with WireMock""" test_id = "manage.v1.projects.keys.get.0" client = get_client(test_id) - client.manage.v1.projects.keys.get(project_id="123456-7890-1234-5678-901234", key_id="123456789012345678901234") + client.manage.v1.projects.keys.get( + project_id="123456-7890-1234-5678-901234", + key_id="123456789012345678901234", + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/keys/123456789012345678901234", None, 1 ) @@ -36,7 +40,10 @@ def test_manage_v1_projects_keys_delete() -> None: """Test delete endpoint with WireMock""" test_id = "manage.v1.projects.keys.delete.0" client = get_client(test_id) - client.manage.v1.projects.keys.delete(project_id="123456-7890-1234-5678-901234", key_id="123456789012345678901234") + client.manage.v1.projects.keys.delete( + project_id="123456-7890-1234-5678-901234", + key_id="123456789012345678901234", + ) verify_request_count( test_id, "DELETE", "/v1/projects/123456-7890-1234-5678-901234/keys/123456789012345678901234", None, 1 ) diff --git a/tests/wire/test_manage_v1_projects_members.py b/tests/wire/test_manage_v1_projects_members.py index a75e908a..b395d9e0 100644 --- a/tests/wire/test_manage_v1_projects_members.py +++ b/tests/wire/test_manage_v1_projects_members.py @@ -5,7 +5,9 @@ def test_manage_v1_projects_members_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.members.list_.0" client = get_client(test_id) - client.manage.v1.projects.members.list(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.members.list( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/members", None, 1) @@ -14,7 +16,8 @@ def test_manage_v1_projects_members_delete() -> None: test_id = "manage.v1.projects.members.delete.0" client = get_client(test_id) client.manage.v1.projects.members.delete( - project_id="123456-7890-1234-5678-901234", member_id="123456789012345678901234" + project_id="123456-7890-1234-5678-901234", + member_id="123456789012345678901234", ) verify_request_count( test_id, "DELETE", "/v1/projects/123456-7890-1234-5678-901234/members/123456789012345678901234", None, 1 diff --git a/tests/wire/test_manage_v1_projects_members_invites.py b/tests/wire/test_manage_v1_projects_members_invites.py index 2f6573e5..e8e29d66 100644 --- a/tests/wire/test_manage_v1_projects_members_invites.py +++ b/tests/wire/test_manage_v1_projects_members_invites.py @@ -5,7 +5,9 @@ def test_manage_v1_projects_members_invites_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.members.invites.list_.0" client = get_client(test_id) - client.manage.v1.projects.members.invites.list(project_id="123456-7890-1234-5678-901234") + client.manage.v1.projects.members.invites.list( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count(test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/invites", None, 1) @@ -14,7 +16,9 @@ def test_manage_v1_projects_members_invites_create() -> None: test_id = "manage.v1.projects.members.invites.create.0" client = get_client(test_id) client.manage.v1.projects.members.invites.create( - project_id="123456-7890-1234-5678-901234", email="email", scope="scope" + project_id="123456-7890-1234-5678-901234", + email="email", + scope="scope", ) verify_request_count(test_id, "POST", "/v1/projects/123456-7890-1234-5678-901234/invites", None, 1) @@ -24,7 +28,8 @@ def test_manage_v1_projects_members_invites_delete() -> None: test_id = "manage.v1.projects.members.invites.delete.0" client = get_client(test_id) client.manage.v1.projects.members.invites.delete( - project_id="123456-7890-1234-5678-901234", email="john.doe@example.com" + project_id="123456-7890-1234-5678-901234", + email="john.doe@example.com", ) verify_request_count( test_id, "DELETE", "/v1/projects/123456-7890-1234-5678-901234/invites/john.doe@example.com", None, 1 diff --git a/tests/wire/test_manage_v1_projects_members_scopes.py b/tests/wire/test_manage_v1_projects_members_scopes.py index 94700ec7..e7f60832 100644 --- a/tests/wire/test_manage_v1_projects_members_scopes.py +++ b/tests/wire/test_manage_v1_projects_members_scopes.py @@ -6,7 +6,8 @@ def test_manage_v1_projects_members_scopes_list_() -> None: test_id = "manage.v1.projects.members.scopes.list_.0" client = get_client(test_id) client.manage.v1.projects.members.scopes.list( - project_id="123456-7890-1234-5678-901234", member_id="123456789012345678901234" + project_id="123456-7890-1234-5678-901234", + member_id="123456789012345678901234", ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/members/123456789012345678901234/scopes", None, 1 @@ -18,7 +19,9 @@ def test_manage_v1_projects_members_scopes_update() -> None: test_id = "manage.v1.projects.members.scopes.update.0" client = get_client(test_id) client.manage.v1.projects.members.scopes.update( - project_id="123456-7890-1234-5678-901234", member_id="123456789012345678901234", scope="admin" + project_id="123456-7890-1234-5678-901234", + member_id="123456789012345678901234", + scope="admin", ) verify_request_count( test_id, "PUT", "/v1/projects/123456-7890-1234-5678-901234/members/123456789012345678901234/scopes", None, 1 diff --git a/tests/wire/test_manage_v1_projects_models.py b/tests/wire/test_manage_v1_projects_models.py index 0cdf66d6..7d87c955 100644 --- a/tests/wire/test_manage_v1_projects_models.py +++ b/tests/wire/test_manage_v1_projects_models.py @@ -5,7 +5,10 @@ def test_manage_v1_projects_models_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.models.list_.0" client = get_client(test_id) - client.manage.v1.projects.models.list(project_id="123456-7890-1234-5678-901234", include_outdated=True) + client.manage.v1.projects.models.list( + project_id="123456-7890-1234-5678-901234", + include_outdated=True, + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/models", {"include_outdated": "true"}, 1 ) @@ -16,7 +19,8 @@ def test_manage_v1_projects_models_get() -> None: test_id = "manage.v1.projects.models.get.0" client = get_client(test_id) client.manage.v1.projects.models.get( - project_id="123456-7890-1234-5678-901234", model_id="af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291" + project_id="123456-7890-1234-5678-901234", + model_id="af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291", ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/models/af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291", None, 1 diff --git a/tests/wire/test_manage_v1_projects_requests.py b/tests/wire/test_manage_v1_projects_requests.py index cfae6837..fd361042 100644 --- a/tests/wire/test_manage_v1_projects_requests.py +++ b/tests/wire/test_manage_v1_projects_requests.py @@ -1,4 +1,4 @@ -from datetime import datetime +import datetime from .conftest import get_client, verify_request_count @@ -9,8 +9,8 @@ def test_manage_v1_projects_requests_list_() -> None: client = get_client(test_id) client.manage.v1.projects.requests.list( project_id="123456-7890-1234-5678-901234", - start=datetime.fromisoformat("2024-01-15T09:30:00+00:00"), - end=datetime.fromisoformat("2024-01-15T09:30:00+00:00"), + start=datetime.datetime.fromisoformat("2024-01-15T09:30:00+00:00"), + end=datetime.datetime.fromisoformat("2024-01-15T09:30:00+00:00"), limit=1.1, page=1.1, accessor="12345678-1234-1234-1234-123456789012", @@ -45,7 +45,8 @@ def test_manage_v1_projects_requests_get() -> None: test_id = "manage.v1.projects.requests.get.0" client = get_client(test_id) client.manage.v1.projects.requests.get( - project_id="123456-7890-1234-5678-901234", request_id="123456-7890-1234-5678-901234" + project_id="123456-7890-1234-5678-901234", + request_id="123456-7890-1234-5678-901234", ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/requests/123456-7890-1234-5678-901234", None, 1 diff --git a/tests/wire/test_manage_v1_projects_usage_fields.py b/tests/wire/test_manage_v1_projects_usage_fields.py index 7b2208fe..c8217a49 100644 --- a/tests/wire/test_manage_v1_projects_usage_fields.py +++ b/tests/wire/test_manage_v1_projects_usage_fields.py @@ -5,7 +5,11 @@ def test_manage_v1_projects_usage_fields_list_() -> None: """Test list endpoint with WireMock""" test_id = "manage.v1.projects.usage.fields.list_.0" client = get_client(test_id) - client.manage.v1.projects.usage.fields.list(project_id="123456-7890-1234-5678-901234", start="start", end="end") + client.manage.v1.projects.usage.fields.list( + project_id="123456-7890-1234-5678-901234", + start="start", + end="end", + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/usage/fields", {"start": "start", "end": "end"}, 1 ) diff --git a/tests/wire/test_read_v1_text.py b/tests/wire/test_read_v1_text.py index 9d8decbd..88af46fe 100644 --- a/tests/wire/test_read_v1_text.py +++ b/tests/wire/test_read_v1_text.py @@ -1,5 +1,7 @@ from .conftest import get_client, verify_request_count +from deepgram import ReadV1RequestUrl + def test_read_v1_text_analyze() -> None: """Test analyze endpoint with WireMock""" @@ -18,7 +20,9 @@ def test_read_v1_text_analyze() -> None: custom_intent=["custom_intent"], custom_intent_mode="extended", language="language", - request={"url": "url"}, + request=ReadV1RequestUrl( + url="url", + ), ) verify_request_count( test_id, diff --git a/tests/wire/test_selfHosted_v1_distributionCredentials.py b/tests/wire/test_selfHosted_v1_distributionCredentials.py index 23b9b240..5bf736d2 100644 --- a/tests/wire/test_selfHosted_v1_distributionCredentials.py +++ b/tests/wire/test_selfHosted_v1_distributionCredentials.py @@ -5,7 +5,9 @@ def test_selfHosted_v1_distributionCredentials_list_() -> None: """Test list endpoint with WireMock""" test_id = "self_hosted.v1.distribution_credentials.list_.0" client = get_client(test_id) - client.self_hosted.v1.distribution_credentials.list(project_id="123456-7890-1234-5678-901234") + client.self_hosted.v1.distribution_credentials.list( + project_id="123456-7890-1234-5678-901234", + ) verify_request_count( test_id, "GET", "/v1/projects/123456-7890-1234-5678-901234/self-hosted/distribution/credentials", None, 1 ) @@ -15,7 +17,10 @@ def test_selfHosted_v1_distributionCredentials_create() -> None: """Test create endpoint with WireMock""" test_id = "self_hosted.v1.distribution_credentials.create.0" client = get_client(test_id) - client.self_hosted.v1.distribution_credentials.create(project_id="123456-7890-1234-5678-901234", provider="quay") + client.self_hosted.v1.distribution_credentials.create( + project_id="123456-7890-1234-5678-901234", + provider="quay", + ) verify_request_count( test_id, "POST", @@ -30,7 +35,8 @@ def test_selfHosted_v1_distributionCredentials_get() -> None: test_id = "self_hosted.v1.distribution_credentials.get.0" client = get_client(test_id) client.self_hosted.v1.distribution_credentials.get( - project_id="123456-7890-1234-5678-901234", distribution_credentials_id="8b36cfd0-472f-4a21-833f-2d6343c3a2f3" + project_id="123456-7890-1234-5678-901234", + distribution_credentials_id="8b36cfd0-472f-4a21-833f-2d6343c3a2f3", ) verify_request_count( test_id, @@ -46,7 +52,8 @@ def test_selfHosted_v1_distributionCredentials_delete() -> None: test_id = "self_hosted.v1.distribution_credentials.delete.0" client = get_client(test_id) client.self_hosted.v1.distribution_credentials.delete( - project_id="123456-7890-1234-5678-901234", distribution_credentials_id="8b36cfd0-472f-4a21-833f-2d6343c3a2f3" + project_id="123456-7890-1234-5678-901234", + distribution_credentials_id="8b36cfd0-472f-4a21-833f-2d6343c3a2f3", ) verify_request_count( test_id, diff --git a/tests/wire/test_speak_v1_audio.py b/tests/wire/test_speak_v1_audio.py index 692f77cb..8048ab4d 100644 --- a/tests/wire/test_speak_v1_audio.py +++ b/tests/wire/test_speak_v1_audio.py @@ -5,6 +5,8 @@ def test_speak_v1_audio_generate() -> None: """Test generate endpoint with WireMock""" test_id = "speak.v1.audio.generate.0" client = get_client(test_id) - for _ in client.speak.v1.audio.generate(text="text"): + for _ in client.speak.v1.audio.generate( + text="text", + ): pass verify_request_count(test_id, "POST", "/v1/speak", None, 1) diff --git a/wiremock/wiremock-mappings.json b/wiremock/wiremock-mappings.json index 384ee619..4248699c 100644 --- a/wiremock/wiremock-mappings.json +++ b/wiremock/wiremock-mappings.json @@ -1 +1 @@ -{"mappings":[{"id":"533b5d52-ab21-4763-aaae-87cf52f49aa5","name":"List Agent Think Models - default","request":{"urlPathTemplate":"/v1/agent/settings/think/models","method":"GET"},"response":{"status":200,"body":"{\n \"models\": [\n {\n \"id\": \"gpt-5\",\n \"name\": \"name\",\n \"provider\": \"open_ai\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"533b5d52-ab21-4763-aaae-87cf52f49aa5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"20e1029e-8bb9-4092-a809-b943e60822ef","name":"Token-based Authentication - default","request":{"urlPathTemplate":"/v1/auth/grant","method":"POST"},"response":{"status":200,"body":"{\n \"access_token\": \"eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzdWIiOiIxMjM0NTY3ODkwIn0.dozjgNryP4J3jVmNHl0w5N_XgL0n3I9PlFUP0THsR8U\",\n \"expires_in\": 30\n}","headers":{"Content-Type":"application/json"}},"uuid":"20e1029e-8bb9-4092-a809-b943e60822ef","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"49d8d51a-7f01-4598-804f-b6f54cdc22da","name":"Transcribe and analyze pre-recorded audio and video - default","request":{"urlPathTemplate":"/v1/listen","method":"POST"},"response":{"status":200,"body":"{\n \"metadata\": {\n \"request_id\": \"a847f427-4ad5-4d67-9b95-db801e58251c\",\n \"sha256\": \"154e291ecfa8be6ab8343560bcc109008fa7853eb5372533e8efdefc9b504c33\",\n \"created\": \"2024-05-12T18:57:13Z\",\n \"duration\": 25.933313,\n \"channels\": 1,\n \"models\": [\n \"30089e05-99d1-4376-b32e-c263170674af\"\n ],\n \"model_info\": {\n \"30089e05-99d1-4376-b32e-c263170674af\": {\n \"name\": \"2-general-nova\",\n \"version\": \"2024-01-09.29447\",\n \"arch\": \"nova-2\"\n }\n },\n \"summary_info\": {\n \"model_uuid\": \"67875a7f-c9c4-48a0-aa55-5bdb8a91c34a\",\n \"input_tokens\": 95,\n \"output_tokens\": 63\n },\n \"sentiment_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 105\n },\n \"topics_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 7\n },\n \"intents_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 4\n },\n \"tags\": [\n \"test\"\n ]\n },\n \"results\": {\n \"channels\": [\n {}\n ],\n \"utterances\": [\n {}\n ],\n \"summary\": {\n \"result\": \"success\",\n \"short\": \"Speaker 0 discusses the significance of the first all-female spacewalk with an all-female team, stating that it is a tribute to the skilled and qualified women who were denied opportunities in the past.\"\n },\n \"sentiments\": {\n \"segments\": [\n {\n \"text\": \"Yeah. As as much as, um, it's worth celebrating, uh, the first, uh, spacewalk, um, with an all-female team, I think many of us are looking forward to it just being normal. And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 0,\n \"end_word\": 69,\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810546875\n }\n ],\n \"average\": {\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810185185185185\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"49d8d51a-7f01-4598-804f-b6f54cdc22da","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"0a0be61b-f024-4120-9c54-23bca3e07c93","name":"List Models - default","request":{"urlPathTemplate":"/v1/models","method":"GET"},"response":{"status":200,"body":"{\n \"stt\": [\n {\n \"name\": \"nova-3\",\n \"canonical_name\": \"nova-3\",\n \"architecture\": \"base\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2021-11-10.1\",\n \"uuid\": \"6b28e919-8427-4f32-9847-492e2efd7daf\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": true\n }\n ],\n \"tts\": [\n {\n \"name\": \"zeus\",\n \"canonical_name\": \"aura-2-zeus-en\",\n \"architecture\": \"aura-2\",\n \"languages\": [\n \"en\",\n \"en-US\"\n ],\n \"version\": \"2025-04-07.0\",\n \"uuid\": \"2baf189d-91ac-481d-b6d1-750888667b31\",\n \"metadata\": {\n \"accent\": \"American\",\n \"age\": \"Adult\",\n \"color\": \"#C58DFF\",\n \"image\": \"https://static.deepgram.com/examples/avatars/zeus.jpg\",\n \"sample\": \"https://static.deepgram.com/examples/Aura-2-zeus.wav\",\n \"tags\": [\n \"masculine\",\n \"deep\",\n \"trustworthy\",\n \"smooth\"\n ],\n \"use_cases\": [\n \"IVR\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"0a0be61b-f024-4120-9c54-23bca3e07c93","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"0f94d3ea-43b6-4a1a-bce4-ab05b85440ae","name":"Get a specific Model - default","request":{"urlPathTemplate":"/v1/models/{model_id}","method":"GET","pathParameters":{"model_id":{"equalTo":"af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291"}}},"response":{"status":200,"body":"{\n \"name\": \"general\",\n \"canonical_name\": \"enhanced-general\",\n \"architecture\": \"polaris\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2022-05-18.1\",\n \"uuid\": \"c7226e9e-ae1c-4057-ae2a-a71a6b0dc588\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": false\n}","headers":{"Content-Type":"application/json"}},"uuid":"0f94d3ea-43b6-4a1a-bce4-ab05b85440ae","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d08cac56-bc4d-4756-8fd1-d508914334d5","name":"List Projects - default","request":{"urlPathTemplate":"/v1/projects","method":"GET"},"response":{"status":200,"body":"{\n \"projects\": [\n {\n \"project_id\": \"project_id\",\n \"name\": \"name\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"d08cac56-bc4d-4756-8fd1-d508914334d5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"6f0163a8-530c-4e25-bbe0-9ca86b9525dc","name":"Get a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"project_id\": \"project_id\",\n \"mip_opt_out\": true,\n \"name\": \"name\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"6f0163a8-530c-4e25-bbe0-9ca86b9525dc","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"68918577-6401-4439-8533-356257ff7bcf","name":"Delete a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"68918577-6401-4439-8533-356257ff7bcf","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"659fc38b-3934-4e43-93bf-d331f547449e","name":"Update a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"PATCH","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"Successfully updated project info.\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"659fc38b-3934-4e43-93bf-d331f547449e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"43ae6622-ad2f-4c81-9bc9-a8bbe17ef9d8","name":"Leave a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/leave","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"43ae6622-ad2f-4c81-9bc9-a8bbe17ef9d8","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"032d9b1c-3b87-40fb-bfab-8c5be92a5d71","name":"List Project Keys - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"api_keys\": [\n {\n \"member\": {\n \"member_id\": \"1000-2000-3000-4000\",\n \"email\": \"john@test.com\"\n },\n \"api_key\": {\n \"api_key_id\": \"1234567890abcdef1234567890abcdef\",\n \"comment\": \"A comment\",\n \"scopes\": [\n \"admin\"\n ],\n \"created\": \"2021-01-01T00:00:00Z\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"032d9b1c-3b87-40fb-bfab-8c5be92a5d71","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"0167c735-0b6f-4715-8df8-32300d4dae72","name":"Create a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys","method":"POST","pathParameters":{"project_id":{"equalTo":"project_id"}}},"response":{"status":200,"body":"{\n \"api_key_id\": \"api_key_id\",\n \"key\": \"key\",\n \"comment\": \"comment\",\n \"scopes\": [\n \"scopes\",\n \"scopes\"\n ],\n \"tags\": [\n \"tags\",\n \"tags\"\n ],\n \"expiration_date\": \"2024-01-15T09:30:00Z\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"0167c735-0b6f-4715-8df8-32300d4dae72","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"c9812dd3-f87e-4798-aec3-af0933330dd5","name":"Get a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys/{key_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"key_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"item\": {\n \"member\": {\n \"member_id\": \"1000-2000-3000-4000\",\n \"email\": \"john@test.com\",\n \"first_name\": \"John\",\n \"last_name\": \"Doe\",\n \"api_key\": {\n \"api_key_id\": \"1000-2000-3000-4000\",\n \"comment\": \"A comment\",\n \"scopes\": [\n \"admin\"\n ],\n \"tags\": [\n \"prod\",\n \"west-region\"\n ],\n \"expiration_date\": \"2021-01-01T00:00:00Z\",\n \"created\": \"2021-01-01T00:00:00Z\"\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"c9812dd3-f87e-4798-aec3-af0933330dd5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"3d2fbc7c-7bac-436f-a6ac-abe1b2c2caac","name":"Delete a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys/{key_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"key_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"3d2fbc7c-7bac-436f-a6ac-abe1b2c2caac","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"91e103d5-72f7-463d-840d-310069e33de9","name":"List Project Members - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"members\": [\n {\n \"member_id\": \"member_id\",\n \"email\": \"email\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"91e103d5-72f7-463d-840d-310069e33de9","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"515c6f7e-09c3-43ea-ad6c-65bc11d20f46","name":"Delete a Project Member - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"515c6f7e-09c3-43ea-ad6c-65bc11d20f46","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a920ad0e-2796-4361-ac16-ac83fb75e32a","name":"List Project Models - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/models","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"stt\": [\n {\n \"name\": \"nova-3\",\n \"canonical_name\": \"nova-3\",\n \"architecture\": \"base\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2021-11-10.1\",\n \"uuid\": \"6b28e919-8427-4f32-9847-492e2efd7daf\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": true\n }\n ],\n \"tts\": [\n {\n \"name\": \"zeus\",\n \"canonical_name\": \"aura-2-zeus-en\",\n \"architecture\": \"aura-2\",\n \"languages\": [\n \"en\",\n \"en-US\"\n ],\n \"version\": \"2025-04-07.0\",\n \"uuid\": \"2baf189d-91ac-481d-b6d1-750888667b31\",\n \"metadata\": {\n \"accent\": \"American\",\n \"age\": \"Adult\",\n \"color\": \"#C58DFF\",\n \"image\": \"https://static.deepgram.com/examples/avatars/zeus.jpg\",\n \"sample\": \"https://static.deepgram.com/examples/Aura-2-zeus.wav\",\n \"tags\": [\n \"masculine\",\n \"deep\",\n \"trustworthy\",\n \"smooth\"\n ],\n \"use_cases\": [\n \"IVR\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"a920ad0e-2796-4361-ac16-ac83fb75e32a","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9f8c6bf2-ebee-4956-b39f-0291b9d64b6e","name":"Get a Project Model - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/models/{model_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"model_id":{"equalTo":"af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291"}}},"response":{"status":200,"body":"{\n \"name\": \"general\",\n \"canonical_name\": \"enhanced-general\",\n \"architecture\": \"polaris\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2022-05-18.1\",\n \"uuid\": \"c7226e9e-ae1c-4057-ae2a-a71a6b0dc588\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": false\n}","headers":{"Content-Type":"application/json"}},"uuid":"9f8c6bf2-ebee-4956-b39f-0291b9d64b6e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d6a14959-05fa-4aec-9f0c-ba2a817c66e5","name":"List Project Requests - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/requests","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"page\": 1.1,\n \"limit\": 1.1,\n \"requests\": [\n {\n \"request_id\": \"request_id\",\n \"project_uuid\": \"project_uuid\",\n \"created\": \"2024-01-15T09:30:00Z\",\n \"path\": \"path\",\n \"api_key_id\": \"api_key_id\",\n \"response\": {\n \"key\": \"value\"\n },\n \"code\": 1.1,\n \"deployment\": \"deployment\",\n \"callback\": \"callback\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"d6a14959-05fa-4aec-9f0c-ba2a817c66e5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"733e39aa-d3ef-4ea7-8062-af080c6288c4","name":"Get a Project Request - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/requests/{request_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"request_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"request\": {\n \"request_id\": \"request_id\",\n \"project_uuid\": \"project_uuid\",\n \"created\": \"2024-01-15T09:30:00Z\",\n \"path\": \"path\",\n \"api_key_id\": \"api_key_id\",\n \"response\": {\n \"key\": \"value\"\n },\n \"code\": 1.1,\n \"deployment\": \"deployment\",\n \"callback\": \"callback\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"733e39aa-d3ef-4ea7-8062-af080c6288c4","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"6309dd55-c993-4ce1-b0b2-01a41c9f08d6","name":"Get Project Usage - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"start\": \"2024-10-16\",\n \"end\": \"2024-10-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"6309dd55-c993-4ce1-b0b2-01a41c9f08d6","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b132121b-4efe-42ad-a268-8acac35c189b","name":"Get Project Balances - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/balances","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"balances\": [\n {\n \"balance_id\": \"balance_id\",\n \"amount\": 1.1,\n \"units\": \"units\",\n \"purchase_order_id\": \"purchase_order_id\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"b132121b-4efe-42ad-a268-8acac35c189b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"4019c244-52d3-4d57-902c-af837631650a","name":"Get a Project Balance - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/balances/{balance_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"balance_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"balance_id\": \"balance_id\",\n \"amount\": 1.1,\n \"units\": \"units\",\n \"purchase_order_id\": \"purchase_order_id\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"4019c244-52d3-4d57-902c-af837631650a","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"555b6751-587f-400c-bf5e-400e108ad6b4","name":"Get Project Billing Breakdown - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/billing/breakdown","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n },\n \"results\": [\n {\n \"dollars\": 0.25,\n \"grouping\": {\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-16\",\n \"accessor\": \"123456789012345678901234\",\n \"deployment\": \"hosted\",\n \"line_item\": \"streaming::nova-3\",\n \"tags\": [\n \"tag1\",\n \"tag2\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"555b6751-587f-400c-bf5e-400e108ad6b4","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a61ae38c-e41f-4726-a55c-88f2135897be","name":"List Project Billing Fields - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/billing/fields","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"accessors\": [\n \"12345678-1234-1234-1234-123456789012\",\n \"87654321-4321-4321-4321-210987654321\"\n ],\n \"deployments\": [\n \"hosted\",\n \"self-hosted\"\n ],\n \"tags\": [\n \"dev\",\n \"production\"\n ],\n \"line_items\": {\n \"streaming::nova-3\": \"Nova - 3 (Stream)\",\n \"sync::aura-2\": \"Aura -2 (Sync)\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"a61ae38c-e41f-4726-a55c-88f2135897be","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"85b4373c-ba39-41b1-84e8-ae1ee6b180ca","name":"List Project Purchases - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/purchases","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"orders\": [\n {\n \"order_id\": \"025e19ba-b6d9-4a04-9f99-4fe715aca5f1\",\n \"expiration\": \"2026-03-04T00:00:00Z\",\n \"created\": \"2023-02-21T21:13:40Z\",\n \"amount\": 150,\n \"units\": \"usd\",\n \"order_type\": \"promotional\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"85b4373c-ba39-41b1-84e8-ae1ee6b180ca","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"2dd14c67-ed4e-4d97-9636-0a712899deb8","name":"List Project Invites - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"invites\": [\n {\n \"email\": \"email\",\n \"scope\": \"scope\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"2dd14c67-ed4e-4d97-9636-0a712899deb8","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"7c109496-adfe-4e85-b007-a6f799ee95cb","name":"Create a Project Invite - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites","method":"POST","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"7c109496-adfe-4e85-b007-a6f799ee95cb","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d6d268d0-d91e-4a65-80e0-339621173db9","name":"Delete a Project Invite - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites/{email}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"email":{"equalTo":"john.doe@example.com"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"d6d268d0-d91e-4a65-80e0-339621173db9","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"1b965d71-c930-4a0b-90f3-2289f80f3634","name":"List Project Member Scopes - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}/scopes","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"scopes\": [\n \"scopes\"\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"1b965d71-c930-4a0b-90f3-2289f80f3634","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"eb2f5de2-b887-47be-abd5-7cb702aca55d","name":"Update Project Member Scopes - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}/scopes","method":"PUT","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"eb2f5de2-b887-47be-abd5-7cb702aca55d","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9bdf51a4-1e10-41b8-8de2-2df650562db3","name":"Get Project Usage Breakdown - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage/breakdown","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n },\n \"results\": [\n {\n \"hours\": 1619.7242069444444,\n \"total_hours\": 1621.7395791666668,\n \"agent_hours\": 41.33564388888889,\n \"tokens_in\": 0,\n \"tokens_out\": 0,\n \"tts_characters\": 9158866,\n \"requests\": 373381,\n \"grouping\": {\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-16\",\n \"accessor\": \"123456789012345678901234\",\n \"endpoint\": \"listen\",\n \"feature_set\": \"punctuate\",\n \"models\": \"Nova-2\",\n \"method\": \"async\",\n \"tags\": \"tag1\",\n \"deployment\": \"self-hosted\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"9bdf51a4-1e10-41b8-8de2-2df650562db3","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"bc9fcd54-076e-48dc-a2dd-d71a8bf8bd4e","name":"List Project Usage Fields - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage/fields","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"tags\": [\n \"tag=dev\",\n \"tag=production\"\n ],\n \"models\": [\n {\n \"name\": \"2-medical-nova\",\n \"language\": \"en-MY\",\n \"version\": \"2024-05-31.13574\",\n \"model_id\": \"1234567890-12345-67890\"\n }\n ],\n \"processing_methods\": [\n \"sync\",\n \"streaming\"\n ],\n \"features\": [\n \"alternatives\",\n \"detect_entities\",\n \"detect_language\"\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"bc9fcd54-076e-48dc-a2dd-d71a8bf8bd4e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b5ac3651-d3b7-4cd7-b0b8-e1a917a16f3b","name":"Analyze text content - default","request":{"urlPathTemplate":"/v1/read","method":"POST"},"response":{"status":200,"body":"{\n \"metadata\": {\n \"metadata\": {\n \"request_id\": \"d04af392-db11-4c1d-83e1-20e34f0b8999\",\n \"created\": \"2024-11-18T23:47:44Z\",\n \"language\": \"en\"\n }\n },\n \"results\": {\n \"sentiments\": {\n \"segments\": [\n {\n \"text\": \"Yeah. As as much as, um, it's worth celebrating, uh, the first, uh, spacewalk, um, with an all-female team, I think many of us are looking forward to it just being normal. And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 0,\n \"end_word\": 69,\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810546875\n }\n ],\n \"average\": {\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810185185185185\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"b5ac3651-d3b7-4cd7-b0b8-e1a917a16f3b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"4110cb96-50e2-4fe6-b8ae-5d69120cee89","name":"List Project Self-Hosted Distribution Credentials - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"distribution_credentials\": [\n {\n \"member\": {\n \"member_id\": \"3376abcd-8e5e-49d3-92d4-876d3a4f0363\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"8b36cfd0-472f-4a21-833f-2d6343c3a2f3\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"4110cb96-50e2-4fe6-b8ae-5d69120cee89","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9c12eea9-6ba6-4d70-bb14-a2742cebc114","name":"Create a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials","method":"POST","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"9c12eea9-6ba6-4d70-bb14-a2742cebc114","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a47cd13f-2314-4190-b2c7-20436ccffbd2","name":"Get a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials/{distribution_credentials_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"distribution_credentials_id":{"equalTo":"8b36cfd0-472f-4a21-833f-2d6343c3a2f3"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"a47cd13f-2314-4190-b2c7-20436ccffbd2","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"8bd46091-0e57-4b3d-9485-a86e6f1eaf17","name":"Delete a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials/{distribution_credentials_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"distribution_credentials_id":{"equalTo":"8b36cfd0-472f-4a21-833f-2d6343c3a2f3"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"8bd46091-0e57-4b3d-9485-a86e6f1eaf17","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b06ec977-02ed-41e2-8fff-2bc45cd2166b","name":"Text to Speech transformation - default","request":{"urlPathTemplate":"/v1/speak","method":"POST"},"response":{"status":200,"body":"{\n \"key\": \"value\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"b06ec977-02ed-41e2-8fff-2bc45cd2166b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}}],"meta":{"total":40}} \ No newline at end of file +{"mappings":[{"id":"533b5d52-ab21-4763-aaae-87cf52f49aa5","name":"List Agent Think Models - default","request":{"urlPathTemplate":"/v1/agent/settings/think/models","method":"GET"},"response":{"status":200,"body":"{\n \"models\": [\n {\n \"id\": \"gpt-5\",\n \"name\": \"name\",\n \"provider\": \"open_ai\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"533b5d52-ab21-4763-aaae-87cf52f49aa5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"20e1029e-8bb9-4092-a809-b943e60822ef","name":"Token-based Authentication - default","request":{"urlPathTemplate":"/v1/auth/grant","method":"POST"},"response":{"status":200,"body":"{\n \"access_token\": \"eyJhbGciOiJIUzI1NiIsInR5cCI6IkpXVCJ9.eyJzdWIiOiIxMjM0NTY3ODkwIn0.dozjgNryP4J3jVmNHl0w5N_XgL0n3I9PlFUP0THsR8U\",\n \"expires_in\": 30\n}","headers":{"Content-Type":"application/json"}},"uuid":"20e1029e-8bb9-4092-a809-b943e60822ef","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"49d8d51a-7f01-4598-804f-b6f54cdc22da","name":"Transcribe and analyze pre-recorded audio and video - default","request":{"urlPathTemplate":"/v1/listen","method":"POST","queryParameters":{"callback":{"equalTo":"callback"},"callback_method":{"equalTo":"POST"},"extra":{"equalTo":"extra"},"sentiment":{"equalTo":"true"},"summarize":{"equalTo":"v2"},"tag":{"equalTo":"tag"},"topics":{"equalTo":"true"},"custom_topic":{"equalTo":"custom_topic"},"custom_topic_mode":{"equalTo":"extended"},"intents":{"equalTo":"true"},"custom_intent":{"equalTo":"custom_intent"},"custom_intent_mode":{"equalTo":"extended"},"detect_entities":{"equalTo":"true"},"detect_language":{"equalTo":"true"},"diarize":{"equalTo":"true"},"dictation":{"equalTo":"true"},"encoding":{"equalTo":"linear16"},"filler_words":{"equalTo":"true"},"keywords":{"equalTo":"keywords"},"language":{"equalTo":"language"},"measurements":{"equalTo":"true"},"model":{"equalTo":"nova-3"},"multichannel":{"equalTo":"true"},"numerals":{"equalTo":"true"},"paragraphs":{"equalTo":"true"},"profanity_filter":{"equalTo":"true"},"punctuate":{"equalTo":"true"},"redact":{"equalTo":"redact"},"replace":{"equalTo":"replace"},"search":{"equalTo":"search"},"smart_format":{"equalTo":"true"},"utterances":{"equalTo":"true"},"utt_split":{"equalTo":"1.1"},"version":{"equalTo":"latest"},"mip_opt_out":{"equalTo":"true"}}},"response":{"status":200,"body":"{\n \"metadata\": {\n \"request_id\": \"a847f427-4ad5-4d67-9b95-db801e58251c\",\n \"sha256\": \"154e291ecfa8be6ab8343560bcc109008fa7853eb5372533e8efdefc9b504c33\",\n \"created\": \"2024-05-12T18:57:13Z\",\n \"duration\": 25.933313,\n \"channels\": 1,\n \"models\": [\n \"30089e05-99d1-4376-b32e-c263170674af\"\n ],\n \"model_info\": {\n \"30089e05-99d1-4376-b32e-c263170674af\": {\n \"name\": \"2-general-nova\",\n \"version\": \"2024-01-09.29447\",\n \"arch\": \"nova-2\"\n }\n },\n \"summary_info\": {\n \"model_uuid\": \"67875a7f-c9c4-48a0-aa55-5bdb8a91c34a\",\n \"input_tokens\": 95,\n \"output_tokens\": 63\n },\n \"sentiment_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 105\n },\n \"topics_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 7\n },\n \"intents_info\": {\n \"model_uuid\": \"80ab3179-d113-4254-bd6b-4a2f96498695\",\n \"input_tokens\": 105,\n \"output_tokens\": 4\n },\n \"tags\": [\n \"test\"\n ]\n },\n \"results\": {\n \"channels\": [\n {}\n ],\n \"utterances\": [\n {}\n ],\n \"summary\": {\n \"result\": \"success\",\n \"short\": \"Speaker 0 discusses the significance of the first all-female spacewalk with an all-female team, stating that it is a tribute to the skilled and qualified women who were denied opportunities in the past.\"\n },\n \"topics\": {\n \"results\": {\n \"topics\": {\n \"segments\": [\n {\n \"text\": \"And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 32,\n \"end_word\": 69,\n \"topics\": [\n {\n \"topic\": \"Spacewalk\",\n \"confidence_score\": 0.91581345\n }\n ]\n }\n ]\n }\n }\n },\n \"intents\": {\n \"results\": {\n \"intents\": {\n \"segments\": [\n {\n \"text\": \"If you found this valuable, you can subscribe to the show on spotify or your favorite podcast app.\",\n \"start_word\": 354,\n \"end_word\": 414,\n \"intents\": [\n {\n \"intent\": \"Encourage podcasting\",\n \"confidence_score\": 0.0038975573\n }\n ]\n }\n ]\n }\n }\n },\n \"sentiments\": {\n \"segments\": [\n {\n \"text\": \"Yeah. As as much as, um, it's worth celebrating, uh, the first, uh, spacewalk, um, with an all-female team, I think many of us are looking forward to it just being normal. And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 0,\n \"end_word\": 69,\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810546875\n }\n ],\n \"average\": {\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810185185185185\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"49d8d51a-7f01-4598-804f-b6f54cdc22da","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"0a0be61b-f024-4120-9c54-23bca3e07c93","name":"List Models - default","request":{"urlPathTemplate":"/v1/models","method":"GET","queryParameters":{"include_outdated":{"equalTo":"true"}}},"response":{"status":200,"body":"{\n \"stt\": [\n {\n \"name\": \"nova-3\",\n \"canonical_name\": \"nova-3\",\n \"architecture\": \"base\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2021-11-10.1\",\n \"uuid\": \"6b28e919-8427-4f32-9847-492e2efd7daf\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": true\n }\n ],\n \"tts\": [\n {\n \"name\": \"zeus\",\n \"canonical_name\": \"aura-2-zeus-en\",\n \"architecture\": \"aura-2\",\n \"languages\": [\n \"en\",\n \"en-US\"\n ],\n \"version\": \"2025-04-07.0\",\n \"uuid\": \"2baf189d-91ac-481d-b6d1-750888667b31\",\n \"metadata\": {\n \"accent\": \"American\",\n \"age\": \"Adult\",\n \"color\": \"#C58DFF\",\n \"image\": \"https://static.deepgram.com/examples/avatars/zeus.jpg\",\n \"sample\": \"https://static.deepgram.com/examples/Aura-2-zeus.wav\",\n \"tags\": [\n \"masculine\",\n \"deep\",\n \"trustworthy\",\n \"smooth\"\n ],\n \"use_cases\": [\n \"IVR\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"0a0be61b-f024-4120-9c54-23bca3e07c93","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"0f94d3ea-43b6-4a1a-bce4-ab05b85440ae","name":"Get a specific Model - default","request":{"urlPathTemplate":"/v1/models/{model_id}","method":"GET","pathParameters":{"model_id":{"equalTo":"af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291"}}},"response":{"status":200,"body":"{\n \"name\": \"general\",\n \"canonical_name\": \"enhanced-general\",\n \"architecture\": \"polaris\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2022-05-18.1\",\n \"uuid\": \"c7226e9e-ae1c-4057-ae2a-a71a6b0dc588\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": false\n}","headers":{"Content-Type":"application/json"}},"uuid":"0f94d3ea-43b6-4a1a-bce4-ab05b85440ae","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d08cac56-bc4d-4756-8fd1-d508914334d5","name":"List Projects - default","request":{"urlPathTemplate":"/v1/projects","method":"GET"},"response":{"status":200,"body":"{\n \"projects\": [\n {\n \"project_id\": \"project_id\",\n \"name\": \"name\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"d08cac56-bc4d-4756-8fd1-d508914334d5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}},"postServeActions":[]},{"id":"6f0163a8-530c-4e25-bbe0-9ca86b9525dc","name":"Get a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"limit":{"equalTo":"1.1"},"page":{"equalTo":"1.1"}}},"response":{"status":200,"body":"{\n \"project_id\": \"project_id\",\n \"mip_opt_out\": true,\n \"name\": \"name\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"6f0163a8-530c-4e25-bbe0-9ca86b9525dc","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"68918577-6401-4439-8533-356257ff7bcf","name":"Delete a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"68918577-6401-4439-8533-356257ff7bcf","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"659fc38b-3934-4e43-93bf-d331f547449e","name":"Update a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}","method":"PATCH","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"Successfully updated project info.\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"659fc38b-3934-4e43-93bf-d331f547449e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"43ae6622-ad2f-4c81-9bc9-a8bbe17ef9d8","name":"Leave a Project - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/leave","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"43ae6622-ad2f-4c81-9bc9-a8bbe17ef9d8","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"032d9b1c-3b87-40fb-bfab-8c5be92a5d71","name":"List Project Keys - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"status":{"equalTo":"active"}}},"response":{"status":200,"body":"{\n \"api_keys\": [\n {\n \"member\": {\n \"member_id\": \"1000-2000-3000-4000\",\n \"email\": \"john@test.com\"\n },\n \"api_key\": {\n \"api_key_id\": \"1234567890abcdef1234567890abcdef\",\n \"comment\": \"A comment\",\n \"scopes\": [\n \"admin\"\n ],\n \"created\": \"2021-01-01T00:00:00Z\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"032d9b1c-3b87-40fb-bfab-8c5be92a5d71","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"0167c735-0b6f-4715-8df8-32300d4dae72","name":"Create a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys","method":"POST","pathParameters":{"project_id":{"equalTo":"project_id"}}},"response":{"status":200,"body":"{\n \"api_key_id\": \"api_key_id\",\n \"key\": \"key\",\n \"comment\": \"comment\",\n \"scopes\": [\n \"scopes\",\n \"scopes\"\n ],\n \"tags\": [\n \"tags\",\n \"tags\"\n ],\n \"expiration_date\": \"2024-01-15T09:30:00Z\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"0167c735-0b6f-4715-8df8-32300d4dae72","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"c9812dd3-f87e-4798-aec3-af0933330dd5","name":"Get a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys/{key_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"key_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"item\": {\n \"member\": {\n \"member_id\": \"1000-2000-3000-4000\",\n \"email\": \"john@test.com\",\n \"first_name\": \"John\",\n \"last_name\": \"Doe\",\n \"api_key\": {\n \"api_key_id\": \"1000-2000-3000-4000\",\n \"comment\": \"A comment\",\n \"scopes\": [\n \"admin\"\n ],\n \"tags\": [\n \"prod\",\n \"west-region\"\n ],\n \"expiration_date\": \"2021-01-01T00:00:00Z\",\n \"created\": \"2021-01-01T00:00:00Z\"\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"c9812dd3-f87e-4798-aec3-af0933330dd5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"3d2fbc7c-7bac-436f-a6ac-abe1b2c2caac","name":"Delete a Project Key - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/keys/{key_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"key_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"3d2fbc7c-7bac-436f-a6ac-abe1b2c2caac","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"91e103d5-72f7-463d-840d-310069e33de9","name":"List Project Members - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"members\": [\n {\n \"member_id\": \"member_id\",\n \"email\": \"email\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"91e103d5-72f7-463d-840d-310069e33de9","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"515c6f7e-09c3-43ea-ad6c-65bc11d20f46","name":"Delete a Project Member - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"515c6f7e-09c3-43ea-ad6c-65bc11d20f46","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a920ad0e-2796-4361-ac16-ac83fb75e32a","name":"List Project Models - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/models","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"include_outdated":{"equalTo":"true"}}},"response":{"status":200,"body":"{\n \"stt\": [\n {\n \"name\": \"nova-3\",\n \"canonical_name\": \"nova-3\",\n \"architecture\": \"base\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2021-11-10.1\",\n \"uuid\": \"6b28e919-8427-4f32-9847-492e2efd7daf\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": true\n }\n ],\n \"tts\": [\n {\n \"name\": \"zeus\",\n \"canonical_name\": \"aura-2-zeus-en\",\n \"architecture\": \"aura-2\",\n \"languages\": [\n \"en\",\n \"en-US\"\n ],\n \"version\": \"2025-04-07.0\",\n \"uuid\": \"2baf189d-91ac-481d-b6d1-750888667b31\",\n \"metadata\": {\n \"accent\": \"American\",\n \"age\": \"Adult\",\n \"color\": \"#C58DFF\",\n \"image\": \"https://static.deepgram.com/examples/avatars/zeus.jpg\",\n \"sample\": \"https://static.deepgram.com/examples/Aura-2-zeus.wav\",\n \"tags\": [\n \"masculine\",\n \"deep\",\n \"trustworthy\",\n \"smooth\"\n ],\n \"use_cases\": [\n \"IVR\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"a920ad0e-2796-4361-ac16-ac83fb75e32a","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9f8c6bf2-ebee-4956-b39f-0291b9d64b6e","name":"Get a Project Model - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/models/{model_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"model_id":{"equalTo":"af6e9977-99f6-4d8f-b6f5-dfdf6fb6e291"}}},"response":{"status":200,"body":"{\n \"name\": \"general\",\n \"canonical_name\": \"enhanced-general\",\n \"architecture\": \"polaris\",\n \"languages\": [\n \"en\",\n \"en-us\"\n ],\n \"version\": \"2022-05-18.1\",\n \"uuid\": \"c7226e9e-ae1c-4057-ae2a-a71a6b0dc588\",\n \"batch\": true,\n \"streaming\": true,\n \"formatted_output\": false\n}","headers":{"Content-Type":"application/json"}},"uuid":"9f8c6bf2-ebee-4956-b39f-0291b9d64b6e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d6a14959-05fa-4aec-9f0c-ba2a817c66e5","name":"List Project Requests - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/requests","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"2024-01-15T09:30:00Z"},"end":{"equalTo":"2024-01-15T09:30:00Z"},"limit":{"equalTo":"1.1"},"page":{"equalTo":"1.1"},"accessor":{"equalTo":"12345678-1234-1234-1234-123456789012"},"request_id":{"equalTo":"12345678-1234-1234-1234-123456789012"},"deployment":{"equalTo":"hosted"},"endpoint":{"equalTo":"listen"},"method":{"equalTo":"sync"},"status":{"equalTo":"succeeded"}}},"response":{"status":200,"body":"{\n \"page\": 1.1,\n \"limit\": 1.1,\n \"requests\": [\n {\n \"request_id\": \"request_id\",\n \"project_uuid\": \"project_uuid\",\n \"created\": \"2024-01-15T09:30:00Z\",\n \"path\": \"path\",\n \"api_key_id\": \"api_key_id\",\n \"response\": {\n \"key\": \"value\"\n },\n \"code\": 1.1,\n \"deployment\": \"deployment\",\n \"callback\": \"callback\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"d6a14959-05fa-4aec-9f0c-ba2a817c66e5","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"733e39aa-d3ef-4ea7-8062-af080c6288c4","name":"Get a Project Request - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/requests/{request_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"request_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"request\": {\n \"request_id\": \"request_id\",\n \"project_uuid\": \"project_uuid\",\n \"created\": \"2024-01-15T09:30:00Z\",\n \"path\": \"path\",\n \"api_key_id\": \"api_key_id\",\n \"response\": {\n \"key\": \"value\"\n },\n \"code\": 1.1,\n \"deployment\": \"deployment\",\n \"callback\": \"callback\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"733e39aa-d3ef-4ea7-8062-af080c6288c4","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"6309dd55-c993-4ce1-b0b2-01a41c9f08d6","name":"Get Project Usage - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"start"},"end":{"equalTo":"end"},"accessor":{"equalTo":"12345678-1234-1234-1234-123456789012"},"alternatives":{"equalTo":"true"},"callback_method":{"equalTo":"true"},"callback":{"equalTo":"true"},"channels":{"equalTo":"true"},"custom_intent_mode":{"equalTo":"true"},"custom_intent":{"equalTo":"true"},"custom_topic_mode":{"equalTo":"true"},"custom_topic":{"equalTo":"true"},"deployment":{"equalTo":"hosted"},"detect_entities":{"equalTo":"true"},"detect_language":{"equalTo":"true"},"diarize":{"equalTo":"true"},"dictation":{"equalTo":"true"},"encoding":{"equalTo":"true"},"endpoint":{"equalTo":"listen"},"extra":{"equalTo":"true"},"filler_words":{"equalTo":"true"},"intents":{"equalTo":"true"},"keyterm":{"equalTo":"true"},"keywords":{"equalTo":"true"},"language":{"equalTo":"true"},"measurements":{"equalTo":"true"},"method":{"equalTo":"sync"},"model":{"equalTo":"6f548761-c9c0-429a-9315-11a1d28499c8"},"multichannel":{"equalTo":"true"},"numerals":{"equalTo":"true"},"paragraphs":{"equalTo":"true"},"profanity_filter":{"equalTo":"true"},"punctuate":{"equalTo":"true"},"redact":{"equalTo":"true"},"replace":{"equalTo":"true"},"sample_rate":{"equalTo":"true"},"search":{"equalTo":"true"},"sentiment":{"equalTo":"true"},"smart_format":{"equalTo":"true"},"summarize":{"equalTo":"true"},"tag":{"equalTo":"tag1"},"topics":{"equalTo":"true"},"utt_split":{"equalTo":"true"},"utterances":{"equalTo":"true"},"version":{"equalTo":"true"}}},"response":{"status":200,"body":"{\n \"start\": \"2024-10-16\",\n \"end\": \"2024-10-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"6309dd55-c993-4ce1-b0b2-01a41c9f08d6","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b132121b-4efe-42ad-a268-8acac35c189b","name":"Get Project Balances - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/balances","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"balances\": [\n {\n \"balance_id\": \"balance_id\",\n \"amount\": 1.1,\n \"units\": \"units\",\n \"purchase_order_id\": \"purchase_order_id\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"b132121b-4efe-42ad-a268-8acac35c189b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"4019c244-52d3-4d57-902c-af837631650a","name":"Get a Project Balance - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/balances/{balance_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"balance_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"balance_id\": \"balance_id\",\n \"amount\": 1.1,\n \"units\": \"units\",\n \"purchase_order_id\": \"purchase_order_id\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"4019c244-52d3-4d57-902c-af837631650a","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"555b6751-587f-400c-bf5e-400e108ad6b4","name":"Get Project Billing Breakdown - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/billing/breakdown","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"start"},"end":{"equalTo":"end"},"accessor":{"equalTo":"12345678-1234-1234-1234-123456789012"},"deployment":{"equalTo":"hosted"},"tag":{"equalTo":"tag1"},"line_item":{"equalTo":"streaming::nova-3"}}},"response":{"status":200,"body":"{\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n },\n \"results\": [\n {\n \"dollars\": 0.25,\n \"grouping\": {\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-16\",\n \"accessor\": \"123456789012345678901234\",\n \"deployment\": \"hosted\",\n \"line_item\": \"streaming::nova-3\",\n \"tags\": [\n \"tag1\",\n \"tag2\"\n ]\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"555b6751-587f-400c-bf5e-400e108ad6b4","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a61ae38c-e41f-4726-a55c-88f2135897be","name":"List Project Billing Fields - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/billing/fields","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"start"},"end":{"equalTo":"end"}}},"response":{"status":200,"body":"{\n \"accessors\": [\n \"12345678-1234-1234-1234-123456789012\",\n \"87654321-4321-4321-4321-210987654321\"\n ],\n \"deployments\": [\n \"hosted\",\n \"self-hosted\"\n ],\n \"tags\": [\n \"dev\",\n \"production\"\n ],\n \"line_items\": {\n \"streaming::nova-3\": \"Nova - 3 (Stream)\",\n \"sync::aura-2\": \"Aura -2 (Sync)\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"a61ae38c-e41f-4726-a55c-88f2135897be","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"85b4373c-ba39-41b1-84e8-ae1ee6b180ca","name":"List Project Purchases - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/purchases","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"limit":{"equalTo":"1.1"}}},"response":{"status":200,"body":"{\n \"orders\": [\n {\n \"order_id\": \"025e19ba-b6d9-4a04-9f99-4fe715aca5f1\",\n \"expiration\": \"2026-03-04T00:00:00Z\",\n \"created\": \"2023-02-21T21:13:40Z\",\n \"amount\": 150,\n \"units\": \"usd\",\n \"order_type\": \"promotional\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"85b4373c-ba39-41b1-84e8-ae1ee6b180ca","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"2dd14c67-ed4e-4d97-9636-0a712899deb8","name":"List Project Invites - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"invites\": [\n {\n \"email\": \"email\",\n \"scope\": \"scope\"\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"2dd14c67-ed4e-4d97-9636-0a712899deb8","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"7c109496-adfe-4e85-b007-a6f799ee95cb","name":"Create a Project Invite - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites","method":"POST","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"7c109496-adfe-4e85-b007-a6f799ee95cb","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"d6d268d0-d91e-4a65-80e0-339621173db9","name":"Delete a Project Invite - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/invites/{email}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"email":{"equalTo":"john.doe@example.com"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"d6d268d0-d91e-4a65-80e0-339621173db9","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"1b965d71-c930-4a0b-90f3-2289f80f3634","name":"List Project Member Scopes - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}/scopes","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"scopes\": [\n \"scopes\"\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"1b965d71-c930-4a0b-90f3-2289f80f3634","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"eb2f5de2-b887-47be-abd5-7cb702aca55d","name":"Update Project Member Scopes - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/members/{member_id}/scopes","method":"PUT","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"member_id":{"equalTo":"123456789012345678901234"}}},"response":{"status":200,"body":"{\n \"message\": \"message\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"eb2f5de2-b887-47be-abd5-7cb702aca55d","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9bdf51a4-1e10-41b8-8de2-2df650562db3","name":"Get Project Usage Breakdown - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage/breakdown","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"start"},"end":{"equalTo":"end"},"grouping":{"equalTo":"accessor"},"accessor":{"equalTo":"12345678-1234-1234-1234-123456789012"},"alternatives":{"equalTo":"true"},"callback_method":{"equalTo":"true"},"callback":{"equalTo":"true"},"channels":{"equalTo":"true"},"custom_intent_mode":{"equalTo":"true"},"custom_intent":{"equalTo":"true"},"custom_topic_mode":{"equalTo":"true"},"custom_topic":{"equalTo":"true"},"deployment":{"equalTo":"hosted"},"detect_entities":{"equalTo":"true"},"detect_language":{"equalTo":"true"},"diarize":{"equalTo":"true"},"dictation":{"equalTo":"true"},"encoding":{"equalTo":"true"},"endpoint":{"equalTo":"listen"},"extra":{"equalTo":"true"},"filler_words":{"equalTo":"true"},"intents":{"equalTo":"true"},"keyterm":{"equalTo":"true"},"keywords":{"equalTo":"true"},"language":{"equalTo":"true"},"measurements":{"equalTo":"true"},"method":{"equalTo":"sync"},"model":{"equalTo":"6f548761-c9c0-429a-9315-11a1d28499c8"},"multichannel":{"equalTo":"true"},"numerals":{"equalTo":"true"},"paragraphs":{"equalTo":"true"},"profanity_filter":{"equalTo":"true"},"punctuate":{"equalTo":"true"},"redact":{"equalTo":"true"},"replace":{"equalTo":"true"},"sample_rate":{"equalTo":"true"},"search":{"equalTo":"true"},"sentiment":{"equalTo":"true"},"smart_format":{"equalTo":"true"},"summarize":{"equalTo":"true"},"tag":{"equalTo":"tag1"},"topics":{"equalTo":"true"},"utt_split":{"equalTo":"true"},"utterances":{"equalTo":"true"},"version":{"equalTo":"true"}}},"response":{"status":200,"body":"{\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-23\",\n \"resolution\": {\n \"units\": \"day\",\n \"amount\": 1\n },\n \"results\": [\n {\n \"hours\": 1619.7242069444444,\n \"total_hours\": 1621.7395791666668,\n \"agent_hours\": 41.33564388888889,\n \"tokens_in\": 0,\n \"tokens_out\": 0,\n \"tts_characters\": 9158866,\n \"requests\": 373381,\n \"grouping\": {\n \"start\": \"2025-01-16\",\n \"end\": \"2025-01-16\",\n \"accessor\": \"123456789012345678901234\",\n \"endpoint\": \"listen\",\n \"feature_set\": \"punctuate\",\n \"models\": [\n \"Nova-2\"\n ],\n \"method\": \"async\",\n \"tags\": \"tag1\",\n \"deployment\": \"self-hosted\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"9bdf51a4-1e10-41b8-8de2-2df650562db3","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"bc9fcd54-076e-48dc-a2dd-d71a8bf8bd4e","name":"List Project Usage Fields - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/usage/fields","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"start":{"equalTo":"start"},"end":{"equalTo":"end"}}},"response":{"status":200,"body":"{\n \"tags\": [\n \"tag=dev\",\n \"tag=production\"\n ],\n \"models\": [\n {\n \"name\": \"2-medical-nova\",\n \"language\": \"en-MY\",\n \"version\": \"2024-05-31.13574\",\n \"model_id\": \"1234567890-12345-67890\"\n }\n ],\n \"processing_methods\": [\n \"sync\",\n \"streaming\"\n ],\n \"features\": [\n \"alternatives\",\n \"detect_entities\",\n \"detect_language\"\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"bc9fcd54-076e-48dc-a2dd-d71a8bf8bd4e","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b5ac3651-d3b7-4cd7-b0b8-e1a917a16f3b","name":"Analyze text content - default","request":{"urlPathTemplate":"/v1/read","method":"POST","queryParameters":{"callback":{"equalTo":"callback"},"callback_method":{"equalTo":"POST"},"sentiment":{"equalTo":"true"},"summarize":{"equalTo":"v2"},"tag":{"equalTo":"tag"},"topics":{"equalTo":"true"},"custom_topic":{"equalTo":"custom_topic"},"custom_topic_mode":{"equalTo":"extended"},"intents":{"equalTo":"true"},"custom_intent":{"equalTo":"custom_intent"},"custom_intent_mode":{"equalTo":"extended"},"language":{"equalTo":"language"}}},"response":{"status":200,"body":"{\n \"metadata\": {\n \"metadata\": {\n \"request_id\": \"d04af392-db11-4c1d-83e1-20e34f0b8999\",\n \"created\": \"2024-11-18T23:47:44Z\",\n \"language\": \"en\"\n }\n },\n \"results\": {\n \"summary\": {\n \"results\": {\n \"summary\": {\n \"text\": \"The summary of the text submitted.\"\n }\n }\n },\n \"topics\": {\n \"results\": {\n \"topics\": {\n \"segments\": [\n {\n \"text\": \"And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 32,\n \"end_word\": 69,\n \"topics\": [\n {\n \"topic\": \"Spacewalk\",\n \"confidence_score\": 0.91581345\n }\n ]\n }\n ]\n }\n }\n },\n \"intents\": {\n \"results\": {\n \"intents\": {\n \"segments\": [\n {\n \"text\": \"If you found this valuable, you can subscribe to the show on spotify or your favorite podcast app.\",\n \"start_word\": 354,\n \"end_word\": 414,\n \"intents\": [\n {\n \"intent\": \"Encourage podcasting\",\n \"confidence_score\": 0.0038975573\n }\n ]\n }\n ]\n }\n }\n },\n \"sentiments\": {\n \"segments\": [\n {\n \"text\": \"Yeah. As as much as, um, it's worth celebrating, uh, the first, uh, spacewalk, um, with an all-female team, I think many of us are looking forward to it just being normal. And, um, I think if it signifies anything, it is, uh, to honor the the women who came before us who, um, were skilled and qualified, um, and didn't get the the same opportunities that we have today.\",\n \"start_word\": 0,\n \"end_word\": 69,\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810546875\n }\n ],\n \"average\": {\n \"sentiment\": \"positive\",\n \"sentiment_score\": 0.5810185185185185\n }\n }\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"b5ac3651-d3b7-4cd7-b0b8-e1a917a16f3b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"4110cb96-50e2-4fe6-b8ae-5d69120cee89","name":"List Project Self-Hosted Distribution Credentials - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}}},"response":{"status":200,"body":"{\n \"distribution_credentials\": [\n {\n \"member\": {\n \"member_id\": \"3376abcd-8e5e-49d3-92d4-876d3a4f0363\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"8b36cfd0-472f-4a21-833f-2d6343c3a2f3\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n }\n ]\n}","headers":{"Content-Type":"application/json"}},"uuid":"4110cb96-50e2-4fe6-b8ae-5d69120cee89","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"9c12eea9-6ba6-4d70-bb14-a2742cebc114","name":"Create a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials","method":"POST","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"}},"queryParameters":{"provider":{"equalTo":"quay"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"9c12eea9-6ba6-4d70-bb14-a2742cebc114","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"a47cd13f-2314-4190-b2c7-20436ccffbd2","name":"Get a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials/{distribution_credentials_id}","method":"GET","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"distribution_credentials_id":{"equalTo":"8b36cfd0-472f-4a21-833f-2d6343c3a2f3"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"a47cd13f-2314-4190-b2c7-20436ccffbd2","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"8bd46091-0e57-4b3d-9485-a86e6f1eaf17","name":"Delete a Project Self-Hosted Distribution Credential - default","request":{"urlPathTemplate":"/v1/projects/{project_id}/self-hosted/distribution/credentials/{distribution_credentials_id}","method":"DELETE","pathParameters":{"project_id":{"equalTo":"123456-7890-1234-5678-901234"},"distribution_credentials_id":{"equalTo":"8b36cfd0-472f-4a21-833f-2d6343c3a2f3"}}},"response":{"status":200,"body":"{\n \"member\": {\n \"member_id\": \"c7b9b131-73f3-11d9-8665-0b00d2e44b83\",\n \"email\": \"email@example.com\"\n },\n \"distribution_credentials\": {\n \"distribution_credentials_id\": \"82c32c10-53b2-4d23-993f-864b3d44502a\",\n \"provider\": \"quay\",\n \"comment\": \"My Self-Hosted Distribution Credentials\",\n \"scopes\": [\n \"self-hosted:product:api\",\n \"self-hosted:product:engine\"\n ],\n \"created\": \"2023-06-28T15:36:59Z\"\n }\n}","headers":{"Content-Type":"application/json"}},"uuid":"8bd46091-0e57-4b3d-9485-a86e6f1eaf17","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}},{"id":"b06ec977-02ed-41e2-8fff-2bc45cd2166b","name":"Text to Speech transformation - default","request":{"urlPathTemplate":"/v1/speak","method":"POST"},"response":{"status":200,"body":"{\n \"key\": \"value\"\n}","headers":{"Content-Type":"application/json"}},"uuid":"b06ec977-02ed-41e2-8fff-2bc45cd2166b","persistent":true,"priority":3,"metadata":{"mocklab":{"created":{"at":"2020-01-01T00:00:00.000Z","via":"SYSTEM"}}}}],"meta":{"total":39}} \ No newline at end of file