From 82930d1a816370b3c4798f1e0f9cbe47181fd848 Mon Sep 17 00:00:00 2001 From: "stainless-sdks[bot]" <167585319+stainless-sdks[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:25:17 +0000 Subject: [PATCH 01/47] Initial commit --- README.md | 1 + 1 file changed, 1 insertion(+) create mode 100644 README.md diff --git a/README.md b/README.md new file mode 100644 index 0000000..7258cf5 --- /dev/null +++ b/README.md @@ -0,0 +1 @@ +# brapi-python \ No newline at end of file From 484467779bb9f7cb448da605d4bd97707a713068 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:25:34 +0000 Subject: [PATCH 02/47] chore: configure new SDK language --- .devcontainer/Dockerfile | 9 + .devcontainer/devcontainer.json | 43 + .github/workflows/ci.yml | 98 + .gitignore | 15 + .python-version | 1 + .stats.yml | 4 + .vscode/settings.json | 3 + Brewfile | 2 + CONTRIBUTING.md | 128 ++ LICENSE | 201 ++ README.md | 387 +++- SECURITY.md | 27 + api.md | 87 + bin/publish-pypi | 6 + examples/.keep | 4 + noxfile.py | 9 + pyproject.toml | 267 +++ requirements-dev.lock | 137 ++ requirements.lock | 75 + scripts/bootstrap | 27 + scripts/format | 8 + scripts/lint | 11 + scripts/mock | 41 + scripts/test | 61 + scripts/utils/ruffen-docs.py | 167 ++ scripts/utils/upload-artifact.sh | 27 + src/brapi/__init__.py | 104 + src/brapi/_base_client.py | 1995 +++++++++++++++++ src/brapi/_client.py | 478 ++++ src/brapi/_compat.py | 219 ++ src/brapi/_constants.py | 14 + src/brapi/_exceptions.py | 108 + src/brapi/_files.py | 123 + src/brapi/_models.py | 835 +++++++ src/brapi/_qs.py | 150 ++ src/brapi/_resource.py | 43 + src/brapi/_response.py | 830 +++++++ src/brapi/_streaming.py | 333 +++ src/brapi/_types.py | 260 +++ src/brapi/_utils/__init__.py | 64 + src/brapi/_utils/_compat.py | 45 + src/brapi/_utils/_datetime_parse.py | 136 ++ src/brapi/_utils/_logs.py | 25 + src/brapi/_utils/_proxy.py | 65 + src/brapi/_utils/_reflection.py | 42 + src/brapi/_utils/_resources_proxy.py | 24 + src/brapi/_utils/_streams.py | 12 + src/brapi/_utils/_sync.py | 86 + src/brapi/_utils/_transform.py | 457 ++++ src/brapi/_utils/_typing.py | 156 ++ src/brapi/_utils/_utils.py | 421 ++++ src/brapi/_version.py | 4 + src/brapi/lib/.keep | 4 + src/brapi/py.typed | 0 src/brapi/resources/__init__.py | 47 + src/brapi/resources/available.py | 283 +++ src/brapi/resources/quote.py | 959 ++++++++ src/brapi/resources/v2/__init__.py | 75 + src/brapi/resources/v2/crypto.py | 524 +++++ src/brapi/resources/v2/currency.py | 456 ++++ src/brapi/resources/v2/inflation.py | 530 +++++ src/brapi/resources/v2/prime_rate.py | 490 ++++ src/brapi/resources/v2/v2.py | 198 ++ src/brapi/types/__init__.py | 16 + src/brapi/types/available_list_params.py | 30 + src/brapi/types/available_list_response.py | 18 + src/brapi/types/balance_sheet_entry.py | 455 ++++ src/brapi/types/cashflow_entry.py | 89 + .../types/default_key_statistics_entry.py | 138 ++ src/brapi/types/financial_data_entry.py | 133 ++ src/brapi/types/income_statement_entry.py | 198 ++ src/brapi/types/quote_list_params.py | 86 + src/brapi/types/quote_list_response.py | 99 + src/brapi/types/quote_retrieve_params.py | 113 + src/brapi/types/quote_retrieve_response.py | 476 ++++ src/brapi/types/v2/__init__.py | 20 + .../types/v2/crypto_list_available_params.py | 31 + .../v2/crypto_list_available_response.py | 15 + src/brapi/types/v2/crypto_retrieve_params.py | 59 + .../types/v2/crypto_retrieve_response.py | 117 + .../v2/currency_list_available_params.py | 30 + .../v2/currency_list_available_response.py | 24 + .../types/v2/currency_retrieve_params.py | 35 + .../types/v2/currency_retrieve_response.py | 84 + .../v2/inflation_list_available_params.py | 30 + .../v2/inflation_list_available_response.py | 15 + .../types/v2/inflation_retrieve_params.py | 63 + .../types/v2/inflation_retrieve_response.py | 34 + .../v2/prime_rate_list_available_params.py | 30 + .../v2/prime_rate_list_available_response.py | 15 + .../types/v2/prime_rate_retrieve_params.py | 67 + .../types/v2/prime_rate_retrieve_response.py | 28 + src/brapi/types/value_added_entry.py | 240 ++ tests/__init__.py | 1 + tests/api_resources/__init__.py | 1 + tests/api_resources/test_available.py | 98 + tests/api_resources/test_quote.py | 222 ++ tests/api_resources/v2/__init__.py | 1 + tests/api_resources/v2/test_crypto.py | 193 ++ tests/api_resources/v2/test_currency.py | 187 ++ tests/api_resources/v2/test_inflation.py | 186 ++ tests/api_resources/v2/test_prime_rate.py | 186 ++ tests/conftest.py | 84 + tests/sample_file.txt | 1 + tests/test_client.py | 1699 ++++++++++++++ tests/test_deepcopy.py | 58 + tests/test_extract_files.py | 64 + tests/test_files.py | 51 + tests/test_models.py | 963 ++++++++ tests/test_qs.py | 78 + tests/test_required_args.py | 111 + tests/test_response.py | 277 +++ tests/test_streaming.py | 248 ++ tests/test_transform.py | 460 ++++ tests/test_utils/test_datetime_parse.py | 110 + tests/test_utils/test_proxy.py | 34 + tests/test_utils/test_typing.py | 73 + tests/utils.py | 167 ++ 118 files changed, 20780 insertions(+), 1 deletion(-) create mode 100644 .devcontainer/Dockerfile create mode 100644 .devcontainer/devcontainer.json create mode 100644 .github/workflows/ci.yml create mode 100644 .gitignore create mode 100644 .python-version create mode 100644 .stats.yml create mode 100644 .vscode/settings.json create mode 100644 Brewfile create mode 100644 CONTRIBUTING.md create mode 100644 LICENSE create mode 100644 SECURITY.md create mode 100644 api.md create mode 100644 bin/publish-pypi create mode 100644 examples/.keep create mode 100644 noxfile.py create mode 100644 pyproject.toml create mode 100644 requirements-dev.lock create mode 100644 requirements.lock create mode 100755 scripts/bootstrap create mode 100755 scripts/format create mode 100755 scripts/lint create mode 100755 scripts/mock create mode 100755 scripts/test create mode 100644 scripts/utils/ruffen-docs.py create mode 100755 scripts/utils/upload-artifact.sh create mode 100644 src/brapi/__init__.py create mode 100644 src/brapi/_base_client.py create mode 100644 src/brapi/_client.py create mode 100644 src/brapi/_compat.py create mode 100644 src/brapi/_constants.py create mode 100644 src/brapi/_exceptions.py create mode 100644 src/brapi/_files.py create mode 100644 src/brapi/_models.py create mode 100644 src/brapi/_qs.py create mode 100644 src/brapi/_resource.py create mode 100644 src/brapi/_response.py create mode 100644 src/brapi/_streaming.py create mode 100644 src/brapi/_types.py create mode 100644 src/brapi/_utils/__init__.py create mode 100644 src/brapi/_utils/_compat.py create mode 100644 src/brapi/_utils/_datetime_parse.py create mode 100644 src/brapi/_utils/_logs.py create mode 100644 src/brapi/_utils/_proxy.py create mode 100644 src/brapi/_utils/_reflection.py create mode 100644 src/brapi/_utils/_resources_proxy.py create mode 100644 src/brapi/_utils/_streams.py create mode 100644 src/brapi/_utils/_sync.py create mode 100644 src/brapi/_utils/_transform.py create mode 100644 src/brapi/_utils/_typing.py create mode 100644 src/brapi/_utils/_utils.py create mode 100644 src/brapi/_version.py create mode 100644 src/brapi/lib/.keep create mode 100644 src/brapi/py.typed create mode 100644 src/brapi/resources/__init__.py create mode 100644 src/brapi/resources/available.py create mode 100644 src/brapi/resources/quote.py create mode 100644 src/brapi/resources/v2/__init__.py create mode 100644 src/brapi/resources/v2/crypto.py create mode 100644 src/brapi/resources/v2/currency.py create mode 100644 src/brapi/resources/v2/inflation.py create mode 100644 src/brapi/resources/v2/prime_rate.py create mode 100644 src/brapi/resources/v2/v2.py create mode 100644 src/brapi/types/__init__.py create mode 100644 src/brapi/types/available_list_params.py create mode 100644 src/brapi/types/available_list_response.py create mode 100644 src/brapi/types/balance_sheet_entry.py create mode 100644 src/brapi/types/cashflow_entry.py create mode 100644 src/brapi/types/default_key_statistics_entry.py create mode 100644 src/brapi/types/financial_data_entry.py create mode 100644 src/brapi/types/income_statement_entry.py create mode 100644 src/brapi/types/quote_list_params.py create mode 100644 src/brapi/types/quote_list_response.py create mode 100644 src/brapi/types/quote_retrieve_params.py create mode 100644 src/brapi/types/quote_retrieve_response.py create mode 100644 src/brapi/types/v2/__init__.py create mode 100644 src/brapi/types/v2/crypto_list_available_params.py create mode 100644 src/brapi/types/v2/crypto_list_available_response.py create mode 100644 src/brapi/types/v2/crypto_retrieve_params.py create mode 100644 src/brapi/types/v2/crypto_retrieve_response.py create mode 100644 src/brapi/types/v2/currency_list_available_params.py create mode 100644 src/brapi/types/v2/currency_list_available_response.py create mode 100644 src/brapi/types/v2/currency_retrieve_params.py create mode 100644 src/brapi/types/v2/currency_retrieve_response.py create mode 100644 src/brapi/types/v2/inflation_list_available_params.py create mode 100644 src/brapi/types/v2/inflation_list_available_response.py create mode 100644 src/brapi/types/v2/inflation_retrieve_params.py create mode 100644 src/brapi/types/v2/inflation_retrieve_response.py create mode 100644 src/brapi/types/v2/prime_rate_list_available_params.py create mode 100644 src/brapi/types/v2/prime_rate_list_available_response.py create mode 100644 src/brapi/types/v2/prime_rate_retrieve_params.py create mode 100644 src/brapi/types/v2/prime_rate_retrieve_response.py create mode 100644 src/brapi/types/value_added_entry.py create mode 100644 tests/__init__.py create mode 100644 tests/api_resources/__init__.py create mode 100644 tests/api_resources/test_available.py create mode 100644 tests/api_resources/test_quote.py create mode 100644 tests/api_resources/v2/__init__.py create mode 100644 tests/api_resources/v2/test_crypto.py create mode 100644 tests/api_resources/v2/test_currency.py create mode 100644 tests/api_resources/v2/test_inflation.py create mode 100644 tests/api_resources/v2/test_prime_rate.py create mode 100644 tests/conftest.py create mode 100644 tests/sample_file.txt create mode 100644 tests/test_client.py create mode 100644 tests/test_deepcopy.py create mode 100644 tests/test_extract_files.py create mode 100644 tests/test_files.py create mode 100644 tests/test_models.py create mode 100644 tests/test_qs.py create mode 100644 tests/test_required_args.py create mode 100644 tests/test_response.py create mode 100644 tests/test_streaming.py create mode 100644 tests/test_transform.py create mode 100644 tests/test_utils/test_datetime_parse.py create mode 100644 tests/test_utils/test_proxy.py create mode 100644 tests/test_utils/test_typing.py create mode 100644 tests/utils.py diff --git a/.devcontainer/Dockerfile b/.devcontainer/Dockerfile new file mode 100644 index 0000000..ff261ba --- /dev/null +++ b/.devcontainer/Dockerfile @@ -0,0 +1,9 @@ +ARG VARIANT="3.9" +FROM mcr.microsoft.com/vscode/devcontainers/python:0-${VARIANT} + +USER vscode + +RUN curl -sSf https://rye.astral.sh/get | RYE_VERSION="0.44.0" RYE_INSTALL_OPTION="--yes" bash +ENV PATH=/home/vscode/.rye/shims:$PATH + +RUN echo "[[ -d .venv ]] && source .venv/bin/activate || export PATH=\$PATH" >> /home/vscode/.bashrc diff --git a/.devcontainer/devcontainer.json b/.devcontainer/devcontainer.json new file mode 100644 index 0000000..c17fdc1 --- /dev/null +++ b/.devcontainer/devcontainer.json @@ -0,0 +1,43 @@ +// For format details, see https://aka.ms/devcontainer.json. For config options, see the +// README at: https://github.com/devcontainers/templates/tree/main/src/debian +{ + "name": "Debian", + "build": { + "dockerfile": "Dockerfile", + "context": ".." + }, + + "postStartCommand": "rye sync --all-features", + + "customizations": { + "vscode": { + "extensions": [ + "ms-python.python" + ], + "settings": { + "terminal.integrated.shell.linux": "/bin/bash", + "python.pythonPath": ".venv/bin/python", + "python.defaultInterpreterPath": ".venv/bin/python", + "python.typeChecking": "basic", + "terminal.integrated.env.linux": { + "PATH": "/home/vscode/.rye/shims:${env:PATH}" + } + } + } + }, + "features": { + "ghcr.io/devcontainers/features/node:1": {} + } + + // Features to add to the dev container. More info: https://containers.dev/features. + // "features": {}, + + // Use 'forwardPorts' to make a list of ports inside the container available locally. + // "forwardPorts": [], + + // Configure tool-specific properties. + // "customizations": {}, + + // Uncomment to connect as root instead. More info: https://aka.ms/dev-containers-non-root. + // "remoteUser": "root" +} diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml new file mode 100644 index 0000000..78bafd2 --- /dev/null +++ b/.github/workflows/ci.yml @@ -0,0 +1,98 @@ +name: CI +on: + push: + branches-ignore: + - 'generated' + - 'codegen/**' + - 'integrated/**' + - 'stl-preview-head/**' + - 'stl-preview-base/**' + pull_request: + branches-ignore: + - 'stl-preview-head/**' + - 'stl-preview-base/**' + +jobs: + lint: + timeout-minutes: 10 + name: lint + runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} + if: github.event_name == 'push' || github.event.pull_request.head.repo.fork + steps: + - uses: actions/checkout@v4 + + - name: Install Rye + run: | + curl -sSf https://rye.astral.sh/get | bash + echo "$HOME/.rye/shims" >> $GITHUB_PATH + env: + RYE_VERSION: '0.44.0' + RYE_INSTALL_OPTION: '--yes' + + - name: Install dependencies + run: rye sync --all-features + + - name: Run lints + run: ./scripts/lint + + build: + if: github.event_name == 'push' || github.event.pull_request.head.repo.fork + timeout-minutes: 10 + name: build + permissions: + contents: read + id-token: write + runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} + steps: + - uses: actions/checkout@v4 + + - name: Install Rye + run: | + curl -sSf https://rye.astral.sh/get | bash + echo "$HOME/.rye/shims" >> $GITHUB_PATH + env: + RYE_VERSION: '0.44.0' + RYE_INSTALL_OPTION: '--yes' + + - name: Install dependencies + run: rye sync --all-features + + - name: Run build + run: rye build + + - name: Get GitHub OIDC Token + if: github.repository == 'stainless-sdks/brapi-python' + id: github-oidc + uses: actions/github-script@v6 + with: + script: core.setOutput('github_token', await core.getIDToken()); + + - name: Upload tarball + if: github.repository == 'stainless-sdks/brapi-python' + env: + URL: https://pkg.stainless.com/s + AUTH: ${{ steps.github-oidc.outputs.github_token }} + SHA: ${{ github.sha }} + run: ./scripts/utils/upload-artifact.sh + + test: + timeout-minutes: 10 + name: test + runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} + if: github.event_name == 'push' || github.event.pull_request.head.repo.fork + steps: + - uses: actions/checkout@v4 + + - name: Install Rye + run: | + curl -sSf https://rye.astral.sh/get | bash + echo "$HOME/.rye/shims" >> $GITHUB_PATH + env: + RYE_VERSION: '0.44.0' + RYE_INSTALL_OPTION: '--yes' + + - name: Bootstrap + run: ./scripts/bootstrap + + - name: Run tests + run: ./scripts/test diff --git a/.gitignore b/.gitignore new file mode 100644 index 0000000..95ceb18 --- /dev/null +++ b/.gitignore @@ -0,0 +1,15 @@ +.prism.log +_dev + +__pycache__ +.mypy_cache + +dist + +.venv +.idea + +.env +.envrc +codegen.log +Brewfile.lock.json diff --git a/.python-version b/.python-version new file mode 100644 index 0000000..43077b2 --- /dev/null +++ b/.python-version @@ -0,0 +1 @@ +3.9.18 diff --git a/.stats.yml b/.stats.yml new file mode 100644 index 0000000..ec152aa --- /dev/null +++ b/.stats.yml @@ -0,0 +1,4 @@ +configured_endpoints: 11 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml +openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 +config_hash: 6de3ea33802724ce95047bc7f2a45358 diff --git a/.vscode/settings.json b/.vscode/settings.json new file mode 100644 index 0000000..5b01030 --- /dev/null +++ b/.vscode/settings.json @@ -0,0 +1,3 @@ +{ + "python.analysis.importFormat": "relative", +} diff --git a/Brewfile b/Brewfile new file mode 100644 index 0000000..492ca37 --- /dev/null +++ b/Brewfile @@ -0,0 +1,2 @@ +brew "rye" + diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md new file mode 100644 index 0000000..8266ffe --- /dev/null +++ b/CONTRIBUTING.md @@ -0,0 +1,128 @@ +## Setting up the environment + +### With Rye + +We use [Rye](https://rye.astral.sh/) to manage dependencies because it will automatically provision a Python environment with the expected Python version. To set it up, run: + +```sh +$ ./scripts/bootstrap +``` + +Or [install Rye manually](https://rye.astral.sh/guide/installation/) and run: + +```sh +$ rye sync --all-features +``` + +You can then run scripts using `rye run python script.py` or by activating the virtual environment: + +```sh +# Activate the virtual environment - https://docs.python.org/3/library/venv.html#how-venvs-work +$ source .venv/bin/activate + +# now you can omit the `rye run` prefix +$ python script.py +``` + +### Without Rye + +Alternatively if you don't want to install `Rye`, you can stick with the standard `pip` setup by ensuring you have the Python version specified in `.python-version`, create a virtual environment however you desire and then install dependencies using this command: + +```sh +$ pip install -r requirements-dev.lock +``` + +## Modifying/Adding code + +Most of the SDK is generated code. Modifications to code will be persisted between generations, but may +result in merge conflicts between manual patches and changes from the generator. The generator will never +modify the contents of the `src/brapi/lib/` and `examples/` directories. + +## Adding and running examples + +All files in the `examples/` directory are not modified by the generator and can be freely edited or added to. + +```py +# add an example to examples/.py + +#!/usr/bin/env -S rye run python +… +``` + +```sh +$ chmod +x examples/.py +# run the example against your api +$ ./examples/.py +``` + +## Using the repository from source + +If you’d like to use the repository from source, you can either install from git or link to a cloned repository: + +To install via git: + +```sh +$ pip install git+ssh://git@github.com/stainless-sdks/brapi-python.git +``` + +Alternatively, you can build from source and install the wheel file: + +Building this package will create two files in the `dist/` directory, a `.tar.gz` containing the source files and a `.whl` that can be used to install the package efficiently. + +To create a distributable version of the library, all you have to do is run this command: + +```sh +$ rye build +# or +$ python -m build +``` + +Then to install: + +```sh +$ pip install ./path-to-wheel-file.whl +``` + +## Running tests + +Most tests require you to [set up a mock server](https://github.com/stoplightio/prism) against the OpenAPI spec to run the tests. + +```sh +# you will need npm installed +$ npx prism mock path/to/your/openapi.yml +``` + +```sh +$ ./scripts/test +``` + +## Linting and formatting + +This repository uses [ruff](https://github.com/astral-sh/ruff) and +[black](https://github.com/psf/black) to format the code in the repository. + +To lint: + +```sh +$ ./scripts/lint +``` + +To format and fix all ruff issues automatically: + +```sh +$ ./scripts/format +``` + +## Publishing and releases + +Changes made to this repository via the automated release PR pipeline should publish to PyPI automatically. If +the changes aren't made through the automated pipeline, you may want to make releases manually. + +### Publish with a GitHub workflow + +You can release to package managers by using [the `Publish PyPI` GitHub action](https://www.github.com/stainless-sdks/brapi-python/actions/workflows/publish-pypi.yml). This requires a setup organization or repository secret to be set up. + +### Publish manually + +If you need to manually release a package, you can run the `bin/publish-pypi` script with a `PYPI_TOKEN` set on +the environment. diff --git a/LICENSE b/LICENSE new file mode 100644 index 0000000..00b551f --- /dev/null +++ b/LICENSE @@ -0,0 +1,201 @@ + Apache License + Version 2.0, January 2004 + http://www.apache.org/licenses/ + + TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION + + 1. Definitions. + + "License" shall mean the terms and conditions for use, reproduction, + and distribution as defined by Sections 1 through 9 of this document. + + "Licensor" shall mean the copyright owner or entity authorized by + the copyright owner that is granting the License. + + "Legal Entity" shall mean the union of the acting entity and all + other entities that control, are controlled by, or are under common + control with that entity. For the purposes of this definition, + "control" means (i) the power, direct or indirect, to cause the + direction or management of such entity, whether by contract or + otherwise, or (ii) ownership of fifty percent (50%) or more of the + outstanding shares, or (iii) beneficial ownership of such entity. + + "You" (or "Your") shall mean an individual or Legal Entity + exercising permissions granted by this License. + + "Source" form shall mean the preferred form for making modifications, + including but not limited to software source code, documentation + source, and configuration files. + + "Object" form shall mean any form resulting from mechanical + transformation or translation of a Source form, including but + not limited to compiled object code, generated documentation, + and conversions to other media types. + + "Work" shall mean the work of authorship, whether in Source or + Object form, made available under the License, as indicated by a + copyright notice that is included in or attached to the work + (an example is provided in the Appendix below). + + "Derivative Works" shall mean any work, whether in Source or Object + form, that is based on (or derived from) the Work and for which the + editorial revisions, annotations, elaborations, or other modifications + represent, as a whole, an original work of authorship. For the purposes + of this License, Derivative Works shall not include works that remain + separable from, or merely link (or bind by name) to the interfaces of, + the Work and Derivative Works thereof. + + "Contribution" shall mean any work of authorship, including + the original version of the Work and any modifications or additions + to that Work or Derivative Works thereof, that is intentionally + submitted to Licensor for inclusion in the Work by the copyright owner + or by an individual or Legal Entity authorized to submit on behalf of + the copyright owner. For the purposes of this definition, "submitted" + means any form of electronic, verbal, or written communication sent + to the Licensor or its representatives, including but not limited to + communication on electronic mailing lists, source code control systems, + and issue tracking systems that are managed by, or on behalf of, the + Licensor for the purpose of discussing and improving the Work, but + excluding communication that is conspicuously marked or otherwise + designated in writing by the copyright owner as "Not a Contribution." + + "Contributor" shall mean Licensor and any individual or Legal Entity + on behalf of whom a Contribution has been received by Licensor and + subsequently incorporated within the Work. + + 2. Grant of Copyright License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + copyright license to reproduce, prepare Derivative Works of, + publicly display, publicly perform, sublicense, and distribute the + Work and such Derivative Works in Source or Object form. + + 3. Grant of Patent License. Subject to the terms and conditions of + this License, each Contributor hereby grants to You a perpetual, + worldwide, non-exclusive, no-charge, royalty-free, irrevocable + (except as stated in this section) patent license to make, have made, + use, offer to sell, sell, import, and otherwise transfer the Work, + where such license applies only to those patent claims licensable + by such Contributor that are necessarily infringed by their + Contribution(s) alone or by combination of their Contribution(s) + with the Work to which such Contribution(s) was submitted. If You + institute patent litigation against any entity (including a + cross-claim or counterclaim in a lawsuit) alleging that the Work + or a Contribution incorporated within the Work constitutes direct + or contributory patent infringement, then any patent licenses + granted to You under this License for that Work shall terminate + as of the date such litigation is filed. + + 4. Redistribution. You may reproduce and distribute copies of the + Work or Derivative Works thereof in any medium, with or without + modifications, and in Source or Object form, provided that You + meet the following conditions: + + (a) You must give any other recipients of the Work or + Derivative Works a copy of this License; and + + (b) You must cause any modified files to carry prominent notices + stating that You changed the files; and + + (c) You must retain, in the Source form of any Derivative Works + that You distribute, all copyright, patent, trademark, and + attribution notices from the Source form of the Work, + excluding those notices that do not pertain to any part of + the Derivative Works; and + + (d) If the Work includes a "NOTICE" text file as part of its + distribution, then any Derivative Works that You distribute must + include a readable copy of the attribution notices contained + within such NOTICE file, excluding those notices that do not + pertain to any part of the Derivative Works, in at least one + of the following places: within a NOTICE text file distributed + as part of the Derivative Works; within the Source form or + documentation, if provided along with the Derivative Works; or, + within a display generated by the Derivative Works, if and + wherever such third-party notices normally appear. The contents + of the NOTICE file are for informational purposes only and + do not modify the License. You may add Your own attribution + notices within Derivative Works that You distribute, alongside + or as an addendum to the NOTICE text from the Work, provided + that such additional attribution notices cannot be construed + as modifying the License. + + You may add Your own copyright statement to Your modifications and + may provide additional or different license terms and conditions + for use, reproduction, or distribution of Your modifications, or + for any such Derivative Works as a whole, provided Your use, + reproduction, and distribution of the Work otherwise complies with + the conditions stated in this License. + + 5. Submission of Contributions. Unless You explicitly state otherwise, + any Contribution intentionally submitted for inclusion in the Work + by You to the Licensor shall be under the terms and conditions of + this License, without any additional terms or conditions. + Notwithstanding the above, nothing herein shall supersede or modify + the terms of any separate license agreement you may have executed + with Licensor regarding such Contributions. + + 6. Trademarks. This License does not grant permission to use the trade + names, trademarks, service marks, or product names of the Licensor, + except as required for reasonable and customary use in describing the + origin of the Work and reproducing the content of the NOTICE file. + + 7. Disclaimer of Warranty. Unless required by applicable law or + agreed to in writing, Licensor provides the Work (and each + Contributor provides its Contributions) on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or + implied, including, without limitation, any warranties or conditions + of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A + PARTICULAR PURPOSE. You are solely responsible for determining the + appropriateness of using or redistributing the Work and assume any + risks associated with Your exercise of permissions under this License. + + 8. Limitation of Liability. In no event and under no legal theory, + whether in tort (including negligence), contract, or otherwise, + unless required by applicable law (such as deliberate and grossly + negligent acts) or agreed to in writing, shall any Contributor be + liable to You for damages, including any direct, indirect, special, + incidental, or consequential damages of any character arising as a + result of this License or out of the use or inability to use the + Work (including but not limited to damages for loss of goodwill, + work stoppage, computer failure or malfunction, or any and all + other commercial damages or losses), even if such Contributor + has been advised of the possibility of such damages. + + 9. Accepting Warranty or Additional Liability. While redistributing + the Work or Derivative Works thereof, You may choose to offer, + and charge a fee for, acceptance of support, warranty, indemnity, + or other liability obligations and/or rights consistent with this + License. However, in accepting such obligations, You may act only + on Your own behalf and on Your sole responsibility, not on behalf + of any other Contributor, and only if You agree to indemnify, + defend, and hold each Contributor harmless for any liability + incurred by, or claims asserted against, such Contributor by reason + of your accepting any such warranty or additional liability. + + END OF TERMS AND CONDITIONS + + APPENDIX: How to apply the Apache License to your work. + + To apply the Apache License to your work, attach the following + boilerplate notice, with the fields enclosed by brackets "[]" + replaced with your own identifying information. (Don't include + the brackets!) The text should be enclosed in the appropriate + comment syntax for the file format. We also recommend that a + file or class name and description of purpose be included on the + same "printed page" as the copyright notice for easier + identification within third-party archives. + + Copyright 2025 Brapi + + Licensed under the Apache License, Version 2.0 (the "License"); + you may not use this file except in compliance with the License. + You may obtain a copy of the License at + + http://www.apache.org/licenses/LICENSE-2.0 + + Unless required by applicable law or agreed to in writing, software + distributed under the License is distributed on an "AS IS" BASIS, + WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. + See the License for the specific language governing permissions and + limitations under the License. diff --git a/README.md b/README.md index 7258cf5..b0480b6 100644 --- a/README.md +++ b/README.md @@ -1 +1,386 @@ -# brapi-python \ No newline at end of file +# Brapi Python API library + + +[![PyPI version](https://img.shields.io/pypi/v/brapi.svg?label=pypi%20(stable))](https://pypi.org/project/brapi/) + +The Brapi Python library provides convenient access to the Brapi REST API from any Python 3.8+ +application. The library includes type definitions for all request params and response fields, +and offers both synchronous and asynchronous clients powered by [httpx](https://github.com/encode/httpx). + +It is generated with [Stainless](https://www.stainless.com/). + +## Documentation + +The REST API documentation can be found on [brapi.dev](https://brapi.dev). The full API of this library can be found in [api.md](api.md). + +## Installation + +```sh +# install from this staging repo +pip install git+ssh://git@github.com/stainless-sdks/brapi-python.git +``` + +> [!NOTE] +> Once this package is [published to PyPI](https://www.stainless.com/docs/guides/publish), this will become: `pip install brapi` + +## Usage + +The full API of this library can be found in [api.md](api.md). + +```python +import os +from brapi import Brapi + +client = Brapi( + api_key=os.environ.get("BRAPI_API_KEY"), # This is the default and can be omitted + # defaults to "production". + environment="environment_1", +) + +quote = client.quote.retrieve( + tickers="REPLACE_ME", +) +print(quote.requested_at) +``` + +While you can provide an `api_key` keyword argument, +we recommend using [python-dotenv](https://pypi.org/project/python-dotenv/) +to add `BRAPI_API_KEY="My API Key"` to your `.env` file +so that your API Key is not stored in source control. + +## Async usage + +Simply import `AsyncBrapi` instead of `Brapi` and use `await` with each API call: + +```python +import os +import asyncio +from brapi import AsyncBrapi + +client = AsyncBrapi( + api_key=os.environ.get("BRAPI_API_KEY"), # This is the default and can be omitted + # defaults to "production". + environment="environment_1", +) + + +async def main() -> None: + quote = await client.quote.retrieve( + tickers="REPLACE_ME", + ) + print(quote.requested_at) + + +asyncio.run(main()) +``` + +Functionality between the synchronous and asynchronous clients is otherwise identical. + +### With aiohttp + +By default, the async client uses `httpx` for HTTP requests. However, for improved concurrency performance you may also use `aiohttp` as the HTTP backend. + +You can enable this by installing `aiohttp`: + +```sh +# install from this staging repo +pip install 'brapi[aiohttp] @ git+ssh://git@github.com/stainless-sdks/brapi-python.git' +``` + +Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: + +```python +import asyncio +from brapi import DefaultAioHttpClient +from brapi import AsyncBrapi + + +async def main() -> None: + async with AsyncBrapi( + api_key="My API Key", + http_client=DefaultAioHttpClient(), + ) as client: + quote = await client.quote.retrieve( + tickers="REPLACE_ME", + ) + print(quote.requested_at) + + +asyncio.run(main()) +``` + +## Using types + +Nested request parameters are [TypedDicts](https://docs.python.org/3/library/typing.html#typing.TypedDict). Responses are [Pydantic models](https://docs.pydantic.dev) which also provide helper methods for things like: + +- Serializing back into JSON, `model.to_json()` +- Converting to a dictionary, `model.to_dict()` + +Typed requests and responses provide autocomplete and documentation within your editor. If you would like to see type errors in VS Code to help catch bugs earlier, set `python.analysis.typeCheckingMode` to `basic`. + +## Handling errors + +When the library is unable to connect to the API (for example, due to network connection problems or a timeout), a subclass of `brapi.APIConnectionError` is raised. + +When the API returns a non-success status code (that is, 4xx or 5xx +response), a subclass of `brapi.APIStatusError` is raised, containing `status_code` and `response` properties. + +All errors inherit from `brapi.APIError`. + +```python +import brapi +from brapi import Brapi + +client = Brapi() + +try: + client.quote.retrieve( + tickers="REPLACE_ME", + ) +except brapi.APIConnectionError as e: + print("The server could not be reached") + print(e.__cause__) # an underlying Exception, likely raised within httpx. +except brapi.RateLimitError as e: + print("A 429 status code was received; we should back off a bit.") +except brapi.APIStatusError as e: + print("Another non-200-range status code was received") + print(e.status_code) + print(e.response) +``` + +Error codes are as follows: + +| Status Code | Error Type | +| ----------- | -------------------------- | +| 400 | `BadRequestError` | +| 401 | `AuthenticationError` | +| 403 | `PermissionDeniedError` | +| 404 | `NotFoundError` | +| 422 | `UnprocessableEntityError` | +| 429 | `RateLimitError` | +| >=500 | `InternalServerError` | +| N/A | `APIConnectionError` | + +### Retries + +Certain errors are automatically retried 2 times by default, with a short exponential backoff. +Connection errors (for example, due to a network connectivity problem), 408 Request Timeout, 409 Conflict, +429 Rate Limit, and >=500 Internal errors are all retried by default. + +You can use the `max_retries` option to configure or disable retry settings: + +```python +from brapi import Brapi + +# Configure the default for all requests: +client = Brapi( + # default is 2 + max_retries=0, +) + +# Or, configure per-request: +client.with_options(max_retries=5).quote.retrieve( + tickers="REPLACE_ME", +) +``` + +### Timeouts + +By default requests time out after 1 minute. You can configure this with a `timeout` option, +which accepts a float or an [`httpx.Timeout`](https://www.python-httpx.org/advanced/timeouts/#fine-tuning-the-configuration) object: + +```python +from brapi import Brapi + +# Configure the default for all requests: +client = Brapi( + # 20 seconds (default is 1 minute) + timeout=20.0, +) + +# More granular control: +client = Brapi( + timeout=httpx.Timeout(60.0, read=5.0, write=10.0, connect=2.0), +) + +# Override per-request: +client.with_options(timeout=5.0).quote.retrieve( + tickers="REPLACE_ME", +) +``` + +On timeout, an `APITimeoutError` is thrown. + +Note that requests that time out are [retried twice by default](#retries). + +## Advanced + +### Logging + +We use the standard library [`logging`](https://docs.python.org/3/library/logging.html) module. + +You can enable logging by setting the environment variable `BRAPI_LOG` to `info`. + +```shell +$ export BRAPI_LOG=info +``` + +Or to `debug` for more verbose logging. + +### How to tell whether `None` means `null` or missing + +In an API response, a field may be explicitly `null`, or missing entirely; in either case, its value is `None` in this library. You can differentiate the two cases with `.model_fields_set`: + +```py +if response.my_field is None: + if 'my_field' not in response.model_fields_set: + print('Got json like {}, without a "my_field" key present at all.') + else: + print('Got json like {"my_field": null}.') +``` + +### Accessing raw response data (e.g. headers) + +The "raw" Response object can be accessed by prefixing `.with_raw_response.` to any HTTP method call, e.g., + +```py +from brapi import Brapi + +client = Brapi() +response = client.quote.with_raw_response.retrieve( + tickers="REPLACE_ME", +) +print(response.headers.get('X-My-Header')) + +quote = response.parse() # get the object that `quote.retrieve()` would have returned +print(quote.requested_at) +``` + +These methods return an [`APIResponse`](https://github.com/stainless-sdks/brapi-python/tree/main/src/brapi/_response.py) object. + +The async client returns an [`AsyncAPIResponse`](https://github.com/stainless-sdks/brapi-python/tree/main/src/brapi/_response.py) with the same structure, the only difference being `await`able methods for reading the response content. + +#### `.with_streaming_response` + +The above interface eagerly reads the full response body when you make the request, which may not always be what you want. + +To stream the response body, use `.with_streaming_response` instead, which requires a context manager and only reads the response body once you call `.read()`, `.text()`, `.json()`, `.iter_bytes()`, `.iter_text()`, `.iter_lines()` or `.parse()`. In the async client, these are async methods. + +```python +with client.quote.with_streaming_response.retrieve( + tickers="REPLACE_ME", +) as response: + print(response.headers.get("X-My-Header")) + + for line in response.iter_lines(): + print(line) +``` + +The context manager is required so that the response will reliably be closed. + +### Making custom/undocumented requests + +This library is typed for convenient access to the documented API. + +If you need to access undocumented endpoints, params, or response properties, the library can still be used. + +#### Undocumented endpoints + +To make requests to undocumented endpoints, you can make requests using `client.get`, `client.post`, and other +http verbs. Options on the client will be respected (such as retries) when making this request. + +```py +import httpx + +response = client.post( + "/foo", + cast_to=httpx.Response, + body={"my_param": True}, +) + +print(response.headers.get("x-foo")) +``` + +#### Undocumented request params + +If you want to explicitly send an extra param, you can do so with the `extra_query`, `extra_body`, and `extra_headers` request +options. + +#### Undocumented response properties + +To access undocumented response properties, you can access the extra fields like `response.unknown_prop`. You +can also get all the extra fields on the Pydantic model as a dict with +[`response.model_extra`](https://docs.pydantic.dev/latest/api/base_model/#pydantic.BaseModel.model_extra). + +### Configuring the HTTP client + +You can directly override the [httpx client](https://www.python-httpx.org/api/#client) to customize it for your use case, including: + +- Support for [proxies](https://www.python-httpx.org/advanced/proxies/) +- Custom [transports](https://www.python-httpx.org/advanced/transports/) +- Additional [advanced](https://www.python-httpx.org/advanced/clients/) functionality + +```python +import httpx +from brapi import Brapi, DefaultHttpxClient + +client = Brapi( + # Or use the `BRAPI_BASE_URL` env var + base_url="http://my.test.server.example.com:8083", + http_client=DefaultHttpxClient( + proxy="http://my.test.proxy.example.com", + transport=httpx.HTTPTransport(local_address="0.0.0.0"), + ), +) +``` + +You can also customize the client on a per-request basis by using `with_options()`: + +```python +client.with_options(http_client=DefaultHttpxClient(...)) +``` + +### Managing HTTP resources + +By default the library closes underlying HTTP connections whenever the client is [garbage collected](https://docs.python.org/3/reference/datamodel.html#object.__del__). You can manually close the client using the `.close()` method if desired, or with a context manager that closes when exiting. + +```py +from brapi import Brapi + +with Brapi() as client: + # make requests here + ... + +# HTTP client is now closed +``` + +## Versioning + +This package generally follows [SemVer](https://semver.org/spec/v2.0.0.html) conventions, though certain backwards-incompatible changes may be released as minor versions: + +1. Changes that only affect static types, without breaking runtime behavior. +2. Changes to library internals which are technically public but not intended or documented for external use. _(Please open a GitHub issue to let us know if you are relying on such internals.)_ +3. Changes that we do not expect to impact the vast majority of users in practice. + +We take backwards-compatibility seriously and work hard to ensure you can rely on a smooth upgrade experience. + +We are keen for your feedback; please open an [issue](https://www.github.com/stainless-sdks/brapi-python/issues) with questions, bugs, or suggestions. + +### Determining the installed version + +If you've upgraded to the latest version but aren't seeing any new features you were expecting then your python environment is likely still using an older version. + +You can determine the version that is being used at runtime with: + +```py +import brapi +print(brapi.__version__) +``` + +## Requirements + +Python 3.8 or higher. + +## Contributing + +See [the contributing documentation](./CONTRIBUTING.md). diff --git a/SECURITY.md b/SECURITY.md new file mode 100644 index 0000000..84be452 --- /dev/null +++ b/SECURITY.md @@ -0,0 +1,27 @@ +# Security Policy + +## Reporting Security Issues + +This SDK is generated by [Stainless Software Inc](http://stainless.com). Stainless takes security seriously, and encourages you to report any security vulnerability promptly so that appropriate action can be taken. + +To report a security issue, please contact the Stainless team at security@stainless.com. + +## Responsible Disclosure + +We appreciate the efforts of security researchers and individuals who help us maintain the security of +SDKs we generate. If you believe you have found a security vulnerability, please adhere to responsible +disclosure practices by allowing us a reasonable amount of time to investigate and address the issue +before making any information public. + +## Reporting Non-SDK Related Security Issues + +If you encounter security issues that are not directly related to SDKs but pertain to the services +or products provided by Brapi, please follow the respective company's security reporting guidelines. + +### Brapi Terms and Policies + +Please contact contact@brapi.dev for any questions or concerns regarding the security of our services. + +--- + +Thank you for helping us keep the SDKs and systems they interact with secure. diff --git a/api.md b/api.md new file mode 100644 index 0000000..5d2bfbf --- /dev/null +++ b/api.md @@ -0,0 +1,87 @@ +# Quote + +Types: + +```python +from brapi.types import ( + BalanceSheetEntry, + CashflowEntry, + DefaultKeyStatisticsEntry, + FinancialDataEntry, + IncomeStatementEntry, + ValueAddedEntry, + QuoteRetrieveResponse, + QuoteListResponse, +) +``` + +Methods: + +- client.quote.retrieve(tickers, \*\*params) -> QuoteRetrieveResponse +- client.quote.list(\*\*params) -> QuoteListResponse + +# Available + +Types: + +```python +from brapi.types import AvailableListResponse +``` + +Methods: + +- client.available.list(\*\*params) -> AvailableListResponse + +# V2 + +## Crypto + +Types: + +```python +from brapi.types.v2 import CryptoRetrieveResponse, CryptoListAvailableResponse +``` + +Methods: + +- client.v2.crypto.retrieve(\*\*params) -> CryptoRetrieveResponse +- client.v2.crypto.list_available(\*\*params) -> CryptoListAvailableResponse + +## Currency + +Types: + +```python +from brapi.types.v2 import CurrencyRetrieveResponse, CurrencyListAvailableResponse +``` + +Methods: + +- client.v2.currency.retrieve(\*\*params) -> CurrencyRetrieveResponse +- client.v2.currency.list_available(\*\*params) -> CurrencyListAvailableResponse + +## Inflation + +Types: + +```python +from brapi.types.v2 import InflationRetrieveResponse, InflationListAvailableResponse +``` + +Methods: + +- client.v2.inflation.retrieve(\*\*params) -> InflationRetrieveResponse +- client.v2.inflation.list_available(\*\*params) -> InflationListAvailableResponse + +## PrimeRate + +Types: + +```python +from brapi.types.v2 import PrimeRateRetrieveResponse, PrimeRateListAvailableResponse +``` + +Methods: + +- client.v2.prime_rate.retrieve(\*\*params) -> PrimeRateRetrieveResponse +- client.v2.prime_rate.list_available(\*\*params) -> PrimeRateListAvailableResponse diff --git a/bin/publish-pypi b/bin/publish-pypi new file mode 100644 index 0000000..826054e --- /dev/null +++ b/bin/publish-pypi @@ -0,0 +1,6 @@ +#!/usr/bin/env bash + +set -eux +mkdir -p dist +rye build --clean +rye publish --yes --token=$PYPI_TOKEN diff --git a/examples/.keep b/examples/.keep new file mode 100644 index 0000000..d8c73e9 --- /dev/null +++ b/examples/.keep @@ -0,0 +1,4 @@ +File generated from our OpenAPI spec by Stainless. + +This directory can be used to store example files demonstrating usage of this SDK. +It is ignored by Stainless code generation and its content (other than this keep file) won't be touched. \ No newline at end of file diff --git a/noxfile.py b/noxfile.py new file mode 100644 index 0000000..53bca7f --- /dev/null +++ b/noxfile.py @@ -0,0 +1,9 @@ +import nox + + +@nox.session(reuse_venv=True, name="test-pydantic-v1") +def test_pydantic_v1(session: nox.Session) -> None: + session.install("-r", "requirements-dev.lock") + session.install("pydantic<2") + + session.run("pytest", "--showlocals", "--ignore=tests/functional", *session.posargs) diff --git a/pyproject.toml b/pyproject.toml new file mode 100644 index 0000000..968a4bf --- /dev/null +++ b/pyproject.toml @@ -0,0 +1,267 @@ +[project] +name = "brapi" +version = "0.0.1" +description = "The official Python library for the brapi API" +dynamic = ["readme"] +license = "Apache-2.0" +authors = [ +{ name = "Brapi", email = "contact@brapi.dev" }, +] +dependencies = [ + "httpx>=0.23.0, <1", + "pydantic>=1.9.0, <3", + "typing-extensions>=4.10, <5", + "anyio>=3.5.0, <5", + "distro>=1.7.0, <2", + "sniffio", +] +requires-python = ">= 3.8" +classifiers = [ + "Typing :: Typed", + "Intended Audience :: Developers", + "Programming Language :: Python :: 3.8", + "Programming Language :: Python :: 3.9", + "Programming Language :: Python :: 3.10", + "Programming Language :: Python :: 3.11", + "Programming Language :: Python :: 3.12", + "Programming Language :: Python :: 3.13", + "Operating System :: OS Independent", + "Operating System :: POSIX", + "Operating System :: MacOS", + "Operating System :: POSIX :: Linux", + "Operating System :: Microsoft :: Windows", + "Topic :: Software Development :: Libraries :: Python Modules", + "License :: OSI Approved :: Apache Software License" +] + +[project.urls] +Homepage = "https://github.com/stainless-sdks/brapi-python" +Repository = "https://github.com/stainless-sdks/brapi-python" + +[project.optional-dependencies] +aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.8"] + +[tool.rye] +managed = true +# version pins are in requirements-dev.lock +dev-dependencies = [ + "pyright==1.1.399", + "mypy", + "respx", + "pytest", + "pytest-asyncio", + "ruff", + "time-machine", + "nox", + "dirty-equals>=0.6.0", + "importlib-metadata>=6.7.0", + "rich>=13.7.1", + "pytest-xdist>=3.6.1", +] + +[tool.rye.scripts] +format = { chain = [ + "format:ruff", + "format:docs", + "fix:ruff", + # run formatting again to fix any inconsistencies when imports are stripped + "format:ruff", +]} +"format:docs" = "python scripts/utils/ruffen-docs.py README.md api.md" +"format:ruff" = "ruff format" + +"lint" = { chain = [ + "check:ruff", + "typecheck", + "check:importable", +]} +"check:ruff" = "ruff check ." +"fix:ruff" = "ruff check --fix ." + +"check:importable" = "python -c 'import brapi'" + +typecheck = { chain = [ + "typecheck:pyright", + "typecheck:mypy" +]} +"typecheck:pyright" = "pyright" +"typecheck:verify-types" = "pyright --verifytypes brapi --ignoreexternal" +"typecheck:mypy" = "mypy ." + +[build-system] +requires = ["hatchling==1.26.3", "hatch-fancy-pypi-readme"] +build-backend = "hatchling.build" + +[tool.hatch.build] +include = [ + "src/*" +] + +[tool.hatch.build.targets.wheel] +packages = ["src/brapi"] + +[tool.hatch.build.targets.sdist] +# Basically everything except hidden files/directories (such as .github, .devcontainers, .python-version, etc) +include = [ + "/*.toml", + "/*.json", + "/*.lock", + "/*.md", + "/mypy.ini", + "/noxfile.py", + "bin/*", + "examples/*", + "src/*", + "tests/*", +] + +[tool.hatch.metadata.hooks.fancy-pypi-readme] +content-type = "text/markdown" + +[[tool.hatch.metadata.hooks.fancy-pypi-readme.fragments]] +path = "README.md" + +[[tool.hatch.metadata.hooks.fancy-pypi-readme.substitutions]] +# replace relative links with absolute links +pattern = '\[(.+?)\]\(((?!https?://)\S+?)\)' +replacement = '[\1](https://github.com/stainless-sdks/brapi-python/tree/main/\g<2>)' + +[tool.pytest.ini_options] +testpaths = ["tests"] +addopts = "--tb=short -n auto" +xfail_strict = true +asyncio_mode = "auto" +asyncio_default_fixture_loop_scope = "session" +filterwarnings = [ + "error" +] + +[tool.pyright] +# this enables practically every flag given by pyright. +# there are a couple of flags that are still disabled by +# default in strict mode as they are experimental and niche. +typeCheckingMode = "strict" +pythonVersion = "3.8" + +exclude = [ + "_dev", + ".venv", + ".nox", + ".git", +] + +reportImplicitOverride = true +reportOverlappingOverload = false + +reportImportCycles = false +reportPrivateUsage = false + +[tool.mypy] +pretty = true +show_error_codes = true + +# Exclude _files.py because mypy isn't smart enough to apply +# the correct type narrowing and as this is an internal module +# it's fine to just use Pyright. +# +# We also exclude our `tests` as mypy doesn't always infer +# types correctly and Pyright will still catch any type errors. +exclude = ['src/brapi/_files.py', '_dev/.*.py', 'tests/.*'] + +strict_equality = true +implicit_reexport = true +check_untyped_defs = true +no_implicit_optional = true + +warn_return_any = true +warn_unreachable = true +warn_unused_configs = true + +# Turn these options off as it could cause conflicts +# with the Pyright options. +warn_unused_ignores = false +warn_redundant_casts = false + +disallow_any_generics = true +disallow_untyped_defs = true +disallow_untyped_calls = true +disallow_subclassing_any = true +disallow_incomplete_defs = true +disallow_untyped_decorators = true +cache_fine_grained = true + +# By default, mypy reports an error if you assign a value to the result +# of a function call that doesn't return anything. We do this in our test +# cases: +# ``` +# result = ... +# assert result is None +# ``` +# Changing this codegen to make mypy happy would increase complexity +# and would not be worth it. +disable_error_code = "func-returns-value,overload-cannot-match" + +# https://github.com/python/mypy/issues/12162 +[[tool.mypy.overrides]] +module = "black.files.*" +ignore_errors = true +ignore_missing_imports = true + + +[tool.ruff] +line-length = 120 +output-format = "grouped" +target-version = "py38" + +[tool.ruff.format] +docstring-code-format = true + +[tool.ruff.lint] +select = [ + # isort + "I", + # bugbear rules + "B", + # remove unused imports + "F401", + # check for missing future annotations + "FA102", + # bare except statements + "E722", + # unused arguments + "ARG", + # print statements + "T201", + "T203", + # misuse of typing.TYPE_CHECKING + "TC004", + # import rules + "TID251", +] +ignore = [ + # mutable defaults + "B006", +] +unfixable = [ + # disable auto fix for print statements + "T201", + "T203", +] + +extend-safe-fixes = ["FA102"] + +[tool.ruff.lint.flake8-tidy-imports.banned-api] +"functools.lru_cache".msg = "This function does not retain type information for the wrapped function's arguments; The `lru_cache` function from `_utils` should be used instead" + +[tool.ruff.lint.isort] +length-sort = true +length-sort-straight = true +combine-as-imports = true +extra-standard-library = ["typing_extensions"] +known-first-party = ["brapi", "tests"] + +[tool.ruff.lint.per-file-ignores] +"bin/**.py" = ["T201", "T203"] +"scripts/**.py" = ["T201", "T203"] +"tests/**.py" = ["T201", "T203"] +"examples/**.py" = ["T201", "T203"] diff --git a/requirements-dev.lock b/requirements-dev.lock new file mode 100644 index 0000000..6f06451 --- /dev/null +++ b/requirements-dev.lock @@ -0,0 +1,137 @@ +# generated by rye +# use `rye lock` or `rye sync` to update this lockfile +# +# last locked with the following flags: +# pre: false +# features: [] +# all-features: true +# with-sources: false +# generate-hashes: false +# universal: false + +-e file:. +aiohappyeyeballs==2.6.1 + # via aiohttp +aiohttp==3.12.8 + # via brapi + # via httpx-aiohttp +aiosignal==1.3.2 + # via aiohttp +annotated-types==0.6.0 + # via pydantic +anyio==4.4.0 + # via brapi + # via httpx +argcomplete==3.1.2 + # via nox +async-timeout==5.0.1 + # via aiohttp +attrs==25.3.0 + # via aiohttp +certifi==2023.7.22 + # via httpcore + # via httpx +colorlog==6.7.0 + # via nox +dirty-equals==0.6.0 +distlib==0.3.7 + # via virtualenv +distro==1.8.0 + # via brapi +exceptiongroup==1.2.2 + # via anyio + # via pytest +execnet==2.1.1 + # via pytest-xdist +filelock==3.12.4 + # via virtualenv +frozenlist==1.6.2 + # via aiohttp + # via aiosignal +h11==0.16.0 + # via httpcore +httpcore==1.0.9 + # via httpx +httpx==0.28.1 + # via brapi + # via httpx-aiohttp + # via respx +httpx-aiohttp==0.1.8 + # via brapi +idna==3.4 + # via anyio + # via httpx + # via yarl +importlib-metadata==7.0.0 +iniconfig==2.0.0 + # via pytest +markdown-it-py==3.0.0 + # via rich +mdurl==0.1.2 + # via markdown-it-py +multidict==6.4.4 + # via aiohttp + # via yarl +mypy==1.14.1 +mypy-extensions==1.0.0 + # via mypy +nodeenv==1.8.0 + # via pyright +nox==2023.4.22 +packaging==23.2 + # via nox + # via pytest +platformdirs==3.11.0 + # via virtualenv +pluggy==1.5.0 + # via pytest +propcache==0.3.1 + # via aiohttp + # via yarl +pydantic==2.11.9 + # via brapi +pydantic-core==2.33.2 + # via pydantic +pygments==2.18.0 + # via rich +pyright==1.1.399 +pytest==8.3.3 + # via pytest-asyncio + # via pytest-xdist +pytest-asyncio==0.24.0 +pytest-xdist==3.7.0 +python-dateutil==2.8.2 + # via time-machine +pytz==2023.3.post1 + # via dirty-equals +respx==0.22.0 +rich==13.7.1 +ruff==0.9.4 +setuptools==68.2.2 + # via nodeenv +six==1.16.0 + # via python-dateutil +sniffio==1.3.0 + # via anyio + # via brapi +time-machine==2.9.0 +tomli==2.0.2 + # via mypy + # via pytest +typing-extensions==4.12.2 + # via anyio + # via brapi + # via multidict + # via mypy + # via pydantic + # via pydantic-core + # via pyright + # via typing-inspection +typing-inspection==0.4.1 + # via pydantic +virtualenv==20.24.5 + # via nox +yarl==1.20.0 + # via aiohttp +zipp==3.17.0 + # via importlib-metadata diff --git a/requirements.lock b/requirements.lock new file mode 100644 index 0000000..6018e55 --- /dev/null +++ b/requirements.lock @@ -0,0 +1,75 @@ +# generated by rye +# use `rye lock` or `rye sync` to update this lockfile +# +# last locked with the following flags: +# pre: false +# features: [] +# all-features: true +# with-sources: false +# generate-hashes: false +# universal: false + +-e file:. +aiohappyeyeballs==2.6.1 + # via aiohttp +aiohttp==3.12.8 + # via brapi + # via httpx-aiohttp +aiosignal==1.3.2 + # via aiohttp +annotated-types==0.6.0 + # via pydantic +anyio==4.4.0 + # via brapi + # via httpx +async-timeout==5.0.1 + # via aiohttp +attrs==25.3.0 + # via aiohttp +certifi==2023.7.22 + # via httpcore + # via httpx +distro==1.8.0 + # via brapi +exceptiongroup==1.2.2 + # via anyio +frozenlist==1.6.2 + # via aiohttp + # via aiosignal +h11==0.16.0 + # via httpcore +httpcore==1.0.9 + # via httpx +httpx==0.28.1 + # via brapi + # via httpx-aiohttp +httpx-aiohttp==0.1.8 + # via brapi +idna==3.4 + # via anyio + # via httpx + # via yarl +multidict==6.4.4 + # via aiohttp + # via yarl +propcache==0.3.1 + # via aiohttp + # via yarl +pydantic==2.11.9 + # via brapi +pydantic-core==2.33.2 + # via pydantic +sniffio==1.3.0 + # via anyio + # via brapi +typing-extensions==4.12.2 + # via anyio + # via brapi + # via multidict + # via pydantic + # via pydantic-core + # via typing-inspection +typing-inspection==0.4.1 + # via pydantic +yarl==1.20.0 + # via aiohttp diff --git a/scripts/bootstrap b/scripts/bootstrap new file mode 100755 index 0000000..b430fee --- /dev/null +++ b/scripts/bootstrap @@ -0,0 +1,27 @@ +#!/usr/bin/env bash + +set -e + +cd "$(dirname "$0")/.." + +if [ -f "Brewfile" ] && [ "$(uname -s)" = "Darwin" ] && [ "$SKIP_BREW" != "1" ] && [ -t 0 ]; then + brew bundle check >/dev/null 2>&1 || { + echo -n "==> Install Homebrew dependencies? (y/N): " + read -r response + case "$response" in + [yY][eE][sS]|[yY]) + brew bundle + ;; + *) + ;; + esac + echo + } +fi + +echo "==> Installing Python dependencies…" + +# experimental uv support makes installations significantly faster +rye config --set-bool behavior.use-uv=true + +rye sync --all-features diff --git a/scripts/format b/scripts/format new file mode 100755 index 0000000..667ec2d --- /dev/null +++ b/scripts/format @@ -0,0 +1,8 @@ +#!/usr/bin/env bash + +set -e + +cd "$(dirname "$0")/.." + +echo "==> Running formatters" +rye run format diff --git a/scripts/lint b/scripts/lint new file mode 100755 index 0000000..e8935a5 --- /dev/null +++ b/scripts/lint @@ -0,0 +1,11 @@ +#!/usr/bin/env bash + +set -e + +cd "$(dirname "$0")/.." + +echo "==> Running lints" +rye run lint + +echo "==> Making sure it imports" +rye run python -c 'import brapi' diff --git a/scripts/mock b/scripts/mock new file mode 100755 index 0000000..0b28f6e --- /dev/null +++ b/scripts/mock @@ -0,0 +1,41 @@ +#!/usr/bin/env bash + +set -e + +cd "$(dirname "$0")/.." + +if [[ -n "$1" && "$1" != '--'* ]]; then + URL="$1" + shift +else + URL="$(grep 'openapi_spec_url' .stats.yml | cut -d' ' -f2)" +fi + +# Check if the URL is empty +if [ -z "$URL" ]; then + echo "Error: No OpenAPI spec path/url provided or found in .stats.yml" + exit 1 +fi + +echo "==> Starting mock server with URL ${URL}" + +# Run prism mock on the given spec +if [ "$1" == "--daemon" ]; then + npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock "$URL" &> .prism.log & + + # Wait for server to come online + echo -n "Waiting for server" + while ! grep -q "✖ fatal\|Prism is listening" ".prism.log" ; do + echo -n "." + sleep 0.1 + done + + if grep -q "✖ fatal" ".prism.log"; then + cat .prism.log + exit 1 + fi + + echo +else + npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock "$URL" +fi diff --git a/scripts/test b/scripts/test new file mode 100755 index 0000000..dbeda2d --- /dev/null +++ b/scripts/test @@ -0,0 +1,61 @@ +#!/usr/bin/env bash + +set -e + +cd "$(dirname "$0")/.." + +RED='\033[0;31m' +GREEN='\033[0;32m' +YELLOW='\033[0;33m' +NC='\033[0m' # No Color + +function prism_is_running() { + curl --silent "http://localhost:4010" >/dev/null 2>&1 +} + +kill_server_on_port() { + pids=$(lsof -t -i tcp:"$1" || echo "") + if [ "$pids" != "" ]; then + kill "$pids" + echo "Stopped $pids." + fi +} + +function is_overriding_api_base_url() { + [ -n "$TEST_API_BASE_URL" ] +} + +if ! is_overriding_api_base_url && ! prism_is_running ; then + # When we exit this script, make sure to kill the background mock server process + trap 'kill_server_on_port 4010' EXIT + + # Start the dev server + ./scripts/mock --daemon +fi + +if is_overriding_api_base_url ; then + echo -e "${GREEN}✔ Running tests against ${TEST_API_BASE_URL}${NC}" + echo +elif ! prism_is_running ; then + echo -e "${RED}ERROR:${NC} The test suite will not run without a mock Prism server" + echo -e "running against your OpenAPI spec." + echo + echo -e "To run the server, pass in the path or url of your OpenAPI" + echo -e "spec to the prism command:" + echo + echo -e " \$ ${YELLOW}npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock path/to/your.openapi.yml${NC}" + echo + + exit 1 +else + echo -e "${GREEN}✔ Mock prism server is running with your OpenAPI spec${NC}" + echo +fi + +export DEFER_PYDANTIC_BUILD=false + +echo "==> Running tests" +rye run pytest "$@" + +echo "==> Running Pydantic v1 tests" +rye run nox -s test-pydantic-v1 -- "$@" diff --git a/scripts/utils/ruffen-docs.py b/scripts/utils/ruffen-docs.py new file mode 100644 index 0000000..0cf2bd2 --- /dev/null +++ b/scripts/utils/ruffen-docs.py @@ -0,0 +1,167 @@ +# fork of https://github.com/asottile/blacken-docs adapted for ruff +from __future__ import annotations + +import re +import sys +import argparse +import textwrap +import contextlib +import subprocess +from typing import Match, Optional, Sequence, Generator, NamedTuple, cast + +MD_RE = re.compile( + r"(?P^(?P *)```\s*python\n)" r"(?P.*?)" r"(?P^(?P=indent)```\s*$)", + re.DOTALL | re.MULTILINE, +) +MD_PYCON_RE = re.compile( + r"(?P^(?P *)```\s*pycon\n)" r"(?P.*?)" r"(?P^(?P=indent)```.*$)", + re.DOTALL | re.MULTILINE, +) +PYCON_PREFIX = ">>> " +PYCON_CONTINUATION_PREFIX = "..." +PYCON_CONTINUATION_RE = re.compile( + rf"^{re.escape(PYCON_CONTINUATION_PREFIX)}( |$)", +) +DEFAULT_LINE_LENGTH = 100 + + +class CodeBlockError(NamedTuple): + offset: int + exc: Exception + + +def format_str( + src: str, +) -> tuple[str, Sequence[CodeBlockError]]: + errors: list[CodeBlockError] = [] + + @contextlib.contextmanager + def _collect_error(match: Match[str]) -> Generator[None, None, None]: + try: + yield + except Exception as e: + errors.append(CodeBlockError(match.start(), e)) + + def _md_match(match: Match[str]) -> str: + code = textwrap.dedent(match["code"]) + with _collect_error(match): + code = format_code_block(code) + code = textwrap.indent(code, match["indent"]) + return f"{match['before']}{code}{match['after']}" + + def _pycon_match(match: Match[str]) -> str: + code = "" + fragment = cast(Optional[str], None) + + def finish_fragment() -> None: + nonlocal code + nonlocal fragment + + if fragment is not None: + with _collect_error(match): + fragment = format_code_block(fragment) + fragment_lines = fragment.splitlines() + code += f"{PYCON_PREFIX}{fragment_lines[0]}\n" + for line in fragment_lines[1:]: + # Skip blank lines to handle Black adding a blank above + # functions within blocks. A blank line would end the REPL + # continuation prompt. + # + # >>> if True: + # ... def f(): + # ... pass + # ... + if line: + code += f"{PYCON_CONTINUATION_PREFIX} {line}\n" + if fragment_lines[-1].startswith(" "): + code += f"{PYCON_CONTINUATION_PREFIX}\n" + fragment = None + + indentation = None + for line in match["code"].splitlines(): + orig_line, line = line, line.lstrip() + if indentation is None and line: + indentation = len(orig_line) - len(line) + continuation_match = PYCON_CONTINUATION_RE.match(line) + if continuation_match and fragment is not None: + fragment += line[continuation_match.end() :] + "\n" + else: + finish_fragment() + if line.startswith(PYCON_PREFIX): + fragment = line[len(PYCON_PREFIX) :] + "\n" + else: + code += orig_line[indentation:] + "\n" + finish_fragment() + return code + + def _md_pycon_match(match: Match[str]) -> str: + code = _pycon_match(match) + code = textwrap.indent(code, match["indent"]) + return f"{match['before']}{code}{match['after']}" + + src = MD_RE.sub(_md_match, src) + src = MD_PYCON_RE.sub(_md_pycon_match, src) + return src, errors + + +def format_code_block(code: str) -> str: + return subprocess.check_output( + [ + sys.executable, + "-m", + "ruff", + "format", + "--stdin-filename=script.py", + f"--line-length={DEFAULT_LINE_LENGTH}", + ], + encoding="utf-8", + input=code, + ) + + +def format_file( + filename: str, + skip_errors: bool, +) -> int: + with open(filename, encoding="UTF-8") as f: + contents = f.read() + new_contents, errors = format_str(contents) + for error in errors: + lineno = contents[: error.offset].count("\n") + 1 + print(f"{filename}:{lineno}: code block parse error {error.exc}") + if errors and not skip_errors: + return 1 + if contents != new_contents: + print(f"{filename}: Rewriting...") + with open(filename, "w", encoding="UTF-8") as f: + f.write(new_contents) + return 0 + else: + return 0 + + +def main(argv: Sequence[str] | None = None) -> int: + parser = argparse.ArgumentParser() + parser.add_argument( + "-l", + "--line-length", + type=int, + default=DEFAULT_LINE_LENGTH, + ) + parser.add_argument( + "-S", + "--skip-string-normalization", + action="store_true", + ) + parser.add_argument("-E", "--skip-errors", action="store_true") + parser.add_argument("filenames", nargs="*") + args = parser.parse_args(argv) + + retv = 0 + for filename in args.filenames: + retv |= format_file(filename, skip_errors=args.skip_errors) + return retv + + +if __name__ == "__main__": + raise SystemExit(main()) diff --git a/scripts/utils/upload-artifact.sh b/scripts/utils/upload-artifact.sh new file mode 100755 index 0000000..78ccd91 --- /dev/null +++ b/scripts/utils/upload-artifact.sh @@ -0,0 +1,27 @@ +#!/usr/bin/env bash +set -exuo pipefail + +FILENAME=$(basename dist/*.whl) + +RESPONSE=$(curl -X POST "$URL?filename=$FILENAME" \ + -H "Authorization: Bearer $AUTH" \ + -H "Content-Type: application/json") + +SIGNED_URL=$(echo "$RESPONSE" | jq -r '.url') + +if [[ "$SIGNED_URL" == "null" ]]; then + echo -e "\033[31mFailed to get signed URL.\033[0m" + exit 1 +fi + +UPLOAD_RESPONSE=$(curl -v -X PUT \ + -H "Content-Type: binary/octet-stream" \ + --data-binary "@dist/$FILENAME" "$SIGNED_URL" 2>&1) + +if echo "$UPLOAD_RESPONSE" | grep -q "HTTP/[0-9.]* 200"; then + echo -e "\033[32mUploaded build to Stainless storage.\033[0m" + echo -e "\033[32mInstallation: pip install 'https://pkg.stainless.com/s/brapi-python/$SHA/$FILENAME'\033[0m" +else + echo -e "\033[31mFailed to upload artifact.\033[0m" + exit 1 +fi diff --git a/src/brapi/__init__.py b/src/brapi/__init__.py new file mode 100644 index 0000000..d940e9f --- /dev/null +++ b/src/brapi/__init__.py @@ -0,0 +1,104 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +import typing as _t + +from . import types +from ._types import NOT_GIVEN, Omit, NoneType, NotGiven, Transport, ProxiesTypes, omit, not_given +from ._utils import file_from_path +from ._client import ( + ENVIRONMENTS, + Brapi, + Client, + Stream, + Timeout, + Transport, + AsyncBrapi, + AsyncClient, + AsyncStream, + RequestOptions, +) +from ._models import BaseModel +from ._version import __title__, __version__ +from ._response import APIResponse as APIResponse, AsyncAPIResponse as AsyncAPIResponse +from ._constants import DEFAULT_TIMEOUT, DEFAULT_MAX_RETRIES, DEFAULT_CONNECTION_LIMITS +from ._exceptions import ( + APIError, + BrapiError, + ConflictError, + NotFoundError, + APIStatusError, + RateLimitError, + APITimeoutError, + BadRequestError, + APIConnectionError, + AuthenticationError, + InternalServerError, + PermissionDeniedError, + UnprocessableEntityError, + APIResponseValidationError, +) +from ._base_client import DefaultHttpxClient, DefaultAioHttpClient, DefaultAsyncHttpxClient +from ._utils._logs import setup_logging as _setup_logging + +__all__ = [ + "types", + "__version__", + "__title__", + "NoneType", + "Transport", + "ProxiesTypes", + "NotGiven", + "NOT_GIVEN", + "not_given", + "Omit", + "omit", + "BrapiError", + "APIError", + "APIStatusError", + "APITimeoutError", + "APIConnectionError", + "APIResponseValidationError", + "BadRequestError", + "AuthenticationError", + "PermissionDeniedError", + "NotFoundError", + "ConflictError", + "UnprocessableEntityError", + "RateLimitError", + "InternalServerError", + "Timeout", + "RequestOptions", + "Client", + "AsyncClient", + "Stream", + "AsyncStream", + "Brapi", + "AsyncBrapi", + "ENVIRONMENTS", + "file_from_path", + "BaseModel", + "DEFAULT_TIMEOUT", + "DEFAULT_MAX_RETRIES", + "DEFAULT_CONNECTION_LIMITS", + "DefaultHttpxClient", + "DefaultAsyncHttpxClient", + "DefaultAioHttpClient", +] + +if not _t.TYPE_CHECKING: + from ._utils._resources_proxy import resources as resources + +_setup_logging() + +# Update the __module__ attribute for exported symbols so that +# error messages point to this module instead of the module +# it was originally defined in, e.g. +# brapi._exceptions.NotFoundError -> brapi.NotFoundError +__locals = locals() +for __name in __all__: + if not __name.startswith("__"): + try: + __locals[__name].__module__ = "brapi" + except (TypeError, AttributeError): + # Some of our exported symbols are builtins which we can't set attributes for. + pass diff --git a/src/brapi/_base_client.py b/src/brapi/_base_client.py new file mode 100644 index 0000000..2c643d3 --- /dev/null +++ b/src/brapi/_base_client.py @@ -0,0 +1,1995 @@ +from __future__ import annotations + +import sys +import json +import time +import uuid +import email +import asyncio +import inspect +import logging +import platform +import email.utils +from types import TracebackType +from random import random +from typing import ( + TYPE_CHECKING, + Any, + Dict, + Type, + Union, + Generic, + Mapping, + TypeVar, + Iterable, + Iterator, + Optional, + Generator, + AsyncIterator, + cast, + overload, +) +from typing_extensions import Literal, override, get_origin + +import anyio +import httpx +import distro +import pydantic +from httpx import URL +from pydantic import PrivateAttr + +from . import _exceptions +from ._qs import Querystring +from ._files import to_httpx_files, async_to_httpx_files +from ._types import ( + Body, + Omit, + Query, + Headers, + Timeout, + NotGiven, + ResponseT, + AnyMapping, + PostParser, + RequestFiles, + HttpxSendArgs, + RequestOptions, + HttpxRequestFiles, + ModelBuilderProtocol, + not_given, +) +from ._utils import is_dict, is_list, asyncify, is_given, lru_cache, is_mapping +from ._compat import PYDANTIC_V1, model_copy, model_dump +from ._models import GenericModel, FinalRequestOptions, validate_type, construct_type +from ._response import ( + APIResponse, + BaseAPIResponse, + AsyncAPIResponse, + extract_response_type, +) +from ._constants import ( + DEFAULT_TIMEOUT, + MAX_RETRY_DELAY, + DEFAULT_MAX_RETRIES, + INITIAL_RETRY_DELAY, + RAW_RESPONSE_HEADER, + OVERRIDE_CAST_TO_HEADER, + DEFAULT_CONNECTION_LIMITS, +) +from ._streaming import Stream, SSEDecoder, AsyncStream, SSEBytesDecoder +from ._exceptions import ( + APIStatusError, + APITimeoutError, + APIConnectionError, + APIResponseValidationError, +) + +log: logging.Logger = logging.getLogger(__name__) + +# TODO: make base page type vars covariant +SyncPageT = TypeVar("SyncPageT", bound="BaseSyncPage[Any]") +AsyncPageT = TypeVar("AsyncPageT", bound="BaseAsyncPage[Any]") + + +_T = TypeVar("_T") +_T_co = TypeVar("_T_co", covariant=True) + +_StreamT = TypeVar("_StreamT", bound=Stream[Any]) +_AsyncStreamT = TypeVar("_AsyncStreamT", bound=AsyncStream[Any]) + +if TYPE_CHECKING: + from httpx._config import ( + DEFAULT_TIMEOUT_CONFIG, # pyright: ignore[reportPrivateImportUsage] + ) + + HTTPX_DEFAULT_TIMEOUT = DEFAULT_TIMEOUT_CONFIG +else: + try: + from httpx._config import DEFAULT_TIMEOUT_CONFIG as HTTPX_DEFAULT_TIMEOUT + except ImportError: + # taken from https://github.com/encode/httpx/blob/3ba5fe0d7ac70222590e759c31442b1cab263791/httpx/_config.py#L366 + HTTPX_DEFAULT_TIMEOUT = Timeout(5.0) + + +class PageInfo: + """Stores the necessary information to build the request to retrieve the next page. + + Either `url` or `params` must be set. + """ + + url: URL | NotGiven + params: Query | NotGiven + json: Body | NotGiven + + @overload + def __init__( + self, + *, + url: URL, + ) -> None: ... + + @overload + def __init__( + self, + *, + params: Query, + ) -> None: ... + + @overload + def __init__( + self, + *, + json: Body, + ) -> None: ... + + def __init__( + self, + *, + url: URL | NotGiven = not_given, + json: Body | NotGiven = not_given, + params: Query | NotGiven = not_given, + ) -> None: + self.url = url + self.json = json + self.params = params + + @override + def __repr__(self) -> str: + if self.url: + return f"{self.__class__.__name__}(url={self.url})" + if self.json: + return f"{self.__class__.__name__}(json={self.json})" + return f"{self.__class__.__name__}(params={self.params})" + + +class BasePage(GenericModel, Generic[_T]): + """ + Defines the core interface for pagination. + + Type Args: + ModelT: The pydantic model that represents an item in the response. + + Methods: + has_next_page(): Check if there is another page available + next_page_info(): Get the necessary information to make a request for the next page + """ + + _options: FinalRequestOptions = PrivateAttr() + _model: Type[_T] = PrivateAttr() + + def has_next_page(self) -> bool: + items = self._get_page_items() + if not items: + return False + return self.next_page_info() is not None + + def next_page_info(self) -> Optional[PageInfo]: ... + + def _get_page_items(self) -> Iterable[_T]: # type: ignore[empty-body] + ... + + def _params_from_url(self, url: URL) -> httpx.QueryParams: + # TODO: do we have to preprocess params here? + return httpx.QueryParams(cast(Any, self._options.params)).merge(url.params) + + def _info_to_options(self, info: PageInfo) -> FinalRequestOptions: + options = model_copy(self._options) + options._strip_raw_response_header() + + if not isinstance(info.params, NotGiven): + options.params = {**options.params, **info.params} + return options + + if not isinstance(info.url, NotGiven): + params = self._params_from_url(info.url) + url = info.url.copy_with(params=params) + options.params = dict(url.params) + options.url = str(url) + return options + + if not isinstance(info.json, NotGiven): + if not is_mapping(info.json): + raise TypeError("Pagination is only supported with mappings") + + if not options.json_data: + options.json_data = {**info.json} + else: + if not is_mapping(options.json_data): + raise TypeError("Pagination is only supported with mappings") + + options.json_data = {**options.json_data, **info.json} + return options + + raise ValueError("Unexpected PageInfo state") + + +class BaseSyncPage(BasePage[_T], Generic[_T]): + _client: SyncAPIClient = pydantic.PrivateAttr() + + def _set_private_attributes( + self, + client: SyncAPIClient, + model: Type[_T], + options: FinalRequestOptions, + ) -> None: + if (not PYDANTIC_V1) and getattr(self, "__pydantic_private__", None) is None: + self.__pydantic_private__ = {} + + self._model = model + self._client = client + self._options = options + + # Pydantic uses a custom `__iter__` method to support casting BaseModels + # to dictionaries. e.g. dict(model). + # As we want to support `for item in page`, this is inherently incompatible + # with the default pydantic behaviour. It is not possible to support both + # use cases at once. Fortunately, this is not a big deal as all other pydantic + # methods should continue to work as expected as there is an alternative method + # to cast a model to a dictionary, model.dict(), which is used internally + # by pydantic. + def __iter__(self) -> Iterator[_T]: # type: ignore + for page in self.iter_pages(): + for item in page._get_page_items(): + yield item + + def iter_pages(self: SyncPageT) -> Iterator[SyncPageT]: + page = self + while True: + yield page + if page.has_next_page(): + page = page.get_next_page() + else: + return + + def get_next_page(self: SyncPageT) -> SyncPageT: + info = self.next_page_info() + if not info: + raise RuntimeError( + "No next page expected; please check `.has_next_page()` before calling `.get_next_page()`." + ) + + options = self._info_to_options(info) + return self._client._request_api_list(self._model, page=self.__class__, options=options) + + +class AsyncPaginator(Generic[_T, AsyncPageT]): + def __init__( + self, + client: AsyncAPIClient, + options: FinalRequestOptions, + page_cls: Type[AsyncPageT], + model: Type[_T], + ) -> None: + self._model = model + self._client = client + self._options = options + self._page_cls = page_cls + + def __await__(self) -> Generator[Any, None, AsyncPageT]: + return self._get_page().__await__() + + async def _get_page(self) -> AsyncPageT: + def _parser(resp: AsyncPageT) -> AsyncPageT: + resp._set_private_attributes( + model=self._model, + options=self._options, + client=self._client, + ) + return resp + + self._options.post_parser = _parser + + return await self._client.request(self._page_cls, self._options) + + async def __aiter__(self) -> AsyncIterator[_T]: + # https://github.com/microsoft/pyright/issues/3464 + page = cast( + AsyncPageT, + await self, # type: ignore + ) + async for item in page: + yield item + + +class BaseAsyncPage(BasePage[_T], Generic[_T]): + _client: AsyncAPIClient = pydantic.PrivateAttr() + + def _set_private_attributes( + self, + model: Type[_T], + client: AsyncAPIClient, + options: FinalRequestOptions, + ) -> None: + if (not PYDANTIC_V1) and getattr(self, "__pydantic_private__", None) is None: + self.__pydantic_private__ = {} + + self._model = model + self._client = client + self._options = options + + async def __aiter__(self) -> AsyncIterator[_T]: + async for page in self.iter_pages(): + for item in page._get_page_items(): + yield item + + async def iter_pages(self: AsyncPageT) -> AsyncIterator[AsyncPageT]: + page = self + while True: + yield page + if page.has_next_page(): + page = await page.get_next_page() + else: + return + + async def get_next_page(self: AsyncPageT) -> AsyncPageT: + info = self.next_page_info() + if not info: + raise RuntimeError( + "No next page expected; please check `.has_next_page()` before calling `.get_next_page()`." + ) + + options = self._info_to_options(info) + return await self._client._request_api_list(self._model, page=self.__class__, options=options) + + +_HttpxClientT = TypeVar("_HttpxClientT", bound=Union[httpx.Client, httpx.AsyncClient]) +_DefaultStreamT = TypeVar("_DefaultStreamT", bound=Union[Stream[Any], AsyncStream[Any]]) + + +class BaseClient(Generic[_HttpxClientT, _DefaultStreamT]): + _client: _HttpxClientT + _version: str + _base_url: URL + max_retries: int + timeout: Union[float, Timeout, None] + _strict_response_validation: bool + _idempotency_header: str | None + _default_stream_cls: type[_DefaultStreamT] | None = None + + def __init__( + self, + *, + version: str, + base_url: str | URL, + _strict_response_validation: bool, + max_retries: int = DEFAULT_MAX_RETRIES, + timeout: float | Timeout | None = DEFAULT_TIMEOUT, + custom_headers: Mapping[str, str] | None = None, + custom_query: Mapping[str, object] | None = None, + ) -> None: + self._version = version + self._base_url = self._enforce_trailing_slash(URL(base_url)) + self.max_retries = max_retries + self.timeout = timeout + self._custom_headers = custom_headers or {} + self._custom_query = custom_query or {} + self._strict_response_validation = _strict_response_validation + self._idempotency_header = None + self._platform: Platform | None = None + + if max_retries is None: # pyright: ignore[reportUnnecessaryComparison] + raise TypeError( + "max_retries cannot be None. If you want to disable retries, pass `0`; if you want unlimited retries, pass `math.inf` or a very high number; if you want the default behavior, pass `brapi.DEFAULT_MAX_RETRIES`" + ) + + def _enforce_trailing_slash(self, url: URL) -> URL: + if url.raw_path.endswith(b"/"): + return url + return url.copy_with(raw_path=url.raw_path + b"/") + + def _make_status_error_from_response( + self, + response: httpx.Response, + ) -> APIStatusError: + if response.is_closed and not response.is_stream_consumed: + # We can't read the response body as it has been closed + # before it was read. This can happen if an event hook + # raises a status error. + body = None + err_msg = f"Error code: {response.status_code}" + else: + err_text = response.text.strip() + body = err_text + + try: + body = json.loads(err_text) + err_msg = f"Error code: {response.status_code} - {body}" + except Exception: + err_msg = err_text or f"Error code: {response.status_code}" + + return self._make_status_error(err_msg, body=body, response=response) + + def _make_status_error( + self, + err_msg: str, + *, + body: object, + response: httpx.Response, + ) -> _exceptions.APIStatusError: + raise NotImplementedError() + + def _build_headers(self, options: FinalRequestOptions, *, retries_taken: int = 0) -> httpx.Headers: + custom_headers = options.headers or {} + headers_dict = _merge_mappings(self.default_headers, custom_headers) + self._validate_headers(headers_dict, custom_headers) + + # headers are case-insensitive while dictionaries are not. + headers = httpx.Headers(headers_dict) + + idempotency_header = self._idempotency_header + if idempotency_header and options.idempotency_key and idempotency_header not in headers: + headers[idempotency_header] = options.idempotency_key + + # Don't set these headers if they were already set or removed by the caller. We check + # `custom_headers`, which can contain `Omit()`, instead of `headers` to account for the removal case. + lower_custom_headers = [header.lower() for header in custom_headers] + if "x-stainless-retry-count" not in lower_custom_headers: + headers["x-stainless-retry-count"] = str(retries_taken) + if "x-stainless-read-timeout" not in lower_custom_headers: + timeout = self.timeout if isinstance(options.timeout, NotGiven) else options.timeout + if isinstance(timeout, Timeout): + timeout = timeout.read + if timeout is not None: + headers["x-stainless-read-timeout"] = str(timeout) + + return headers + + def _prepare_url(self, url: str) -> URL: + """ + Merge a URL argument together with any 'base_url' on the client, + to create the URL used for the outgoing request. + """ + # Copied from httpx's `_merge_url` method. + merge_url = URL(url) + if merge_url.is_relative_url: + merge_raw_path = self.base_url.raw_path + merge_url.raw_path.lstrip(b"/") + return self.base_url.copy_with(raw_path=merge_raw_path) + + return merge_url + + def _make_sse_decoder(self) -> SSEDecoder | SSEBytesDecoder: + return SSEDecoder() + + def _build_request( + self, + options: FinalRequestOptions, + *, + retries_taken: int = 0, + ) -> httpx.Request: + if log.isEnabledFor(logging.DEBUG): + log.debug("Request options: %s", model_dump(options, exclude_unset=True)) + + kwargs: dict[str, Any] = {} + + json_data = options.json_data + if options.extra_json is not None: + if json_data is None: + json_data = cast(Body, options.extra_json) + elif is_mapping(json_data): + json_data = _merge_mappings(json_data, options.extra_json) + else: + raise RuntimeError(f"Unexpected JSON data type, {type(json_data)}, cannot merge with `extra_body`") + + headers = self._build_headers(options, retries_taken=retries_taken) + params = _merge_mappings(self.default_query, options.params) + content_type = headers.get("Content-Type") + files = options.files + + # If the given Content-Type header is multipart/form-data then it + # has to be removed so that httpx can generate the header with + # additional information for us as it has to be in this form + # for the server to be able to correctly parse the request: + # multipart/form-data; boundary=---abc-- + if content_type is not None and content_type.startswith("multipart/form-data"): + if "boundary" not in content_type: + # only remove the header if the boundary hasn't been explicitly set + # as the caller doesn't want httpx to come up with their own boundary + headers.pop("Content-Type") + + # As we are now sending multipart/form-data instead of application/json + # we need to tell httpx to use it, https://www.python-httpx.org/advanced/clients/#multipart-file-encoding + if json_data: + if not is_dict(json_data): + raise TypeError( + f"Expected query input to be a dictionary for multipart requests but got {type(json_data)} instead." + ) + kwargs["data"] = self._serialize_multipartform(json_data) + + # httpx determines whether or not to send a "multipart/form-data" + # request based on the truthiness of the "files" argument. + # This gets around that issue by generating a dict value that + # evaluates to true. + # + # https://github.com/encode/httpx/discussions/2399#discussioncomment-3814186 + if not files: + files = cast(HttpxRequestFiles, ForceMultipartDict()) + + prepared_url = self._prepare_url(options.url) + if "_" in prepared_url.host: + # work around https://github.com/encode/httpx/discussions/2880 + kwargs["extensions"] = {"sni_hostname": prepared_url.host.replace("_", "-")} + + is_body_allowed = options.method.lower() != "get" + + if is_body_allowed: + if isinstance(json_data, bytes): + kwargs["content"] = json_data + else: + kwargs["json"] = json_data if is_given(json_data) else None + kwargs["files"] = files + else: + headers.pop("Content-Type", None) + kwargs.pop("data", None) + + # TODO: report this error to httpx + return self._client.build_request( # pyright: ignore[reportUnknownMemberType] + headers=headers, + timeout=self.timeout if isinstance(options.timeout, NotGiven) else options.timeout, + method=options.method, + url=prepared_url, + # the `Query` type that we use is incompatible with qs' + # `Params` type as it needs to be typed as `Mapping[str, object]` + # so that passing a `TypedDict` doesn't cause an error. + # https://github.com/microsoft/pyright/issues/3526#event-6715453066 + params=self.qs.stringify(cast(Mapping[str, Any], params)) if params else None, + **kwargs, + ) + + def _serialize_multipartform(self, data: Mapping[object, object]) -> dict[str, object]: + items = self.qs.stringify_items( + # TODO: type ignore is required as stringify_items is well typed but we can't be + # well typed without heavy validation. + data, # type: ignore + array_format="brackets", + ) + serialized: dict[str, object] = {} + for key, value in items: + existing = serialized.get(key) + + if not existing: + serialized[key] = value + continue + + # If a value has already been set for this key then that + # means we're sending data like `array[]=[1, 2, 3]` and we + # need to tell httpx that we want to send multiple values with + # the same key which is done by using a list or a tuple. + # + # Note: 2d arrays should never result in the same key at both + # levels so it's safe to assume that if the value is a list, + # it was because we changed it to be a list. + if is_list(existing): + existing.append(value) + else: + serialized[key] = [existing, value] + + return serialized + + def _maybe_override_cast_to(self, cast_to: type[ResponseT], options: FinalRequestOptions) -> type[ResponseT]: + if not is_given(options.headers): + return cast_to + + # make a copy of the headers so we don't mutate user-input + headers = dict(options.headers) + + # we internally support defining a temporary header to override the + # default `cast_to` type for use with `.with_raw_response` and `.with_streaming_response` + # see _response.py for implementation details + override_cast_to = headers.pop(OVERRIDE_CAST_TO_HEADER, not_given) + if is_given(override_cast_to): + options.headers = headers + return cast(Type[ResponseT], override_cast_to) + + return cast_to + + def _should_stream_response_body(self, request: httpx.Request) -> bool: + return request.headers.get(RAW_RESPONSE_HEADER) == "stream" # type: ignore[no-any-return] + + def _process_response_data( + self, + *, + data: object, + cast_to: type[ResponseT], + response: httpx.Response, + ) -> ResponseT: + if data is None: + return cast(ResponseT, None) + + if cast_to is object: + return cast(ResponseT, data) + + try: + if inspect.isclass(cast_to) and issubclass(cast_to, ModelBuilderProtocol): + return cast(ResponseT, cast_to.build(response=response, data=data)) + + if self._strict_response_validation: + return cast(ResponseT, validate_type(type_=cast_to, value=data)) + + return cast(ResponseT, construct_type(type_=cast_to, value=data)) + except pydantic.ValidationError as err: + raise APIResponseValidationError(response=response, body=data) from err + + @property + def qs(self) -> Querystring: + return Querystring() + + @property + def custom_auth(self) -> httpx.Auth | None: + return None + + @property + def auth_headers(self) -> dict[str, str]: + return {} + + @property + def default_headers(self) -> dict[str, str | Omit]: + return { + "Accept": "application/json", + "Content-Type": "application/json", + "User-Agent": self.user_agent, + **self.platform_headers(), + **self.auth_headers, + **self._custom_headers, + } + + @property + def default_query(self) -> dict[str, object]: + return { + **self._custom_query, + } + + def _validate_headers( + self, + headers: Headers, # noqa: ARG002 + custom_headers: Headers, # noqa: ARG002 + ) -> None: + """Validate the given default headers and custom headers. + + Does nothing by default. + """ + return + + @property + def user_agent(self) -> str: + return f"{self.__class__.__name__}/Python {self._version}" + + @property + def base_url(self) -> URL: + return self._base_url + + @base_url.setter + def base_url(self, url: URL | str) -> None: + self._base_url = self._enforce_trailing_slash(url if isinstance(url, URL) else URL(url)) + + def platform_headers(self) -> Dict[str, str]: + # the actual implementation is in a separate `lru_cache` decorated + # function because adding `lru_cache` to methods will leak memory + # https://github.com/python/cpython/issues/88476 + return platform_headers(self._version, platform=self._platform) + + def _parse_retry_after_header(self, response_headers: Optional[httpx.Headers] = None) -> float | None: + """Returns a float of the number of seconds (not milliseconds) to wait after retrying, or None if unspecified. + + About the Retry-After header: https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Retry-After + See also https://developer.mozilla.org/en-US/docs/Web/HTTP/Headers/Retry-After#syntax + """ + if response_headers is None: + return None + + # First, try the non-standard `retry-after-ms` header for milliseconds, + # which is more precise than integer-seconds `retry-after` + try: + retry_ms_header = response_headers.get("retry-after-ms", None) + return float(retry_ms_header) / 1000 + except (TypeError, ValueError): + pass + + # Next, try parsing `retry-after` header as seconds (allowing nonstandard floats). + retry_header = response_headers.get("retry-after") + try: + # note: the spec indicates that this should only ever be an integer + # but if someone sends a float there's no reason for us to not respect it + return float(retry_header) + except (TypeError, ValueError): + pass + + # Last, try parsing `retry-after` as a date. + retry_date_tuple = email.utils.parsedate_tz(retry_header) + if retry_date_tuple is None: + return None + + retry_date = email.utils.mktime_tz(retry_date_tuple) + return float(retry_date - time.time()) + + def _calculate_retry_timeout( + self, + remaining_retries: int, + options: FinalRequestOptions, + response_headers: Optional[httpx.Headers] = None, + ) -> float: + max_retries = options.get_max_retries(self.max_retries) + + # If the API asks us to wait a certain amount of time (and it's a reasonable amount), just do what it says. + retry_after = self._parse_retry_after_header(response_headers) + if retry_after is not None and 0 < retry_after <= 60: + return retry_after + + # Also cap retry count to 1000 to avoid any potential overflows with `pow` + nb_retries = min(max_retries - remaining_retries, 1000) + + # Apply exponential backoff, but not more than the max. + sleep_seconds = min(INITIAL_RETRY_DELAY * pow(2.0, nb_retries), MAX_RETRY_DELAY) + + # Apply some jitter, plus-or-minus half a second. + jitter = 1 - 0.25 * random() + timeout = sleep_seconds * jitter + return timeout if timeout >= 0 else 0 + + def _should_retry(self, response: httpx.Response) -> bool: + # Note: this is not a standard header + should_retry_header = response.headers.get("x-should-retry") + + # If the server explicitly says whether or not to retry, obey. + if should_retry_header == "true": + log.debug("Retrying as header `x-should-retry` is set to `true`") + return True + if should_retry_header == "false": + log.debug("Not retrying as header `x-should-retry` is set to `false`") + return False + + # Retry on request timeouts. + if response.status_code == 408: + log.debug("Retrying due to status code %i", response.status_code) + return True + + # Retry on lock timeouts. + if response.status_code == 409: + log.debug("Retrying due to status code %i", response.status_code) + return True + + # Retry on rate limits. + if response.status_code == 429: + log.debug("Retrying due to status code %i", response.status_code) + return True + + # Retry internal errors. + if response.status_code >= 500: + log.debug("Retrying due to status code %i", response.status_code) + return True + + log.debug("Not retrying") + return False + + def _idempotency_key(self) -> str: + return f"stainless-python-retry-{uuid.uuid4()}" + + +class _DefaultHttpxClient(httpx.Client): + def __init__(self, **kwargs: Any) -> None: + kwargs.setdefault("timeout", DEFAULT_TIMEOUT) + kwargs.setdefault("limits", DEFAULT_CONNECTION_LIMITS) + kwargs.setdefault("follow_redirects", True) + super().__init__(**kwargs) + + +if TYPE_CHECKING: + DefaultHttpxClient = httpx.Client + """An alias to `httpx.Client` that provides the same defaults that this SDK + uses internally. + + This is useful because overriding the `http_client` with your own instance of + `httpx.Client` will result in httpx's defaults being used, not ours. + """ +else: + DefaultHttpxClient = _DefaultHttpxClient + + +class SyncHttpxClientWrapper(DefaultHttpxClient): + def __del__(self) -> None: + if self.is_closed: + return + + try: + self.close() + except Exception: + pass + + +class SyncAPIClient(BaseClient[httpx.Client, Stream[Any]]): + _client: httpx.Client + _default_stream_cls: type[Stream[Any]] | None = None + + def __init__( + self, + *, + version: str, + base_url: str | URL, + max_retries: int = DEFAULT_MAX_RETRIES, + timeout: float | Timeout | None | NotGiven = not_given, + http_client: httpx.Client | None = None, + custom_headers: Mapping[str, str] | None = None, + custom_query: Mapping[str, object] | None = None, + _strict_response_validation: bool, + ) -> None: + if not is_given(timeout): + # if the user passed in a custom http client with a non-default + # timeout set then we use that timeout. + # + # note: there is an edge case here where the user passes in a client + # where they've explicitly set the timeout to match the default timeout + # as this check is structural, meaning that we'll think they didn't + # pass in a timeout and will ignore it + if http_client and http_client.timeout != HTTPX_DEFAULT_TIMEOUT: + timeout = http_client.timeout + else: + timeout = DEFAULT_TIMEOUT + + if http_client is not None and not isinstance(http_client, httpx.Client): # pyright: ignore[reportUnnecessaryIsInstance] + raise TypeError( + f"Invalid `http_client` argument; Expected an instance of `httpx.Client` but got {type(http_client)}" + ) + + super().__init__( + version=version, + # cast to a valid type because mypy doesn't understand our type narrowing + timeout=cast(Timeout, timeout), + base_url=base_url, + max_retries=max_retries, + custom_query=custom_query, + custom_headers=custom_headers, + _strict_response_validation=_strict_response_validation, + ) + self._client = http_client or SyncHttpxClientWrapper( + base_url=base_url, + # cast to a valid type because mypy doesn't understand our type narrowing + timeout=cast(Timeout, timeout), + ) + + def is_closed(self) -> bool: + return self._client.is_closed + + def close(self) -> None: + """Close the underlying HTTPX client. + + The client will *not* be usable after this. + """ + # If an error is thrown while constructing a client, self._client + # may not be present + if hasattr(self, "_client"): + self._client.close() + + def __enter__(self: _T) -> _T: + return self + + def __exit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + self.close() + + def _prepare_options( + self, + options: FinalRequestOptions, # noqa: ARG002 + ) -> FinalRequestOptions: + """Hook for mutating the given options""" + return options + + def _prepare_request( + self, + request: httpx.Request, # noqa: ARG002 + ) -> None: + """This method is used as a callback for mutating the `Request` object + after it has been constructed. + This is useful for cases where you want to add certain headers based off of + the request properties, e.g. `url`, `method` etc. + """ + return None + + @overload + def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: Literal[True], + stream_cls: Type[_StreamT], + ) -> _StreamT: ... + + @overload + def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: bool = False, + stream_cls: Type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: ... + + def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: bool = False, + stream_cls: type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: + cast_to = self._maybe_override_cast_to(cast_to, options) + + # create a copy of the options we were given so that if the + # options are mutated later & we then retry, the retries are + # given the original options + input_options = model_copy(options) + if input_options.idempotency_key is None and input_options.method.lower() != "get": + # ensure the idempotency key is reused between requests + input_options.idempotency_key = self._idempotency_key() + + response: httpx.Response | None = None + max_retries = input_options.get_max_retries(self.max_retries) + + retries_taken = 0 + for retries_taken in range(max_retries + 1): + options = model_copy(input_options) + options = self._prepare_options(options) + + remaining_retries = max_retries - retries_taken + request = self._build_request(options, retries_taken=retries_taken) + self._prepare_request(request) + + kwargs: HttpxSendArgs = {} + if self.custom_auth is not None: + kwargs["auth"] = self.custom_auth + + if options.follow_redirects is not None: + kwargs["follow_redirects"] = options.follow_redirects + + log.debug("Sending HTTP Request: %s %s", request.method, request.url) + + response = None + try: + response = self._client.send( + request, + stream=stream or self._should_stream_response_body(request=request), + **kwargs, + ) + except httpx.TimeoutException as err: + log.debug("Encountered httpx.TimeoutException", exc_info=True) + + if remaining_retries > 0: + self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=None, + ) + continue + + log.debug("Raising timeout error") + raise APITimeoutError(request=request) from err + except Exception as err: + log.debug("Encountered Exception", exc_info=True) + + if remaining_retries > 0: + self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=None, + ) + continue + + log.debug("Raising connection error") + raise APIConnectionError(request=request) from err + + log.debug( + 'HTTP Response: %s %s "%i %s" %s', + request.method, + request.url, + response.status_code, + response.reason_phrase, + response.headers, + ) + + try: + response.raise_for_status() + except httpx.HTTPStatusError as err: # thrown on 4xx and 5xx status code + log.debug("Encountered httpx.HTTPStatusError", exc_info=True) + + if remaining_retries > 0 and self._should_retry(err.response): + err.response.close() + self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=response, + ) + continue + + # If the response is streamed then we need to explicitly read the response + # to completion before attempting to access the response text. + if not err.response.is_closed: + err.response.read() + + log.debug("Re-raising status error") + raise self._make_status_error_from_response(err.response) from None + + break + + assert response is not None, "could not resolve response (should never happen)" + return self._process_response( + cast_to=cast_to, + options=options, + response=response, + stream=stream, + stream_cls=stream_cls, + retries_taken=retries_taken, + ) + + def _sleep_for_retry( + self, *, retries_taken: int, max_retries: int, options: FinalRequestOptions, response: httpx.Response | None + ) -> None: + remaining_retries = max_retries - retries_taken + if remaining_retries == 1: + log.debug("1 retry left") + else: + log.debug("%i retries left", remaining_retries) + + timeout = self._calculate_retry_timeout(remaining_retries, options, response.headers if response else None) + log.info("Retrying request to %s in %f seconds", options.url, timeout) + + time.sleep(timeout) + + def _process_response( + self, + *, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + response: httpx.Response, + stream: bool, + stream_cls: type[Stream[Any]] | type[AsyncStream[Any]] | None, + retries_taken: int = 0, + ) -> ResponseT: + origin = get_origin(cast_to) or cast_to + + if ( + inspect.isclass(origin) + and issubclass(origin, BaseAPIResponse) + # we only want to actually return the custom BaseAPIResponse class if we're + # returning the raw response, or if we're not streaming SSE, as if we're streaming + # SSE then `cast_to` doesn't actively reflect the type we need to parse into + and (not stream or bool(response.request.headers.get(RAW_RESPONSE_HEADER))) + ): + if not issubclass(origin, APIResponse): + raise TypeError(f"API Response types must subclass {APIResponse}; Received {origin}") + + response_cls = cast("type[BaseAPIResponse[Any]]", cast_to) + return cast( + ResponseT, + response_cls( + raw=response, + client=self, + cast_to=extract_response_type(response_cls), + stream=stream, + stream_cls=stream_cls, + options=options, + retries_taken=retries_taken, + ), + ) + + if cast_to == httpx.Response: + return cast(ResponseT, response) + + api_response = APIResponse( + raw=response, + client=self, + cast_to=cast("type[ResponseT]", cast_to), # pyright: ignore[reportUnnecessaryCast] + stream=stream, + stream_cls=stream_cls, + options=options, + retries_taken=retries_taken, + ) + if bool(response.request.headers.get(RAW_RESPONSE_HEADER)): + return cast(ResponseT, api_response) + + return api_response.parse() + + def _request_api_list( + self, + model: Type[object], + page: Type[SyncPageT], + options: FinalRequestOptions, + ) -> SyncPageT: + def _parser(resp: SyncPageT) -> SyncPageT: + resp._set_private_attributes( + client=self, + model=model, + options=options, + ) + return resp + + options.post_parser = _parser + + return self.request(page, options, stream=False) + + @overload + def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: Literal[True], + stream_cls: type[_StreamT], + ) -> _StreamT: ... + + @overload + def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: bool, + stream_cls: type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: ... + + def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: bool = False, + stream_cls: type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: + opts = FinalRequestOptions.construct(method="get", url=path, **options) + # cast is required because mypy complains about returning Any even though + # it understands the type variables + return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) + + @overload + def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + files: RequestFiles | None = None, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + files: RequestFiles | None = None, + stream: Literal[True], + stream_cls: type[_StreamT], + ) -> _StreamT: ... + + @overload + def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + files: RequestFiles | None = None, + stream: bool, + stream_cls: type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: ... + + def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + files: RequestFiles | None = None, + stream: bool = False, + stream_cls: type[_StreamT] | None = None, + ) -> ResponseT | _StreamT: + opts = FinalRequestOptions.construct( + method="post", url=path, json_data=body, files=to_httpx_files(files), **options + ) + return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) + + def patch( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct(method="patch", url=path, json_data=body, **options) + return self.request(cast_to, opts) + + def put( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct( + method="put", url=path, json_data=body, files=to_httpx_files(files), **options + ) + return self.request(cast_to, opts) + + def delete( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, **options) + return self.request(cast_to, opts) + + def get_api_list( + self, + path: str, + *, + model: Type[object], + page: Type[SyncPageT], + body: Body | None = None, + options: RequestOptions = {}, + method: str = "get", + ) -> SyncPageT: + opts = FinalRequestOptions.construct(method=method, url=path, json_data=body, **options) + return self._request_api_list(model, page, opts) + + +class _DefaultAsyncHttpxClient(httpx.AsyncClient): + def __init__(self, **kwargs: Any) -> None: + kwargs.setdefault("timeout", DEFAULT_TIMEOUT) + kwargs.setdefault("limits", DEFAULT_CONNECTION_LIMITS) + kwargs.setdefault("follow_redirects", True) + super().__init__(**kwargs) + + +try: + import httpx_aiohttp +except ImportError: + + class _DefaultAioHttpClient(httpx.AsyncClient): + def __init__(self, **_kwargs: Any) -> None: + raise RuntimeError("To use the aiohttp client you must have installed the package with the `aiohttp` extra") +else: + + class _DefaultAioHttpClient(httpx_aiohttp.HttpxAiohttpClient): # type: ignore + def __init__(self, **kwargs: Any) -> None: + kwargs.setdefault("timeout", DEFAULT_TIMEOUT) + kwargs.setdefault("limits", DEFAULT_CONNECTION_LIMITS) + kwargs.setdefault("follow_redirects", True) + + super().__init__(**kwargs) + + +if TYPE_CHECKING: + DefaultAsyncHttpxClient = httpx.AsyncClient + """An alias to `httpx.AsyncClient` that provides the same defaults that this SDK + uses internally. + + This is useful because overriding the `http_client` with your own instance of + `httpx.AsyncClient` will result in httpx's defaults being used, not ours. + """ + + DefaultAioHttpClient = httpx.AsyncClient + """An alias to `httpx.AsyncClient` that changes the default HTTP transport to `aiohttp`.""" +else: + DefaultAsyncHttpxClient = _DefaultAsyncHttpxClient + DefaultAioHttpClient = _DefaultAioHttpClient + + +class AsyncHttpxClientWrapper(DefaultAsyncHttpxClient): + def __del__(self) -> None: + if self.is_closed: + return + + try: + # TODO(someday): support non asyncio runtimes here + asyncio.get_running_loop().create_task(self.aclose()) + except Exception: + pass + + +class AsyncAPIClient(BaseClient[httpx.AsyncClient, AsyncStream[Any]]): + _client: httpx.AsyncClient + _default_stream_cls: type[AsyncStream[Any]] | None = None + + def __init__( + self, + *, + version: str, + base_url: str | URL, + _strict_response_validation: bool, + max_retries: int = DEFAULT_MAX_RETRIES, + timeout: float | Timeout | None | NotGiven = not_given, + http_client: httpx.AsyncClient | None = None, + custom_headers: Mapping[str, str] | None = None, + custom_query: Mapping[str, object] | None = None, + ) -> None: + if not is_given(timeout): + # if the user passed in a custom http client with a non-default + # timeout set then we use that timeout. + # + # note: there is an edge case here where the user passes in a client + # where they've explicitly set the timeout to match the default timeout + # as this check is structural, meaning that we'll think they didn't + # pass in a timeout and will ignore it + if http_client and http_client.timeout != HTTPX_DEFAULT_TIMEOUT: + timeout = http_client.timeout + else: + timeout = DEFAULT_TIMEOUT + + if http_client is not None and not isinstance(http_client, httpx.AsyncClient): # pyright: ignore[reportUnnecessaryIsInstance] + raise TypeError( + f"Invalid `http_client` argument; Expected an instance of `httpx.AsyncClient` but got {type(http_client)}" + ) + + super().__init__( + version=version, + base_url=base_url, + # cast to a valid type because mypy doesn't understand our type narrowing + timeout=cast(Timeout, timeout), + max_retries=max_retries, + custom_query=custom_query, + custom_headers=custom_headers, + _strict_response_validation=_strict_response_validation, + ) + self._client = http_client or AsyncHttpxClientWrapper( + base_url=base_url, + # cast to a valid type because mypy doesn't understand our type narrowing + timeout=cast(Timeout, timeout), + ) + + def is_closed(self) -> bool: + return self._client.is_closed + + async def close(self) -> None: + """Close the underlying HTTPX client. + + The client will *not* be usable after this. + """ + await self._client.aclose() + + async def __aenter__(self: _T) -> _T: + return self + + async def __aexit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + await self.close() + + async def _prepare_options( + self, + options: FinalRequestOptions, # noqa: ARG002 + ) -> FinalRequestOptions: + """Hook for mutating the given options""" + return options + + async def _prepare_request( + self, + request: httpx.Request, # noqa: ARG002 + ) -> None: + """This method is used as a callback for mutating the `Request` object + after it has been constructed. + This is useful for cases where you want to add certain headers based off of + the request properties, e.g. `url`, `method` etc. + """ + return None + + @overload + async def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + async def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: Literal[True], + stream_cls: type[_AsyncStreamT], + ) -> _AsyncStreamT: ... + + @overload + async def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: bool, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: ... + + async def request( + self, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + *, + stream: bool = False, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: + if self._platform is None: + # `get_platform` can make blocking IO calls so we + # execute it earlier while we are in an async context + self._platform = await asyncify(get_platform)() + + cast_to = self._maybe_override_cast_to(cast_to, options) + + # create a copy of the options we were given so that if the + # options are mutated later & we then retry, the retries are + # given the original options + input_options = model_copy(options) + if input_options.idempotency_key is None and input_options.method.lower() != "get": + # ensure the idempotency key is reused between requests + input_options.idempotency_key = self._idempotency_key() + + response: httpx.Response | None = None + max_retries = input_options.get_max_retries(self.max_retries) + + retries_taken = 0 + for retries_taken in range(max_retries + 1): + options = model_copy(input_options) + options = await self._prepare_options(options) + + remaining_retries = max_retries - retries_taken + request = self._build_request(options, retries_taken=retries_taken) + await self._prepare_request(request) + + kwargs: HttpxSendArgs = {} + if self.custom_auth is not None: + kwargs["auth"] = self.custom_auth + + if options.follow_redirects is not None: + kwargs["follow_redirects"] = options.follow_redirects + + log.debug("Sending HTTP Request: %s %s", request.method, request.url) + + response = None + try: + response = await self._client.send( + request, + stream=stream or self._should_stream_response_body(request=request), + **kwargs, + ) + except httpx.TimeoutException as err: + log.debug("Encountered httpx.TimeoutException", exc_info=True) + + if remaining_retries > 0: + await self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=None, + ) + continue + + log.debug("Raising timeout error") + raise APITimeoutError(request=request) from err + except Exception as err: + log.debug("Encountered Exception", exc_info=True) + + if remaining_retries > 0: + await self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=None, + ) + continue + + log.debug("Raising connection error") + raise APIConnectionError(request=request) from err + + log.debug( + 'HTTP Response: %s %s "%i %s" %s', + request.method, + request.url, + response.status_code, + response.reason_phrase, + response.headers, + ) + + try: + response.raise_for_status() + except httpx.HTTPStatusError as err: # thrown on 4xx and 5xx status code + log.debug("Encountered httpx.HTTPStatusError", exc_info=True) + + if remaining_retries > 0 and self._should_retry(err.response): + await err.response.aclose() + await self._sleep_for_retry( + retries_taken=retries_taken, + max_retries=max_retries, + options=input_options, + response=response, + ) + continue + + # If the response is streamed then we need to explicitly read the response + # to completion before attempting to access the response text. + if not err.response.is_closed: + await err.response.aread() + + log.debug("Re-raising status error") + raise self._make_status_error_from_response(err.response) from None + + break + + assert response is not None, "could not resolve response (should never happen)" + return await self._process_response( + cast_to=cast_to, + options=options, + response=response, + stream=stream, + stream_cls=stream_cls, + retries_taken=retries_taken, + ) + + async def _sleep_for_retry( + self, *, retries_taken: int, max_retries: int, options: FinalRequestOptions, response: httpx.Response | None + ) -> None: + remaining_retries = max_retries - retries_taken + if remaining_retries == 1: + log.debug("1 retry left") + else: + log.debug("%i retries left", remaining_retries) + + timeout = self._calculate_retry_timeout(remaining_retries, options, response.headers if response else None) + log.info("Retrying request to %s in %f seconds", options.url, timeout) + + await anyio.sleep(timeout) + + async def _process_response( + self, + *, + cast_to: Type[ResponseT], + options: FinalRequestOptions, + response: httpx.Response, + stream: bool, + stream_cls: type[Stream[Any]] | type[AsyncStream[Any]] | None, + retries_taken: int = 0, + ) -> ResponseT: + origin = get_origin(cast_to) or cast_to + + if ( + inspect.isclass(origin) + and issubclass(origin, BaseAPIResponse) + # we only want to actually return the custom BaseAPIResponse class if we're + # returning the raw response, or if we're not streaming SSE, as if we're streaming + # SSE then `cast_to` doesn't actively reflect the type we need to parse into + and (not stream or bool(response.request.headers.get(RAW_RESPONSE_HEADER))) + ): + if not issubclass(origin, AsyncAPIResponse): + raise TypeError(f"API Response types must subclass {AsyncAPIResponse}; Received {origin}") + + response_cls = cast("type[BaseAPIResponse[Any]]", cast_to) + return cast( + "ResponseT", + response_cls( + raw=response, + client=self, + cast_to=extract_response_type(response_cls), + stream=stream, + stream_cls=stream_cls, + options=options, + retries_taken=retries_taken, + ), + ) + + if cast_to == httpx.Response: + return cast(ResponseT, response) + + api_response = AsyncAPIResponse( + raw=response, + client=self, + cast_to=cast("type[ResponseT]", cast_to), # pyright: ignore[reportUnnecessaryCast] + stream=stream, + stream_cls=stream_cls, + options=options, + retries_taken=retries_taken, + ) + if bool(response.request.headers.get(RAW_RESPONSE_HEADER)): + return cast(ResponseT, api_response) + + return await api_response.parse() + + def _request_api_list( + self, + model: Type[_T], + page: Type[AsyncPageT], + options: FinalRequestOptions, + ) -> AsyncPaginator[_T, AsyncPageT]: + return AsyncPaginator(client=self, options=options, page_cls=page, model=model) + + @overload + async def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + async def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: Literal[True], + stream_cls: type[_AsyncStreamT], + ) -> _AsyncStreamT: ... + + @overload + async def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: bool, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: ... + + async def get( + self, + path: str, + *, + cast_to: Type[ResponseT], + options: RequestOptions = {}, + stream: bool = False, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: + opts = FinalRequestOptions.construct(method="get", url=path, **options) + return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) + + @overload + async def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + stream: Literal[False] = False, + ) -> ResponseT: ... + + @overload + async def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + stream: Literal[True], + stream_cls: type[_AsyncStreamT], + ) -> _AsyncStreamT: ... + + @overload + async def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + stream: bool, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: ... + + async def post( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + stream: bool = False, + stream_cls: type[_AsyncStreamT] | None = None, + ) -> ResponseT | _AsyncStreamT: + opts = FinalRequestOptions.construct( + method="post", url=path, json_data=body, files=await async_to_httpx_files(files), **options + ) + return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) + + async def patch( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct(method="patch", url=path, json_data=body, **options) + return await self.request(cast_to, opts) + + async def put( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + files: RequestFiles | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct( + method="put", url=path, json_data=body, files=await async_to_httpx_files(files), **options + ) + return await self.request(cast_to, opts) + + async def delete( + self, + path: str, + *, + cast_to: Type[ResponseT], + body: Body | None = None, + options: RequestOptions = {}, + ) -> ResponseT: + opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, **options) + return await self.request(cast_to, opts) + + def get_api_list( + self, + path: str, + *, + model: Type[_T], + page: Type[AsyncPageT], + body: Body | None = None, + options: RequestOptions = {}, + method: str = "get", + ) -> AsyncPaginator[_T, AsyncPageT]: + opts = FinalRequestOptions.construct(method=method, url=path, json_data=body, **options) + return self._request_api_list(model, page, opts) + + +def make_request_options( + *, + query: Query | None = None, + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + idempotency_key: str | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + post_parser: PostParser | NotGiven = not_given, +) -> RequestOptions: + """Create a dict of type RequestOptions without keys of NotGiven values.""" + options: RequestOptions = {} + if extra_headers is not None: + options["headers"] = extra_headers + + if extra_body is not None: + options["extra_json"] = cast(AnyMapping, extra_body) + + if query is not None: + options["params"] = query + + if extra_query is not None: + options["params"] = {**options.get("params", {}), **extra_query} + + if not isinstance(timeout, NotGiven): + options["timeout"] = timeout + + if idempotency_key is not None: + options["idempotency_key"] = idempotency_key + + if is_given(post_parser): + # internal + options["post_parser"] = post_parser # type: ignore + + return options + + +class ForceMultipartDict(Dict[str, None]): + def __bool__(self) -> bool: + return True + + +class OtherPlatform: + def __init__(self, name: str) -> None: + self.name = name + + @override + def __str__(self) -> str: + return f"Other:{self.name}" + + +Platform = Union[ + OtherPlatform, + Literal[ + "MacOS", + "Linux", + "Windows", + "FreeBSD", + "OpenBSD", + "iOS", + "Android", + "Unknown", + ], +] + + +def get_platform() -> Platform: + try: + system = platform.system().lower() + platform_name = platform.platform().lower() + except Exception: + return "Unknown" + + if "iphone" in platform_name or "ipad" in platform_name: + # Tested using Python3IDE on an iPhone 11 and Pythonista on an iPad 7 + # system is Darwin and platform_name is a string like: + # - Darwin-21.6.0-iPhone12,1-64bit + # - Darwin-21.6.0-iPad7,11-64bit + return "iOS" + + if system == "darwin": + return "MacOS" + + if system == "windows": + return "Windows" + + if "android" in platform_name: + # Tested using Pydroid 3 + # system is Linux and platform_name is a string like 'Linux-5.10.81-android12-9-00001-geba40aecb3b7-ab8534902-aarch64-with-libc' + return "Android" + + if system == "linux": + # https://distro.readthedocs.io/en/latest/#distro.id + distro_id = distro.id() + if distro_id == "freebsd": + return "FreeBSD" + + if distro_id == "openbsd": + return "OpenBSD" + + return "Linux" + + if platform_name: + return OtherPlatform(platform_name) + + return "Unknown" + + +@lru_cache(maxsize=None) +def platform_headers(version: str, *, platform: Platform | None) -> Dict[str, str]: + return { + "X-Stainless-Lang": "python", + "X-Stainless-Package-Version": version, + "X-Stainless-OS": str(platform or get_platform()), + "X-Stainless-Arch": str(get_architecture()), + "X-Stainless-Runtime": get_python_runtime(), + "X-Stainless-Runtime-Version": get_python_version(), + } + + +class OtherArch: + def __init__(self, name: str) -> None: + self.name = name + + @override + def __str__(self) -> str: + return f"other:{self.name}" + + +Arch = Union[OtherArch, Literal["x32", "x64", "arm", "arm64", "unknown"]] + + +def get_python_runtime() -> str: + try: + return platform.python_implementation() + except Exception: + return "unknown" + + +def get_python_version() -> str: + try: + return platform.python_version() + except Exception: + return "unknown" + + +def get_architecture() -> Arch: + try: + machine = platform.machine().lower() + except Exception: + return "unknown" + + if machine in ("arm64", "aarch64"): + return "arm64" + + # TODO: untested + if machine == "arm": + return "arm" + + if machine == "x86_64": + return "x64" + + # TODO: untested + if sys.maxsize <= 2**32: + return "x32" + + if machine: + return OtherArch(machine) + + return "unknown" + + +def _merge_mappings( + obj1: Mapping[_T_co, Union[_T, Omit]], + obj2: Mapping[_T_co, Union[_T, Omit]], +) -> Dict[_T_co, _T]: + """Merge two mappings of the same type, removing any values that are instances of `Omit`. + + In cases with duplicate keys the second mapping takes precedence. + """ + merged = {**obj1, **obj2} + return {key: value for key, value in merged.items() if not isinstance(value, Omit)} diff --git a/src/brapi/_client.py b/src/brapi/_client.py new file mode 100644 index 0000000..992de0c --- /dev/null +++ b/src/brapi/_client.py @@ -0,0 +1,478 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, Dict, Mapping, cast +from typing_extensions import Self, Literal, override + +import httpx + +from . import _exceptions +from ._qs import Querystring +from ._types import ( + Omit, + Timeout, + NotGiven, + Transport, + ProxiesTypes, + RequestOptions, + not_given, +) +from ._utils import is_given, get_async_library +from ._version import __version__ +from .resources import quote, available +from ._streaming import Stream as Stream, AsyncStream as AsyncStream +from ._exceptions import BrapiError, APIStatusError +from ._base_client import ( + DEFAULT_MAX_RETRIES, + SyncAPIClient, + AsyncAPIClient, +) +from .resources.v2 import v2 + +__all__ = [ + "ENVIRONMENTS", + "Timeout", + "Transport", + "ProxiesTypes", + "RequestOptions", + "Brapi", + "AsyncBrapi", + "Client", + "AsyncClient", +] + +ENVIRONMENTS: Dict[str, str] = { + "production": "https://brapi.dev", + "environment_1": "http://localhost:3000", +} + + +class Brapi(SyncAPIClient): + quote: quote.QuoteResource + available: available.AvailableResource + v2: v2.V2Resource + with_raw_response: BrapiWithRawResponse + with_streaming_response: BrapiWithStreamedResponse + + # client options + api_key: str + + _environment: Literal["production", "environment_1"] | NotGiven + + def __init__( + self, + *, + api_key: str | None = None, + environment: Literal["production", "environment_1"] | NotGiven = not_given, + base_url: str | httpx.URL | None | NotGiven = not_given, + timeout: float | Timeout | None | NotGiven = not_given, + max_retries: int = DEFAULT_MAX_RETRIES, + default_headers: Mapping[str, str] | None = None, + default_query: Mapping[str, object] | None = None, + # Configure a custom httpx client. + # We provide a `DefaultHttpxClient` class that you can pass to retain the default values we use for `limits`, `timeout` & `follow_redirects`. + # See the [httpx documentation](https://www.python-httpx.org/api/#client) for more details. + http_client: httpx.Client | None = None, + # Enable or disable schema validation for data returned by the API. + # When enabled an error APIResponseValidationError is raised + # if the API responds with invalid data for the expected schema. + # + # This parameter may be removed or changed in the future. + # If you rely on this feature, please open a GitHub issue + # outlining your use-case to help us decide if it should be + # part of our public interface in the future. + _strict_response_validation: bool = False, + ) -> None: + """Construct a new synchronous Brapi client instance. + + This automatically infers the `api_key` argument from the `BRAPI_API_KEY` environment variable if it is not provided. + """ + if api_key is None: + api_key = os.environ.get("BRAPI_API_KEY") + if api_key is None: + raise BrapiError( + "The api_key client option must be set either by passing api_key to the client or by setting the BRAPI_API_KEY environment variable" + ) + self.api_key = api_key + + self._environment = environment + + base_url_env = os.environ.get("BRAPI_BASE_URL") + if is_given(base_url) and base_url is not None: + # cast required because mypy doesn't understand the type narrowing + base_url = cast("str | httpx.URL", base_url) # pyright: ignore[reportUnnecessaryCast] + elif is_given(environment): + if base_url_env and base_url is not None: + raise ValueError( + "Ambiguous URL; The `BRAPI_BASE_URL` env var and the `environment` argument are given. If you want to use the environment, you must pass base_url=None", + ) + + try: + base_url = ENVIRONMENTS[environment] + except KeyError as exc: + raise ValueError(f"Unknown environment: {environment}") from exc + elif base_url_env is not None: + base_url = base_url_env + else: + self._environment = environment = "production" + + try: + base_url = ENVIRONMENTS[environment] + except KeyError as exc: + raise ValueError(f"Unknown environment: {environment}") from exc + + super().__init__( + version=__version__, + base_url=base_url, + max_retries=max_retries, + timeout=timeout, + http_client=http_client, + custom_headers=default_headers, + custom_query=default_query, + _strict_response_validation=_strict_response_validation, + ) + + self.quote = quote.QuoteResource(self) + self.available = available.AvailableResource(self) + self.v2 = v2.V2Resource(self) + self.with_raw_response = BrapiWithRawResponse(self) + self.with_streaming_response = BrapiWithStreamedResponse(self) + + @property + @override + def qs(self) -> Querystring: + return Querystring(array_format="comma") + + @property + @override + def auth_headers(self) -> dict[str, str]: + api_key = self.api_key + return {"Authorization": f"Bearer {api_key}"} + + @property + @override + def default_headers(self) -> dict[str, str | Omit]: + return { + **super().default_headers, + "X-Stainless-Async": "false", + **self._custom_headers, + } + + def copy( + self, + *, + api_key: str | None = None, + environment: Literal["production", "environment_1"] | None = None, + base_url: str | httpx.URL | None = None, + timeout: float | Timeout | None | NotGiven = not_given, + http_client: httpx.Client | None = None, + max_retries: int | NotGiven = not_given, + default_headers: Mapping[str, str] | None = None, + set_default_headers: Mapping[str, str] | None = None, + default_query: Mapping[str, object] | None = None, + set_default_query: Mapping[str, object] | None = None, + _extra_kwargs: Mapping[str, Any] = {}, + ) -> Self: + """ + Create a new client instance re-using the same options given to the current client with optional overriding. + """ + if default_headers is not None and set_default_headers is not None: + raise ValueError("The `default_headers` and `set_default_headers` arguments are mutually exclusive") + + if default_query is not None and set_default_query is not None: + raise ValueError("The `default_query` and `set_default_query` arguments are mutually exclusive") + + headers = self._custom_headers + if default_headers is not None: + headers = {**headers, **default_headers} + elif set_default_headers is not None: + headers = set_default_headers + + params = self._custom_query + if default_query is not None: + params = {**params, **default_query} + elif set_default_query is not None: + params = set_default_query + + http_client = http_client or self._client + return self.__class__( + api_key=api_key or self.api_key, + base_url=base_url or self.base_url, + environment=environment or self._environment, + timeout=self.timeout if isinstance(timeout, NotGiven) else timeout, + http_client=http_client, + max_retries=max_retries if is_given(max_retries) else self.max_retries, + default_headers=headers, + default_query=params, + **_extra_kwargs, + ) + + # Alias for `copy` for nicer inline usage, e.g. + # client.with_options(timeout=10).foo.create(...) + with_options = copy + + @override + def _make_status_error( + self, + err_msg: str, + *, + body: object, + response: httpx.Response, + ) -> APIStatusError: + if response.status_code == 400: + return _exceptions.BadRequestError(err_msg, response=response, body=body) + + if response.status_code == 401: + return _exceptions.AuthenticationError(err_msg, response=response, body=body) + + if response.status_code == 403: + return _exceptions.PermissionDeniedError(err_msg, response=response, body=body) + + if response.status_code == 404: + return _exceptions.NotFoundError(err_msg, response=response, body=body) + + if response.status_code == 409: + return _exceptions.ConflictError(err_msg, response=response, body=body) + + if response.status_code == 422: + return _exceptions.UnprocessableEntityError(err_msg, response=response, body=body) + + if response.status_code == 429: + return _exceptions.RateLimitError(err_msg, response=response, body=body) + + if response.status_code >= 500: + return _exceptions.InternalServerError(err_msg, response=response, body=body) + return APIStatusError(err_msg, response=response, body=body) + + +class AsyncBrapi(AsyncAPIClient): + quote: quote.AsyncQuoteResource + available: available.AsyncAvailableResource + v2: v2.AsyncV2Resource + with_raw_response: AsyncBrapiWithRawResponse + with_streaming_response: AsyncBrapiWithStreamedResponse + + # client options + api_key: str + + _environment: Literal["production", "environment_1"] | NotGiven + + def __init__( + self, + *, + api_key: str | None = None, + environment: Literal["production", "environment_1"] | NotGiven = not_given, + base_url: str | httpx.URL | None | NotGiven = not_given, + timeout: float | Timeout | None | NotGiven = not_given, + max_retries: int = DEFAULT_MAX_RETRIES, + default_headers: Mapping[str, str] | None = None, + default_query: Mapping[str, object] | None = None, + # Configure a custom httpx client. + # We provide a `DefaultAsyncHttpxClient` class that you can pass to retain the default values we use for `limits`, `timeout` & `follow_redirects`. + # See the [httpx documentation](https://www.python-httpx.org/api/#asyncclient) for more details. + http_client: httpx.AsyncClient | None = None, + # Enable or disable schema validation for data returned by the API. + # When enabled an error APIResponseValidationError is raised + # if the API responds with invalid data for the expected schema. + # + # This parameter may be removed or changed in the future. + # If you rely on this feature, please open a GitHub issue + # outlining your use-case to help us decide if it should be + # part of our public interface in the future. + _strict_response_validation: bool = False, + ) -> None: + """Construct a new async AsyncBrapi client instance. + + This automatically infers the `api_key` argument from the `BRAPI_API_KEY` environment variable if it is not provided. + """ + if api_key is None: + api_key = os.environ.get("BRAPI_API_KEY") + if api_key is None: + raise BrapiError( + "The api_key client option must be set either by passing api_key to the client or by setting the BRAPI_API_KEY environment variable" + ) + self.api_key = api_key + + self._environment = environment + + base_url_env = os.environ.get("BRAPI_BASE_URL") + if is_given(base_url) and base_url is not None: + # cast required because mypy doesn't understand the type narrowing + base_url = cast("str | httpx.URL", base_url) # pyright: ignore[reportUnnecessaryCast] + elif is_given(environment): + if base_url_env and base_url is not None: + raise ValueError( + "Ambiguous URL; The `BRAPI_BASE_URL` env var and the `environment` argument are given. If you want to use the environment, you must pass base_url=None", + ) + + try: + base_url = ENVIRONMENTS[environment] + except KeyError as exc: + raise ValueError(f"Unknown environment: {environment}") from exc + elif base_url_env is not None: + base_url = base_url_env + else: + self._environment = environment = "production" + + try: + base_url = ENVIRONMENTS[environment] + except KeyError as exc: + raise ValueError(f"Unknown environment: {environment}") from exc + + super().__init__( + version=__version__, + base_url=base_url, + max_retries=max_retries, + timeout=timeout, + http_client=http_client, + custom_headers=default_headers, + custom_query=default_query, + _strict_response_validation=_strict_response_validation, + ) + + self.quote = quote.AsyncQuoteResource(self) + self.available = available.AsyncAvailableResource(self) + self.v2 = v2.AsyncV2Resource(self) + self.with_raw_response = AsyncBrapiWithRawResponse(self) + self.with_streaming_response = AsyncBrapiWithStreamedResponse(self) + + @property + @override + def qs(self) -> Querystring: + return Querystring(array_format="comma") + + @property + @override + def auth_headers(self) -> dict[str, str]: + api_key = self.api_key + return {"Authorization": f"Bearer {api_key}"} + + @property + @override + def default_headers(self) -> dict[str, str | Omit]: + return { + **super().default_headers, + "X-Stainless-Async": f"async:{get_async_library()}", + **self._custom_headers, + } + + def copy( + self, + *, + api_key: str | None = None, + environment: Literal["production", "environment_1"] | None = None, + base_url: str | httpx.URL | None = None, + timeout: float | Timeout | None | NotGiven = not_given, + http_client: httpx.AsyncClient | None = None, + max_retries: int | NotGiven = not_given, + default_headers: Mapping[str, str] | None = None, + set_default_headers: Mapping[str, str] | None = None, + default_query: Mapping[str, object] | None = None, + set_default_query: Mapping[str, object] | None = None, + _extra_kwargs: Mapping[str, Any] = {}, + ) -> Self: + """ + Create a new client instance re-using the same options given to the current client with optional overriding. + """ + if default_headers is not None and set_default_headers is not None: + raise ValueError("The `default_headers` and `set_default_headers` arguments are mutually exclusive") + + if default_query is not None and set_default_query is not None: + raise ValueError("The `default_query` and `set_default_query` arguments are mutually exclusive") + + headers = self._custom_headers + if default_headers is not None: + headers = {**headers, **default_headers} + elif set_default_headers is not None: + headers = set_default_headers + + params = self._custom_query + if default_query is not None: + params = {**params, **default_query} + elif set_default_query is not None: + params = set_default_query + + http_client = http_client or self._client + return self.__class__( + api_key=api_key or self.api_key, + base_url=base_url or self.base_url, + environment=environment or self._environment, + timeout=self.timeout if isinstance(timeout, NotGiven) else timeout, + http_client=http_client, + max_retries=max_retries if is_given(max_retries) else self.max_retries, + default_headers=headers, + default_query=params, + **_extra_kwargs, + ) + + # Alias for `copy` for nicer inline usage, e.g. + # client.with_options(timeout=10).foo.create(...) + with_options = copy + + @override + def _make_status_error( + self, + err_msg: str, + *, + body: object, + response: httpx.Response, + ) -> APIStatusError: + if response.status_code == 400: + return _exceptions.BadRequestError(err_msg, response=response, body=body) + + if response.status_code == 401: + return _exceptions.AuthenticationError(err_msg, response=response, body=body) + + if response.status_code == 403: + return _exceptions.PermissionDeniedError(err_msg, response=response, body=body) + + if response.status_code == 404: + return _exceptions.NotFoundError(err_msg, response=response, body=body) + + if response.status_code == 409: + return _exceptions.ConflictError(err_msg, response=response, body=body) + + if response.status_code == 422: + return _exceptions.UnprocessableEntityError(err_msg, response=response, body=body) + + if response.status_code == 429: + return _exceptions.RateLimitError(err_msg, response=response, body=body) + + if response.status_code >= 500: + return _exceptions.InternalServerError(err_msg, response=response, body=body) + return APIStatusError(err_msg, response=response, body=body) + + +class BrapiWithRawResponse: + def __init__(self, client: Brapi) -> None: + self.quote = quote.QuoteResourceWithRawResponse(client.quote) + self.available = available.AvailableResourceWithRawResponse(client.available) + self.v2 = v2.V2ResourceWithRawResponse(client.v2) + + +class AsyncBrapiWithRawResponse: + def __init__(self, client: AsyncBrapi) -> None: + self.quote = quote.AsyncQuoteResourceWithRawResponse(client.quote) + self.available = available.AsyncAvailableResourceWithRawResponse(client.available) + self.v2 = v2.AsyncV2ResourceWithRawResponse(client.v2) + + +class BrapiWithStreamedResponse: + def __init__(self, client: Brapi) -> None: + self.quote = quote.QuoteResourceWithStreamingResponse(client.quote) + self.available = available.AvailableResourceWithStreamingResponse(client.available) + self.v2 = v2.V2ResourceWithStreamingResponse(client.v2) + + +class AsyncBrapiWithStreamedResponse: + def __init__(self, client: AsyncBrapi) -> None: + self.quote = quote.AsyncQuoteResourceWithStreamingResponse(client.quote) + self.available = available.AsyncAvailableResourceWithStreamingResponse(client.available) + self.v2 = v2.AsyncV2ResourceWithStreamingResponse(client.v2) + + +Client = Brapi + +AsyncClient = AsyncBrapi diff --git a/src/brapi/_compat.py b/src/brapi/_compat.py new file mode 100644 index 0000000..bdef67f --- /dev/null +++ b/src/brapi/_compat.py @@ -0,0 +1,219 @@ +from __future__ import annotations + +from typing import TYPE_CHECKING, Any, Union, Generic, TypeVar, Callable, cast, overload +from datetime import date, datetime +from typing_extensions import Self, Literal + +import pydantic +from pydantic.fields import FieldInfo + +from ._types import IncEx, StrBytesIntFloat + +_T = TypeVar("_T") +_ModelT = TypeVar("_ModelT", bound=pydantic.BaseModel) + +# --------------- Pydantic v2, v3 compatibility --------------- + +# Pyright incorrectly reports some of our functions as overriding a method when they don't +# pyright: reportIncompatibleMethodOverride=false + +PYDANTIC_V1 = pydantic.VERSION.startswith("1.") + +if TYPE_CHECKING: + + def parse_date(value: date | StrBytesIntFloat) -> date: # noqa: ARG001 + ... + + def parse_datetime(value: Union[datetime, StrBytesIntFloat]) -> datetime: # noqa: ARG001 + ... + + def get_args(t: type[Any]) -> tuple[Any, ...]: # noqa: ARG001 + ... + + def is_union(tp: type[Any] | None) -> bool: # noqa: ARG001 + ... + + def get_origin(t: type[Any]) -> type[Any] | None: # noqa: ARG001 + ... + + def is_literal_type(type_: type[Any]) -> bool: # noqa: ARG001 + ... + + def is_typeddict(type_: type[Any]) -> bool: # noqa: ARG001 + ... + +else: + # v1 re-exports + if PYDANTIC_V1: + from pydantic.typing import ( + get_args as get_args, + is_union as is_union, + get_origin as get_origin, + is_typeddict as is_typeddict, + is_literal_type as is_literal_type, + ) + from pydantic.datetime_parse import parse_date as parse_date, parse_datetime as parse_datetime + else: + from ._utils import ( + get_args as get_args, + is_union as is_union, + get_origin as get_origin, + parse_date as parse_date, + is_typeddict as is_typeddict, + parse_datetime as parse_datetime, + is_literal_type as is_literal_type, + ) + + +# refactored config +if TYPE_CHECKING: + from pydantic import ConfigDict as ConfigDict +else: + if PYDANTIC_V1: + # TODO: provide an error message here? + ConfigDict = None + else: + from pydantic import ConfigDict as ConfigDict + + +# renamed methods / properties +def parse_obj(model: type[_ModelT], value: object) -> _ModelT: + if PYDANTIC_V1: + return cast(_ModelT, model.parse_obj(value)) # pyright: ignore[reportDeprecated, reportUnnecessaryCast] + else: + return model.model_validate(value) + + +def field_is_required(field: FieldInfo) -> bool: + if PYDANTIC_V1: + return field.required # type: ignore + return field.is_required() + + +def field_get_default(field: FieldInfo) -> Any: + value = field.get_default() + if PYDANTIC_V1: + return value + from pydantic_core import PydanticUndefined + + if value == PydanticUndefined: + return None + return value + + +def field_outer_type(field: FieldInfo) -> Any: + if PYDANTIC_V1: + return field.outer_type_ # type: ignore + return field.annotation + + +def get_model_config(model: type[pydantic.BaseModel]) -> Any: + if PYDANTIC_V1: + return model.__config__ # type: ignore + return model.model_config + + +def get_model_fields(model: type[pydantic.BaseModel]) -> dict[str, FieldInfo]: + if PYDANTIC_V1: + return model.__fields__ # type: ignore + return model.model_fields + + +def model_copy(model: _ModelT, *, deep: bool = False) -> _ModelT: + if PYDANTIC_V1: + return model.copy(deep=deep) # type: ignore + return model.model_copy(deep=deep) + + +def model_json(model: pydantic.BaseModel, *, indent: int | None = None) -> str: + if PYDANTIC_V1: + return model.json(indent=indent) # type: ignore + return model.model_dump_json(indent=indent) + + +def model_dump( + model: pydantic.BaseModel, + *, + exclude: IncEx | None = None, + exclude_unset: bool = False, + exclude_defaults: bool = False, + warnings: bool = True, + mode: Literal["json", "python"] = "python", +) -> dict[str, Any]: + if (not PYDANTIC_V1) or hasattr(model, "model_dump"): + return model.model_dump( + mode=mode, + exclude=exclude, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + # warnings are not supported in Pydantic v1 + warnings=True if PYDANTIC_V1 else warnings, + ) + return cast( + "dict[str, Any]", + model.dict( # pyright: ignore[reportDeprecated, reportUnnecessaryCast] + exclude=exclude, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + ), + ) + + +def model_parse(model: type[_ModelT], data: Any) -> _ModelT: + if PYDANTIC_V1: + return model.parse_obj(data) # pyright: ignore[reportDeprecated] + return model.model_validate(data) + + +# generic models +if TYPE_CHECKING: + + class GenericModel(pydantic.BaseModel): ... + +else: + if PYDANTIC_V1: + import pydantic.generics + + class GenericModel(pydantic.generics.GenericModel, pydantic.BaseModel): ... + else: + # there no longer needs to be a distinction in v2 but + # we still have to create our own subclass to avoid + # inconsistent MRO ordering errors + class GenericModel(pydantic.BaseModel): ... + + +# cached properties +if TYPE_CHECKING: + cached_property = property + + # we define a separate type (copied from typeshed) + # that represents that `cached_property` is `set`able + # at runtime, which differs from `@property`. + # + # this is a separate type as editors likely special case + # `@property` and we don't want to cause issues just to have + # more helpful internal types. + + class typed_cached_property(Generic[_T]): + func: Callable[[Any], _T] + attrname: str | None + + def __init__(self, func: Callable[[Any], _T]) -> None: ... + + @overload + def __get__(self, instance: None, owner: type[Any] | None = None) -> Self: ... + + @overload + def __get__(self, instance: object, owner: type[Any] | None = None) -> _T: ... + + def __get__(self, instance: object, owner: type[Any] | None = None) -> _T | Self: + raise NotImplementedError() + + def __set_name__(self, owner: type[Any], name: str) -> None: ... + + # __set__ is not defined at runtime, but @cached_property is designed to be settable + def __set__(self, instance: object, value: _T) -> None: ... +else: + from functools import cached_property as cached_property + + typed_cached_property = cached_property diff --git a/src/brapi/_constants.py b/src/brapi/_constants.py new file mode 100644 index 0000000..6ddf2c7 --- /dev/null +++ b/src/brapi/_constants.py @@ -0,0 +1,14 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +import httpx + +RAW_RESPONSE_HEADER = "X-Stainless-Raw-Response" +OVERRIDE_CAST_TO_HEADER = "____stainless_override_cast_to" + +# default timeout is 1 minute +DEFAULT_TIMEOUT = httpx.Timeout(timeout=60, connect=5.0) +DEFAULT_MAX_RETRIES = 2 +DEFAULT_CONNECTION_LIMITS = httpx.Limits(max_connections=100, max_keepalive_connections=20) + +INITIAL_RETRY_DELAY = 0.5 +MAX_RETRY_DELAY = 8.0 diff --git a/src/brapi/_exceptions.py b/src/brapi/_exceptions.py new file mode 100644 index 0000000..3aa815f --- /dev/null +++ b/src/brapi/_exceptions.py @@ -0,0 +1,108 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import Literal + +import httpx + +__all__ = [ + "BadRequestError", + "AuthenticationError", + "PermissionDeniedError", + "NotFoundError", + "ConflictError", + "UnprocessableEntityError", + "RateLimitError", + "InternalServerError", +] + + +class BrapiError(Exception): + pass + + +class APIError(BrapiError): + message: str + request: httpx.Request + + body: object | None + """The API response body. + + If the API responded with a valid JSON structure then this property will be the + decoded result. + + If it isn't a valid JSON structure then this will be the raw response. + + If there was no response associated with this error then it will be `None`. + """ + + def __init__(self, message: str, request: httpx.Request, *, body: object | None) -> None: # noqa: ARG002 + super().__init__(message) + self.request = request + self.message = message + self.body = body + + +class APIResponseValidationError(APIError): + response: httpx.Response + status_code: int + + def __init__(self, response: httpx.Response, body: object | None, *, message: str | None = None) -> None: + super().__init__(message or "Data returned by API invalid for expected schema.", response.request, body=body) + self.response = response + self.status_code = response.status_code + + +class APIStatusError(APIError): + """Raised when an API response has a status code of 4xx or 5xx.""" + + response: httpx.Response + status_code: int + + def __init__(self, message: str, *, response: httpx.Response, body: object | None) -> None: + super().__init__(message, response.request, body=body) + self.response = response + self.status_code = response.status_code + + +class APIConnectionError(APIError): + def __init__(self, *, message: str = "Connection error.", request: httpx.Request) -> None: + super().__init__(message, request, body=None) + + +class APITimeoutError(APIConnectionError): + def __init__(self, request: httpx.Request) -> None: + super().__init__(message="Request timed out.", request=request) + + +class BadRequestError(APIStatusError): + status_code: Literal[400] = 400 # pyright: ignore[reportIncompatibleVariableOverride] + + +class AuthenticationError(APIStatusError): + status_code: Literal[401] = 401 # pyright: ignore[reportIncompatibleVariableOverride] + + +class PermissionDeniedError(APIStatusError): + status_code: Literal[403] = 403 # pyright: ignore[reportIncompatibleVariableOverride] + + +class NotFoundError(APIStatusError): + status_code: Literal[404] = 404 # pyright: ignore[reportIncompatibleVariableOverride] + + +class ConflictError(APIStatusError): + status_code: Literal[409] = 409 # pyright: ignore[reportIncompatibleVariableOverride] + + +class UnprocessableEntityError(APIStatusError): + status_code: Literal[422] = 422 # pyright: ignore[reportIncompatibleVariableOverride] + + +class RateLimitError(APIStatusError): + status_code: Literal[429] = 429 # pyright: ignore[reportIncompatibleVariableOverride] + + +class InternalServerError(APIStatusError): + pass diff --git a/src/brapi/_files.py b/src/brapi/_files.py new file mode 100644 index 0000000..cc14c14 --- /dev/null +++ b/src/brapi/_files.py @@ -0,0 +1,123 @@ +from __future__ import annotations + +import io +import os +import pathlib +from typing import overload +from typing_extensions import TypeGuard + +import anyio + +from ._types import ( + FileTypes, + FileContent, + RequestFiles, + HttpxFileTypes, + Base64FileInput, + HttpxFileContent, + HttpxRequestFiles, +) +from ._utils import is_tuple_t, is_mapping_t, is_sequence_t + + +def is_base64_file_input(obj: object) -> TypeGuard[Base64FileInput]: + return isinstance(obj, io.IOBase) or isinstance(obj, os.PathLike) + + +def is_file_content(obj: object) -> TypeGuard[FileContent]: + return ( + isinstance(obj, bytes) or isinstance(obj, tuple) or isinstance(obj, io.IOBase) or isinstance(obj, os.PathLike) + ) + + +def assert_is_file_content(obj: object, *, key: str | None = None) -> None: + if not is_file_content(obj): + prefix = f"Expected entry at `{key}`" if key is not None else f"Expected file input `{obj!r}`" + raise RuntimeError( + f"{prefix} to be bytes, an io.IOBase instance, PathLike or a tuple but received {type(obj)} instead." + ) from None + + +@overload +def to_httpx_files(files: None) -> None: ... + + +@overload +def to_httpx_files(files: RequestFiles) -> HttpxRequestFiles: ... + + +def to_httpx_files(files: RequestFiles | None) -> HttpxRequestFiles | None: + if files is None: + return None + + if is_mapping_t(files): + files = {key: _transform_file(file) for key, file in files.items()} + elif is_sequence_t(files): + files = [(key, _transform_file(file)) for key, file in files] + else: + raise TypeError(f"Unexpected file type input {type(files)}, expected mapping or sequence") + + return files + + +def _transform_file(file: FileTypes) -> HttpxFileTypes: + if is_file_content(file): + if isinstance(file, os.PathLike): + path = pathlib.Path(file) + return (path.name, path.read_bytes()) + + return file + + if is_tuple_t(file): + return (file[0], read_file_content(file[1]), *file[2:]) + + raise TypeError(f"Expected file types input to be a FileContent type or to be a tuple") + + +def read_file_content(file: FileContent) -> HttpxFileContent: + if isinstance(file, os.PathLike): + return pathlib.Path(file).read_bytes() + return file + + +@overload +async def async_to_httpx_files(files: None) -> None: ... + + +@overload +async def async_to_httpx_files(files: RequestFiles) -> HttpxRequestFiles: ... + + +async def async_to_httpx_files(files: RequestFiles | None) -> HttpxRequestFiles | None: + if files is None: + return None + + if is_mapping_t(files): + files = {key: await _async_transform_file(file) for key, file in files.items()} + elif is_sequence_t(files): + files = [(key, await _async_transform_file(file)) for key, file in files] + else: + raise TypeError("Unexpected file type input {type(files)}, expected mapping or sequence") + + return files + + +async def _async_transform_file(file: FileTypes) -> HttpxFileTypes: + if is_file_content(file): + if isinstance(file, os.PathLike): + path = anyio.Path(file) + return (path.name, await path.read_bytes()) + + return file + + if is_tuple_t(file): + return (file[0], await async_read_file_content(file[1]), *file[2:]) + + raise TypeError(f"Expected file types input to be a FileContent type or to be a tuple") + + +async def async_read_file_content(file: FileContent) -> HttpxFileContent: + if isinstance(file, os.PathLike): + return await anyio.Path(file).read_bytes() + + return file diff --git a/src/brapi/_models.py b/src/brapi/_models.py new file mode 100644 index 0000000..6a3cd1d --- /dev/null +++ b/src/brapi/_models.py @@ -0,0 +1,835 @@ +from __future__ import annotations + +import os +import inspect +from typing import TYPE_CHECKING, Any, Type, Union, Generic, TypeVar, Callable, Optional, cast +from datetime import date, datetime +from typing_extensions import ( + List, + Unpack, + Literal, + ClassVar, + Protocol, + Required, + ParamSpec, + TypedDict, + TypeGuard, + final, + override, + runtime_checkable, +) + +import pydantic +from pydantic.fields import FieldInfo + +from ._types import ( + Body, + IncEx, + Query, + ModelT, + Headers, + Timeout, + NotGiven, + AnyMapping, + HttpxRequestFiles, +) +from ._utils import ( + PropertyInfo, + is_list, + is_given, + json_safe, + lru_cache, + is_mapping, + parse_date, + coerce_boolean, + parse_datetime, + strip_not_given, + extract_type_arg, + is_annotated_type, + is_type_alias_type, + strip_annotated_type, +) +from ._compat import ( + PYDANTIC_V1, + ConfigDict, + GenericModel as BaseGenericModel, + get_args, + is_union, + parse_obj, + get_origin, + is_literal_type, + get_model_config, + get_model_fields, + field_get_default, +) +from ._constants import RAW_RESPONSE_HEADER + +if TYPE_CHECKING: + from pydantic_core.core_schema import ModelField, ModelSchema, LiteralSchema, ModelFieldsSchema + +__all__ = ["BaseModel", "GenericModel"] + +_T = TypeVar("_T") +_BaseModelT = TypeVar("_BaseModelT", bound="BaseModel") + +P = ParamSpec("P") + + +@runtime_checkable +class _ConfigProtocol(Protocol): + allow_population_by_field_name: bool + + +class BaseModel(pydantic.BaseModel): + if PYDANTIC_V1: + + @property + @override + def model_fields_set(self) -> set[str]: + # a forwards-compat shim for pydantic v2 + return self.__fields_set__ # type: ignore + + class Config(pydantic.BaseConfig): # pyright: ignore[reportDeprecated] + extra: Any = pydantic.Extra.allow # type: ignore + else: + model_config: ClassVar[ConfigDict] = ConfigDict( + extra="allow", defer_build=coerce_boolean(os.environ.get("DEFER_PYDANTIC_BUILD", "true")) + ) + + def to_dict( + self, + *, + mode: Literal["json", "python"] = "python", + use_api_names: bool = True, + exclude_unset: bool = True, + exclude_defaults: bool = False, + exclude_none: bool = False, + warnings: bool = True, + ) -> dict[str, object]: + """Recursively generate a dictionary representation of the model, optionally specifying which fields to include or exclude. + + By default, fields that were not set by the API will not be included, + and keys will match the API response, *not* the property names from the model. + + For example, if the API responds with `"fooBar": true` but we've defined a `foo_bar: bool` property, + the output will use the `"fooBar"` key (unless `use_api_names=False` is passed). + + Args: + mode: + If mode is 'json', the dictionary will only contain JSON serializable types. e.g. `datetime` will be turned into a string, `"2024-3-22T18:11:19.117000Z"`. + If mode is 'python', the dictionary may contain any Python objects. e.g. `datetime(2024, 3, 22)` + + use_api_names: Whether to use the key that the API responded with or the property name. Defaults to `True`. + exclude_unset: Whether to exclude fields that have not been explicitly set. + exclude_defaults: Whether to exclude fields that are set to their default value from the output. + exclude_none: Whether to exclude fields that have a value of `None` from the output. + warnings: Whether to log warnings when invalid fields are encountered. This is only supported in Pydantic v2. + """ + return self.model_dump( + mode=mode, + by_alias=use_api_names, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + exclude_none=exclude_none, + warnings=warnings, + ) + + def to_json( + self, + *, + indent: int | None = 2, + use_api_names: bool = True, + exclude_unset: bool = True, + exclude_defaults: bool = False, + exclude_none: bool = False, + warnings: bool = True, + ) -> str: + """Generates a JSON string representing this model as it would be received from or sent to the API (but with indentation). + + By default, fields that were not set by the API will not be included, + and keys will match the API response, *not* the property names from the model. + + For example, if the API responds with `"fooBar": true` but we've defined a `foo_bar: bool` property, + the output will use the `"fooBar"` key (unless `use_api_names=False` is passed). + + Args: + indent: Indentation to use in the JSON output. If `None` is passed, the output will be compact. Defaults to `2` + use_api_names: Whether to use the key that the API responded with or the property name. Defaults to `True`. + exclude_unset: Whether to exclude fields that have not been explicitly set. + exclude_defaults: Whether to exclude fields that have the default value. + exclude_none: Whether to exclude fields that have a value of `None`. + warnings: Whether to show any warnings that occurred during serialization. This is only supported in Pydantic v2. + """ + return self.model_dump_json( + indent=indent, + by_alias=use_api_names, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + exclude_none=exclude_none, + warnings=warnings, + ) + + @override + def __str__(self) -> str: + # mypy complains about an invalid self arg + return f"{self.__repr_name__()}({self.__repr_str__(', ')})" # type: ignore[misc] + + # Override the 'construct' method in a way that supports recursive parsing without validation. + # Based on https://github.com/samuelcolvin/pydantic/issues/1168#issuecomment-817742836. + @classmethod + @override + def construct( # pyright: ignore[reportIncompatibleMethodOverride] + __cls: Type[ModelT], + _fields_set: set[str] | None = None, + **values: object, + ) -> ModelT: + m = __cls.__new__(__cls) + fields_values: dict[str, object] = {} + + config = get_model_config(__cls) + populate_by_name = ( + config.allow_population_by_field_name + if isinstance(config, _ConfigProtocol) + else config.get("populate_by_name") + ) + + if _fields_set is None: + _fields_set = set() + + model_fields = get_model_fields(__cls) + for name, field in model_fields.items(): + key = field.alias + if key is None or (key not in values and populate_by_name): + key = name + + if key in values: + fields_values[name] = _construct_field(value=values[key], field=field, key=key) + _fields_set.add(name) + else: + fields_values[name] = field_get_default(field) + + extra_field_type = _get_extra_fields_type(__cls) + + _extra = {} + for key, value in values.items(): + if key not in model_fields: + parsed = construct_type(value=value, type_=extra_field_type) if extra_field_type is not None else value + + if PYDANTIC_V1: + _fields_set.add(key) + fields_values[key] = parsed + else: + _extra[key] = parsed + + object.__setattr__(m, "__dict__", fields_values) + + if PYDANTIC_V1: + # init_private_attributes() does not exist in v2 + m._init_private_attributes() # type: ignore + + # copied from Pydantic v1's `construct()` method + object.__setattr__(m, "__fields_set__", _fields_set) + else: + # these properties are copied from Pydantic's `model_construct()` method + object.__setattr__(m, "__pydantic_private__", None) + object.__setattr__(m, "__pydantic_extra__", _extra) + object.__setattr__(m, "__pydantic_fields_set__", _fields_set) + + return m + + if not TYPE_CHECKING: + # type checkers incorrectly complain about this assignment + # because the type signatures are technically different + # although not in practice + model_construct = construct + + if PYDANTIC_V1: + # we define aliases for some of the new pydantic v2 methods so + # that we can just document these methods without having to specify + # a specific pydantic version as some users may not know which + # pydantic version they are currently using + + @override + def model_dump( + self, + *, + mode: Literal["json", "python"] | str = "python", + include: IncEx | None = None, + exclude: IncEx | None = None, + by_alias: bool | None = None, + exclude_unset: bool = False, + exclude_defaults: bool = False, + exclude_none: bool = False, + round_trip: bool = False, + warnings: bool | Literal["none", "warn", "error"] = True, + context: dict[str, Any] | None = None, + serialize_as_any: bool = False, + fallback: Callable[[Any], Any] | None = None, + ) -> dict[str, Any]: + """Usage docs: https://docs.pydantic.dev/2.4/concepts/serialization/#modelmodel_dump + + Generate a dictionary representation of the model, optionally specifying which fields to include or exclude. + + Args: + mode: The mode in which `to_python` should run. + If mode is 'json', the dictionary will only contain JSON serializable types. + If mode is 'python', the dictionary may contain any Python objects. + include: A list of fields to include in the output. + exclude: A list of fields to exclude from the output. + by_alias: Whether to use the field's alias in the dictionary key if defined. + exclude_unset: Whether to exclude fields that are unset or None from the output. + exclude_defaults: Whether to exclude fields that are set to their default value from the output. + exclude_none: Whether to exclude fields that have a value of `None` from the output. + round_trip: Whether to enable serialization and deserialization round-trip support. + warnings: Whether to log warnings when invalid fields are encountered. + + Returns: + A dictionary representation of the model. + """ + if mode not in {"json", "python"}: + raise ValueError("mode must be either 'json' or 'python'") + if round_trip != False: + raise ValueError("round_trip is only supported in Pydantic v2") + if warnings != True: + raise ValueError("warnings is only supported in Pydantic v2") + if context is not None: + raise ValueError("context is only supported in Pydantic v2") + if serialize_as_any != False: + raise ValueError("serialize_as_any is only supported in Pydantic v2") + if fallback is not None: + raise ValueError("fallback is only supported in Pydantic v2") + dumped = super().dict( # pyright: ignore[reportDeprecated] + include=include, + exclude=exclude, + by_alias=by_alias if by_alias is not None else False, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + exclude_none=exclude_none, + ) + + return cast("dict[str, Any]", json_safe(dumped)) if mode == "json" else dumped + + @override + def model_dump_json( + self, + *, + indent: int | None = None, + include: IncEx | None = None, + exclude: IncEx | None = None, + by_alias: bool | None = None, + exclude_unset: bool = False, + exclude_defaults: bool = False, + exclude_none: bool = False, + round_trip: bool = False, + warnings: bool | Literal["none", "warn", "error"] = True, + context: dict[str, Any] | None = None, + fallback: Callable[[Any], Any] | None = None, + serialize_as_any: bool = False, + ) -> str: + """Usage docs: https://docs.pydantic.dev/2.4/concepts/serialization/#modelmodel_dump_json + + Generates a JSON representation of the model using Pydantic's `to_json` method. + + Args: + indent: Indentation to use in the JSON output. If None is passed, the output will be compact. + include: Field(s) to include in the JSON output. Can take either a string or set of strings. + exclude: Field(s) to exclude from the JSON output. Can take either a string or set of strings. + by_alias: Whether to serialize using field aliases. + exclude_unset: Whether to exclude fields that have not been explicitly set. + exclude_defaults: Whether to exclude fields that have the default value. + exclude_none: Whether to exclude fields that have a value of `None`. + round_trip: Whether to use serialization/deserialization between JSON and class instance. + warnings: Whether to show any warnings that occurred during serialization. + + Returns: + A JSON string representation of the model. + """ + if round_trip != False: + raise ValueError("round_trip is only supported in Pydantic v2") + if warnings != True: + raise ValueError("warnings is only supported in Pydantic v2") + if context is not None: + raise ValueError("context is only supported in Pydantic v2") + if serialize_as_any != False: + raise ValueError("serialize_as_any is only supported in Pydantic v2") + if fallback is not None: + raise ValueError("fallback is only supported in Pydantic v2") + return super().json( # type: ignore[reportDeprecated] + indent=indent, + include=include, + exclude=exclude, + by_alias=by_alias if by_alias is not None else False, + exclude_unset=exclude_unset, + exclude_defaults=exclude_defaults, + exclude_none=exclude_none, + ) + + +def _construct_field(value: object, field: FieldInfo, key: str) -> object: + if value is None: + return field_get_default(field) + + if PYDANTIC_V1: + type_ = cast(type, field.outer_type_) # type: ignore + else: + type_ = field.annotation # type: ignore + + if type_ is None: + raise RuntimeError(f"Unexpected field type is None for {key}") + + return construct_type(value=value, type_=type_, metadata=getattr(field, "metadata", None)) + + +def _get_extra_fields_type(cls: type[pydantic.BaseModel]) -> type | None: + if PYDANTIC_V1: + # TODO + return None + + schema = cls.__pydantic_core_schema__ + if schema["type"] == "model": + fields = schema["schema"] + if fields["type"] == "model-fields": + extras = fields.get("extras_schema") + if extras and "cls" in extras: + # mypy can't narrow the type + return extras["cls"] # type: ignore[no-any-return] + + return None + + +def is_basemodel(type_: type) -> bool: + """Returns whether or not the given type is either a `BaseModel` or a union of `BaseModel`""" + if is_union(type_): + for variant in get_args(type_): + if is_basemodel(variant): + return True + + return False + + return is_basemodel_type(type_) + + +def is_basemodel_type(type_: type) -> TypeGuard[type[BaseModel] | type[GenericModel]]: + origin = get_origin(type_) or type_ + if not inspect.isclass(origin): + return False + return issubclass(origin, BaseModel) or issubclass(origin, GenericModel) + + +def build( + base_model_cls: Callable[P, _BaseModelT], + *args: P.args, + **kwargs: P.kwargs, +) -> _BaseModelT: + """Construct a BaseModel class without validation. + + This is useful for cases where you need to instantiate a `BaseModel` + from an API response as this provides type-safe params which isn't supported + by helpers like `construct_type()`. + + ```py + build(MyModel, my_field_a="foo", my_field_b=123) + ``` + """ + if args: + raise TypeError( + "Received positional arguments which are not supported; Keyword arguments must be used instead", + ) + + return cast(_BaseModelT, construct_type(type_=base_model_cls, value=kwargs)) + + +def construct_type_unchecked(*, value: object, type_: type[_T]) -> _T: + """Loose coercion to the expected type with construction of nested values. + + Note: the returned value from this function is not guaranteed to match the + given type. + """ + return cast(_T, construct_type(value=value, type_=type_)) + + +def construct_type(*, value: object, type_: object, metadata: Optional[List[Any]] = None) -> object: + """Loose coercion to the expected type with construction of nested values. + + If the given value does not match the expected type then it is returned as-is. + """ + + # store a reference to the original type we were given before we extract any inner + # types so that we can properly resolve forward references in `TypeAliasType` annotations + original_type = None + + # we allow `object` as the input type because otherwise, passing things like + # `Literal['value']` will be reported as a type error by type checkers + type_ = cast("type[object]", type_) + if is_type_alias_type(type_): + original_type = type_ # type: ignore[unreachable] + type_ = type_.__value__ # type: ignore[unreachable] + + # unwrap `Annotated[T, ...]` -> `T` + if metadata is not None and len(metadata) > 0: + meta: tuple[Any, ...] = tuple(metadata) + elif is_annotated_type(type_): + meta = get_args(type_)[1:] + type_ = extract_type_arg(type_, 0) + else: + meta = tuple() + + # we need to use the origin class for any types that are subscripted generics + # e.g. Dict[str, object] + origin = get_origin(type_) or type_ + args = get_args(type_) + + if is_union(origin): + try: + return validate_type(type_=cast("type[object]", original_type or type_), value=value) + except Exception: + pass + + # if the type is a discriminated union then we want to construct the right variant + # in the union, even if the data doesn't match exactly, otherwise we'd break code + # that relies on the constructed class types, e.g. + # + # class FooType: + # kind: Literal['foo'] + # value: str + # + # class BarType: + # kind: Literal['bar'] + # value: int + # + # without this block, if the data we get is something like `{'kind': 'bar', 'value': 'foo'}` then + # we'd end up constructing `FooType` when it should be `BarType`. + discriminator = _build_discriminated_union_meta(union=type_, meta_annotations=meta) + if discriminator and is_mapping(value): + variant_value = value.get(discriminator.field_alias_from or discriminator.field_name) + if variant_value and isinstance(variant_value, str): + variant_type = discriminator.mapping.get(variant_value) + if variant_type: + return construct_type(type_=variant_type, value=value) + + # if the data is not valid, use the first variant that doesn't fail while deserializing + for variant in args: + try: + return construct_type(value=value, type_=variant) + except Exception: + continue + + raise RuntimeError(f"Could not convert data into a valid instance of {type_}") + + if origin == dict: + if not is_mapping(value): + return value + + _, items_type = get_args(type_) # Dict[_, items_type] + return {key: construct_type(value=item, type_=items_type) for key, item in value.items()} + + if ( + not is_literal_type(type_) + and inspect.isclass(origin) + and (issubclass(origin, BaseModel) or issubclass(origin, GenericModel)) + ): + if is_list(value): + return [cast(Any, type_).construct(**entry) if is_mapping(entry) else entry for entry in value] + + if is_mapping(value): + if issubclass(type_, BaseModel): + return type_.construct(**value) # type: ignore[arg-type] + + return cast(Any, type_).construct(**value) + + if origin == list: + if not is_list(value): + return value + + inner_type = args[0] # List[inner_type] + return [construct_type(value=entry, type_=inner_type) for entry in value] + + if origin == float: + if isinstance(value, int): + coerced = float(value) + if coerced != value: + return value + return coerced + + return value + + if type_ == datetime: + try: + return parse_datetime(value) # type: ignore + except Exception: + return value + + if type_ == date: + try: + return parse_date(value) # type: ignore + except Exception: + return value + + return value + + +@runtime_checkable +class CachedDiscriminatorType(Protocol): + __discriminator__: DiscriminatorDetails + + +class DiscriminatorDetails: + field_name: str + """The name of the discriminator field in the variant class, e.g. + + ```py + class Foo(BaseModel): + type: Literal['foo'] + ``` + + Will result in field_name='type' + """ + + field_alias_from: str | None + """The name of the discriminator field in the API response, e.g. + + ```py + class Foo(BaseModel): + type: Literal['foo'] = Field(alias='type_from_api') + ``` + + Will result in field_alias_from='type_from_api' + """ + + mapping: dict[str, type] + """Mapping of discriminator value to variant type, e.g. + + {'foo': FooVariant, 'bar': BarVariant} + """ + + def __init__( + self, + *, + mapping: dict[str, type], + discriminator_field: str, + discriminator_alias: str | None, + ) -> None: + self.mapping = mapping + self.field_name = discriminator_field + self.field_alias_from = discriminator_alias + + +def _build_discriminated_union_meta(*, union: type, meta_annotations: tuple[Any, ...]) -> DiscriminatorDetails | None: + if isinstance(union, CachedDiscriminatorType): + return union.__discriminator__ + + discriminator_field_name: str | None = None + + for annotation in meta_annotations: + if isinstance(annotation, PropertyInfo) and annotation.discriminator is not None: + discriminator_field_name = annotation.discriminator + break + + if not discriminator_field_name: + return None + + mapping: dict[str, type] = {} + discriminator_alias: str | None = None + + for variant in get_args(union): + variant = strip_annotated_type(variant) + if is_basemodel_type(variant): + if PYDANTIC_V1: + field_info = cast("dict[str, FieldInfo]", variant.__fields__).get(discriminator_field_name) # pyright: ignore[reportDeprecated, reportUnnecessaryCast] + if not field_info: + continue + + # Note: if one variant defines an alias then they all should + discriminator_alias = field_info.alias + + if (annotation := getattr(field_info, "annotation", None)) and is_literal_type(annotation): + for entry in get_args(annotation): + if isinstance(entry, str): + mapping[entry] = variant + else: + field = _extract_field_schema_pv2(variant, discriminator_field_name) + if not field: + continue + + # Note: if one variant defines an alias then they all should + discriminator_alias = field.get("serialization_alias") + + field_schema = field["schema"] + + if field_schema["type"] == "literal": + for entry in cast("LiteralSchema", field_schema)["expected"]: + if isinstance(entry, str): + mapping[entry] = variant + + if not mapping: + return None + + details = DiscriminatorDetails( + mapping=mapping, + discriminator_field=discriminator_field_name, + discriminator_alias=discriminator_alias, + ) + cast(CachedDiscriminatorType, union).__discriminator__ = details + return details + + +def _extract_field_schema_pv2(model: type[BaseModel], field_name: str) -> ModelField | None: + schema = model.__pydantic_core_schema__ + if schema["type"] == "definitions": + schema = schema["schema"] + + if schema["type"] != "model": + return None + + schema = cast("ModelSchema", schema) + fields_schema = schema["schema"] + if fields_schema["type"] != "model-fields": + return None + + fields_schema = cast("ModelFieldsSchema", fields_schema) + field = fields_schema["fields"].get(field_name) + if not field: + return None + + return cast("ModelField", field) # pyright: ignore[reportUnnecessaryCast] + + +def validate_type(*, type_: type[_T], value: object) -> _T: + """Strict validation that the given value matches the expected type""" + if inspect.isclass(type_) and issubclass(type_, pydantic.BaseModel): + return cast(_T, parse_obj(type_, value)) + + return cast(_T, _validate_non_model_type(type_=type_, value=value)) + + +def set_pydantic_config(typ: Any, config: pydantic.ConfigDict) -> None: + """Add a pydantic config for the given type. + + Note: this is a no-op on Pydantic v1. + """ + setattr(typ, "__pydantic_config__", config) # noqa: B010 + + +# our use of subclassing here causes weirdness for type checkers, +# so we just pretend that we don't subclass +if TYPE_CHECKING: + GenericModel = BaseModel +else: + + class GenericModel(BaseGenericModel, BaseModel): + pass + + +if not PYDANTIC_V1: + from pydantic import TypeAdapter as _TypeAdapter + + _CachedTypeAdapter = cast("TypeAdapter[object]", lru_cache(maxsize=None)(_TypeAdapter)) + + if TYPE_CHECKING: + from pydantic import TypeAdapter + else: + TypeAdapter = _CachedTypeAdapter + + def _validate_non_model_type(*, type_: type[_T], value: object) -> _T: + return TypeAdapter(type_).validate_python(value) + +elif not TYPE_CHECKING: # TODO: condition is weird + + class RootModel(GenericModel, Generic[_T]): + """Used as a placeholder to easily convert runtime types to a Pydantic format + to provide validation. + + For example: + ```py + validated = RootModel[int](__root__="5").__root__ + # validated: 5 + ``` + """ + + __root__: _T + + def _validate_non_model_type(*, type_: type[_T], value: object) -> _T: + model = _create_pydantic_model(type_).validate(value) + return cast(_T, model.__root__) + + def _create_pydantic_model(type_: _T) -> Type[RootModel[_T]]: + return RootModel[type_] # type: ignore + + +class FinalRequestOptionsInput(TypedDict, total=False): + method: Required[str] + url: Required[str] + params: Query + headers: Headers + max_retries: int + timeout: float | Timeout | None + files: HttpxRequestFiles | None + idempotency_key: str + json_data: Body + extra_json: AnyMapping + follow_redirects: bool + + +@final +class FinalRequestOptions(pydantic.BaseModel): + method: str + url: str + params: Query = {} + headers: Union[Headers, NotGiven] = NotGiven() + max_retries: Union[int, NotGiven] = NotGiven() + timeout: Union[float, Timeout, None, NotGiven] = NotGiven() + files: Union[HttpxRequestFiles, None] = None + idempotency_key: Union[str, None] = None + post_parser: Union[Callable[[Any], Any], NotGiven] = NotGiven() + follow_redirects: Union[bool, None] = None + + # It should be noted that we cannot use `json` here as that would override + # a BaseModel method in an incompatible fashion. + json_data: Union[Body, None] = None + extra_json: Union[AnyMapping, None] = None + + if PYDANTIC_V1: + + class Config(pydantic.BaseConfig): # pyright: ignore[reportDeprecated] + arbitrary_types_allowed: bool = True + else: + model_config: ClassVar[ConfigDict] = ConfigDict(arbitrary_types_allowed=True) + + def get_max_retries(self, max_retries: int) -> int: + if isinstance(self.max_retries, NotGiven): + return max_retries + return self.max_retries + + def _strip_raw_response_header(self) -> None: + if not is_given(self.headers): + return + + if self.headers.get(RAW_RESPONSE_HEADER): + self.headers = {**self.headers} + self.headers.pop(RAW_RESPONSE_HEADER) + + # override the `construct` method so that we can run custom transformations. + # this is necessary as we don't want to do any actual runtime type checking + # (which means we can't use validators) but we do want to ensure that `NotGiven` + # values are not present + # + # type ignore required because we're adding explicit types to `**values` + @classmethod + def construct( # type: ignore + cls, + _fields_set: set[str] | None = None, + **values: Unpack[FinalRequestOptionsInput], + ) -> FinalRequestOptions: + kwargs: dict[str, Any] = { + # we unconditionally call `strip_not_given` on any value + # as it will just ignore any non-mapping types + key: strip_not_given(value) + for key, value in values.items() + } + if PYDANTIC_V1: + return cast(FinalRequestOptions, super().construct(_fields_set, **kwargs)) # pyright: ignore[reportDeprecated] + return super().model_construct(_fields_set, **kwargs) + + if not TYPE_CHECKING: + # type checkers incorrectly complain about this assignment + model_construct = construct diff --git a/src/brapi/_qs.py b/src/brapi/_qs.py new file mode 100644 index 0000000..ada6fd3 --- /dev/null +++ b/src/brapi/_qs.py @@ -0,0 +1,150 @@ +from __future__ import annotations + +from typing import Any, List, Tuple, Union, Mapping, TypeVar +from urllib.parse import parse_qs, urlencode +from typing_extensions import Literal, get_args + +from ._types import NotGiven, not_given +from ._utils import flatten + +_T = TypeVar("_T") + + +ArrayFormat = Literal["comma", "repeat", "indices", "brackets"] +NestedFormat = Literal["dots", "brackets"] + +PrimitiveData = Union[str, int, float, bool, None] +# this should be Data = Union[PrimitiveData, "List[Data]", "Tuple[Data]", "Mapping[str, Data]"] +# https://github.com/microsoft/pyright/issues/3555 +Data = Union[PrimitiveData, List[Any], Tuple[Any], "Mapping[str, Any]"] +Params = Mapping[str, Data] + + +class Querystring: + array_format: ArrayFormat + nested_format: NestedFormat + + def __init__( + self, + *, + array_format: ArrayFormat = "repeat", + nested_format: NestedFormat = "brackets", + ) -> None: + self.array_format = array_format + self.nested_format = nested_format + + def parse(self, query: str) -> Mapping[str, object]: + # Note: custom format syntax is not supported yet + return parse_qs(query) + + def stringify( + self, + params: Params, + *, + array_format: ArrayFormat | NotGiven = not_given, + nested_format: NestedFormat | NotGiven = not_given, + ) -> str: + return urlencode( + self.stringify_items( + params, + array_format=array_format, + nested_format=nested_format, + ) + ) + + def stringify_items( + self, + params: Params, + *, + array_format: ArrayFormat | NotGiven = not_given, + nested_format: NestedFormat | NotGiven = not_given, + ) -> list[tuple[str, str]]: + opts = Options( + qs=self, + array_format=array_format, + nested_format=nested_format, + ) + return flatten([self._stringify_item(key, value, opts) for key, value in params.items()]) + + def _stringify_item( + self, + key: str, + value: Data, + opts: Options, + ) -> list[tuple[str, str]]: + if isinstance(value, Mapping): + items: list[tuple[str, str]] = [] + nested_format = opts.nested_format + for subkey, subvalue in value.items(): + items.extend( + self._stringify_item( + # TODO: error if unknown format + f"{key}.{subkey}" if nested_format == "dots" else f"{key}[{subkey}]", + subvalue, + opts, + ) + ) + return items + + if isinstance(value, (list, tuple)): + array_format = opts.array_format + if array_format == "comma": + return [ + ( + key, + ",".join(self._primitive_value_to_str(item) for item in value if item is not None), + ), + ] + elif array_format == "repeat": + items = [] + for item in value: + items.extend(self._stringify_item(key, item, opts)) + return items + elif array_format == "indices": + raise NotImplementedError("The array indices format is not supported yet") + elif array_format == "brackets": + items = [] + key = key + "[]" + for item in value: + items.extend(self._stringify_item(key, item, opts)) + return items + else: + raise NotImplementedError( + f"Unknown array_format value: {array_format}, choose from {', '.join(get_args(ArrayFormat))}" + ) + + serialised = self._primitive_value_to_str(value) + if not serialised: + return [] + return [(key, serialised)] + + def _primitive_value_to_str(self, value: PrimitiveData) -> str: + # copied from httpx + if value is True: + return "true" + elif value is False: + return "false" + elif value is None: + return "" + return str(value) + + +_qs = Querystring() +parse = _qs.parse +stringify = _qs.stringify +stringify_items = _qs.stringify_items + + +class Options: + array_format: ArrayFormat + nested_format: NestedFormat + + def __init__( + self, + qs: Querystring = _qs, + *, + array_format: ArrayFormat | NotGiven = not_given, + nested_format: NestedFormat | NotGiven = not_given, + ) -> None: + self.array_format = qs.array_format if isinstance(array_format, NotGiven) else array_format + self.nested_format = qs.nested_format if isinstance(nested_format, NotGiven) else nested_format diff --git a/src/brapi/_resource.py b/src/brapi/_resource.py new file mode 100644 index 0000000..57a841c --- /dev/null +++ b/src/brapi/_resource.py @@ -0,0 +1,43 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import time +from typing import TYPE_CHECKING + +import anyio + +if TYPE_CHECKING: + from ._client import Brapi, AsyncBrapi + + +class SyncAPIResource: + _client: Brapi + + def __init__(self, client: Brapi) -> None: + self._client = client + self._get = client.get + self._post = client.post + self._patch = client.patch + self._put = client.put + self._delete = client.delete + self._get_api_list = client.get_api_list + + def _sleep(self, seconds: float) -> None: + time.sleep(seconds) + + +class AsyncAPIResource: + _client: AsyncBrapi + + def __init__(self, client: AsyncBrapi) -> None: + self._client = client + self._get = client.get + self._post = client.post + self._patch = client.patch + self._put = client.put + self._delete = client.delete + self._get_api_list = client.get_api_list + + async def _sleep(self, seconds: float) -> None: + await anyio.sleep(seconds) diff --git a/src/brapi/_response.py b/src/brapi/_response.py new file mode 100644 index 0000000..0f7b71f --- /dev/null +++ b/src/brapi/_response.py @@ -0,0 +1,830 @@ +from __future__ import annotations + +import os +import inspect +import logging +import datetime +import functools +from types import TracebackType +from typing import ( + TYPE_CHECKING, + Any, + Union, + Generic, + TypeVar, + Callable, + Iterator, + AsyncIterator, + cast, + overload, +) +from typing_extensions import Awaitable, ParamSpec, override, get_origin + +import anyio +import httpx +import pydantic + +from ._types import NoneType +from ._utils import is_given, extract_type_arg, is_annotated_type, is_type_alias_type, extract_type_var_from_base +from ._models import BaseModel, is_basemodel +from ._constants import RAW_RESPONSE_HEADER, OVERRIDE_CAST_TO_HEADER +from ._streaming import Stream, AsyncStream, is_stream_class_type, extract_stream_chunk_type +from ._exceptions import BrapiError, APIResponseValidationError + +if TYPE_CHECKING: + from ._models import FinalRequestOptions + from ._base_client import BaseClient + + +P = ParamSpec("P") +R = TypeVar("R") +_T = TypeVar("_T") +_APIResponseT = TypeVar("_APIResponseT", bound="APIResponse[Any]") +_AsyncAPIResponseT = TypeVar("_AsyncAPIResponseT", bound="AsyncAPIResponse[Any]") + +log: logging.Logger = logging.getLogger(__name__) + + +class BaseAPIResponse(Generic[R]): + _cast_to: type[R] + _client: BaseClient[Any, Any] + _parsed_by_type: dict[type[Any], Any] + _is_sse_stream: bool + _stream_cls: type[Stream[Any]] | type[AsyncStream[Any]] | None + _options: FinalRequestOptions + + http_response: httpx.Response + + retries_taken: int + """The number of retries made. If no retries happened this will be `0`""" + + def __init__( + self, + *, + raw: httpx.Response, + cast_to: type[R], + client: BaseClient[Any, Any], + stream: bool, + stream_cls: type[Stream[Any]] | type[AsyncStream[Any]] | None, + options: FinalRequestOptions, + retries_taken: int = 0, + ) -> None: + self._cast_to = cast_to + self._client = client + self._parsed_by_type = {} + self._is_sse_stream = stream + self._stream_cls = stream_cls + self._options = options + self.http_response = raw + self.retries_taken = retries_taken + + @property + def headers(self) -> httpx.Headers: + return self.http_response.headers + + @property + def http_request(self) -> httpx.Request: + """Returns the httpx Request instance associated with the current response.""" + return self.http_response.request + + @property + def status_code(self) -> int: + return self.http_response.status_code + + @property + def url(self) -> httpx.URL: + """Returns the URL for which the request was made.""" + return self.http_response.url + + @property + def method(self) -> str: + return self.http_request.method + + @property + def http_version(self) -> str: + return self.http_response.http_version + + @property + def elapsed(self) -> datetime.timedelta: + """The time taken for the complete request/response cycle to complete.""" + return self.http_response.elapsed + + @property + def is_closed(self) -> bool: + """Whether or not the response body has been closed. + + If this is False then there is response data that has not been read yet. + You must either fully consume the response body or call `.close()` + before discarding the response to prevent resource leaks. + """ + return self.http_response.is_closed + + @override + def __repr__(self) -> str: + return ( + f"<{self.__class__.__name__} [{self.status_code} {self.http_response.reason_phrase}] type={self._cast_to}>" + ) + + def _parse(self, *, to: type[_T] | None = None) -> R | _T: + cast_to = to if to is not None else self._cast_to + + # unwrap `TypeAlias('Name', T)` -> `T` + if is_type_alias_type(cast_to): + cast_to = cast_to.__value__ # type: ignore[unreachable] + + # unwrap `Annotated[T, ...]` -> `T` + if cast_to and is_annotated_type(cast_to): + cast_to = extract_type_arg(cast_to, 0) + + origin = get_origin(cast_to) or cast_to + + if self._is_sse_stream: + if to: + if not is_stream_class_type(to): + raise TypeError(f"Expected custom parse type to be a subclass of {Stream} or {AsyncStream}") + + return cast( + _T, + to( + cast_to=extract_stream_chunk_type( + to, + failure_message="Expected custom stream type to be passed with a type argument, e.g. Stream[ChunkType]", + ), + response=self.http_response, + client=cast(Any, self._client), + ), + ) + + if self._stream_cls: + return cast( + R, + self._stream_cls( + cast_to=extract_stream_chunk_type(self._stream_cls), + response=self.http_response, + client=cast(Any, self._client), + ), + ) + + stream_cls = cast("type[Stream[Any]] | type[AsyncStream[Any]] | None", self._client._default_stream_cls) + if stream_cls is None: + raise MissingStreamClassError() + + return cast( + R, + stream_cls( + cast_to=cast_to, + response=self.http_response, + client=cast(Any, self._client), + ), + ) + + if cast_to is NoneType: + return cast(R, None) + + response = self.http_response + if cast_to == str: + return cast(R, response.text) + + if cast_to == bytes: + return cast(R, response.content) + + if cast_to == int: + return cast(R, int(response.text)) + + if cast_to == float: + return cast(R, float(response.text)) + + if cast_to == bool: + return cast(R, response.text.lower() == "true") + + if origin == APIResponse: + raise RuntimeError("Unexpected state - cast_to is `APIResponse`") + + if inspect.isclass(origin) and issubclass(origin, httpx.Response): + # Because of the invariance of our ResponseT TypeVar, users can subclass httpx.Response + # and pass that class to our request functions. We cannot change the variance to be either + # covariant or contravariant as that makes our usage of ResponseT illegal. We could construct + # the response class ourselves but that is something that should be supported directly in httpx + # as it would be easy to incorrectly construct the Response object due to the multitude of arguments. + if cast_to != httpx.Response: + raise ValueError(f"Subclasses of httpx.Response cannot be passed to `cast_to`") + return cast(R, response) + + if ( + inspect.isclass( + origin # pyright: ignore[reportUnknownArgumentType] + ) + and not issubclass(origin, BaseModel) + and issubclass(origin, pydantic.BaseModel) + ): + raise TypeError("Pydantic models must subclass our base model type, e.g. `from brapi import BaseModel`") + + if ( + cast_to is not object + and not origin is list + and not origin is dict + and not origin is Union + and not issubclass(origin, BaseModel) + ): + raise RuntimeError( + f"Unsupported type, expected {cast_to} to be a subclass of {BaseModel}, {dict}, {list}, {Union}, {NoneType}, {str} or {httpx.Response}." + ) + + # split is required to handle cases where additional information is included + # in the response, e.g. application/json; charset=utf-8 + content_type, *_ = response.headers.get("content-type", "*").split(";") + if not content_type.endswith("json"): + if is_basemodel(cast_to): + try: + data = response.json() + except Exception as exc: + log.debug("Could not read JSON from response data due to %s - %s", type(exc), exc) + else: + return self._client._process_response_data( + data=data, + cast_to=cast_to, # type: ignore + response=response, + ) + + if self._client._strict_response_validation: + raise APIResponseValidationError( + response=response, + message=f"Expected Content-Type response header to be `application/json` but received `{content_type}` instead.", + body=response.text, + ) + + # If the API responds with content that isn't JSON then we just return + # the (decoded) text without performing any parsing so that you can still + # handle the response however you need to. + return response.text # type: ignore + + data = response.json() + + return self._client._process_response_data( + data=data, + cast_to=cast_to, # type: ignore + response=response, + ) + + +class APIResponse(BaseAPIResponse[R]): + @overload + def parse(self, *, to: type[_T]) -> _T: ... + + @overload + def parse(self) -> R: ... + + def parse(self, *, to: type[_T] | None = None) -> R | _T: + """Returns the rich python representation of this response's data. + + For lower-level control, see `.read()`, `.json()`, `.iter_bytes()`. + + You can customise the type that the response is parsed into through + the `to` argument, e.g. + + ```py + from brapi import BaseModel + + + class MyModel(BaseModel): + foo: str + + + obj = response.parse(to=MyModel) + print(obj.foo) + ``` + + We support parsing: + - `BaseModel` + - `dict` + - `list` + - `Union` + - `str` + - `int` + - `float` + - `httpx.Response` + """ + cache_key = to if to is not None else self._cast_to + cached = self._parsed_by_type.get(cache_key) + if cached is not None: + return cached # type: ignore[no-any-return] + + if not self._is_sse_stream: + self.read() + + parsed = self._parse(to=to) + if is_given(self._options.post_parser): + parsed = self._options.post_parser(parsed) + + self._parsed_by_type[cache_key] = parsed + return parsed + + def read(self) -> bytes: + """Read and return the binary response content.""" + try: + return self.http_response.read() + except httpx.StreamConsumed as exc: + # The default error raised by httpx isn't very + # helpful in our case so we re-raise it with + # a different error message. + raise StreamAlreadyConsumed() from exc + + def text(self) -> str: + """Read and decode the response content into a string.""" + self.read() + return self.http_response.text + + def json(self) -> object: + """Read and decode the JSON response content.""" + self.read() + return self.http_response.json() + + def close(self) -> None: + """Close the response and release the connection. + + Automatically called if the response body is read to completion. + """ + self.http_response.close() + + def iter_bytes(self, chunk_size: int | None = None) -> Iterator[bytes]: + """ + A byte-iterator over the decoded response content. + + This automatically handles gzip, deflate and brotli encoded responses. + """ + for chunk in self.http_response.iter_bytes(chunk_size): + yield chunk + + def iter_text(self, chunk_size: int | None = None) -> Iterator[str]: + """A str-iterator over the decoded response content + that handles both gzip, deflate, etc but also detects the content's + string encoding. + """ + for chunk in self.http_response.iter_text(chunk_size): + yield chunk + + def iter_lines(self) -> Iterator[str]: + """Like `iter_text()` but will only yield chunks for each line""" + for chunk in self.http_response.iter_lines(): + yield chunk + + +class AsyncAPIResponse(BaseAPIResponse[R]): + @overload + async def parse(self, *, to: type[_T]) -> _T: ... + + @overload + async def parse(self) -> R: ... + + async def parse(self, *, to: type[_T] | None = None) -> R | _T: + """Returns the rich python representation of this response's data. + + For lower-level control, see `.read()`, `.json()`, `.iter_bytes()`. + + You can customise the type that the response is parsed into through + the `to` argument, e.g. + + ```py + from brapi import BaseModel + + + class MyModel(BaseModel): + foo: str + + + obj = response.parse(to=MyModel) + print(obj.foo) + ``` + + We support parsing: + - `BaseModel` + - `dict` + - `list` + - `Union` + - `str` + - `httpx.Response` + """ + cache_key = to if to is not None else self._cast_to + cached = self._parsed_by_type.get(cache_key) + if cached is not None: + return cached # type: ignore[no-any-return] + + if not self._is_sse_stream: + await self.read() + + parsed = self._parse(to=to) + if is_given(self._options.post_parser): + parsed = self._options.post_parser(parsed) + + self._parsed_by_type[cache_key] = parsed + return parsed + + async def read(self) -> bytes: + """Read and return the binary response content.""" + try: + return await self.http_response.aread() + except httpx.StreamConsumed as exc: + # the default error raised by httpx isn't very + # helpful in our case so we re-raise it with + # a different error message + raise StreamAlreadyConsumed() from exc + + async def text(self) -> str: + """Read and decode the response content into a string.""" + await self.read() + return self.http_response.text + + async def json(self) -> object: + """Read and decode the JSON response content.""" + await self.read() + return self.http_response.json() + + async def close(self) -> None: + """Close the response and release the connection. + + Automatically called if the response body is read to completion. + """ + await self.http_response.aclose() + + async def iter_bytes(self, chunk_size: int | None = None) -> AsyncIterator[bytes]: + """ + A byte-iterator over the decoded response content. + + This automatically handles gzip, deflate and brotli encoded responses. + """ + async for chunk in self.http_response.aiter_bytes(chunk_size): + yield chunk + + async def iter_text(self, chunk_size: int | None = None) -> AsyncIterator[str]: + """A str-iterator over the decoded response content + that handles both gzip, deflate, etc but also detects the content's + string encoding. + """ + async for chunk in self.http_response.aiter_text(chunk_size): + yield chunk + + async def iter_lines(self) -> AsyncIterator[str]: + """Like `iter_text()` but will only yield chunks for each line""" + async for chunk in self.http_response.aiter_lines(): + yield chunk + + +class BinaryAPIResponse(APIResponse[bytes]): + """Subclass of APIResponse providing helpers for dealing with binary data. + + Note: If you want to stream the response data instead of eagerly reading it + all at once then you should use `.with_streaming_response` when making + the API request, e.g. `.with_streaming_response.get_binary_response()` + """ + + def write_to_file( + self, + file: str | os.PathLike[str], + ) -> None: + """Write the output to the given file. + + Accepts a filename or any path-like object, e.g. pathlib.Path + + Note: if you want to stream the data to the file instead of writing + all at once then you should use `.with_streaming_response` when making + the API request, e.g. `.with_streaming_response.get_binary_response()` + """ + with open(file, mode="wb") as f: + for data in self.iter_bytes(): + f.write(data) + + +class AsyncBinaryAPIResponse(AsyncAPIResponse[bytes]): + """Subclass of APIResponse providing helpers for dealing with binary data. + + Note: If you want to stream the response data instead of eagerly reading it + all at once then you should use `.with_streaming_response` when making + the API request, e.g. `.with_streaming_response.get_binary_response()` + """ + + async def write_to_file( + self, + file: str | os.PathLike[str], + ) -> None: + """Write the output to the given file. + + Accepts a filename or any path-like object, e.g. pathlib.Path + + Note: if you want to stream the data to the file instead of writing + all at once then you should use `.with_streaming_response` when making + the API request, e.g. `.with_streaming_response.get_binary_response()` + """ + path = anyio.Path(file) + async with await path.open(mode="wb") as f: + async for data in self.iter_bytes(): + await f.write(data) + + +class StreamedBinaryAPIResponse(APIResponse[bytes]): + def stream_to_file( + self, + file: str | os.PathLike[str], + *, + chunk_size: int | None = None, + ) -> None: + """Streams the output to the given file. + + Accepts a filename or any path-like object, e.g. pathlib.Path + """ + with open(file, mode="wb") as f: + for data in self.iter_bytes(chunk_size): + f.write(data) + + +class AsyncStreamedBinaryAPIResponse(AsyncAPIResponse[bytes]): + async def stream_to_file( + self, + file: str | os.PathLike[str], + *, + chunk_size: int | None = None, + ) -> None: + """Streams the output to the given file. + + Accepts a filename or any path-like object, e.g. pathlib.Path + """ + path = anyio.Path(file) + async with await path.open(mode="wb") as f: + async for data in self.iter_bytes(chunk_size): + await f.write(data) + + +class MissingStreamClassError(TypeError): + def __init__(self) -> None: + super().__init__( + "The `stream` argument was set to `True` but the `stream_cls` argument was not given. See `brapi._streaming` for reference", + ) + + +class StreamAlreadyConsumed(BrapiError): + """ + Attempted to read or stream content, but the content has already + been streamed. + + This can happen if you use a method like `.iter_lines()` and then attempt + to read th entire response body afterwards, e.g. + + ```py + response = await client.post(...) + async for line in response.iter_lines(): + ... # do something with `line` + + content = await response.read() + # ^ error + ``` + + If you want this behaviour you'll need to either manually accumulate the response + content or call `await response.read()` before iterating over the stream. + """ + + def __init__(self) -> None: + message = ( + "Attempted to read or stream some content, but the content has " + "already been streamed. " + "This could be due to attempting to stream the response " + "content more than once." + "\n\n" + "You can fix this by manually accumulating the response content while streaming " + "or by calling `.read()` before starting to stream." + ) + super().__init__(message) + + +class ResponseContextManager(Generic[_APIResponseT]): + """Context manager for ensuring that a request is not made + until it is entered and that the response will always be closed + when the context manager exits + """ + + def __init__(self, request_func: Callable[[], _APIResponseT]) -> None: + self._request_func = request_func + self.__response: _APIResponseT | None = None + + def __enter__(self) -> _APIResponseT: + self.__response = self._request_func() + return self.__response + + def __exit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + if self.__response is not None: + self.__response.close() + + +class AsyncResponseContextManager(Generic[_AsyncAPIResponseT]): + """Context manager for ensuring that a request is not made + until it is entered and that the response will always be closed + when the context manager exits + """ + + def __init__(self, api_request: Awaitable[_AsyncAPIResponseT]) -> None: + self._api_request = api_request + self.__response: _AsyncAPIResponseT | None = None + + async def __aenter__(self) -> _AsyncAPIResponseT: + self.__response = await self._api_request + return self.__response + + async def __aexit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + if self.__response is not None: + await self.__response.close() + + +def to_streamed_response_wrapper(func: Callable[P, R]) -> Callable[P, ResponseContextManager[APIResponse[R]]]: + """Higher order function that takes one of our bound API methods and wraps it + to support streaming and returning the raw `APIResponse` object directly. + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> ResponseContextManager[APIResponse[R]]: + extra_headers: dict[str, str] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "stream" + + kwargs["extra_headers"] = extra_headers + + make_request = functools.partial(func, *args, **kwargs) + + return ResponseContextManager(cast(Callable[[], APIResponse[R]], make_request)) + + return wrapped + + +def async_to_streamed_response_wrapper( + func: Callable[P, Awaitable[R]], +) -> Callable[P, AsyncResponseContextManager[AsyncAPIResponse[R]]]: + """Higher order function that takes one of our bound API methods and wraps it + to support streaming and returning the raw `APIResponse` object directly. + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> AsyncResponseContextManager[AsyncAPIResponse[R]]: + extra_headers: dict[str, str] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "stream" + + kwargs["extra_headers"] = extra_headers + + make_request = func(*args, **kwargs) + + return AsyncResponseContextManager(cast(Awaitable[AsyncAPIResponse[R]], make_request)) + + return wrapped + + +def to_custom_streamed_response_wrapper( + func: Callable[P, object], + response_cls: type[_APIResponseT], +) -> Callable[P, ResponseContextManager[_APIResponseT]]: + """Higher order function that takes one of our bound API methods and an `APIResponse` class + and wraps the method to support streaming and returning the given response class directly. + + Note: the given `response_cls` *must* be concrete, e.g. `class BinaryAPIResponse(APIResponse[bytes])` + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> ResponseContextManager[_APIResponseT]: + extra_headers: dict[str, Any] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "stream" + extra_headers[OVERRIDE_CAST_TO_HEADER] = response_cls + + kwargs["extra_headers"] = extra_headers + + make_request = functools.partial(func, *args, **kwargs) + + return ResponseContextManager(cast(Callable[[], _APIResponseT], make_request)) + + return wrapped + + +def async_to_custom_streamed_response_wrapper( + func: Callable[P, Awaitable[object]], + response_cls: type[_AsyncAPIResponseT], +) -> Callable[P, AsyncResponseContextManager[_AsyncAPIResponseT]]: + """Higher order function that takes one of our bound API methods and an `APIResponse` class + and wraps the method to support streaming and returning the given response class directly. + + Note: the given `response_cls` *must* be concrete, e.g. `class BinaryAPIResponse(APIResponse[bytes])` + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> AsyncResponseContextManager[_AsyncAPIResponseT]: + extra_headers: dict[str, Any] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "stream" + extra_headers[OVERRIDE_CAST_TO_HEADER] = response_cls + + kwargs["extra_headers"] = extra_headers + + make_request = func(*args, **kwargs) + + return AsyncResponseContextManager(cast(Awaitable[_AsyncAPIResponseT], make_request)) + + return wrapped + + +def to_raw_response_wrapper(func: Callable[P, R]) -> Callable[P, APIResponse[R]]: + """Higher order function that takes one of our bound API methods and wraps it + to support returning the raw `APIResponse` object directly. + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> APIResponse[R]: + extra_headers: dict[str, str] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "raw" + + kwargs["extra_headers"] = extra_headers + + return cast(APIResponse[R], func(*args, **kwargs)) + + return wrapped + + +def async_to_raw_response_wrapper(func: Callable[P, Awaitable[R]]) -> Callable[P, Awaitable[AsyncAPIResponse[R]]]: + """Higher order function that takes one of our bound API methods and wraps it + to support returning the raw `APIResponse` object directly. + """ + + @functools.wraps(func) + async def wrapped(*args: P.args, **kwargs: P.kwargs) -> AsyncAPIResponse[R]: + extra_headers: dict[str, str] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "raw" + + kwargs["extra_headers"] = extra_headers + + return cast(AsyncAPIResponse[R], await func(*args, **kwargs)) + + return wrapped + + +def to_custom_raw_response_wrapper( + func: Callable[P, object], + response_cls: type[_APIResponseT], +) -> Callable[P, _APIResponseT]: + """Higher order function that takes one of our bound API methods and an `APIResponse` class + and wraps the method to support returning the given response class directly. + + Note: the given `response_cls` *must* be concrete, e.g. `class BinaryAPIResponse(APIResponse[bytes])` + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> _APIResponseT: + extra_headers: dict[str, Any] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "raw" + extra_headers[OVERRIDE_CAST_TO_HEADER] = response_cls + + kwargs["extra_headers"] = extra_headers + + return cast(_APIResponseT, func(*args, **kwargs)) + + return wrapped + + +def async_to_custom_raw_response_wrapper( + func: Callable[P, Awaitable[object]], + response_cls: type[_AsyncAPIResponseT], +) -> Callable[P, Awaitable[_AsyncAPIResponseT]]: + """Higher order function that takes one of our bound API methods and an `APIResponse` class + and wraps the method to support returning the given response class directly. + + Note: the given `response_cls` *must* be concrete, e.g. `class BinaryAPIResponse(APIResponse[bytes])` + """ + + @functools.wraps(func) + def wrapped(*args: P.args, **kwargs: P.kwargs) -> Awaitable[_AsyncAPIResponseT]: + extra_headers: dict[str, Any] = {**(cast(Any, kwargs.get("extra_headers")) or {})} + extra_headers[RAW_RESPONSE_HEADER] = "raw" + extra_headers[OVERRIDE_CAST_TO_HEADER] = response_cls + + kwargs["extra_headers"] = extra_headers + + return cast(Awaitable[_AsyncAPIResponseT], func(*args, **kwargs)) + + return wrapped + + +def extract_response_type(typ: type[BaseAPIResponse[Any]]) -> type: + """Given a type like `APIResponse[T]`, returns the generic type variable `T`. + + This also handles the case where a concrete subclass is given, e.g. + ```py + class MyResponse(APIResponse[bytes]): + ... + + extract_response_type(MyResponse) -> bytes + ``` + """ + return extract_type_var_from_base( + typ, + generic_bases=cast("tuple[type, ...]", (BaseAPIResponse, APIResponse, AsyncAPIResponse)), + index=0, + ) diff --git a/src/brapi/_streaming.py b/src/brapi/_streaming.py new file mode 100644 index 0000000..b9d6d1c --- /dev/null +++ b/src/brapi/_streaming.py @@ -0,0 +1,333 @@ +# Note: initially copied from https://github.com/florimondmanca/httpx-sse/blob/master/src/httpx_sse/_decoders.py +from __future__ import annotations + +import json +import inspect +from types import TracebackType +from typing import TYPE_CHECKING, Any, Generic, TypeVar, Iterator, AsyncIterator, cast +from typing_extensions import Self, Protocol, TypeGuard, override, get_origin, runtime_checkable + +import httpx + +from ._utils import extract_type_var_from_base + +if TYPE_CHECKING: + from ._client import Brapi, AsyncBrapi + + +_T = TypeVar("_T") + + +class Stream(Generic[_T]): + """Provides the core interface to iterate over a synchronous stream response.""" + + response: httpx.Response + + _decoder: SSEBytesDecoder + + def __init__( + self, + *, + cast_to: type[_T], + response: httpx.Response, + client: Brapi, + ) -> None: + self.response = response + self._cast_to = cast_to + self._client = client + self._decoder = client._make_sse_decoder() + self._iterator = self.__stream__() + + def __next__(self) -> _T: + return self._iterator.__next__() + + def __iter__(self) -> Iterator[_T]: + for item in self._iterator: + yield item + + def _iter_events(self) -> Iterator[ServerSentEvent]: + yield from self._decoder.iter_bytes(self.response.iter_bytes()) + + def __stream__(self) -> Iterator[_T]: + cast_to = cast(Any, self._cast_to) + response = self.response + process_data = self._client._process_response_data + iterator = self._iter_events() + + for sse in iterator: + yield process_data(data=sse.json(), cast_to=cast_to, response=response) + + # Ensure the entire stream is consumed + for _sse in iterator: + ... + + def __enter__(self) -> Self: + return self + + def __exit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + self.close() + + def close(self) -> None: + """ + Close the response and release the connection. + + Automatically called if the response body is read to completion. + """ + self.response.close() + + +class AsyncStream(Generic[_T]): + """Provides the core interface to iterate over an asynchronous stream response.""" + + response: httpx.Response + + _decoder: SSEDecoder | SSEBytesDecoder + + def __init__( + self, + *, + cast_to: type[_T], + response: httpx.Response, + client: AsyncBrapi, + ) -> None: + self.response = response + self._cast_to = cast_to + self._client = client + self._decoder = client._make_sse_decoder() + self._iterator = self.__stream__() + + async def __anext__(self) -> _T: + return await self._iterator.__anext__() + + async def __aiter__(self) -> AsyncIterator[_T]: + async for item in self._iterator: + yield item + + async def _iter_events(self) -> AsyncIterator[ServerSentEvent]: + async for sse in self._decoder.aiter_bytes(self.response.aiter_bytes()): + yield sse + + async def __stream__(self) -> AsyncIterator[_T]: + cast_to = cast(Any, self._cast_to) + response = self.response + process_data = self._client._process_response_data + iterator = self._iter_events() + + async for sse in iterator: + yield process_data(data=sse.json(), cast_to=cast_to, response=response) + + # Ensure the entire stream is consumed + async for _sse in iterator: + ... + + async def __aenter__(self) -> Self: + return self + + async def __aexit__( + self, + exc_type: type[BaseException] | None, + exc: BaseException | None, + exc_tb: TracebackType | None, + ) -> None: + await self.close() + + async def close(self) -> None: + """ + Close the response and release the connection. + + Automatically called if the response body is read to completion. + """ + await self.response.aclose() + + +class ServerSentEvent: + def __init__( + self, + *, + event: str | None = None, + data: str | None = None, + id: str | None = None, + retry: int | None = None, + ) -> None: + if data is None: + data = "" + + self._id = id + self._data = data + self._event = event or None + self._retry = retry + + @property + def event(self) -> str | None: + return self._event + + @property + def id(self) -> str | None: + return self._id + + @property + def retry(self) -> int | None: + return self._retry + + @property + def data(self) -> str: + return self._data + + def json(self) -> Any: + return json.loads(self.data) + + @override + def __repr__(self) -> str: + return f"ServerSentEvent(event={self.event}, data={self.data}, id={self.id}, retry={self.retry})" + + +class SSEDecoder: + _data: list[str] + _event: str | None + _retry: int | None + _last_event_id: str | None + + def __init__(self) -> None: + self._event = None + self._data = [] + self._last_event_id = None + self._retry = None + + def iter_bytes(self, iterator: Iterator[bytes]) -> Iterator[ServerSentEvent]: + """Given an iterator that yields raw binary data, iterate over it & yield every event encountered""" + for chunk in self._iter_chunks(iterator): + # Split before decoding so splitlines() only uses \r and \n + for raw_line in chunk.splitlines(): + line = raw_line.decode("utf-8") + sse = self.decode(line) + if sse: + yield sse + + def _iter_chunks(self, iterator: Iterator[bytes]) -> Iterator[bytes]: + """Given an iterator that yields raw binary data, iterate over it and yield individual SSE chunks""" + data = b"" + for chunk in iterator: + for line in chunk.splitlines(keepends=True): + data += line + if data.endswith((b"\r\r", b"\n\n", b"\r\n\r\n")): + yield data + data = b"" + if data: + yield data + + async def aiter_bytes(self, iterator: AsyncIterator[bytes]) -> AsyncIterator[ServerSentEvent]: + """Given an iterator that yields raw binary data, iterate over it & yield every event encountered""" + async for chunk in self._aiter_chunks(iterator): + # Split before decoding so splitlines() only uses \r and \n + for raw_line in chunk.splitlines(): + line = raw_line.decode("utf-8") + sse = self.decode(line) + if sse: + yield sse + + async def _aiter_chunks(self, iterator: AsyncIterator[bytes]) -> AsyncIterator[bytes]: + """Given an iterator that yields raw binary data, iterate over it and yield individual SSE chunks""" + data = b"" + async for chunk in iterator: + for line in chunk.splitlines(keepends=True): + data += line + if data.endswith((b"\r\r", b"\n\n", b"\r\n\r\n")): + yield data + data = b"" + if data: + yield data + + def decode(self, line: str) -> ServerSentEvent | None: + # See: https://html.spec.whatwg.org/multipage/server-sent-events.html#event-stream-interpretation # noqa: E501 + + if not line: + if not self._event and not self._data and not self._last_event_id and self._retry is None: + return None + + sse = ServerSentEvent( + event=self._event, + data="\n".join(self._data), + id=self._last_event_id, + retry=self._retry, + ) + + # NOTE: as per the SSE spec, do not reset last_event_id. + self._event = None + self._data = [] + self._retry = None + + return sse + + if line.startswith(":"): + return None + + fieldname, _, value = line.partition(":") + + if value.startswith(" "): + value = value[1:] + + if fieldname == "event": + self._event = value + elif fieldname == "data": + self._data.append(value) + elif fieldname == "id": + if "\0" in value: + pass + else: + self._last_event_id = value + elif fieldname == "retry": + try: + self._retry = int(value) + except (TypeError, ValueError): + pass + else: + pass # Field is ignored. + + return None + + +@runtime_checkable +class SSEBytesDecoder(Protocol): + def iter_bytes(self, iterator: Iterator[bytes]) -> Iterator[ServerSentEvent]: + """Given an iterator that yields raw binary data, iterate over it & yield every event encountered""" + ... + + def aiter_bytes(self, iterator: AsyncIterator[bytes]) -> AsyncIterator[ServerSentEvent]: + """Given an async iterator that yields raw binary data, iterate over it & yield every event encountered""" + ... + + +def is_stream_class_type(typ: type) -> TypeGuard[type[Stream[object]] | type[AsyncStream[object]]]: + """TypeGuard for determining whether or not the given type is a subclass of `Stream` / `AsyncStream`""" + origin = get_origin(typ) or typ + return inspect.isclass(origin) and issubclass(origin, (Stream, AsyncStream)) + + +def extract_stream_chunk_type( + stream_cls: type, + *, + failure_message: str | None = None, +) -> type: + """Given a type like `Stream[T]`, returns the generic type variable `T`. + + This also handles the case where a concrete subclass is given, e.g. + ```py + class MyStream(Stream[bytes]): + ... + + extract_stream_chunk_type(MyStream) -> bytes + ``` + """ + from ._base_client import Stream, AsyncStream + + return extract_type_var_from_base( + stream_cls, + index=0, + generic_bases=cast("tuple[type, ...]", (Stream, AsyncStream)), + failure_message=failure_message, + ) diff --git a/src/brapi/_types.py b/src/brapi/_types.py new file mode 100644 index 0000000..a3e695f --- /dev/null +++ b/src/brapi/_types.py @@ -0,0 +1,260 @@ +from __future__ import annotations + +from os import PathLike +from typing import ( + IO, + TYPE_CHECKING, + Any, + Dict, + List, + Type, + Tuple, + Union, + Mapping, + TypeVar, + Callable, + Iterator, + Optional, + Sequence, +) +from typing_extensions import ( + Set, + Literal, + Protocol, + TypeAlias, + TypedDict, + SupportsIndex, + overload, + override, + runtime_checkable, +) + +import httpx +import pydantic +from httpx import URL, Proxy, Timeout, Response, BaseTransport, AsyncBaseTransport + +if TYPE_CHECKING: + from ._models import BaseModel + from ._response import APIResponse, AsyncAPIResponse + +Transport = BaseTransport +AsyncTransport = AsyncBaseTransport +Query = Mapping[str, object] +Body = object +AnyMapping = Mapping[str, object] +ModelT = TypeVar("ModelT", bound=pydantic.BaseModel) +_T = TypeVar("_T") + + +# Approximates httpx internal ProxiesTypes and RequestFiles types +# while adding support for `PathLike` instances +ProxiesDict = Dict["str | URL", Union[None, str, URL, Proxy]] +ProxiesTypes = Union[str, Proxy, ProxiesDict] +if TYPE_CHECKING: + Base64FileInput = Union[IO[bytes], PathLike[str]] + FileContent = Union[IO[bytes], bytes, PathLike[str]] +else: + Base64FileInput = Union[IO[bytes], PathLike] + FileContent = Union[IO[bytes], bytes, PathLike] # PathLike is not subscriptable in Python 3.8. +FileTypes = Union[ + # file (or bytes) + FileContent, + # (filename, file (or bytes)) + Tuple[Optional[str], FileContent], + # (filename, file (or bytes), content_type) + Tuple[Optional[str], FileContent, Optional[str]], + # (filename, file (or bytes), content_type, headers) + Tuple[Optional[str], FileContent, Optional[str], Mapping[str, str]], +] +RequestFiles = Union[Mapping[str, FileTypes], Sequence[Tuple[str, FileTypes]]] + +# duplicate of the above but without our custom file support +HttpxFileContent = Union[IO[bytes], bytes] +HttpxFileTypes = Union[ + # file (or bytes) + HttpxFileContent, + # (filename, file (or bytes)) + Tuple[Optional[str], HttpxFileContent], + # (filename, file (or bytes), content_type) + Tuple[Optional[str], HttpxFileContent, Optional[str]], + # (filename, file (or bytes), content_type, headers) + Tuple[Optional[str], HttpxFileContent, Optional[str], Mapping[str, str]], +] +HttpxRequestFiles = Union[Mapping[str, HttpxFileTypes], Sequence[Tuple[str, HttpxFileTypes]]] + +# Workaround to support (cast_to: Type[ResponseT]) -> ResponseT +# where ResponseT includes `None`. In order to support directly +# passing `None`, overloads would have to be defined for every +# method that uses `ResponseT` which would lead to an unacceptable +# amount of code duplication and make it unreadable. See _base_client.py +# for example usage. +# +# This unfortunately means that you will either have +# to import this type and pass it explicitly: +# +# from brapi import NoneType +# client.get('/foo', cast_to=NoneType) +# +# or build it yourself: +# +# client.get('/foo', cast_to=type(None)) +if TYPE_CHECKING: + NoneType: Type[None] +else: + NoneType = type(None) + + +class RequestOptions(TypedDict, total=False): + headers: Headers + max_retries: int + timeout: float | Timeout | None + params: Query + extra_json: AnyMapping + idempotency_key: str + follow_redirects: bool + + +# Sentinel class used until PEP 0661 is accepted +class NotGiven: + """ + For parameters with a meaningful None value, we need to distinguish between + the user explicitly passing None, and the user not passing the parameter at + all. + + User code shouldn't need to use not_given directly. + + For example: + + ```py + def create(timeout: Timeout | None | NotGiven = not_given): ... + + + create(timeout=1) # 1s timeout + create(timeout=None) # No timeout + create() # Default timeout behavior + ``` + """ + + def __bool__(self) -> Literal[False]: + return False + + @override + def __repr__(self) -> str: + return "NOT_GIVEN" + + +not_given = NotGiven() +# for backwards compatibility: +NOT_GIVEN = NotGiven() + + +class Omit: + """ + To explicitly omit something from being sent in a request, use `omit`. + + ```py + # as the default `Content-Type` header is `application/json` that will be sent + client.post("/upload/files", files={"file": b"my raw file content"}) + + # you can't explicitly override the header as it has to be dynamically generated + # to look something like: 'multipart/form-data; boundary=0d8382fcf5f8c3be01ca2e11002d2983' + client.post(..., headers={"Content-Type": "multipart/form-data"}) + + # instead you can remove the default `application/json` header by passing omit + client.post(..., headers={"Content-Type": omit}) + ``` + """ + + def __bool__(self) -> Literal[False]: + return False + + +omit = Omit() + + +@runtime_checkable +class ModelBuilderProtocol(Protocol): + @classmethod + def build( + cls: type[_T], + *, + response: Response, + data: object, + ) -> _T: ... + + +Headers = Mapping[str, Union[str, Omit]] + + +class HeadersLikeProtocol(Protocol): + def get(self, __key: str) -> str | None: ... + + +HeadersLike = Union[Headers, HeadersLikeProtocol] + +ResponseT = TypeVar( + "ResponseT", + bound=Union[ + object, + str, + None, + "BaseModel", + List[Any], + Dict[str, Any], + Response, + ModelBuilderProtocol, + "APIResponse[Any]", + "AsyncAPIResponse[Any]", + ], +) + +StrBytesIntFloat = Union[str, bytes, int, float] + +# Note: copied from Pydantic +# https://github.com/pydantic/pydantic/blob/6f31f8f68ef011f84357330186f603ff295312fd/pydantic/main.py#L79 +IncEx: TypeAlias = Union[Set[int], Set[str], Mapping[int, Union["IncEx", bool]], Mapping[str, Union["IncEx", bool]]] + +PostParser = Callable[[Any], Any] + + +@runtime_checkable +class InheritsGeneric(Protocol): + """Represents a type that has inherited from `Generic` + + The `__orig_bases__` property can be used to determine the resolved + type variable for a given base class. + """ + + __orig_bases__: tuple[_GenericAlias] + + +class _GenericAlias(Protocol): + __origin__: type[object] + + +class HttpxSendArgs(TypedDict, total=False): + auth: httpx.Auth + follow_redirects: bool + + +_T_co = TypeVar("_T_co", covariant=True) + + +if TYPE_CHECKING: + # This works because str.__contains__ does not accept object (either in typeshed or at runtime) + # https://github.com/hauntsaninja/useful_types/blob/5e9710f3875107d068e7679fd7fec9cfab0eff3b/useful_types/__init__.py#L285 + class SequenceNotStr(Protocol[_T_co]): + @overload + def __getitem__(self, index: SupportsIndex, /) -> _T_co: ... + @overload + def __getitem__(self, index: slice, /) -> Sequence[_T_co]: ... + def __contains__(self, value: object, /) -> bool: ... + def __len__(self) -> int: ... + def __iter__(self) -> Iterator[_T_co]: ... + def index(self, value: Any, start: int = 0, stop: int = ..., /) -> int: ... + def count(self, value: Any, /) -> int: ... + def __reversed__(self) -> Iterator[_T_co]: ... +else: + # just point this to a normal `Sequence` at runtime to avoid having to special case + # deserializing our custom sequence type + SequenceNotStr = Sequence diff --git a/src/brapi/_utils/__init__.py b/src/brapi/_utils/__init__.py new file mode 100644 index 0000000..dc64e29 --- /dev/null +++ b/src/brapi/_utils/__init__.py @@ -0,0 +1,64 @@ +from ._sync import asyncify as asyncify +from ._proxy import LazyProxy as LazyProxy +from ._utils import ( + flatten as flatten, + is_dict as is_dict, + is_list as is_list, + is_given as is_given, + is_tuple as is_tuple, + json_safe as json_safe, + lru_cache as lru_cache, + is_mapping as is_mapping, + is_tuple_t as is_tuple_t, + is_iterable as is_iterable, + is_sequence as is_sequence, + coerce_float as coerce_float, + is_mapping_t as is_mapping_t, + removeprefix as removeprefix, + removesuffix as removesuffix, + extract_files as extract_files, + is_sequence_t as is_sequence_t, + required_args as required_args, + coerce_boolean as coerce_boolean, + coerce_integer as coerce_integer, + file_from_path as file_from_path, + strip_not_given as strip_not_given, + deepcopy_minimal as deepcopy_minimal, + get_async_library as get_async_library, + maybe_coerce_float as maybe_coerce_float, + get_required_header as get_required_header, + maybe_coerce_boolean as maybe_coerce_boolean, + maybe_coerce_integer as maybe_coerce_integer, +) +from ._compat import ( + get_args as get_args, + is_union as is_union, + get_origin as get_origin, + is_typeddict as is_typeddict, + is_literal_type as is_literal_type, +) +from ._typing import ( + is_list_type as is_list_type, + is_union_type as is_union_type, + extract_type_arg as extract_type_arg, + is_iterable_type as is_iterable_type, + is_required_type as is_required_type, + is_sequence_type as is_sequence_type, + is_annotated_type as is_annotated_type, + is_type_alias_type as is_type_alias_type, + strip_annotated_type as strip_annotated_type, + extract_type_var_from_base as extract_type_var_from_base, +) +from ._streams import consume_sync_iterator as consume_sync_iterator, consume_async_iterator as consume_async_iterator +from ._transform import ( + PropertyInfo as PropertyInfo, + transform as transform, + async_transform as async_transform, + maybe_transform as maybe_transform, + async_maybe_transform as async_maybe_transform, +) +from ._reflection import ( + function_has_argument as function_has_argument, + assert_signatures_in_sync as assert_signatures_in_sync, +) +from ._datetime_parse import parse_date as parse_date, parse_datetime as parse_datetime diff --git a/src/brapi/_utils/_compat.py b/src/brapi/_utils/_compat.py new file mode 100644 index 0000000..dd70323 --- /dev/null +++ b/src/brapi/_utils/_compat.py @@ -0,0 +1,45 @@ +from __future__ import annotations + +import sys +import typing_extensions +from typing import Any, Type, Union, Literal, Optional +from datetime import date, datetime +from typing_extensions import get_args as _get_args, get_origin as _get_origin + +from .._types import StrBytesIntFloat +from ._datetime_parse import parse_date as _parse_date, parse_datetime as _parse_datetime + +_LITERAL_TYPES = {Literal, typing_extensions.Literal} + + +def get_args(tp: type[Any]) -> tuple[Any, ...]: + return _get_args(tp) + + +def get_origin(tp: type[Any]) -> type[Any] | None: + return _get_origin(tp) + + +def is_union(tp: Optional[Type[Any]]) -> bool: + if sys.version_info < (3, 10): + return tp is Union # type: ignore[comparison-overlap] + else: + import types + + return tp is Union or tp is types.UnionType + + +def is_typeddict(tp: Type[Any]) -> bool: + return typing_extensions.is_typeddict(tp) + + +def is_literal_type(tp: Type[Any]) -> bool: + return get_origin(tp) in _LITERAL_TYPES + + +def parse_date(value: Union[date, StrBytesIntFloat]) -> date: + return _parse_date(value) + + +def parse_datetime(value: Union[datetime, StrBytesIntFloat]) -> datetime: + return _parse_datetime(value) diff --git a/src/brapi/_utils/_datetime_parse.py b/src/brapi/_utils/_datetime_parse.py new file mode 100644 index 0000000..7cb9d9e --- /dev/null +++ b/src/brapi/_utils/_datetime_parse.py @@ -0,0 +1,136 @@ +""" +This file contains code from https://github.com/pydantic/pydantic/blob/main/pydantic/v1/datetime_parse.py +without the Pydantic v1 specific errors. +""" + +from __future__ import annotations + +import re +from typing import Dict, Union, Optional +from datetime import date, datetime, timezone, timedelta + +from .._types import StrBytesIntFloat + +date_expr = r"(?P\d{4})-(?P\d{1,2})-(?P\d{1,2})" +time_expr = ( + r"(?P\d{1,2}):(?P\d{1,2})" + r"(?::(?P\d{1,2})(?:\.(?P\d{1,6})\d{0,6})?)?" + r"(?PZ|[+-]\d{2}(?::?\d{2})?)?$" +) + +date_re = re.compile(f"{date_expr}$") +datetime_re = re.compile(f"{date_expr}[T ]{time_expr}") + + +EPOCH = datetime(1970, 1, 1) +# if greater than this, the number is in ms, if less than or equal it's in seconds +# (in seconds this is 11th October 2603, in ms it's 20th August 1970) +MS_WATERSHED = int(2e10) +# slightly more than datetime.max in ns - (datetime.max - EPOCH).total_seconds() * 1e9 +MAX_NUMBER = int(3e20) + + +def _get_numeric(value: StrBytesIntFloat, native_expected_type: str) -> Union[None, int, float]: + if isinstance(value, (int, float)): + return value + try: + return float(value) + except ValueError: + return None + except TypeError: + raise TypeError(f"invalid type; expected {native_expected_type}, string, bytes, int or float") from None + + +def _from_unix_seconds(seconds: Union[int, float]) -> datetime: + if seconds > MAX_NUMBER: + return datetime.max + elif seconds < -MAX_NUMBER: + return datetime.min + + while abs(seconds) > MS_WATERSHED: + seconds /= 1000 + dt = EPOCH + timedelta(seconds=seconds) + return dt.replace(tzinfo=timezone.utc) + + +def _parse_timezone(value: Optional[str]) -> Union[None, int, timezone]: + if value == "Z": + return timezone.utc + elif value is not None: + offset_mins = int(value[-2:]) if len(value) > 3 else 0 + offset = 60 * int(value[1:3]) + offset_mins + if value[0] == "-": + offset = -offset + return timezone(timedelta(minutes=offset)) + else: + return None + + +def parse_datetime(value: Union[datetime, StrBytesIntFloat]) -> datetime: + """ + Parse a datetime/int/float/string and return a datetime.datetime. + + This function supports time zone offsets. When the input contains one, + the output uses a timezone with a fixed offset from UTC. + + Raise ValueError if the input is well formatted but not a valid datetime. + Raise ValueError if the input isn't well formatted. + """ + if isinstance(value, datetime): + return value + + number = _get_numeric(value, "datetime") + if number is not None: + return _from_unix_seconds(number) + + if isinstance(value, bytes): + value = value.decode() + + assert not isinstance(value, (float, int)) + + match = datetime_re.match(value) + if match is None: + raise ValueError("invalid datetime format") + + kw = match.groupdict() + if kw["microsecond"]: + kw["microsecond"] = kw["microsecond"].ljust(6, "0") + + tzinfo = _parse_timezone(kw.pop("tzinfo")) + kw_: Dict[str, Union[None, int, timezone]] = {k: int(v) for k, v in kw.items() if v is not None} + kw_["tzinfo"] = tzinfo + + return datetime(**kw_) # type: ignore + + +def parse_date(value: Union[date, StrBytesIntFloat]) -> date: + """ + Parse a date/int/float/string and return a datetime.date. + + Raise ValueError if the input is well formatted but not a valid date. + Raise ValueError if the input isn't well formatted. + """ + if isinstance(value, date): + if isinstance(value, datetime): + return value.date() + else: + return value + + number = _get_numeric(value, "date") + if number is not None: + return _from_unix_seconds(number).date() + + if isinstance(value, bytes): + value = value.decode() + + assert not isinstance(value, (float, int)) + match = date_re.match(value) + if match is None: + raise ValueError("invalid date format") + + kw = {k: int(v) for k, v in match.groupdict().items()} + + try: + return date(**kw) + except ValueError: + raise ValueError("invalid date format") from None diff --git a/src/brapi/_utils/_logs.py b/src/brapi/_utils/_logs.py new file mode 100644 index 0000000..ad3f965 --- /dev/null +++ b/src/brapi/_utils/_logs.py @@ -0,0 +1,25 @@ +import os +import logging + +logger: logging.Logger = logging.getLogger("brapi") +httpx_logger: logging.Logger = logging.getLogger("httpx") + + +def _basic_config() -> None: + # e.g. [2023-10-05 14:12:26 - brapi._base_client:818 - DEBUG] HTTP Request: POST http://127.0.0.1:4010/foo/bar "200 OK" + logging.basicConfig( + format="[%(asctime)s - %(name)s:%(lineno)d - %(levelname)s] %(message)s", + datefmt="%Y-%m-%d %H:%M:%S", + ) + + +def setup_logging() -> None: + env = os.environ.get("BRAPI_LOG") + if env == "debug": + _basic_config() + logger.setLevel(logging.DEBUG) + httpx_logger.setLevel(logging.DEBUG) + elif env == "info": + _basic_config() + logger.setLevel(logging.INFO) + httpx_logger.setLevel(logging.INFO) diff --git a/src/brapi/_utils/_proxy.py b/src/brapi/_utils/_proxy.py new file mode 100644 index 0000000..0f239a3 --- /dev/null +++ b/src/brapi/_utils/_proxy.py @@ -0,0 +1,65 @@ +from __future__ import annotations + +from abc import ABC, abstractmethod +from typing import Generic, TypeVar, Iterable, cast +from typing_extensions import override + +T = TypeVar("T") + + +class LazyProxy(Generic[T], ABC): + """Implements data methods to pretend that an instance is another instance. + + This includes forwarding attribute access and other methods. + """ + + # Note: we have to special case proxies that themselves return proxies + # to support using a proxy as a catch-all for any random access, e.g. `proxy.foo.bar.baz` + + def __getattr__(self, attr: str) -> object: + proxied = self.__get_proxied__() + if isinstance(proxied, LazyProxy): + return proxied # pyright: ignore + return getattr(proxied, attr) + + @override + def __repr__(self) -> str: + proxied = self.__get_proxied__() + if isinstance(proxied, LazyProxy): + return proxied.__class__.__name__ + return repr(self.__get_proxied__()) + + @override + def __str__(self) -> str: + proxied = self.__get_proxied__() + if isinstance(proxied, LazyProxy): + return proxied.__class__.__name__ + return str(proxied) + + @override + def __dir__(self) -> Iterable[str]: + proxied = self.__get_proxied__() + if isinstance(proxied, LazyProxy): + return [] + return proxied.__dir__() + + @property # type: ignore + @override + def __class__(self) -> type: # pyright: ignore + try: + proxied = self.__get_proxied__() + except Exception: + return type(self) + if issubclass(type(proxied), LazyProxy): + return type(proxied) + return proxied.__class__ + + def __get_proxied__(self) -> T: + return self.__load__() + + def __as_proxied__(self) -> T: + """Helper method that returns the current proxy, typed as the loaded object""" + return cast(T, self) + + @abstractmethod + def __load__(self) -> T: ... diff --git a/src/brapi/_utils/_reflection.py b/src/brapi/_utils/_reflection.py new file mode 100644 index 0000000..89aa712 --- /dev/null +++ b/src/brapi/_utils/_reflection.py @@ -0,0 +1,42 @@ +from __future__ import annotations + +import inspect +from typing import Any, Callable + + +def function_has_argument(func: Callable[..., Any], arg_name: str) -> bool: + """Returns whether or not the given function has a specific parameter""" + sig = inspect.signature(func) + return arg_name in sig.parameters + + +def assert_signatures_in_sync( + source_func: Callable[..., Any], + check_func: Callable[..., Any], + *, + exclude_params: set[str] = set(), +) -> None: + """Ensure that the signature of the second function matches the first.""" + + check_sig = inspect.signature(check_func) + source_sig = inspect.signature(source_func) + + errors: list[str] = [] + + for name, source_param in source_sig.parameters.items(): + if name in exclude_params: + continue + + custom_param = check_sig.parameters.get(name) + if not custom_param: + errors.append(f"the `{name}` param is missing") + continue + + if custom_param.annotation != source_param.annotation: + errors.append( + f"types for the `{name}` param are do not match; source={repr(source_param.annotation)} checking={repr(custom_param.annotation)}" + ) + continue + + if errors: + raise AssertionError(f"{len(errors)} errors encountered when comparing signatures:\n\n" + "\n\n".join(errors)) diff --git a/src/brapi/_utils/_resources_proxy.py b/src/brapi/_utils/_resources_proxy.py new file mode 100644 index 0000000..afea909 --- /dev/null +++ b/src/brapi/_utils/_resources_proxy.py @@ -0,0 +1,24 @@ +from __future__ import annotations + +from typing import Any +from typing_extensions import override + +from ._proxy import LazyProxy + + +class ResourcesProxy(LazyProxy[Any]): + """A proxy for the `brapi.resources` module. + + This is used so that we can lazily import `brapi.resources` only when + needed *and* so that users can just import `brapi` and reference `brapi.resources` + """ + + @override + def __load__(self) -> Any: + import importlib + + mod = importlib.import_module("brapi.resources") + return mod + + +resources = ResourcesProxy().__as_proxied__() diff --git a/src/brapi/_utils/_streams.py b/src/brapi/_utils/_streams.py new file mode 100644 index 0000000..f4a0208 --- /dev/null +++ b/src/brapi/_utils/_streams.py @@ -0,0 +1,12 @@ +from typing import Any +from typing_extensions import Iterator, AsyncIterator + + +def consume_sync_iterator(iterator: Iterator[Any]) -> None: + for _ in iterator: + ... + + +async def consume_async_iterator(iterator: AsyncIterator[Any]) -> None: + async for _ in iterator: + ... diff --git a/src/brapi/_utils/_sync.py b/src/brapi/_utils/_sync.py new file mode 100644 index 0000000..ad7ec71 --- /dev/null +++ b/src/brapi/_utils/_sync.py @@ -0,0 +1,86 @@ +from __future__ import annotations + +import sys +import asyncio +import functools +import contextvars +from typing import Any, TypeVar, Callable, Awaitable +from typing_extensions import ParamSpec + +import anyio +import sniffio +import anyio.to_thread + +T_Retval = TypeVar("T_Retval") +T_ParamSpec = ParamSpec("T_ParamSpec") + + +if sys.version_info >= (3, 9): + _asyncio_to_thread = asyncio.to_thread +else: + # backport of https://docs.python.org/3/library/asyncio-task.html#asyncio.to_thread + # for Python 3.8 support + async def _asyncio_to_thread( + func: Callable[T_ParamSpec, T_Retval], /, *args: T_ParamSpec.args, **kwargs: T_ParamSpec.kwargs + ) -> Any: + """Asynchronously run function *func* in a separate thread. + + Any *args and **kwargs supplied for this function are directly passed + to *func*. Also, the current :class:`contextvars.Context` is propagated, + allowing context variables from the main thread to be accessed in the + separate thread. + + Returns a coroutine that can be awaited to get the eventual result of *func*. + """ + loop = asyncio.events.get_running_loop() + ctx = contextvars.copy_context() + func_call = functools.partial(ctx.run, func, *args, **kwargs) + return await loop.run_in_executor(None, func_call) + + +async def to_thread( + func: Callable[T_ParamSpec, T_Retval], /, *args: T_ParamSpec.args, **kwargs: T_ParamSpec.kwargs +) -> T_Retval: + if sniffio.current_async_library() == "asyncio": + return await _asyncio_to_thread(func, *args, **kwargs) + + return await anyio.to_thread.run_sync( + functools.partial(func, *args, **kwargs), + ) + + +# inspired by `asyncer`, https://github.com/tiangolo/asyncer +def asyncify(function: Callable[T_ParamSpec, T_Retval]) -> Callable[T_ParamSpec, Awaitable[T_Retval]]: + """ + Take a blocking function and create an async one that receives the same + positional and keyword arguments. For python version 3.9 and above, it uses + asyncio.to_thread to run the function in a separate thread. For python version + 3.8, it uses locally defined copy of the asyncio.to_thread function which was + introduced in python 3.9. + + Usage: + + ```python + def blocking_func(arg1, arg2, kwarg1=None): + # blocking code + return result + + + result = asyncify(blocking_function)(arg1, arg2, kwarg1=value1) + ``` + + ## Arguments + + `function`: a blocking regular callable (e.g. a function) + + ## Return + + An async function that takes the same positional and keyword arguments as the + original one, that when called runs the same original function in a thread worker + and returns the result. + """ + + async def wrapper(*args: T_ParamSpec.args, **kwargs: T_ParamSpec.kwargs) -> T_Retval: + return await to_thread(function, *args, **kwargs) + + return wrapper diff --git a/src/brapi/_utils/_transform.py b/src/brapi/_utils/_transform.py new file mode 100644 index 0000000..5207549 --- /dev/null +++ b/src/brapi/_utils/_transform.py @@ -0,0 +1,457 @@ +from __future__ import annotations + +import io +import base64 +import pathlib +from typing import Any, Mapping, TypeVar, cast +from datetime import date, datetime +from typing_extensions import Literal, get_args, override, get_type_hints as _get_type_hints + +import anyio +import pydantic + +from ._utils import ( + is_list, + is_given, + lru_cache, + is_mapping, + is_iterable, + is_sequence, +) +from .._files import is_base64_file_input +from ._compat import get_origin, is_typeddict +from ._typing import ( + is_list_type, + is_union_type, + extract_type_arg, + is_iterable_type, + is_required_type, + is_sequence_type, + is_annotated_type, + strip_annotated_type, +) + +_T = TypeVar("_T") + + +# TODO: support for drilling globals() and locals() +# TODO: ensure works correctly with forward references in all cases + + +PropertyFormat = Literal["iso8601", "base64", "custom"] + + +class PropertyInfo: + """Metadata class to be used in Annotated types to provide information about a given type. + + For example: + + class MyParams(TypedDict): + account_holder_name: Annotated[str, PropertyInfo(alias='accountHolderName')] + + This means that {'account_holder_name': 'Robert'} will be transformed to {'accountHolderName': 'Robert'} before being sent to the API. + """ + + alias: str | None + format: PropertyFormat | None + format_template: str | None + discriminator: str | None + + def __init__( + self, + *, + alias: str | None = None, + format: PropertyFormat | None = None, + format_template: str | None = None, + discriminator: str | None = None, + ) -> None: + self.alias = alias + self.format = format + self.format_template = format_template + self.discriminator = discriminator + + @override + def __repr__(self) -> str: + return f"{self.__class__.__name__}(alias='{self.alias}', format={self.format}, format_template='{self.format_template}', discriminator='{self.discriminator}')" + + +def maybe_transform( + data: object, + expected_type: object, +) -> Any | None: + """Wrapper over `transform()` that allows `None` to be passed. + + See `transform()` for more details. + """ + if data is None: + return None + return transform(data, expected_type) + + +# Wrapper over _transform_recursive providing fake types +def transform( + data: _T, + expected_type: object, +) -> _T: + """Transform dictionaries based off of type information from the given type, for example: + + ```py + class Params(TypedDict, total=False): + card_id: Required[Annotated[str, PropertyInfo(alias="cardID")]] + + + transformed = transform({"card_id": ""}, Params) + # {'cardID': ''} + ``` + + Any keys / data that does not have type information given will be included as is. + + It should be noted that the transformations that this function does are not represented in the type system. + """ + transformed = _transform_recursive(data, annotation=cast(type, expected_type)) + return cast(_T, transformed) + + +@lru_cache(maxsize=8096) +def _get_annotated_type(type_: type) -> type | None: + """If the given type is an `Annotated` type then it is returned, if not `None` is returned. + + This also unwraps the type when applicable, e.g. `Required[Annotated[T, ...]]` + """ + if is_required_type(type_): + # Unwrap `Required[Annotated[T, ...]]` to `Annotated[T, ...]` + type_ = get_args(type_)[0] + + if is_annotated_type(type_): + return type_ + + return None + + +def _maybe_transform_key(key: str, type_: type) -> str: + """Transform the given `data` based on the annotations provided in `type_`. + + Note: this function only looks at `Annotated` types that contain `PropertyInfo` metadata. + """ + annotated_type = _get_annotated_type(type_) + if annotated_type is None: + # no `Annotated` definition for this type, no transformation needed + return key + + # ignore the first argument as it is the actual type + annotations = get_args(annotated_type)[1:] + for annotation in annotations: + if isinstance(annotation, PropertyInfo) and annotation.alias is not None: + return annotation.alias + + return key + + +def _no_transform_needed(annotation: type) -> bool: + return annotation == float or annotation == int + + +def _transform_recursive( + data: object, + *, + annotation: type, + inner_type: type | None = None, +) -> object: + """Transform the given data against the expected type. + + Args: + annotation: The direct type annotation given to the particular piece of data. + This may or may not be wrapped in metadata types, e.g. `Required[T]`, `Annotated[T, ...]` etc + + inner_type: If applicable, this is the "inside" type. This is useful in certain cases where the outside type + is a container type such as `List[T]`. In that case `inner_type` should be set to `T` so that each entry in + the list can be transformed using the metadata from the container type. + + Defaults to the same value as the `annotation` argument. + """ + from .._compat import model_dump + + if inner_type is None: + inner_type = annotation + + stripped_type = strip_annotated_type(inner_type) + origin = get_origin(stripped_type) or stripped_type + if is_typeddict(stripped_type) and is_mapping(data): + return _transform_typeddict(data, stripped_type) + + if origin == dict and is_mapping(data): + items_type = get_args(stripped_type)[1] + return {key: _transform_recursive(value, annotation=items_type) for key, value in data.items()} + + if ( + # List[T] + (is_list_type(stripped_type) and is_list(data)) + # Iterable[T] + or (is_iterable_type(stripped_type) and is_iterable(data) and not isinstance(data, str)) + # Sequence[T] + or (is_sequence_type(stripped_type) and is_sequence(data) and not isinstance(data, str)) + ): + # dicts are technically iterable, but it is an iterable on the keys of the dict and is not usually + # intended as an iterable, so we don't transform it. + if isinstance(data, dict): + return cast(object, data) + + inner_type = extract_type_arg(stripped_type, 0) + if _no_transform_needed(inner_type): + # for some types there is no need to transform anything, so we can get a small + # perf boost from skipping that work. + # + # but we still need to convert to a list to ensure the data is json-serializable + if is_list(data): + return data + return list(data) + + return [_transform_recursive(d, annotation=annotation, inner_type=inner_type) for d in data] + + if is_union_type(stripped_type): + # For union types we run the transformation against all subtypes to ensure that everything is transformed. + # + # TODO: there may be edge cases where the same normalized field name will transform to two different names + # in different subtypes. + for subtype in get_args(stripped_type): + data = _transform_recursive(data, annotation=annotation, inner_type=subtype) + return data + + if isinstance(data, pydantic.BaseModel): + return model_dump(data, exclude_unset=True, mode="json") + + annotated_type = _get_annotated_type(annotation) + if annotated_type is None: + return data + + # ignore the first argument as it is the actual type + annotations = get_args(annotated_type)[1:] + for annotation in annotations: + if isinstance(annotation, PropertyInfo) and annotation.format is not None: + return _format_data(data, annotation.format, annotation.format_template) + + return data + + +def _format_data(data: object, format_: PropertyFormat, format_template: str | None) -> object: + if isinstance(data, (date, datetime)): + if format_ == "iso8601": + return data.isoformat() + + if format_ == "custom" and format_template is not None: + return data.strftime(format_template) + + if format_ == "base64" and is_base64_file_input(data): + binary: str | bytes | None = None + + if isinstance(data, pathlib.Path): + binary = data.read_bytes() + elif isinstance(data, io.IOBase): + binary = data.read() + + if isinstance(binary, str): # type: ignore[unreachable] + binary = binary.encode() + + if not isinstance(binary, bytes): + raise RuntimeError(f"Could not read bytes from {data}; Received {type(binary)}") + + return base64.b64encode(binary).decode("ascii") + + return data + + +def _transform_typeddict( + data: Mapping[str, object], + expected_type: type, +) -> Mapping[str, object]: + result: dict[str, object] = {} + annotations = get_type_hints(expected_type, include_extras=True) + for key, value in data.items(): + if not is_given(value): + # we don't need to include omitted values here as they'll + # be stripped out before the request is sent anyway + continue + + type_ = annotations.get(key) + if type_ is None: + # we do not have a type annotation for this field, leave it as is + result[key] = value + else: + result[_maybe_transform_key(key, type_)] = _transform_recursive(value, annotation=type_) + return result + + +async def async_maybe_transform( + data: object, + expected_type: object, +) -> Any | None: + """Wrapper over `async_transform()` that allows `None` to be passed. + + See `async_transform()` for more details. + """ + if data is None: + return None + return await async_transform(data, expected_type) + + +async def async_transform( + data: _T, + expected_type: object, +) -> _T: + """Transform dictionaries based off of type information from the given type, for example: + + ```py + class Params(TypedDict, total=False): + card_id: Required[Annotated[str, PropertyInfo(alias="cardID")]] + + + transformed = transform({"card_id": ""}, Params) + # {'cardID': ''} + ``` + + Any keys / data that does not have type information given will be included as is. + + It should be noted that the transformations that this function does are not represented in the type system. + """ + transformed = await _async_transform_recursive(data, annotation=cast(type, expected_type)) + return cast(_T, transformed) + + +async def _async_transform_recursive( + data: object, + *, + annotation: type, + inner_type: type | None = None, +) -> object: + """Transform the given data against the expected type. + + Args: + annotation: The direct type annotation given to the particular piece of data. + This may or may not be wrapped in metadata types, e.g. `Required[T]`, `Annotated[T, ...]` etc + + inner_type: If applicable, this is the "inside" type. This is useful in certain cases where the outside type + is a container type such as `List[T]`. In that case `inner_type` should be set to `T` so that each entry in + the list can be transformed using the metadata from the container type. + + Defaults to the same value as the `annotation` argument. + """ + from .._compat import model_dump + + if inner_type is None: + inner_type = annotation + + stripped_type = strip_annotated_type(inner_type) + origin = get_origin(stripped_type) or stripped_type + if is_typeddict(stripped_type) and is_mapping(data): + return await _async_transform_typeddict(data, stripped_type) + + if origin == dict and is_mapping(data): + items_type = get_args(stripped_type)[1] + return {key: _transform_recursive(value, annotation=items_type) for key, value in data.items()} + + if ( + # List[T] + (is_list_type(stripped_type) and is_list(data)) + # Iterable[T] + or (is_iterable_type(stripped_type) and is_iterable(data) and not isinstance(data, str)) + # Sequence[T] + or (is_sequence_type(stripped_type) and is_sequence(data) and not isinstance(data, str)) + ): + # dicts are technically iterable, but it is an iterable on the keys of the dict and is not usually + # intended as an iterable, so we don't transform it. + if isinstance(data, dict): + return cast(object, data) + + inner_type = extract_type_arg(stripped_type, 0) + if _no_transform_needed(inner_type): + # for some types there is no need to transform anything, so we can get a small + # perf boost from skipping that work. + # + # but we still need to convert to a list to ensure the data is json-serializable + if is_list(data): + return data + return list(data) + + return [await _async_transform_recursive(d, annotation=annotation, inner_type=inner_type) for d in data] + + if is_union_type(stripped_type): + # For union types we run the transformation against all subtypes to ensure that everything is transformed. + # + # TODO: there may be edge cases where the same normalized field name will transform to two different names + # in different subtypes. + for subtype in get_args(stripped_type): + data = await _async_transform_recursive(data, annotation=annotation, inner_type=subtype) + return data + + if isinstance(data, pydantic.BaseModel): + return model_dump(data, exclude_unset=True, mode="json") + + annotated_type = _get_annotated_type(annotation) + if annotated_type is None: + return data + + # ignore the first argument as it is the actual type + annotations = get_args(annotated_type)[1:] + for annotation in annotations: + if isinstance(annotation, PropertyInfo) and annotation.format is not None: + return await _async_format_data(data, annotation.format, annotation.format_template) + + return data + + +async def _async_format_data(data: object, format_: PropertyFormat, format_template: str | None) -> object: + if isinstance(data, (date, datetime)): + if format_ == "iso8601": + return data.isoformat() + + if format_ == "custom" and format_template is not None: + return data.strftime(format_template) + + if format_ == "base64" and is_base64_file_input(data): + binary: str | bytes | None = None + + if isinstance(data, pathlib.Path): + binary = await anyio.Path(data).read_bytes() + elif isinstance(data, io.IOBase): + binary = data.read() + + if isinstance(binary, str): # type: ignore[unreachable] + binary = binary.encode() + + if not isinstance(binary, bytes): + raise RuntimeError(f"Could not read bytes from {data}; Received {type(binary)}") + + return base64.b64encode(binary).decode("ascii") + + return data + + +async def _async_transform_typeddict( + data: Mapping[str, object], + expected_type: type, +) -> Mapping[str, object]: + result: dict[str, object] = {} + annotations = get_type_hints(expected_type, include_extras=True) + for key, value in data.items(): + if not is_given(value): + # we don't need to include omitted values here as they'll + # be stripped out before the request is sent anyway + continue + + type_ = annotations.get(key) + if type_ is None: + # we do not have a type annotation for this field, leave it as is + result[key] = value + else: + result[_maybe_transform_key(key, type_)] = await _async_transform_recursive(value, annotation=type_) + return result + + +@lru_cache(maxsize=8096) +def get_type_hints( + obj: Any, + globalns: dict[str, Any] | None = None, + localns: Mapping[str, Any] | None = None, + include_extras: bool = False, +) -> dict[str, Any]: + return _get_type_hints(obj, globalns=globalns, localns=localns, include_extras=include_extras) diff --git a/src/brapi/_utils/_typing.py b/src/brapi/_utils/_typing.py new file mode 100644 index 0000000..193109f --- /dev/null +++ b/src/brapi/_utils/_typing.py @@ -0,0 +1,156 @@ +from __future__ import annotations + +import sys +import typing +import typing_extensions +from typing import Any, TypeVar, Iterable, cast +from collections import abc as _c_abc +from typing_extensions import ( + TypeIs, + Required, + Annotated, + get_args, + get_origin, +) + +from ._utils import lru_cache +from .._types import InheritsGeneric +from ._compat import is_union as _is_union + + +def is_annotated_type(typ: type) -> bool: + return get_origin(typ) == Annotated + + +def is_list_type(typ: type) -> bool: + return (get_origin(typ) or typ) == list + + +def is_sequence_type(typ: type) -> bool: + origin = get_origin(typ) or typ + return origin == typing_extensions.Sequence or origin == typing.Sequence or origin == _c_abc.Sequence + + +def is_iterable_type(typ: type) -> bool: + """If the given type is `typing.Iterable[T]`""" + origin = get_origin(typ) or typ + return origin == Iterable or origin == _c_abc.Iterable + + +def is_union_type(typ: type) -> bool: + return _is_union(get_origin(typ)) + + +def is_required_type(typ: type) -> bool: + return get_origin(typ) == Required + + +def is_typevar(typ: type) -> bool: + # type ignore is required because type checkers + # think this expression will always return False + return type(typ) == TypeVar # type: ignore + + +_TYPE_ALIAS_TYPES: tuple[type[typing_extensions.TypeAliasType], ...] = (typing_extensions.TypeAliasType,) +if sys.version_info >= (3, 12): + _TYPE_ALIAS_TYPES = (*_TYPE_ALIAS_TYPES, typing.TypeAliasType) + + +def is_type_alias_type(tp: Any, /) -> TypeIs[typing_extensions.TypeAliasType]: + """Return whether the provided argument is an instance of `TypeAliasType`. + + ```python + type Int = int + is_type_alias_type(Int) + # > True + Str = TypeAliasType("Str", str) + is_type_alias_type(Str) + # > True + ``` + """ + return isinstance(tp, _TYPE_ALIAS_TYPES) + + +# Extracts T from Annotated[T, ...] or from Required[Annotated[T, ...]] +@lru_cache(maxsize=8096) +def strip_annotated_type(typ: type) -> type: + if is_required_type(typ) or is_annotated_type(typ): + return strip_annotated_type(cast(type, get_args(typ)[0])) + + return typ + + +def extract_type_arg(typ: type, index: int) -> type: + args = get_args(typ) + try: + return cast(type, args[index]) + except IndexError as err: + raise RuntimeError(f"Expected type {typ} to have a type argument at index {index} but it did not") from err + + +def extract_type_var_from_base( + typ: type, + *, + generic_bases: tuple[type, ...], + index: int, + failure_message: str | None = None, +) -> type: + """Given a type like `Foo[T]`, returns the generic type variable `T`. + + This also handles the case where a concrete subclass is given, e.g. + ```py + class MyResponse(Foo[bytes]): + ... + + extract_type_var(MyResponse, bases=(Foo,), index=0) -> bytes + ``` + + And where a generic subclass is given: + ```py + _T = TypeVar('_T') + class MyResponse(Foo[_T]): + ... + + extract_type_var(MyResponse[bytes], bases=(Foo,), index=0) -> bytes + ``` + """ + cls = cast(object, get_origin(typ) or typ) + if cls in generic_bases: # pyright: ignore[reportUnnecessaryContains] + # we're given the class directly + return extract_type_arg(typ, index) + + # if a subclass is given + # --- + # this is needed as __orig_bases__ is not present in the typeshed stubs + # because it is intended to be for internal use only, however there does + # not seem to be a way to resolve generic TypeVars for inherited subclasses + # without using it. + if isinstance(cls, InheritsGeneric): + target_base_class: Any | None = None + for base in cls.__orig_bases__: + if base.__origin__ in generic_bases: + target_base_class = base + break + + if target_base_class is None: + raise RuntimeError( + "Could not find the generic base class;\n" + "This should never happen;\n" + f"Does {cls} inherit from one of {generic_bases} ?" + ) + + extracted = extract_type_arg(target_base_class, index) + if is_typevar(extracted): + # If the extracted type argument is itself a type variable + # then that means the subclass itself is generic, so we have + # to resolve the type argument from the class itself, not + # the base class. + # + # Note: if there is more than 1 type argument, the subclass could + # change the ordering of the type arguments, this is not currently + # supported. + return extract_type_arg(typ, index) + + return extracted + + raise RuntimeError(failure_message or f"Could not resolve inner type variable at index {index} for {typ}") diff --git a/src/brapi/_utils/_utils.py b/src/brapi/_utils/_utils.py new file mode 100644 index 0000000..50d5926 --- /dev/null +++ b/src/brapi/_utils/_utils.py @@ -0,0 +1,421 @@ +from __future__ import annotations + +import os +import re +import inspect +import functools +from typing import ( + Any, + Tuple, + Mapping, + TypeVar, + Callable, + Iterable, + Sequence, + cast, + overload, +) +from pathlib import Path +from datetime import date, datetime +from typing_extensions import TypeGuard + +import sniffio + +from .._types import Omit, NotGiven, FileTypes, HeadersLike + +_T = TypeVar("_T") +_TupleT = TypeVar("_TupleT", bound=Tuple[object, ...]) +_MappingT = TypeVar("_MappingT", bound=Mapping[str, object]) +_SequenceT = TypeVar("_SequenceT", bound=Sequence[object]) +CallableT = TypeVar("CallableT", bound=Callable[..., Any]) + + +def flatten(t: Iterable[Iterable[_T]]) -> list[_T]: + return [item for sublist in t for item in sublist] + + +def extract_files( + # TODO: this needs to take Dict but variance issues..... + # create protocol type ? + query: Mapping[str, object], + *, + paths: Sequence[Sequence[str]], +) -> list[tuple[str, FileTypes]]: + """Recursively extract files from the given dictionary based on specified paths. + + A path may look like this ['foo', 'files', '', 'data']. + + Note: this mutates the given dictionary. + """ + files: list[tuple[str, FileTypes]] = [] + for path in paths: + files.extend(_extract_items(query, path, index=0, flattened_key=None)) + return files + + +def _extract_items( + obj: object, + path: Sequence[str], + *, + index: int, + flattened_key: str | None, +) -> list[tuple[str, FileTypes]]: + try: + key = path[index] + except IndexError: + if not is_given(obj): + # no value was provided - we can safely ignore + return [] + + # cyclical import + from .._files import assert_is_file_content + + # We have exhausted the path, return the entry we found. + assert flattened_key is not None + + if is_list(obj): + files: list[tuple[str, FileTypes]] = [] + for entry in obj: + assert_is_file_content(entry, key=flattened_key + "[]" if flattened_key else "") + files.append((flattened_key + "[]", cast(FileTypes, entry))) + return files + + assert_is_file_content(obj, key=flattened_key) + return [(flattened_key, cast(FileTypes, obj))] + + index += 1 + if is_dict(obj): + try: + # We are at the last entry in the path so we must remove the field + if (len(path)) == index: + item = obj.pop(key) + else: + item = obj[key] + except KeyError: + # Key was not present in the dictionary, this is not indicative of an error + # as the given path may not point to a required field. We also do not want + # to enforce required fields as the API may differ from the spec in some cases. + return [] + if flattened_key is None: + flattened_key = key + else: + flattened_key += f"[{key}]" + return _extract_items( + item, + path, + index=index, + flattened_key=flattened_key, + ) + elif is_list(obj): + if key != "": + return [] + + return flatten( + [ + _extract_items( + item, + path, + index=index, + flattened_key=flattened_key + "[]" if flattened_key is not None else "[]", + ) + for item in obj + ] + ) + + # Something unexpected was passed, just ignore it. + return [] + + +def is_given(obj: _T | NotGiven | Omit) -> TypeGuard[_T]: + return not isinstance(obj, NotGiven) and not isinstance(obj, Omit) + + +# Type safe methods for narrowing types with TypeVars. +# The default narrowing for isinstance(obj, dict) is dict[unknown, unknown], +# however this cause Pyright to rightfully report errors. As we know we don't +# care about the contained types we can safely use `object` in it's place. +# +# There are two separate functions defined, `is_*` and `is_*_t` for different use cases. +# `is_*` is for when you're dealing with an unknown input +# `is_*_t` is for when you're narrowing a known union type to a specific subset + + +def is_tuple(obj: object) -> TypeGuard[tuple[object, ...]]: + return isinstance(obj, tuple) + + +def is_tuple_t(obj: _TupleT | object) -> TypeGuard[_TupleT]: + return isinstance(obj, tuple) + + +def is_sequence(obj: object) -> TypeGuard[Sequence[object]]: + return isinstance(obj, Sequence) + + +def is_sequence_t(obj: _SequenceT | object) -> TypeGuard[_SequenceT]: + return isinstance(obj, Sequence) + + +def is_mapping(obj: object) -> TypeGuard[Mapping[str, object]]: + return isinstance(obj, Mapping) + + +def is_mapping_t(obj: _MappingT | object) -> TypeGuard[_MappingT]: + return isinstance(obj, Mapping) + + +def is_dict(obj: object) -> TypeGuard[dict[object, object]]: + return isinstance(obj, dict) + + +def is_list(obj: object) -> TypeGuard[list[object]]: + return isinstance(obj, list) + + +def is_iterable(obj: object) -> TypeGuard[Iterable[object]]: + return isinstance(obj, Iterable) + + +def deepcopy_minimal(item: _T) -> _T: + """Minimal reimplementation of copy.deepcopy() that will only copy certain object types: + + - mappings, e.g. `dict` + - list + + This is done for performance reasons. + """ + if is_mapping(item): + return cast(_T, {k: deepcopy_minimal(v) for k, v in item.items()}) + if is_list(item): + return cast(_T, [deepcopy_minimal(entry) for entry in item]) + return item + + +# copied from https://github.com/Rapptz/RoboDanny +def human_join(seq: Sequence[str], *, delim: str = ", ", final: str = "or") -> str: + size = len(seq) + if size == 0: + return "" + + if size == 1: + return seq[0] + + if size == 2: + return f"{seq[0]} {final} {seq[1]}" + + return delim.join(seq[:-1]) + f" {final} {seq[-1]}" + + +def quote(string: str) -> str: + """Add single quotation marks around the given string. Does *not* do any escaping.""" + return f"'{string}'" + + +def required_args(*variants: Sequence[str]) -> Callable[[CallableT], CallableT]: + """Decorator to enforce a given set of arguments or variants of arguments are passed to the decorated function. + + Useful for enforcing runtime validation of overloaded functions. + + Example usage: + ```py + @overload + def foo(*, a: str) -> str: ... + + + @overload + def foo(*, b: bool) -> str: ... + + + # This enforces the same constraints that a static type checker would + # i.e. that either a or b must be passed to the function + @required_args(["a"], ["b"]) + def foo(*, a: str | None = None, b: bool | None = None) -> str: ... + ``` + """ + + def inner(func: CallableT) -> CallableT: + params = inspect.signature(func).parameters + positional = [ + name + for name, param in params.items() + if param.kind + in { + param.POSITIONAL_ONLY, + param.POSITIONAL_OR_KEYWORD, + } + ] + + @functools.wraps(func) + def wrapper(*args: object, **kwargs: object) -> object: + given_params: set[str] = set() + for i, _ in enumerate(args): + try: + given_params.add(positional[i]) + except IndexError: + raise TypeError( + f"{func.__name__}() takes {len(positional)} argument(s) but {len(args)} were given" + ) from None + + for key in kwargs.keys(): + given_params.add(key) + + for variant in variants: + matches = all((param in given_params for param in variant)) + if matches: + break + else: # no break + if len(variants) > 1: + variations = human_join( + ["(" + human_join([quote(arg) for arg in variant], final="and") + ")" for variant in variants] + ) + msg = f"Missing required arguments; Expected either {variations} arguments to be given" + else: + assert len(variants) > 0 + + # TODO: this error message is not deterministic + missing = list(set(variants[0]) - given_params) + if len(missing) > 1: + msg = f"Missing required arguments: {human_join([quote(arg) for arg in missing])}" + else: + msg = f"Missing required argument: {quote(missing[0])}" + raise TypeError(msg) + return func(*args, **kwargs) + + return wrapper # type: ignore + + return inner + + +_K = TypeVar("_K") +_V = TypeVar("_V") + + +@overload +def strip_not_given(obj: None) -> None: ... + + +@overload +def strip_not_given(obj: Mapping[_K, _V | NotGiven]) -> dict[_K, _V]: ... + + +@overload +def strip_not_given(obj: object) -> object: ... + + +def strip_not_given(obj: object | None) -> object: + """Remove all top-level keys where their values are instances of `NotGiven`""" + if obj is None: + return None + + if not is_mapping(obj): + return obj + + return {key: value for key, value in obj.items() if not isinstance(value, NotGiven)} + + +def coerce_integer(val: str) -> int: + return int(val, base=10) + + +def coerce_float(val: str) -> float: + return float(val) + + +def coerce_boolean(val: str) -> bool: + return val == "true" or val == "1" or val == "on" + + +def maybe_coerce_integer(val: str | None) -> int | None: + if val is None: + return None + return coerce_integer(val) + + +def maybe_coerce_float(val: str | None) -> float | None: + if val is None: + return None + return coerce_float(val) + + +def maybe_coerce_boolean(val: str | None) -> bool | None: + if val is None: + return None + return coerce_boolean(val) + + +def removeprefix(string: str, prefix: str) -> str: + """Remove a prefix from a string. + + Backport of `str.removeprefix` for Python < 3.9 + """ + if string.startswith(prefix): + return string[len(prefix) :] + return string + + +def removesuffix(string: str, suffix: str) -> str: + """Remove a suffix from a string. + + Backport of `str.removesuffix` for Python < 3.9 + """ + if string.endswith(suffix): + return string[: -len(suffix)] + return string + + +def file_from_path(path: str) -> FileTypes: + contents = Path(path).read_bytes() + file_name = os.path.basename(path) + return (file_name, contents) + + +def get_required_header(headers: HeadersLike, header: str) -> str: + lower_header = header.lower() + if is_mapping_t(headers): + # mypy doesn't understand the type narrowing here + for k, v in headers.items(): # type: ignore + if k.lower() == lower_header and isinstance(v, str): + return v + + # to deal with the case where the header looks like Stainless-Event-Id + intercaps_header = re.sub(r"([^\w])(\w)", lambda pat: pat.group(1) + pat.group(2).upper(), header.capitalize()) + + for normalized_header in [header, lower_header, header.upper(), intercaps_header]: + value = headers.get(normalized_header) + if value: + return value + + raise ValueError(f"Could not find {header} header") + + +def get_async_library() -> str: + try: + return sniffio.current_async_library() + except Exception: + return "false" + + +def lru_cache(*, maxsize: int | None = 128) -> Callable[[CallableT], CallableT]: + """A version of functools.lru_cache that retains the type signature + for the wrapped function arguments. + """ + wrapper = functools.lru_cache( # noqa: TID251 + maxsize=maxsize, + ) + return cast(Any, wrapper) # type: ignore[no-any-return] + + +def json_safe(data: object) -> object: + """Translates a mapping / sequence recursively in the same fashion + as `pydantic` v2's `model_dump(mode="json")`. + """ + if is_mapping(data): + return {json_safe(key): json_safe(value) for key, value in data.items()} + + if is_iterable(data) and not isinstance(data, (str, bytes, bytearray)): + return [json_safe(item) for item in data] + + if isinstance(data, (datetime, date)): + return data.isoformat() + + return data diff --git a/src/brapi/_version.py b/src/brapi/_version.py new file mode 100644 index 0000000..905eebf --- /dev/null +++ b/src/brapi/_version.py @@ -0,0 +1,4 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +__title__ = "brapi" +__version__ = "0.0.1" diff --git a/src/brapi/lib/.keep b/src/brapi/lib/.keep new file mode 100644 index 0000000..5e2c99f --- /dev/null +++ b/src/brapi/lib/.keep @@ -0,0 +1,4 @@ +File generated from our OpenAPI spec by Stainless. + +This directory can be used to store custom files to expand the SDK. +It is ignored by Stainless code generation and its content (other than this keep file) won't be touched. \ No newline at end of file diff --git a/src/brapi/py.typed b/src/brapi/py.typed new file mode 100644 index 0000000..e69de29 diff --git a/src/brapi/resources/__init__.py b/src/brapi/resources/__init__.py new file mode 100644 index 0000000..a4d2d01 --- /dev/null +++ b/src/brapi/resources/__init__.py @@ -0,0 +1,47 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from .v2 import ( + V2Resource, + AsyncV2Resource, + V2ResourceWithRawResponse, + AsyncV2ResourceWithRawResponse, + V2ResourceWithStreamingResponse, + AsyncV2ResourceWithStreamingResponse, +) +from .quote import ( + QuoteResource, + AsyncQuoteResource, + QuoteResourceWithRawResponse, + AsyncQuoteResourceWithRawResponse, + QuoteResourceWithStreamingResponse, + AsyncQuoteResourceWithStreamingResponse, +) +from .available import ( + AvailableResource, + AsyncAvailableResource, + AvailableResourceWithRawResponse, + AsyncAvailableResourceWithRawResponse, + AvailableResourceWithStreamingResponse, + AsyncAvailableResourceWithStreamingResponse, +) + +__all__ = [ + "QuoteResource", + "AsyncQuoteResource", + "QuoteResourceWithRawResponse", + "AsyncQuoteResourceWithRawResponse", + "QuoteResourceWithStreamingResponse", + "AsyncQuoteResourceWithStreamingResponse", + "AvailableResource", + "AsyncAvailableResource", + "AvailableResourceWithRawResponse", + "AsyncAvailableResourceWithRawResponse", + "AvailableResourceWithStreamingResponse", + "AsyncAvailableResourceWithStreamingResponse", + "V2Resource", + "AsyncV2Resource", + "V2ResourceWithRawResponse", + "AsyncV2ResourceWithRawResponse", + "V2ResourceWithStreamingResponse", + "AsyncV2ResourceWithStreamingResponse", +] diff --git a/src/brapi/resources/available.py b/src/brapi/resources/available.py new file mode 100644 index 0000000..7c65696 --- /dev/null +++ b/src/brapi/resources/available.py @@ -0,0 +1,283 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import httpx + +from ..types import available_list_params +from .._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from .._utils import maybe_transform, async_maybe_transform +from .._compat import cached_property +from .._resource import SyncAPIResource, AsyncAPIResource +from .._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from .._base_client import make_request_options +from ..types.available_list_response import AvailableListResponse + +__all__ = ["AvailableResource", "AsyncAvailableResource"] + + +class AvailableResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> AvailableResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AvailableResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AvailableResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AvailableResourceWithStreamingResponse(self) + + def list( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> AvailableListResponse: + """ + Obtenha uma lista completa de todos os tickers (identificadores) de ativos + financeiros (ações, FIIs, BDRs, ETFs, índices) que a API Brapi tem dados + disponíveis para consulta no endpoint `/api/quote/{tickers}`. + + ### Funcionalidade: + + - Retorna arrays separados para `indexes` (índices) e `stocks` (outros ativos). + - Pode ser filtrado usando o parâmetro `search` para encontrar tickers + específicos. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todos os tickers disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/available?token=SEU_TOKEN" + ``` + + **Buscar tickers que contenham 'BBDC':** + + ```bash + curl -X GET "https://brapi.dev/api/available?search=BBDC&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com duas chaves: + + - `indexes`: Array de strings contendo os tickers dos índices disponíveis (ex: + `["^BVSP", "^IFIX"]`). + - `stocks`: Array de strings contendo os tickers das ações, FIIs, BDRs e ETFs + disponíveis (ex: `["PETR4", "VALE3", "ITSA4", "MXRF11"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de tickers (correspondência parcial, + case-insensitive). Se omitido, retorna todos os tickers. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "search": search, + }, + available_list_params.AvailableListParams, + ), + ), + cast_to=AvailableListResponse, + ) + + +class AsyncAvailableResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncAvailableResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncAvailableResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncAvailableResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncAvailableResourceWithStreamingResponse(self) + + async def list( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> AvailableListResponse: + """ + Obtenha uma lista completa de todos os tickers (identificadores) de ativos + financeiros (ações, FIIs, BDRs, ETFs, índices) que a API Brapi tem dados + disponíveis para consulta no endpoint `/api/quote/{tickers}`. + + ### Funcionalidade: + + - Retorna arrays separados para `indexes` (índices) e `stocks` (outros ativos). + - Pode ser filtrado usando o parâmetro `search` para encontrar tickers + específicos. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todos os tickers disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/available?token=SEU_TOKEN" + ``` + + **Buscar tickers que contenham 'BBDC':** + + ```bash + curl -X GET "https://brapi.dev/api/available?search=BBDC&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com duas chaves: + + - `indexes`: Array de strings contendo os tickers dos índices disponíveis (ex: + `["^BVSP", "^IFIX"]`). + - `stocks`: Array de strings contendo os tickers das ações, FIIs, BDRs e ETFs + disponíveis (ex: `["PETR4", "VALE3", "ITSA4", "MXRF11"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de tickers (correspondência parcial, + case-insensitive). Se omitido, retorna todos os tickers. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "search": search, + }, + available_list_params.AvailableListParams, + ), + ), + cast_to=AvailableListResponse, + ) + + +class AvailableResourceWithRawResponse: + def __init__(self, available: AvailableResource) -> None: + self._available = available + + self.list = to_raw_response_wrapper( + available.list, + ) + + +class AsyncAvailableResourceWithRawResponse: + def __init__(self, available: AsyncAvailableResource) -> None: + self._available = available + + self.list = async_to_raw_response_wrapper( + available.list, + ) + + +class AvailableResourceWithStreamingResponse: + def __init__(self, available: AvailableResource) -> None: + self._available = available + + self.list = to_streamed_response_wrapper( + available.list, + ) + + +class AsyncAvailableResourceWithStreamingResponse: + def __init__(self, available: AsyncAvailableResource) -> None: + self._available = available + + self.list = async_to_streamed_response_wrapper( + available.list, + ) diff --git a/src/brapi/resources/quote.py b/src/brapi/resources/quote.py new file mode 100644 index 0000000..48da3f8 --- /dev/null +++ b/src/brapi/resources/quote.py @@ -0,0 +1,959 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import List +from typing_extensions import Literal + +import httpx + +from ..types import quote_list_params, quote_retrieve_params +from .._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from .._utils import maybe_transform, async_maybe_transform +from .._compat import cached_property +from .._resource import SyncAPIResource, AsyncAPIResource +from .._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from .._base_client import make_request_options +from ..types.quote_list_response import QuoteListResponse +from ..types.quote_retrieve_response import QuoteRetrieveResponse + +__all__ = ["QuoteResource", "AsyncQuoteResource"] + + +class QuoteResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> QuoteResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return QuoteResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> QuoteResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return QuoteResourceWithStreamingResponse(self) + + def retrieve( + self, + tickers: str, + *, + token: str | Omit = omit, + dividends: bool | Omit = omit, + fundamental: bool | Omit = omit, + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + | Omit = omit, + modules: List[ + Literal[ + "summaryProfile", + "balanceSheetHistory", + "defaultKeyStatistics", + "balanceSheetHistoryQuarterly", + "incomeStatementHistory", + "incomeStatementHistoryQuarterly", + "financialData", + "financialDataHistory", + "financialDataHistoryQuarterly", + "defaultKeyStatisticsHistory", + "defaultKeyStatisticsHistoryQuarterly", + "valueAddedHistory", + "valueAddedHistoryQuarterly", + "cashflowHistory", + "cashflowHistoryQuarterly", + ] + ] + | Omit = omit, + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> QuoteRetrieveResponse: + """ + Este endpoint é a principal forma de obter informações detalhadas sobre um ou + mais ativos financeiros (ações, FIIs, ETFs, BDRs, índices) listados na B3, + identificados pelos seus respectivos **tickers**. + + ### Funcionalidades Principais: + + - **Cotação Atual:** Retorna o preço mais recente, variação diária, máximas, + mínimas, volume, etc. + - **Dados Históricos:** Permite solicitar séries históricas de preços usando os + parâmetros `range` e `interval`. + - **Dados Fundamentalistas:** Opcionalmente, inclui dados fundamentalistas + básicos (P/L, LPA) com o parâmetro `fundamental=true`. + - **Dividendos:** Opcionalmente, inclui histórico de dividendos e JCP com + `dividends=true`. + - **Módulos Adicionais:** Permite requisitar conjuntos de dados financeiros mais + aprofundados através do parâmetro `modules` (veja detalhes abaixo). + + ### 🧪 Ações de Teste (Sem Autenticação): + + Para facilitar o desenvolvimento e teste, as seguintes **4 ações têm acesso + irrestrito** e **não requerem autenticação**: + + - **PETR4** (Petrobras PN) + - **MGLU3** (Magazine Luiza ON) + - **VALE3** (Vale ON) + - **ITUB4** (Itaú Unibanco PN) + + **Importante:** Você pode consultar essas ações sem token e com acesso a todos + os recursos (históricos, módulos, dividendos). Porém, se misturar essas ações + com outras na mesma requisição, a autenticação será obrigatória. + + ### Autenticação: + + Para **outras ações** (além das 4 de teste), é **obrigatório** fornecer um token + de autenticação válido, seja via query parameter `token` ou via header + `Authorization: Bearer seu_token`. + + ### Exemplos de Requisição: + + **1. Cotação simples de PETR4 e VALE3 (ações de teste - sem token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/PETR4,VALE3" + ``` + + **2. Cotação de MGLU3 com dados históricos do último mês (ação de teste - sem + token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/MGLU3?range=1mo&interval=1d" + ``` + + **3. Cotação de ITUB4 incluindo dividendos e dados fundamentalistas (ação de + teste - sem token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/ITUB4?fundamental=true÷nds=true" + ``` + + **4. Cotação de WEGE3 com Resumo da Empresa e Balanço Patrimonial Anual (via + módulos - requer token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/WEGE3?modules=summaryProfile,balanceSheetHistory&token=SEU_TOKEN" + ``` + + **5. Exemplo de requisição mista (requer token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/PETR4,BBAS3?token=SEU_TOKEN" + ``` + + _Nota: Como BBAS3 não é uma ação de teste, toda a requisição requer + autenticação, mesmo contendo PETR4._ + + ### Parâmetro `modules` (Detalhado): + + O parâmetro `modules` é extremamente poderoso para enriquecer a resposta com + dados financeiros detalhados. Você pode solicitar um ou mais módulos, separados + por vírgula. + + **Módulos Disponíveis:** + + - `summaryProfile`: Informações cadastrais da empresa (endereço, setor, + descrição do negócio, website, número de funcionários). + - `balanceSheetHistory`: Histórico **anual** do Balanço Patrimonial. + - `balanceSheetHistoryQuarterly`: Histórico **trimestral** do Balanço + Patrimonial. + - `defaultKeyStatistics`: Principais estatísticas da empresa (Valor de Mercado, + P/L, ROE, Dividend Yield, etc.) - **TTM (Trailing Twelve Months)**. + - `defaultKeyStatisticsHistory`: Histórico **anual** das Principais + Estatísticas. + - `defaultKeyStatisticsHistoryQuarterly`: Histórico **trimestral** das + Principais Estatísticas. + - `incomeStatementHistory`: Histórico **anual** da Demonstração do Resultado do + Exercício (DRE). + - `incomeStatementHistoryQuarterly`: Histórico **trimestral** da Demonstração do + Resultado do Exercício (DRE). + - `financialData`: Dados financeiros selecionados (Receita, Lucro Bruto, EBITDA, + Dívida Líquida, Fluxo de Caixa Livre, Margens) - **TTM (Trailing Twelve + Months)**. + - `financialDataHistory`: Histórico **anual** dos Dados Financeiros. + - `financialDataHistoryQuarterly`: Histórico **trimestral** dos Dados + Financeiros. + - `valueAddedHistory`: Histórico **anual** da Demonstração do Valor Adicionado + (DVA). + - `valueAddedHistoryQuarterly`: Histórico **trimestral** da Demonstração do + Valor Adicionado (DVA). + - `cashflowHistory`: Histórico **anual** da Demonstração do Fluxo de Caixa + (DFC). + - `cashflowHistoryQuarterly`: Histórico **trimestral** da Demonstração do Fluxo + de Caixa (DFC). + + **Exemplo de Uso do `modules`:** + + Para obter a cotação de BBDC4 junto com seu DRE trimestral e Fluxo de Caixa + anual: + + ```bash + curl -X GET "https://brapi.dev/api/quote/BBDC4?modules=incomeStatementHistoryQuarterly,cashflowHistory&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON contendo a chave `results`, que é um array. Cada + elemento do array corresponde a um ticker solicitado e contém os dados da + cotação e os módulos adicionais requisitados. + + - **Sucesso (200 OK):** Retorna os dados conforme solicitado. + - **Bad Request (400 Bad Request):** Ocorre se um parâmetro for inválido (ex: + `range=invalid`) ou se a formatação estiver incorreta. + - **Unauthorized (401 Unauthorized):** Token inválido ou ausente. + - **Payment Required (402 Payment Required):** Limite de requisições do plano + atual excedido. + - **Not Found (404 Not Found):** Um ou mais tickers solicitados não foram + encontrados. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + dividends: **Opcional.** Booleano (`true` ou `false`). Se `true`, inclui informações sobre + dividendos e JCP (Juros sobre Capital Próprio) pagos historicamente pelo ativo + na chave `dividendsData`. + + fundamental: **Opcional.** Booleano (`true` ou `false`). Se `true`, inclui dados + fundamentalistas básicos na resposta, como Preço/Lucro (P/L) e Lucro Por Ação + (LPA). + + **Nota:** Para dados fundamentalistas mais completos, utilize o parâmetro + `modules`. + + interval: **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. + + **Valores Possíveis:** + + - `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`: Intervalos intraday + (minutos/horas). **Atenção:** Disponibilidade pode variar conforme o `range` e + o ativo. + - `1d`: Diário (padrão se `range` for especificado e `interval` omitido). + - `5d`: 5 dias. + - `1wk`: Semanal. + - `1mo`: Mensal. + - `3mo`: Trimestral. + + modules: **Opcional.** Uma lista de módulos de dados adicionais, separados por vírgula + (`,`), para incluir na resposta. Permite buscar dados financeiros detalhados. + + **Exemplos:** + + - `modules=summaryProfile` (retorna perfil da empresa) + - `modules=balanceSheetHistory,incomeStatementHistory` (retorna histórico anual + do BP e DRE) + + Veja a descrição principal do endpoint para a lista completa de módulos e seus + conteúdos. + + range: **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Se omitido, apenas a cotação mais recente é retornada + (a menos que `interval` seja usado). + + **Valores Possíveis:** + + - `1d`: Último dia de pregão (intraday se `interval` for minutos/horas). + - `5d`: Últimos 5 dias. + - `1mo`: Último mês. + - `3mo`: Últimos 3 meses. + - `6mo`: Últimos 6 meses. + - `1y`: Último ano. + - `2y`: Últimos 2 anos. + - `5y`: Últimos 5 anos. + - `10y`: Últimos 10 anos. + - `ytd`: Desde o início do ano atual (Year-to-Date). + - `max`: Todo o período histórico disponível. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + if not tickers: + raise ValueError(f"Expected a non-empty value for `tickers` but received {tickers!r}") + return self._get( + f"/api/quote/{tickers}", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "dividends": dividends, + "fundamental": fundamental, + "interval": interval, + "modules": modules, + "range": range, + }, + quote_retrieve_params.QuoteRetrieveParams, + ), + ), + cast_to=QuoteRetrieveResponse, + ) + + def list( + self, + *, + token: str | Omit = omit, + limit: int | Omit = omit, + page: int | Omit = omit, + search: str | Omit = omit, + sector: Literal[ + "Retail Trade", + "Energy Minerals", + "Health Services", + "Utilities", + "Finance", + "Consumer Services", + "Consumer Non-Durables", + "Non-Energy Minerals", + "Commercial Services", + "Distribution Services", + "Transportation", + "Technology Services", + "Process Industries", + "Communications", + "Producer Manufacturing", + "Miscellaneous", + "Electronic Technology", + "Industrial Services", + "Health Technology", + "Consumer Durables", + ] + | Omit = omit, + sort_by: Literal["name", "close", "change", "change_abs", "volume", "market_cap_basic", "sector"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + type: Literal["stock", "fund", "bdr"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> QuoteListResponse: + """ + Obtenha uma lista paginada de cotações de diversos ativos (ações, FIIs, BDRs) + negociados na B3, com opções avançadas de busca, filtragem e ordenação. + + ### Funcionalidades: + + - **Busca por Ticker:** Filtre por parte do ticker usando `search`. + - **Filtragem por Tipo:** Restrinja a lista a `stock`, `fund` (FII) ou `bdr` com + o parâmetro `type`. + - **Filtragem por Setor:** Selecione ativos de um setor específico usando + `sector`. + - **Ordenação:** Ordene os resultados por diversos campos (preço, variação, + volume, etc.) usando `sortBy` e `sortOrder`. + - **Paginação:** Controle o número de resultados por página (`limit`) e a página + desejada (`page`). + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar as 10 ações do setor Financeiro com maior volume, ordenadas de forma + decrescente:** + + ```bash + curl -X GET "https://brapi.dev/api/quote/list?sector=Finance&sortBy=volume&sortOrder=desc&limit=10&page=1&token=SEU_TOKEN" + ``` + + **Buscar por ativos cujo ticker contenha 'ITUB' e ordenar por nome ascendente:** + + ```bash + curl -X GET "https://brapi.dev/api/quote/list?search=ITUB&sortBy=name&sortOrder=asc&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém a lista de `stocks` (e `indexes` relevantes), informações + sobre os filtros aplicados, detalhes da paginação (`currentPage`, `totalPages`, + `itemsPerPage`, `totalCount`, `hasNextPage`) e listas de setores + (`availableSectors`) e tipos (`availableStockTypes`) disponíveis para filtragem. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + limit: **Opcional.** Número máximo de ativos a serem retornados por página. O valor + padrão pode variar. + + page: **Opcional.** Número da página dos resultados a ser retornada, considerando o + `limit` especificado. Começa em 1. + + search: + **Opcional.** Termo para buscar ativos por ticker (correspondência parcial). Ex: + `PETR` encontrará `PETR4`, `PETR3`. + + sector: **Opcional.** Filtra os resultados por setor de atuação da empresa. Utilize um + dos valores retornados em `availableSectors`. + + sort_by: **Opcional.** Campo pelo qual os resultados serão ordenados. + + sort_order: **Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + Requer que `sortBy` seja especificado. + + type: **Opcional.** Filtra os resultados por tipo de ativo. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/quote/list", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "limit": limit, + "page": page, + "search": search, + "sector": sector, + "sort_by": sort_by, + "sort_order": sort_order, + "type": type, + }, + quote_list_params.QuoteListParams, + ), + ), + cast_to=QuoteListResponse, + ) + + +class AsyncQuoteResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncQuoteResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncQuoteResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncQuoteResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncQuoteResourceWithStreamingResponse(self) + + async def retrieve( + self, + tickers: str, + *, + token: str | Omit = omit, + dividends: bool | Omit = omit, + fundamental: bool | Omit = omit, + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + | Omit = omit, + modules: List[ + Literal[ + "summaryProfile", + "balanceSheetHistory", + "defaultKeyStatistics", + "balanceSheetHistoryQuarterly", + "incomeStatementHistory", + "incomeStatementHistoryQuarterly", + "financialData", + "financialDataHistory", + "financialDataHistoryQuarterly", + "defaultKeyStatisticsHistory", + "defaultKeyStatisticsHistoryQuarterly", + "valueAddedHistory", + "valueAddedHistoryQuarterly", + "cashflowHistory", + "cashflowHistoryQuarterly", + ] + ] + | Omit = omit, + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> QuoteRetrieveResponse: + """ + Este endpoint é a principal forma de obter informações detalhadas sobre um ou + mais ativos financeiros (ações, FIIs, ETFs, BDRs, índices) listados na B3, + identificados pelos seus respectivos **tickers**. + + ### Funcionalidades Principais: + + - **Cotação Atual:** Retorna o preço mais recente, variação diária, máximas, + mínimas, volume, etc. + - **Dados Históricos:** Permite solicitar séries históricas de preços usando os + parâmetros `range` e `interval`. + - **Dados Fundamentalistas:** Opcionalmente, inclui dados fundamentalistas + básicos (P/L, LPA) com o parâmetro `fundamental=true`. + - **Dividendos:** Opcionalmente, inclui histórico de dividendos e JCP com + `dividends=true`. + - **Módulos Adicionais:** Permite requisitar conjuntos de dados financeiros mais + aprofundados através do parâmetro `modules` (veja detalhes abaixo). + + ### 🧪 Ações de Teste (Sem Autenticação): + + Para facilitar o desenvolvimento e teste, as seguintes **4 ações têm acesso + irrestrito** e **não requerem autenticação**: + + - **PETR4** (Petrobras PN) + - **MGLU3** (Magazine Luiza ON) + - **VALE3** (Vale ON) + - **ITUB4** (Itaú Unibanco PN) + + **Importante:** Você pode consultar essas ações sem token e com acesso a todos + os recursos (históricos, módulos, dividendos). Porém, se misturar essas ações + com outras na mesma requisição, a autenticação será obrigatória. + + ### Autenticação: + + Para **outras ações** (além das 4 de teste), é **obrigatório** fornecer um token + de autenticação válido, seja via query parameter `token` ou via header + `Authorization: Bearer seu_token`. + + ### Exemplos de Requisição: + + **1. Cotação simples de PETR4 e VALE3 (ações de teste - sem token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/PETR4,VALE3" + ``` + + **2. Cotação de MGLU3 com dados históricos do último mês (ação de teste - sem + token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/MGLU3?range=1mo&interval=1d" + ``` + + **3. Cotação de ITUB4 incluindo dividendos e dados fundamentalistas (ação de + teste - sem token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/ITUB4?fundamental=true÷nds=true" + ``` + + **4. Cotação de WEGE3 com Resumo da Empresa e Balanço Patrimonial Anual (via + módulos - requer token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/WEGE3?modules=summaryProfile,balanceSheetHistory&token=SEU_TOKEN" + ``` + + **5. Exemplo de requisição mista (requer token):** + + ```bash + curl -X GET "https://brapi.dev/api/quote/PETR4,BBAS3?token=SEU_TOKEN" + ``` + + _Nota: Como BBAS3 não é uma ação de teste, toda a requisição requer + autenticação, mesmo contendo PETR4._ + + ### Parâmetro `modules` (Detalhado): + + O parâmetro `modules` é extremamente poderoso para enriquecer a resposta com + dados financeiros detalhados. Você pode solicitar um ou mais módulos, separados + por vírgula. + + **Módulos Disponíveis:** + + - `summaryProfile`: Informações cadastrais da empresa (endereço, setor, + descrição do negócio, website, número de funcionários). + - `balanceSheetHistory`: Histórico **anual** do Balanço Patrimonial. + - `balanceSheetHistoryQuarterly`: Histórico **trimestral** do Balanço + Patrimonial. + - `defaultKeyStatistics`: Principais estatísticas da empresa (Valor de Mercado, + P/L, ROE, Dividend Yield, etc.) - **TTM (Trailing Twelve Months)**. + - `defaultKeyStatisticsHistory`: Histórico **anual** das Principais + Estatísticas. + - `defaultKeyStatisticsHistoryQuarterly`: Histórico **trimestral** das + Principais Estatísticas. + - `incomeStatementHistory`: Histórico **anual** da Demonstração do Resultado do + Exercício (DRE). + - `incomeStatementHistoryQuarterly`: Histórico **trimestral** da Demonstração do + Resultado do Exercício (DRE). + - `financialData`: Dados financeiros selecionados (Receita, Lucro Bruto, EBITDA, + Dívida Líquida, Fluxo de Caixa Livre, Margens) - **TTM (Trailing Twelve + Months)**. + - `financialDataHistory`: Histórico **anual** dos Dados Financeiros. + - `financialDataHistoryQuarterly`: Histórico **trimestral** dos Dados + Financeiros. + - `valueAddedHistory`: Histórico **anual** da Demonstração do Valor Adicionado + (DVA). + - `valueAddedHistoryQuarterly`: Histórico **trimestral** da Demonstração do + Valor Adicionado (DVA). + - `cashflowHistory`: Histórico **anual** da Demonstração do Fluxo de Caixa + (DFC). + - `cashflowHistoryQuarterly`: Histórico **trimestral** da Demonstração do Fluxo + de Caixa (DFC). + + **Exemplo de Uso do `modules`:** + + Para obter a cotação de BBDC4 junto com seu DRE trimestral e Fluxo de Caixa + anual: + + ```bash + curl -X GET "https://brapi.dev/api/quote/BBDC4?modules=incomeStatementHistoryQuarterly,cashflowHistory&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON contendo a chave `results`, que é um array. Cada + elemento do array corresponde a um ticker solicitado e contém os dados da + cotação e os módulos adicionais requisitados. + + - **Sucesso (200 OK):** Retorna os dados conforme solicitado. + - **Bad Request (400 Bad Request):** Ocorre se um parâmetro for inválido (ex: + `range=invalid`) ou se a formatação estiver incorreta. + - **Unauthorized (401 Unauthorized):** Token inválido ou ausente. + - **Payment Required (402 Payment Required):** Limite de requisições do plano + atual excedido. + - **Not Found (404 Not Found):** Um ou mais tickers solicitados não foram + encontrados. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + dividends: **Opcional.** Booleano (`true` ou `false`). Se `true`, inclui informações sobre + dividendos e JCP (Juros sobre Capital Próprio) pagos historicamente pelo ativo + na chave `dividendsData`. + + fundamental: **Opcional.** Booleano (`true` ou `false`). Se `true`, inclui dados + fundamentalistas básicos na resposta, como Preço/Lucro (P/L) e Lucro Por Ação + (LPA). + + **Nota:** Para dados fundamentalistas mais completos, utilize o parâmetro + `modules`. + + interval: **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. + + **Valores Possíveis:** + + - `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`: Intervalos intraday + (minutos/horas). **Atenção:** Disponibilidade pode variar conforme o `range` e + o ativo. + - `1d`: Diário (padrão se `range` for especificado e `interval` omitido). + - `5d`: 5 dias. + - `1wk`: Semanal. + - `1mo`: Mensal. + - `3mo`: Trimestral. + + modules: **Opcional.** Uma lista de módulos de dados adicionais, separados por vírgula + (`,`), para incluir na resposta. Permite buscar dados financeiros detalhados. + + **Exemplos:** + + - `modules=summaryProfile` (retorna perfil da empresa) + - `modules=balanceSheetHistory,incomeStatementHistory` (retorna histórico anual + do BP e DRE) + + Veja a descrição principal do endpoint para a lista completa de módulos e seus + conteúdos. + + range: **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Se omitido, apenas a cotação mais recente é retornada + (a menos que `interval` seja usado). + + **Valores Possíveis:** + + - `1d`: Último dia de pregão (intraday se `interval` for minutos/horas). + - `5d`: Últimos 5 dias. + - `1mo`: Último mês. + - `3mo`: Últimos 3 meses. + - `6mo`: Últimos 6 meses. + - `1y`: Último ano. + - `2y`: Últimos 2 anos. + - `5y`: Últimos 5 anos. + - `10y`: Últimos 10 anos. + - `ytd`: Desde o início do ano atual (Year-to-Date). + - `max`: Todo o período histórico disponível. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + if not tickers: + raise ValueError(f"Expected a non-empty value for `tickers` but received {tickers!r}") + return await self._get( + f"/api/quote/{tickers}", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "dividends": dividends, + "fundamental": fundamental, + "interval": interval, + "modules": modules, + "range": range, + }, + quote_retrieve_params.QuoteRetrieveParams, + ), + ), + cast_to=QuoteRetrieveResponse, + ) + + async def list( + self, + *, + token: str | Omit = omit, + limit: int | Omit = omit, + page: int | Omit = omit, + search: str | Omit = omit, + sector: Literal[ + "Retail Trade", + "Energy Minerals", + "Health Services", + "Utilities", + "Finance", + "Consumer Services", + "Consumer Non-Durables", + "Non-Energy Minerals", + "Commercial Services", + "Distribution Services", + "Transportation", + "Technology Services", + "Process Industries", + "Communications", + "Producer Manufacturing", + "Miscellaneous", + "Electronic Technology", + "Industrial Services", + "Health Technology", + "Consumer Durables", + ] + | Omit = omit, + sort_by: Literal["name", "close", "change", "change_abs", "volume", "market_cap_basic", "sector"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + type: Literal["stock", "fund", "bdr"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> QuoteListResponse: + """ + Obtenha uma lista paginada de cotações de diversos ativos (ações, FIIs, BDRs) + negociados na B3, com opções avançadas de busca, filtragem e ordenação. + + ### Funcionalidades: + + - **Busca por Ticker:** Filtre por parte do ticker usando `search`. + - **Filtragem por Tipo:** Restrinja a lista a `stock`, `fund` (FII) ou `bdr` com + o parâmetro `type`. + - **Filtragem por Setor:** Selecione ativos de um setor específico usando + `sector`. + - **Ordenação:** Ordene os resultados por diversos campos (preço, variação, + volume, etc.) usando `sortBy` e `sortOrder`. + - **Paginação:** Controle o número de resultados por página (`limit`) e a página + desejada (`page`). + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar as 10 ações do setor Financeiro com maior volume, ordenadas de forma + decrescente:** + + ```bash + curl -X GET "https://brapi.dev/api/quote/list?sector=Finance&sortBy=volume&sortOrder=desc&limit=10&page=1&token=SEU_TOKEN" + ``` + + **Buscar por ativos cujo ticker contenha 'ITUB' e ordenar por nome ascendente:** + + ```bash + curl -X GET "https://brapi.dev/api/quote/list?search=ITUB&sortBy=name&sortOrder=asc&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém a lista de `stocks` (e `indexes` relevantes), informações + sobre os filtros aplicados, detalhes da paginação (`currentPage`, `totalPages`, + `itemsPerPage`, `totalCount`, `hasNextPage`) e listas de setores + (`availableSectors`) e tipos (`availableStockTypes`) disponíveis para filtragem. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + limit: **Opcional.** Número máximo de ativos a serem retornados por página. O valor + padrão pode variar. + + page: **Opcional.** Número da página dos resultados a ser retornada, considerando o + `limit` especificado. Começa em 1. + + search: + **Opcional.** Termo para buscar ativos por ticker (correspondência parcial). Ex: + `PETR` encontrará `PETR4`, `PETR3`. + + sector: **Opcional.** Filtra os resultados por setor de atuação da empresa. Utilize um + dos valores retornados em `availableSectors`. + + sort_by: **Opcional.** Campo pelo qual os resultados serão ordenados. + + sort_order: **Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + Requer que `sortBy` seja especificado. + + type: **Opcional.** Filtra os resultados por tipo de ativo. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/quote/list", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "limit": limit, + "page": page, + "search": search, + "sector": sector, + "sort_by": sort_by, + "sort_order": sort_order, + "type": type, + }, + quote_list_params.QuoteListParams, + ), + ), + cast_to=QuoteListResponse, + ) + + +class QuoteResourceWithRawResponse: + def __init__(self, quote: QuoteResource) -> None: + self._quote = quote + + self.retrieve = to_raw_response_wrapper( + quote.retrieve, + ) + self.list = to_raw_response_wrapper( + quote.list, + ) + + +class AsyncQuoteResourceWithRawResponse: + def __init__(self, quote: AsyncQuoteResource) -> None: + self._quote = quote + + self.retrieve = async_to_raw_response_wrapper( + quote.retrieve, + ) + self.list = async_to_raw_response_wrapper( + quote.list, + ) + + +class QuoteResourceWithStreamingResponse: + def __init__(self, quote: QuoteResource) -> None: + self._quote = quote + + self.retrieve = to_streamed_response_wrapper( + quote.retrieve, + ) + self.list = to_streamed_response_wrapper( + quote.list, + ) + + +class AsyncQuoteResourceWithStreamingResponse: + def __init__(self, quote: AsyncQuoteResource) -> None: + self._quote = quote + + self.retrieve = async_to_streamed_response_wrapper( + quote.retrieve, + ) + self.list = async_to_streamed_response_wrapper( + quote.list, + ) diff --git a/src/brapi/resources/v2/__init__.py b/src/brapi/resources/v2/__init__.py new file mode 100644 index 0000000..e6c512e --- /dev/null +++ b/src/brapi/resources/v2/__init__.py @@ -0,0 +1,75 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from .v2 import ( + V2Resource, + AsyncV2Resource, + V2ResourceWithRawResponse, + AsyncV2ResourceWithRawResponse, + V2ResourceWithStreamingResponse, + AsyncV2ResourceWithStreamingResponse, +) +from .crypto import ( + CryptoResource, + AsyncCryptoResource, + CryptoResourceWithRawResponse, + AsyncCryptoResourceWithRawResponse, + CryptoResourceWithStreamingResponse, + AsyncCryptoResourceWithStreamingResponse, +) +from .currency import ( + CurrencyResource, + AsyncCurrencyResource, + CurrencyResourceWithRawResponse, + AsyncCurrencyResourceWithRawResponse, + CurrencyResourceWithStreamingResponse, + AsyncCurrencyResourceWithStreamingResponse, +) +from .inflation import ( + InflationResource, + AsyncInflationResource, + InflationResourceWithRawResponse, + AsyncInflationResourceWithRawResponse, + InflationResourceWithStreamingResponse, + AsyncInflationResourceWithStreamingResponse, +) +from .prime_rate import ( + PrimeRateResource, + AsyncPrimeRateResource, + PrimeRateResourceWithRawResponse, + AsyncPrimeRateResourceWithRawResponse, + PrimeRateResourceWithStreamingResponse, + AsyncPrimeRateResourceWithStreamingResponse, +) + +__all__ = [ + "CryptoResource", + "AsyncCryptoResource", + "CryptoResourceWithRawResponse", + "AsyncCryptoResourceWithRawResponse", + "CryptoResourceWithStreamingResponse", + "AsyncCryptoResourceWithStreamingResponse", + "CurrencyResource", + "AsyncCurrencyResource", + "CurrencyResourceWithRawResponse", + "AsyncCurrencyResourceWithRawResponse", + "CurrencyResourceWithStreamingResponse", + "AsyncCurrencyResourceWithStreamingResponse", + "InflationResource", + "AsyncInflationResource", + "InflationResourceWithRawResponse", + "AsyncInflationResourceWithRawResponse", + "InflationResourceWithStreamingResponse", + "AsyncInflationResourceWithStreamingResponse", + "PrimeRateResource", + "AsyncPrimeRateResource", + "PrimeRateResourceWithRawResponse", + "AsyncPrimeRateResourceWithRawResponse", + "PrimeRateResourceWithStreamingResponse", + "AsyncPrimeRateResourceWithStreamingResponse", + "V2Resource", + "AsyncV2Resource", + "V2ResourceWithRawResponse", + "AsyncV2ResourceWithRawResponse", + "V2ResourceWithStreamingResponse", + "AsyncV2ResourceWithStreamingResponse", +] diff --git a/src/brapi/resources/v2/crypto.py b/src/brapi/resources/v2/crypto.py new file mode 100644 index 0000000..c4db8df --- /dev/null +++ b/src/brapi/resources/v2/crypto.py @@ -0,0 +1,524 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import Literal + +import httpx + +from ..._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from ..._utils import maybe_transform, async_maybe_transform +from ..._compat import cached_property +from ...types.v2 import crypto_retrieve_params, crypto_list_available_params +from ..._resource import SyncAPIResource, AsyncAPIResource +from ..._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from ..._base_client import make_request_options +from ...types.v2.crypto_retrieve_response import CryptoRetrieveResponse +from ...types.v2.crypto_list_available_response import CryptoListAvailableResponse + +__all__ = ["CryptoResource", "AsyncCryptoResource"] + + +class CryptoResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> CryptoResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return CryptoResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> CryptoResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return CryptoResourceWithStreamingResponse(self) + + def retrieve( + self, + *, + coin: str, + token: str | Omit = omit, + currency: str | Omit = omit, + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + | Omit = omit, + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CryptoRetrieveResponse: + """ + Obtenha cotações atualizadas e dados históricos para uma ou mais criptomoedas. + + ### Funcionalidades: + + - **Cotação Múltipla:** Consulte várias criptomoedas em uma única requisição + usando o parâmetro `coin`. + - **Moeda de Referência:** Especifique a moeda fiduciária para a cotação com + `currency` (padrão: BRL). + - **Dados Históricos:** Solicite séries históricas usando `range` e `interval` + (similar ao endpoint de ações). + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Cotação de Bitcoin (BTC) e Ethereum (ETH) em Dólar Americano (USD):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto?coin=BTC,ETH¤cy=USD&token=SEU_TOKEN" + ``` + + **Cotação de Cardano (ADA) em Real (BRL) com histórico do último mês (intervalo + diário):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto?coin=ADA¤cy=BRL&range=1mo&interval=1d&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém um array `coins`, onde cada objeto representa uma criptomoeda + solicitada, incluindo sua cotação atual, dados de mercado e, opcionalmente, a + série histórica (`historicalDataPrice`). + + Args: + coin: **Obrigatório.** Uma ou mais siglas (tickers) de criptomoedas que você deseja + consultar. Separe múltiplas siglas por vírgula (`,`). + + - **Exemplos:** `BTC`, `ETH,ADA`, `SOL`. + + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + currency: **Opcional.** A sigla da moeda fiduciária na qual a cotação da(s) criptomoeda(s) + deve ser retornada. Se omitido, o padrão é `BRL` (Real Brasileiro). + + interval: **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. Funciona + de forma análoga ao endpoint de ações. + + - Valores: `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`, `1d`, `5d`, + `1wk`, `1mo`, `3mo`. + + range: **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Funciona de forma análoga ao endpoint de ações. Se + omitido, apenas a cotação mais recente é retornada (a menos que `interval` seja + usado). + + - Valores: `1d`, `5d`, `1mo`, `3mo`, `6mo`, `1y`, `2y`, `5y`, `10y`, `ytd`, + `max`. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/crypto", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "coin": coin, + "token": token, + "currency": currency, + "interval": interval, + "range": range, + }, + crypto_retrieve_params.CryptoRetrieveParams, + ), + ), + cast_to=CryptoRetrieveResponse, + ) + + def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CryptoListAvailableResponse: + """ + Obtenha a lista completa de todas as siglas (tickers) de criptomoedas que a API + Brapi suporta para consulta no endpoint `/api/v2/crypto`. + + ### Funcionalidade: + + - Retorna um array `coins` com as siglas. + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todas as criptomoedas disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto/available?token=SEU_TOKEN" + ``` + + **Buscar criptomoedas cujo ticker contenha 'DOGE':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto/available?search=DOGE&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `coins`, contendo um array de strings + com as siglas das criptomoedas (ex: `["BTC", "ETH", "LTC", "XRP"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de siglas de criptomoedas + (correspondência parcial, case-insensitive). Se omitido, retorna todas as + siglas. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/crypto/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "search": search, + }, + crypto_list_available_params.CryptoListAvailableParams, + ), + ), + cast_to=CryptoListAvailableResponse, + ) + + +class AsyncCryptoResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncCryptoResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncCryptoResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncCryptoResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncCryptoResourceWithStreamingResponse(self) + + async def retrieve( + self, + *, + coin: str, + token: str | Omit = omit, + currency: str | Omit = omit, + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + | Omit = omit, + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CryptoRetrieveResponse: + """ + Obtenha cotações atualizadas e dados históricos para uma ou mais criptomoedas. + + ### Funcionalidades: + + - **Cotação Múltipla:** Consulte várias criptomoedas em uma única requisição + usando o parâmetro `coin`. + - **Moeda de Referência:** Especifique a moeda fiduciária para a cotação com + `currency` (padrão: BRL). + - **Dados Históricos:** Solicite séries históricas usando `range` e `interval` + (similar ao endpoint de ações). + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Cotação de Bitcoin (BTC) e Ethereum (ETH) em Dólar Americano (USD):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto?coin=BTC,ETH¤cy=USD&token=SEU_TOKEN" + ``` + + **Cotação de Cardano (ADA) em Real (BRL) com histórico do último mês (intervalo + diário):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto?coin=ADA¤cy=BRL&range=1mo&interval=1d&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém um array `coins`, onde cada objeto representa uma criptomoeda + solicitada, incluindo sua cotação atual, dados de mercado e, opcionalmente, a + série histórica (`historicalDataPrice`). + + Args: + coin: **Obrigatório.** Uma ou mais siglas (tickers) de criptomoedas que você deseja + consultar. Separe múltiplas siglas por vírgula (`,`). + + - **Exemplos:** `BTC`, `ETH,ADA`, `SOL`. + + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + currency: **Opcional.** A sigla da moeda fiduciária na qual a cotação da(s) criptomoeda(s) + deve ser retornada. Se omitido, o padrão é `BRL` (Real Brasileiro). + + interval: **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. Funciona + de forma análoga ao endpoint de ações. + + - Valores: `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`, `1d`, `5d`, + `1wk`, `1mo`, `3mo`. + + range: **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Funciona de forma análoga ao endpoint de ações. Se + omitido, apenas a cotação mais recente é retornada (a menos que `interval` seja + usado). + + - Valores: `1d`, `5d`, `1mo`, `3mo`, `6mo`, `1y`, `2y`, `5y`, `10y`, `ytd`, + `max`. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/crypto", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "coin": coin, + "token": token, + "currency": currency, + "interval": interval, + "range": range, + }, + crypto_retrieve_params.CryptoRetrieveParams, + ), + ), + cast_to=CryptoRetrieveResponse, + ) + + async def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CryptoListAvailableResponse: + """ + Obtenha a lista completa de todas as siglas (tickers) de criptomoedas que a API + Brapi suporta para consulta no endpoint `/api/v2/crypto`. + + ### Funcionalidade: + + - Retorna um array `coins` com as siglas. + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todas as criptomoedas disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto/available?token=SEU_TOKEN" + ``` + + **Buscar criptomoedas cujo ticker contenha 'DOGE':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/crypto/available?search=DOGE&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `coins`, contendo um array de strings + com as siglas das criptomoedas (ex: `["BTC", "ETH", "LTC", "XRP"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de siglas de criptomoedas + (correspondência parcial, case-insensitive). Se omitido, retorna todas as + siglas. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/crypto/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "search": search, + }, + crypto_list_available_params.CryptoListAvailableParams, + ), + ), + cast_to=CryptoListAvailableResponse, + ) + + +class CryptoResourceWithRawResponse: + def __init__(self, crypto: CryptoResource) -> None: + self._crypto = crypto + + self.retrieve = to_raw_response_wrapper( + crypto.retrieve, + ) + self.list_available = to_raw_response_wrapper( + crypto.list_available, + ) + + +class AsyncCryptoResourceWithRawResponse: + def __init__(self, crypto: AsyncCryptoResource) -> None: + self._crypto = crypto + + self.retrieve = async_to_raw_response_wrapper( + crypto.retrieve, + ) + self.list_available = async_to_raw_response_wrapper( + crypto.list_available, + ) + + +class CryptoResourceWithStreamingResponse: + def __init__(self, crypto: CryptoResource) -> None: + self._crypto = crypto + + self.retrieve = to_streamed_response_wrapper( + crypto.retrieve, + ) + self.list_available = to_streamed_response_wrapper( + crypto.list_available, + ) + + +class AsyncCryptoResourceWithStreamingResponse: + def __init__(self, crypto: AsyncCryptoResource) -> None: + self._crypto = crypto + + self.retrieve = async_to_streamed_response_wrapper( + crypto.retrieve, + ) + self.list_available = async_to_streamed_response_wrapper( + crypto.list_available, + ) diff --git a/src/brapi/resources/v2/currency.py b/src/brapi/resources/v2/currency.py new file mode 100644 index 0000000..e082a6d --- /dev/null +++ b/src/brapi/resources/v2/currency.py @@ -0,0 +1,456 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import httpx + +from ..._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from ..._utils import maybe_transform, async_maybe_transform +from ..._compat import cached_property +from ...types.v2 import currency_retrieve_params, currency_list_available_params +from ..._resource import SyncAPIResource, AsyncAPIResource +from ..._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from ..._base_client import make_request_options +from ...types.v2.currency_retrieve_response import CurrencyRetrieveResponse +from ...types.v2.currency_list_available_response import CurrencyListAvailableResponse + +__all__ = ["CurrencyResource", "AsyncCurrencyResource"] + + +class CurrencyResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> CurrencyResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return CurrencyResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> CurrencyResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return CurrencyResourceWithStreamingResponse(self) + + def retrieve( + self, + *, + currency: str, + token: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CurrencyRetrieveResponse: + """ + Obtenha cotações atualizadas para um ou mais pares de moedas fiduciárias (ex: + USD-BRL, EUR-USD). + + ### Funcionalidades: + + - **Cotação Múltipla:** Consulte vários pares de moedas em uma única requisição + usando o parâmetro `currency`. + - **Dados Retornados:** Inclui nome do par, preços de compra (bid) e venda + (ask), variação, máximas e mínimas, e timestamp da atualização. + + ### Parâmetros: + + - **`currency` (Obrigatório):** Uma lista de pares de moedas separados por + vírgula, no formato `MOEDA_ORIGEM-MOEDA_DESTINO` (ex: `USD-BRL`, `EUR-USD`). + Consulte os pares disponíveis em + [`/api/v2/currency/available`](#/Moedas/getAvailableCurrencies). + - **`token` (Obrigatório):** Seu token de autenticação. + + ### Autenticação: + + Requer token de autenticação válido via `token` (query) ou `Authorization` + (header). + + Args: + currency: **Obrigatório.** Uma lista de um ou mais pares de moedas a serem consultados, + separados por vírgula (`,`). + + - **Formato:** `MOEDA_ORIGEM-MOEDA_DESTINO` (ex: `USD-BRL`). + - **Disponibilidade:** Consulte os pares válidos usando o endpoint + [`/api/v2/currency/available`](#/Moedas/getAvailableCurrencies). + - **Exemplo:** `USD-BRL,EUR-BRL,BTC-BRL` + + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/currency", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "currency": currency, + "token": token, + }, + currency_retrieve_params.CurrencyRetrieveParams, + ), + ), + cast_to=CurrencyRetrieveResponse, + ) + + def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CurrencyListAvailableResponse: + """ + Obtenha a lista completa de todas as moedas fiduciárias suportadas pela API, + geralmente utilizadas no parâmetro `currency` de outros endpoints (como o de + criptomoedas) ou para futuras funcionalidades de conversão. + + ### Funcionalidade: + + - Retorna um array `currencies` com os nomes das moedas. + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todas as moedas disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/currency/available?token=SEU_TOKEN" + ``` + + **Buscar moedas cujo nome contenha 'Euro':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/currency/available?search=Euro&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `currencies`, contendo um array de + objetos. Cada objeto possui uma chave `currency` com o nome completo da moeda + (ex: `"Dólar Americano/Real Brasileiro"`). **Nota:** O formato do nome pode + indicar um par de moedas, dependendo do contexto interno da API. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista pelo nome da moeda (correspondência + parcial, case-insensitive). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/currency/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "search": search, + }, + currency_list_available_params.CurrencyListAvailableParams, + ), + ), + cast_to=CurrencyListAvailableResponse, + ) + + +class AsyncCurrencyResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncCurrencyResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncCurrencyResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncCurrencyResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncCurrencyResourceWithStreamingResponse(self) + + async def retrieve( + self, + *, + currency: str, + token: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CurrencyRetrieveResponse: + """ + Obtenha cotações atualizadas para um ou mais pares de moedas fiduciárias (ex: + USD-BRL, EUR-USD). + + ### Funcionalidades: + + - **Cotação Múltipla:** Consulte vários pares de moedas em uma única requisição + usando o parâmetro `currency`. + - **Dados Retornados:** Inclui nome do par, preços de compra (bid) e venda + (ask), variação, máximas e mínimas, e timestamp da atualização. + + ### Parâmetros: + + - **`currency` (Obrigatório):** Uma lista de pares de moedas separados por + vírgula, no formato `MOEDA_ORIGEM-MOEDA_DESTINO` (ex: `USD-BRL`, `EUR-USD`). + Consulte os pares disponíveis em + [`/api/v2/currency/available`](#/Moedas/getAvailableCurrencies). + - **`token` (Obrigatório):** Seu token de autenticação. + + ### Autenticação: + + Requer token de autenticação válido via `token` (query) ou `Authorization` + (header). + + Args: + currency: **Obrigatório.** Uma lista de um ou mais pares de moedas a serem consultados, + separados por vírgula (`,`). + + - **Formato:** `MOEDA_ORIGEM-MOEDA_DESTINO` (ex: `USD-BRL`). + - **Disponibilidade:** Consulte os pares válidos usando o endpoint + [`/api/v2/currency/available`](#/Moedas/getAvailableCurrencies). + - **Exemplo:** `USD-BRL,EUR-BRL,BTC-BRL` + + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/currency", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "currency": currency, + "token": token, + }, + currency_retrieve_params.CurrencyRetrieveParams, + ), + ), + cast_to=CurrencyRetrieveResponse, + ) + + async def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> CurrencyListAvailableResponse: + """ + Obtenha a lista completa de todas as moedas fiduciárias suportadas pela API, + geralmente utilizadas no parâmetro `currency` de outros endpoints (como o de + criptomoedas) ou para futuras funcionalidades de conversão. + + ### Funcionalidade: + + - Retorna um array `currencies` com os nomes das moedas. + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todas as moedas disponíveis:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/currency/available?token=SEU_TOKEN" + ``` + + **Buscar moedas cujo nome contenha 'Euro':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/currency/available?search=Euro&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `currencies`, contendo um array de + objetos. Cada objeto possui uma chave `currency` com o nome completo da moeda + (ex: `"Dólar Americano/Real Brasileiro"`). **Nota:** O formato do nome pode + indicar um par de moedas, dependendo do contexto interno da API. + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista pelo nome da moeda (correspondência + parcial, case-insensitive). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/currency/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "search": search, + }, + currency_list_available_params.CurrencyListAvailableParams, + ), + ), + cast_to=CurrencyListAvailableResponse, + ) + + +class CurrencyResourceWithRawResponse: + def __init__(self, currency: CurrencyResource) -> None: + self._currency = currency + + self.retrieve = to_raw_response_wrapper( + currency.retrieve, + ) + self.list_available = to_raw_response_wrapper( + currency.list_available, + ) + + +class AsyncCurrencyResourceWithRawResponse: + def __init__(self, currency: AsyncCurrencyResource) -> None: + self._currency = currency + + self.retrieve = async_to_raw_response_wrapper( + currency.retrieve, + ) + self.list_available = async_to_raw_response_wrapper( + currency.list_available, + ) + + +class CurrencyResourceWithStreamingResponse: + def __init__(self, currency: CurrencyResource) -> None: + self._currency = currency + + self.retrieve = to_streamed_response_wrapper( + currency.retrieve, + ) + self.list_available = to_streamed_response_wrapper( + currency.list_available, + ) + + +class AsyncCurrencyResourceWithStreamingResponse: + def __init__(self, currency: AsyncCurrencyResource) -> None: + self._currency = currency + + self.retrieve = async_to_streamed_response_wrapper( + currency.retrieve, + ) + self.list_available = async_to_streamed_response_wrapper( + currency.list_available, + ) diff --git a/src/brapi/resources/v2/inflation.py b/src/brapi/resources/v2/inflation.py new file mode 100644 index 0000000..b0e1138 --- /dev/null +++ b/src/brapi/resources/v2/inflation.py @@ -0,0 +1,530 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import Union +from datetime import date +from typing_extensions import Literal + +import httpx + +from ..._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from ..._utils import maybe_transform, async_maybe_transform +from ..._compat import cached_property +from ...types.v2 import inflation_retrieve_params, inflation_list_available_params +from ..._resource import SyncAPIResource, AsyncAPIResource +from ..._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from ..._base_client import make_request_options +from ...types.v2.inflation_retrieve_response import InflationRetrieveResponse +from ...types.v2.inflation_list_available_response import InflationListAvailableResponse + +__all__ = ["InflationResource", "AsyncInflationResource"] + + +class InflationResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> InflationResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return InflationResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> InflationResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return InflationResourceWithStreamingResponse(self) + + def retrieve( + self, + *, + token: str | Omit = omit, + country: str | Omit = omit, + end: Union[str, date] | Omit = omit, + historical: bool | Omit = omit, + sort_by: Literal["date", "value"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + start: Union[str, date] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> InflationRetrieveResponse: + """ + Obtenha dados históricos sobre índices de inflação para um país específico. + + ### Funcionalidades: + + - **Seleção de País:** Especifique o país desejado com o parâmetro `country` + (padrão: `brazil`). + - **Filtragem por Período:** Defina um intervalo de datas com `start` e `end` + (formato DD/MM/YYYY). + - **Inclusão de Histórico:** O parâmetro `historical` (booleano) parece + controlar a inclusão de dados históricos (verificar comportamento exato, pode + ser redundante com `start`/`end`). + - **Ordenação:** Ordene os resultados por data (`date`) ou valor (`value`) + usando `sortBy` e `sortOrder`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Buscar dados de inflação do Brasil para o ano de 2022, ordenados por valor + ascendente:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation?country=brazil&start=01/01/2022&end=31/12/2022&sortBy=value&sortOrder=asc&token=SEU_TOKEN" + ``` + + **Buscar os dados mais recentes de inflação (sem período definido, ordenação + padrão):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation?country=brazil&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém um array `inflation`, onde cada objeto representa um ponto de + dado de inflação com sua `date` (DD/MM/YYYY), `value` (o índice de inflação como + string) e `epochDate` (timestamp UNIX). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + country: **Opcional.** Nome do país para o qual buscar os dados de inflação. Use nomes em + minúsculas. O padrão é `brazil`. Consulte `/api/v2/inflation/available` para a + lista de países suportados. + + end: **Opcional.** Data final do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `start` for especificado. + + historical: **Opcional.** Booleano (`true` ou `false`). Define se dados históricos devem ser + incluídos. O comportamento exato em conjunto com `start`/`end` deve ser + verificado. Padrão: `false`. + + sort_by: **Opcional.** Campo pelo qual os resultados da inflação serão ordenados. + + sort_order: **Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + Padrão: `desc`. Requer que `sortBy` seja especificado. + + start: **Opcional.** Data de início do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `end` for especificado. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/inflation", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "country": country, + "end": end, + "historical": historical, + "sort_by": sort_by, + "sort_order": sort_order, + "start": start, + }, + inflation_retrieve_params.InflationRetrieveParams, + ), + ), + cast_to=InflationRetrieveResponse, + ) + + def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> InflationListAvailableResponse: + """ + Obtenha a lista completa de todos os países para os quais a API Brapi possui + dados de inflação disponíveis para consulta no endpoint `/api/v2/inflation`. + + ### Funcionalidade: + + - Retorna um array `countries` com os nomes dos países (em minúsculas). + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todos os países com dados de inflação:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation/available?token=SEU_TOKEN" + ``` + + **Buscar países cujo nome contenha 'arg':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation/available?search=arg&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `countries`, contendo um array de + strings com os nomes dos países (ex: `["brazil", "argentina", "usa"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista pelo nome do país (correspondência + parcial, case-insensitive). Se omitido, retorna todos os países. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/inflation/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "search": search, + }, + inflation_list_available_params.InflationListAvailableParams, + ), + ), + cast_to=InflationListAvailableResponse, + ) + + +class AsyncInflationResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncInflationResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncInflationResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncInflationResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncInflationResourceWithStreamingResponse(self) + + async def retrieve( + self, + *, + token: str | Omit = omit, + country: str | Omit = omit, + end: Union[str, date] | Omit = omit, + historical: bool | Omit = omit, + sort_by: Literal["date", "value"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + start: Union[str, date] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> InflationRetrieveResponse: + """ + Obtenha dados históricos sobre índices de inflação para um país específico. + + ### Funcionalidades: + + - **Seleção de País:** Especifique o país desejado com o parâmetro `country` + (padrão: `brazil`). + - **Filtragem por Período:** Defina um intervalo de datas com `start` e `end` + (formato DD/MM/YYYY). + - **Inclusão de Histórico:** O parâmetro `historical` (booleano) parece + controlar a inclusão de dados históricos (verificar comportamento exato, pode + ser redundante com `start`/`end`). + - **Ordenação:** Ordene os resultados por data (`date`) ou valor (`value`) + usando `sortBy` e `sortOrder`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Buscar dados de inflação do Brasil para o ano de 2022, ordenados por valor + ascendente:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation?country=brazil&start=01/01/2022&end=31/12/2022&sortBy=value&sortOrder=asc&token=SEU_TOKEN" + ``` + + **Buscar os dados mais recentes de inflação (sem período definido, ordenação + padrão):** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation?country=brazil&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta contém um array `inflation`, onde cada objeto representa um ponto de + dado de inflação com sua `date` (DD/MM/YYYY), `value` (o índice de inflação como + string) e `epochDate` (timestamp UNIX). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + country: **Opcional.** Nome do país para o qual buscar os dados de inflação. Use nomes em + minúsculas. O padrão é `brazil`. Consulte `/api/v2/inflation/available` para a + lista de países suportados. + + end: **Opcional.** Data final do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `start` for especificado. + + historical: **Opcional.** Booleano (`true` ou `false`). Define se dados históricos devem ser + incluídos. O comportamento exato em conjunto com `start`/`end` deve ser + verificado. Padrão: `false`. + + sort_by: **Opcional.** Campo pelo qual os resultados da inflação serão ordenados. + + sort_order: **Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + Padrão: `desc`. Requer que `sortBy` seja especificado. + + start: **Opcional.** Data de início do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `end` for especificado. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/inflation", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "country": country, + "end": end, + "historical": historical, + "sort_by": sort_by, + "sort_order": sort_order, + "start": start, + }, + inflation_retrieve_params.InflationRetrieveParams, + ), + ), + cast_to=InflationRetrieveResponse, + ) + + async def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> InflationListAvailableResponse: + """ + Obtenha a lista completa de todos os países para os quais a API Brapi possui + dados de inflação disponíveis para consulta no endpoint `/api/v2/inflation`. + + ### Funcionalidade: + + - Retorna um array `countries` com os nomes dos países (em minúsculas). + - Pode ser filtrado usando o parâmetro `search`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar todos os países com dados de inflação:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation/available?token=SEU_TOKEN" + ``` + + **Buscar países cujo nome contenha 'arg':** + + ```bash + curl -X GET "https://brapi.dev/api/v2/inflation/available?search=arg&token=SEU_TOKEN" + ``` + + ### Resposta: + + A resposta é um objeto JSON com a chave `countries`, contendo um array de + strings com os nomes dos países (ex: `["brazil", "argentina", "usa"]`). + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista pelo nome do país (correspondência + parcial, case-insensitive). Se omitido, retorna todos os países. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/inflation/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "search": search, + }, + inflation_list_available_params.InflationListAvailableParams, + ), + ), + cast_to=InflationListAvailableResponse, + ) + + +class InflationResourceWithRawResponse: + def __init__(self, inflation: InflationResource) -> None: + self._inflation = inflation + + self.retrieve = to_raw_response_wrapper( + inflation.retrieve, + ) + self.list_available = to_raw_response_wrapper( + inflation.list_available, + ) + + +class AsyncInflationResourceWithRawResponse: + def __init__(self, inflation: AsyncInflationResource) -> None: + self._inflation = inflation + + self.retrieve = async_to_raw_response_wrapper( + inflation.retrieve, + ) + self.list_available = async_to_raw_response_wrapper( + inflation.list_available, + ) + + +class InflationResourceWithStreamingResponse: + def __init__(self, inflation: InflationResource) -> None: + self._inflation = inflation + + self.retrieve = to_streamed_response_wrapper( + inflation.retrieve, + ) + self.list_available = to_streamed_response_wrapper( + inflation.list_available, + ) + + +class AsyncInflationResourceWithStreamingResponse: + def __init__(self, inflation: AsyncInflationResource) -> None: + self._inflation = inflation + + self.retrieve = async_to_streamed_response_wrapper( + inflation.retrieve, + ) + self.list_available = async_to_streamed_response_wrapper( + inflation.list_available, + ) diff --git a/src/brapi/resources/v2/prime_rate.py b/src/brapi/resources/v2/prime_rate.py new file mode 100644 index 0000000..21d9e81 --- /dev/null +++ b/src/brapi/resources/v2/prime_rate.py @@ -0,0 +1,490 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import Union +from datetime import date +from typing_extensions import Literal + +import httpx + +from ..._types import Body, Omit, Query, Headers, NotGiven, omit, not_given +from ..._utils import maybe_transform, async_maybe_transform +from ..._compat import cached_property +from ...types.v2 import prime_rate_retrieve_params, prime_rate_list_available_params +from ..._resource import SyncAPIResource, AsyncAPIResource +from ..._response import ( + to_raw_response_wrapper, + to_streamed_response_wrapper, + async_to_raw_response_wrapper, + async_to_streamed_response_wrapper, +) +from ..._base_client import make_request_options +from ...types.v2.prime_rate_retrieve_response import PrimeRateRetrieveResponse +from ...types.v2.prime_rate_list_available_response import PrimeRateListAvailableResponse + +__all__ = ["PrimeRateResource", "AsyncPrimeRateResource"] + + +class PrimeRateResource(SyncAPIResource): + @cached_property + def with_raw_response(self) -> PrimeRateResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return PrimeRateResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> PrimeRateResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return PrimeRateResourceWithStreamingResponse(self) + + def retrieve( + self, + *, + token: str | Omit = omit, + country: str | Omit = omit, + end: Union[str, date] | Omit = omit, + historical: bool | Omit = omit, + sort_by: Literal["date", "value"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + start: Union[str, date] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> PrimeRateRetrieveResponse: + """ + Obtenha informações atualizadas sobre a taxa básica de juros (SELIC) de um país + por um período determinado. + + ### Funcionalidades: + + - **Seleção por País:** Especifique o país desejado usando o parâmetro `country` + (padrão: brazil). + - **Período Customizado:** Defina datas de início e fim com `start` e `end` para + consultar um intervalo específico. + - **Ordenação:** Ordene os resultados por data ou valor com os parâmetros + `sortBy` e `sortOrder`. + - **Dados Históricos:** Solicite o histórico completo ou apenas o valor mais + recente com o parâmetro `historical`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Taxa de juros do Brasil entre dezembro/2021 e janeiro/2022:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/prime-rate?country=brazil&start=01/12/2021&end=01/01/2022&sortBy=date&sortOrder=desc&token=SEU_TOKEN" + ``` + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + country: **Opcional.** O país do qual você deseja obter informações sobre a taxa básica + de juros. Por padrão, o país é definido como brazil. Você pode consultar a lista + de países disponíveis através do endpoint `/api/v2/prime-rate/available`. + + end: **Opcional.** Data final do período para busca no formato DD/MM/YYYY. Por padrão + é a data atual. Útil quando `historical=true` para restringir o período da série + histórica. + + historical: **Opcional.** Define se os dados históricos serão retornados. Se definido como + `true`, retorna a série histórica completa. Se `false` (padrão) ou omitido, + retorna apenas o valor mais recente. + + sort_by: **Opcional.** Campo pelo qual os resultados serão ordenados. Por padrão, ordena + por `date` (data). + + sort_order: **Opcional.** Define se a ordenação será crescente (`asc`) ou decrescente + (`desc`). Por padrão, é `desc` (decrescente). + + start: **Opcional.** Data inicial do período para busca no formato DD/MM/YYYY. Útil + quando `historical=true` para restringir o período da série histórica. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/prime-rate", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "country": country, + "end": end, + "historical": historical, + "sort_by": sort_by, + "sort_order": sort_order, + "start": start, + }, + prime_rate_retrieve_params.PrimeRateRetrieveParams, + ), + ), + cast_to=PrimeRateRetrieveResponse, + ) + + def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> PrimeRateListAvailableResponse: + """ + Liste todos os países disponíveis com dados de taxa básica de juros (SELIC) na + API brapi. Este endpoint facilita a descoberta de quais países possuem dados + disponíveis para consulta através do endpoint principal `/api/v2/prime-rate`. + + ### Funcionalidades: + + - **Busca Filtrada:** Utilize o parâmetro `search` para filtrar países por nome + ou parte do nome. + - **Ideal para Autocomplete:** Perfeito para implementar campos de busca com + autocompletar em interfaces de usuário. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar países que contenham "BR" no nome:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/prime-rate/available?search=BR&token=SEU_TOKEN" + ``` + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de países por nome. Retorna países + cujos nomes contenham o termo especificado (case insensitive). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return self._get( + "/api/v2/prime-rate/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=maybe_transform( + { + "token": token, + "search": search, + }, + prime_rate_list_available_params.PrimeRateListAvailableParams, + ), + ), + cast_to=PrimeRateListAvailableResponse, + ) + + +class AsyncPrimeRateResource(AsyncAPIResource): + @cached_property + def with_raw_response(self) -> AsyncPrimeRateResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncPrimeRateResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncPrimeRateResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncPrimeRateResourceWithStreamingResponse(self) + + async def retrieve( + self, + *, + token: str | Omit = omit, + country: str | Omit = omit, + end: Union[str, date] | Omit = omit, + historical: bool | Omit = omit, + sort_by: Literal["date", "value"] | Omit = omit, + sort_order: Literal["asc", "desc"] | Omit = omit, + start: Union[str, date] | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> PrimeRateRetrieveResponse: + """ + Obtenha informações atualizadas sobre a taxa básica de juros (SELIC) de um país + por um período determinado. + + ### Funcionalidades: + + - **Seleção por País:** Especifique o país desejado usando o parâmetro `country` + (padrão: brazil). + - **Período Customizado:** Defina datas de início e fim com `start` e `end` para + consultar um intervalo específico. + - **Ordenação:** Ordene os resultados por data ou valor com os parâmetros + `sortBy` e `sortOrder`. + - **Dados Históricos:** Solicite o histórico completo ou apenas o valor mais + recente com o parâmetro `historical`. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Taxa de juros do Brasil entre dezembro/2021 e janeiro/2022:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/prime-rate?country=brazil&start=01/12/2021&end=01/01/2022&sortBy=date&sortOrder=desc&token=SEU_TOKEN" + ``` + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + country: **Opcional.** O país do qual você deseja obter informações sobre a taxa básica + de juros. Por padrão, o país é definido como brazil. Você pode consultar a lista + de países disponíveis através do endpoint `/api/v2/prime-rate/available`. + + end: **Opcional.** Data final do período para busca no formato DD/MM/YYYY. Por padrão + é a data atual. Útil quando `historical=true` para restringir o período da série + histórica. + + historical: **Opcional.** Define se os dados históricos serão retornados. Se definido como + `true`, retorna a série histórica completa. Se `false` (padrão) ou omitido, + retorna apenas o valor mais recente. + + sort_by: **Opcional.** Campo pelo qual os resultados serão ordenados. Por padrão, ordena + por `date` (data). + + sort_order: **Opcional.** Define se a ordenação será crescente (`asc`) ou decrescente + (`desc`). Por padrão, é `desc` (decrescente). + + start: **Opcional.** Data inicial do período para busca no formato DD/MM/YYYY. Útil + quando `historical=true` para restringir o período da série histórica. + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/prime-rate", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "country": country, + "end": end, + "historical": historical, + "sort_by": sort_by, + "sort_order": sort_order, + "start": start, + }, + prime_rate_retrieve_params.PrimeRateRetrieveParams, + ), + ), + cast_to=PrimeRateRetrieveResponse, + ) + + async def list_available( + self, + *, + token: str | Omit = omit, + search: str | Omit = omit, + # Use the following arguments if you need to pass additional parameters to the API that aren't available via kwargs. + # The extra values given here take precedence over values defined on the client or passed to this method. + extra_headers: Headers | None = None, + extra_query: Query | None = None, + extra_body: Body | None = None, + timeout: float | httpx.Timeout | None | NotGiven = not_given, + ) -> PrimeRateListAvailableResponse: + """ + Liste todos os países disponíveis com dados de taxa básica de juros (SELIC) na + API brapi. Este endpoint facilita a descoberta de quais países possuem dados + disponíveis para consulta através do endpoint principal `/api/v2/prime-rate`. + + ### Funcionalidades: + + - **Busca Filtrada:** Utilize o parâmetro `search` para filtrar países por nome + ou parte do nome. + - **Ideal para Autocomplete:** Perfeito para implementar campos de busca com + autocompletar em interfaces de usuário. + + ### Autenticação: + + Requer token de autenticação via `token` (query) ou `Authorization` (header). + + ### Exemplo de Requisição: + + **Listar países que contenham "BR" no nome:** + + ```bash + curl -X GET "https://brapi.dev/api/v2/prime-rate/available?search=BR&token=SEU_TOKEN" + ``` + + Args: + token: **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + + search: **Opcional.** Termo para filtrar a lista de países por nome. Retorna países + cujos nomes contenham o termo especificado (case insensitive). + + extra_headers: Send extra headers + + extra_query: Add additional query parameters to the request + + extra_body: Add additional JSON properties to the request + + timeout: Override the client-level default timeout for this request, in seconds + """ + return await self._get( + "/api/v2/prime-rate/available", + options=make_request_options( + extra_headers=extra_headers, + extra_query=extra_query, + extra_body=extra_body, + timeout=timeout, + query=await async_maybe_transform( + { + "token": token, + "search": search, + }, + prime_rate_list_available_params.PrimeRateListAvailableParams, + ), + ), + cast_to=PrimeRateListAvailableResponse, + ) + + +class PrimeRateResourceWithRawResponse: + def __init__(self, prime_rate: PrimeRateResource) -> None: + self._prime_rate = prime_rate + + self.retrieve = to_raw_response_wrapper( + prime_rate.retrieve, + ) + self.list_available = to_raw_response_wrapper( + prime_rate.list_available, + ) + + +class AsyncPrimeRateResourceWithRawResponse: + def __init__(self, prime_rate: AsyncPrimeRateResource) -> None: + self._prime_rate = prime_rate + + self.retrieve = async_to_raw_response_wrapper( + prime_rate.retrieve, + ) + self.list_available = async_to_raw_response_wrapper( + prime_rate.list_available, + ) + + +class PrimeRateResourceWithStreamingResponse: + def __init__(self, prime_rate: PrimeRateResource) -> None: + self._prime_rate = prime_rate + + self.retrieve = to_streamed_response_wrapper( + prime_rate.retrieve, + ) + self.list_available = to_streamed_response_wrapper( + prime_rate.list_available, + ) + + +class AsyncPrimeRateResourceWithStreamingResponse: + def __init__(self, prime_rate: AsyncPrimeRateResource) -> None: + self._prime_rate = prime_rate + + self.retrieve = async_to_streamed_response_wrapper( + prime_rate.retrieve, + ) + self.list_available = async_to_streamed_response_wrapper( + prime_rate.list_available, + ) diff --git a/src/brapi/resources/v2/v2.py b/src/brapi/resources/v2/v2.py new file mode 100644 index 0000000..54e46f1 --- /dev/null +++ b/src/brapi/resources/v2/v2.py @@ -0,0 +1,198 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from .crypto import ( + CryptoResource, + AsyncCryptoResource, + CryptoResourceWithRawResponse, + AsyncCryptoResourceWithRawResponse, + CryptoResourceWithStreamingResponse, + AsyncCryptoResourceWithStreamingResponse, +) +from .currency import ( + CurrencyResource, + AsyncCurrencyResource, + CurrencyResourceWithRawResponse, + AsyncCurrencyResourceWithRawResponse, + CurrencyResourceWithStreamingResponse, + AsyncCurrencyResourceWithStreamingResponse, +) +from ..._compat import cached_property +from .inflation import ( + InflationResource, + AsyncInflationResource, + InflationResourceWithRawResponse, + AsyncInflationResourceWithRawResponse, + InflationResourceWithStreamingResponse, + AsyncInflationResourceWithStreamingResponse, +) +from .prime_rate import ( + PrimeRateResource, + AsyncPrimeRateResource, + PrimeRateResourceWithRawResponse, + AsyncPrimeRateResourceWithRawResponse, + PrimeRateResourceWithStreamingResponse, + AsyncPrimeRateResourceWithStreamingResponse, +) +from ..._resource import SyncAPIResource, AsyncAPIResource + +__all__ = ["V2Resource", "AsyncV2Resource"] + + +class V2Resource(SyncAPIResource): + @cached_property + def crypto(self) -> CryptoResource: + return CryptoResource(self._client) + + @cached_property + def currency(self) -> CurrencyResource: + return CurrencyResource(self._client) + + @cached_property + def inflation(self) -> InflationResource: + return InflationResource(self._client) + + @cached_property + def prime_rate(self) -> PrimeRateResource: + return PrimeRateResource(self._client) + + @cached_property + def with_raw_response(self) -> V2ResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return V2ResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> V2ResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return V2ResourceWithStreamingResponse(self) + + +class AsyncV2Resource(AsyncAPIResource): + @cached_property + def crypto(self) -> AsyncCryptoResource: + return AsyncCryptoResource(self._client) + + @cached_property + def currency(self) -> AsyncCurrencyResource: + return AsyncCurrencyResource(self._client) + + @cached_property + def inflation(self) -> AsyncInflationResource: + return AsyncInflationResource(self._client) + + @cached_property + def prime_rate(self) -> AsyncPrimeRateResource: + return AsyncPrimeRateResource(self._client) + + @cached_property + def with_raw_response(self) -> AsyncV2ResourceWithRawResponse: + """ + This property can be used as a prefix for any HTTP method call to return + the raw response object instead of the parsed content. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + """ + return AsyncV2ResourceWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncV2ResourceWithStreamingResponse: + """ + An alternative to `.with_raw_response` that doesn't eagerly read the response body. + + For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + """ + return AsyncV2ResourceWithStreamingResponse(self) + + +class V2ResourceWithRawResponse: + def __init__(self, v2: V2Resource) -> None: + self._v2 = v2 + + @cached_property + def crypto(self) -> CryptoResourceWithRawResponse: + return CryptoResourceWithRawResponse(self._v2.crypto) + + @cached_property + def currency(self) -> CurrencyResourceWithRawResponse: + return CurrencyResourceWithRawResponse(self._v2.currency) + + @cached_property + def inflation(self) -> InflationResourceWithRawResponse: + return InflationResourceWithRawResponse(self._v2.inflation) + + @cached_property + def prime_rate(self) -> PrimeRateResourceWithRawResponse: + return PrimeRateResourceWithRawResponse(self._v2.prime_rate) + + +class AsyncV2ResourceWithRawResponse: + def __init__(self, v2: AsyncV2Resource) -> None: + self._v2 = v2 + + @cached_property + def crypto(self) -> AsyncCryptoResourceWithRawResponse: + return AsyncCryptoResourceWithRawResponse(self._v2.crypto) + + @cached_property + def currency(self) -> AsyncCurrencyResourceWithRawResponse: + return AsyncCurrencyResourceWithRawResponse(self._v2.currency) + + @cached_property + def inflation(self) -> AsyncInflationResourceWithRawResponse: + return AsyncInflationResourceWithRawResponse(self._v2.inflation) + + @cached_property + def prime_rate(self) -> AsyncPrimeRateResourceWithRawResponse: + return AsyncPrimeRateResourceWithRawResponse(self._v2.prime_rate) + + +class V2ResourceWithStreamingResponse: + def __init__(self, v2: V2Resource) -> None: + self._v2 = v2 + + @cached_property + def crypto(self) -> CryptoResourceWithStreamingResponse: + return CryptoResourceWithStreamingResponse(self._v2.crypto) + + @cached_property + def currency(self) -> CurrencyResourceWithStreamingResponse: + return CurrencyResourceWithStreamingResponse(self._v2.currency) + + @cached_property + def inflation(self) -> InflationResourceWithStreamingResponse: + return InflationResourceWithStreamingResponse(self._v2.inflation) + + @cached_property + def prime_rate(self) -> PrimeRateResourceWithStreamingResponse: + return PrimeRateResourceWithStreamingResponse(self._v2.prime_rate) + + +class AsyncV2ResourceWithStreamingResponse: + def __init__(self, v2: AsyncV2Resource) -> None: + self._v2 = v2 + + @cached_property + def crypto(self) -> AsyncCryptoResourceWithStreamingResponse: + return AsyncCryptoResourceWithStreamingResponse(self._v2.crypto) + + @cached_property + def currency(self) -> AsyncCurrencyResourceWithStreamingResponse: + return AsyncCurrencyResourceWithStreamingResponse(self._v2.currency) + + @cached_property + def inflation(self) -> AsyncInflationResourceWithStreamingResponse: + return AsyncInflationResourceWithStreamingResponse(self._v2.inflation) + + @cached_property + def prime_rate(self) -> AsyncPrimeRateResourceWithStreamingResponse: + return AsyncPrimeRateResourceWithStreamingResponse(self._v2.prime_rate) diff --git a/src/brapi/types/__init__.py b/src/brapi/types/__init__.py new file mode 100644 index 0000000..91dde6b --- /dev/null +++ b/src/brapi/types/__init__.py @@ -0,0 +1,16 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from .cashflow_entry import CashflowEntry as CashflowEntry +from .quote_list_params import QuoteListParams as QuoteListParams +from .value_added_entry import ValueAddedEntry as ValueAddedEntry +from .balance_sheet_entry import BalanceSheetEntry as BalanceSheetEntry +from .quote_list_response import QuoteListResponse as QuoteListResponse +from .financial_data_entry import FinancialDataEntry as FinancialDataEntry +from .available_list_params import AvailableListParams as AvailableListParams +from .quote_retrieve_params import QuoteRetrieveParams as QuoteRetrieveParams +from .income_statement_entry import IncomeStatementEntry as IncomeStatementEntry +from .available_list_response import AvailableListResponse as AvailableListResponse +from .quote_retrieve_response import QuoteRetrieveResponse as QuoteRetrieveResponse +from .default_key_statistics_entry import DefaultKeyStatisticsEntry as DefaultKeyStatisticsEntry diff --git a/src/brapi/types/available_list_params.py b/src/brapi/types/available_list_params.py new file mode 100644 index 0000000..631921d --- /dev/null +++ b/src/brapi/types/available_list_params.py @@ -0,0 +1,30 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import TypedDict + +__all__ = ["AvailableListParams"] + + +class AvailableListParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + search: str + """ + **Opcional.** Termo para filtrar a lista de tickers (correspondência parcial, + case-insensitive). Se omitido, retorna todos os tickers. + """ diff --git a/src/brapi/types/available_list_response.py b/src/brapi/types/available_list_response.py new file mode 100644 index 0000000..20f59dd --- /dev/null +++ b/src/brapi/types/available_list_response.py @@ -0,0 +1,18 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List + +from .._models import BaseModel + +__all__ = ["AvailableListResponse"] + + +class AvailableListResponse(BaseModel): + indexes: List[str] + """Lista de tickers de **índices** disponíveis (ex: `^BVSP`, `^IFIX`).""" + + stocks: List[str] + """ + Lista de tickers de **ações, FIIs, BDRs e ETFs** disponíveis (ex: `PETR4`, + `VALE3`, `MXRF11`). + """ diff --git a/src/brapi/types/balance_sheet_entry.py b/src/brapi/types/balance_sheet_entry.py new file mode 100644 index 0000000..f40cd15 --- /dev/null +++ b/src/brapi/types/balance_sheet_entry.py @@ -0,0 +1,455 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["BalanceSheetEntry"] + + +class BalanceSheetEntry(BaseModel): + accounts_payable: Optional[float] = FieldInfo(alias="accountsPayable", default=None) + """Contas a pagar (fornecedores).""" + + accounts_receivable_from_clients: Optional[float] = FieldInfo(alias="accountsReceivableFromClients", default=None) + """Contas a receber de clientes (bruto).""" + + accumulated_profits_or_losses: Optional[float] = FieldInfo(alias="accumulatedProfitsOrLosses", default=None) + """Lucros ou prejuízos acumulados.""" + + advance_for_future_capital_increase: Optional[float] = FieldInfo( + alias="advanceForFutureCapitalIncrease", default=None + ) + """Adiantamento para futuro aumento de capital (AFAC).""" + + biological_assets: Optional[float] = FieldInfo(alias="biologicalAssets", default=None) + """Ativos biológicos.""" + + capitalization: Optional[float] = None + """Obrigações de capitalização.""" + + capital_reserves: Optional[float] = FieldInfo(alias="capitalReserves", default=None) + """Reservas de capital (sinônimo de `capitalSurplus`).""" + + capital_surplus: Optional[float] = FieldInfo(alias="capitalSurplus", default=None) + """Reservas de capital.""" + + cash: Optional[float] = None + """Caixa e equivalentes de caixa.""" + + central_bank_compulsory_deposit: Optional[float] = FieldInfo(alias="centralBankCompulsoryDeposit", default=None) + """Depósitos compulsórios no Banco Central.""" + + common_stock: Optional[float] = FieldInfo(alias="commonStock", default=None) + """Capital social realizado.""" + + complementary_pension: Optional[float] = FieldInfo(alias="complementaryPension", default=None) + """Obrigações de previdência complementar.""" + + compulsory_loans_and_deposits: Optional[float] = FieldInfo(alias="compulsoryLoansAndDeposits", default=None) + """Empréstimos e depósitos compulsórios.""" + + controller_shareholders_equity: Optional[float] = FieldInfo(alias="controllerShareholdersEquity", default=None) + """Patrimônio líquido atribuível aos controladores.""" + + credits_from_operations: Optional[float] = FieldInfo(alias="creditsFromOperations", default=None) + """Créditos oriundos de operações (instituições financeiras/seguradoras).""" + + credits_with_related_parties: Optional[float] = FieldInfo(alias="creditsWithRelatedParties", default=None) + """Créditos com partes relacionadas.""" + + cumulative_conversion_adjustments: Optional[float] = FieldInfo( + alias="cumulativeConversionAdjustments", default=None + ) + """Ajustes acumulados de conversão.""" + + current_and_deferred_taxes: Optional[float] = FieldInfo(alias="currentAndDeferredTaxes", default=None) + """Tributos correntes e diferidos no ativo.""" + + current_liabilities: Optional[float] = FieldInfo(alias="currentLiabilities", default=None) + """Total do passivo circulante (sinônimo de `totalCurrentLiabilities`).""" + + debentures: Optional[float] = None + """Debêntures (passivo circulante).""" + + debits_from_capitalization: Optional[float] = FieldInfo(alias="debitsFromCapitalization", default=None) + """Débitos de operações de capitalização.""" + + debits_from_complementary_pension: Optional[float] = FieldInfo(alias="debitsFromComplementaryPension", default=None) + """Débitos de operações de previdência complementar.""" + + debits_from_insurance_and_reinsurance: Optional[float] = FieldInfo( + alias="debitsFromInsuranceAndReinsurance", default=None + ) + """Débitos de operações de seguros e resseguros.""" + + debits_from_operations: Optional[float] = FieldInfo(alias="debitsFromOperations", default=None) + """Débitos oriundos de operações.""" + + debits_from_other_operations: Optional[float] = FieldInfo(alias="debitsFromOtherOperations", default=None) + """Débitos de outras operações.""" + + deferred_long_term_asset_charges: Optional[float] = FieldInfo(alias="deferredLongTermAssetCharges", default=None) + """Encargos diferidos de ativos de longo prazo.""" + + deferred_long_term_liab: Optional[float] = FieldInfo(alias="deferredLongTermLiab", default=None) + """Passivos fiscais diferidos (longo prazo).""" + + deferred_selling_expenses: Optional[float] = FieldInfo(alias="deferredSellingExpenses", default=None) + """Despesas de comercialização diferidas.""" + + deferred_taxes: Optional[float] = FieldInfo(alias="deferredTaxes", default=None) + """Tributos diferidos no ativo.""" + + end_date: Optional[date] = FieldInfo(alias="endDate", default=None) + """Data de término do período fiscal ao qual o balanço se refere (YYYY-MM-DD).""" + + equity_valuation_adjustments: Optional[float] = FieldInfo(alias="equityValuationAdjustments", default=None) + """Ajustes de avaliação patrimonial.""" + + financial_assets: Optional[float] = FieldInfo(alias="financialAssets", default=None) + """Ativos financeiros (agregado de instrumentos financeiros no ativo).""" + + financial_assets_at_amortized_cost: Optional[float] = FieldInfo( + alias="financialAssetsAtAmortizedCost", default=None + ) + """Ativos financeiros ao custo amortizado.""" + + financial_assets_measured_at_fair_value_through_other_comprehensive_income: Optional[float] = FieldInfo( + alias="financialAssetsMeasuredAtFairValueThroughOtherComprehensiveIncome", default=None + ) + """ + Ativos financeiros mensurados a valor justo por outros resultados abrangentes + (FVOCI). + """ + + financial_assets_measured_at_fair_value_through_profit_or_loss: Optional[float] = FieldInfo( + alias="financialAssetsMeasuredAtFairValueThroughProfitOrLoss", default=None + ) + """Ativos financeiros mensurados a valor justo por meio do resultado (FVTPL).""" + + financial_investments_measured_at_amortized_cost: Optional[float] = FieldInfo( + alias="financialInvestmentsMeasuredAtAmortizedCost", default=None + ) + """Investimentos financeiros mensurados ao custo amortizado.""" + + financial_investments_measured_at_fair_value_through_other_comprehensive_income: Optional[float] = FieldInfo( + alias="financialInvestmentsMeasuredAtFairValueThroughOtherComprehensiveIncome", default=None + ) + """ + Investimentos financeiros mensurados a valor justo por outros resultados + abrangentes. + """ + + financial_liabilities_at_amortized_cost: Optional[float] = FieldInfo( + alias="financialLiabilitiesAtAmortizedCost", default=None + ) + """Passivos financeiros ao custo amortizado.""" + + financial_liabilities_measured_at_fair_value_through_income: Optional[float] = FieldInfo( + alias="financialLiabilitiesMeasuredAtFairValueThroughIncome", default=None + ) + """Passivos financeiros mensurados a valor justo por meio do resultado.""" + + foreign_suppliers: Optional[float] = FieldInfo(alias="foreignSuppliers", default=None) + """Fornecedores estrangeiros.""" + + good_will: Optional[float] = FieldInfo(alias="goodWill", default=None) + """Ágio por expectativa de rentabilidade futura (Goodwill).""" + + insurance_and_reinsurance: Optional[float] = FieldInfo(alias="insuranceAndReinsurance", default=None) + """Provisões/obrigações de seguros e resseguros.""" + + intangible_asset: Optional[float] = FieldInfo(alias="intangibleAsset", default=None) + """Ativo intangível (valor agregado).""" + + intangible_assets: Optional[float] = FieldInfo(alias="intangibleAssets", default=None) + """Ativos intangíveis (marcas, patentes, etc.).""" + + inventory: Optional[float] = None + """Estoques.""" + + investment_properties: Optional[float] = FieldInfo(alias="investmentProperties", default=None) + """Propriedades para investimento.""" + + investments: Optional[float] = None + """Investimentos (participações e outros).""" + + lease_financing: Optional[float] = FieldInfo(alias="leaseFinancing", default=None) + """Financiamento por arrendamento mercantil (circulante).""" + + loans_and_financing: Optional[float] = FieldInfo(alias="loansAndFinancing", default=None) + """Empréstimos e financiamentos (circulante).""" + + loans_and_financing_in_foreign_currency: Optional[float] = FieldInfo( + alias="loansAndFinancingInForeignCurrency", default=None + ) + """Empréstimos e financiamentos em moeda estrangeira (circulante).""" + + loans_and_financing_in_national_currency: Optional[float] = FieldInfo( + alias="loansAndFinancingInNationalCurrency", default=None + ) + """Empréstimos e financiamentos em moeda nacional (circulante).""" + + long_term_accounts_payable: Optional[float] = FieldInfo(alias="longTermAccountsPayable", default=None) + """Fornecedores/contas a pagar de longo prazo.""" + + long_term_accounts_receivable_from_clients: Optional[float] = FieldInfo( + alias="longTermAccountsReceivableFromClients", default=None + ) + """Contas a receber de clientes - longo prazo.""" + + long_term_assets: Optional[float] = FieldInfo(alias="longTermAssets", default=None) + """Total do ativo não circulante (agregado).""" + + long_term_biological_assets: Optional[float] = FieldInfo(alias="longTermBiologicalAssets", default=None) + """Ativos biológicos de longo prazo.""" + + long_term_capitalization: Optional[float] = FieldInfo(alias="longTermCapitalization", default=None) + """Obrigações de capitalização de longo prazo.""" + + long_term_complementary_pension: Optional[float] = FieldInfo(alias="longTermComplementaryPension", default=None) + """Obrigações de previdência complementar de longo prazo.""" + + long_term_debentures: Optional[float] = FieldInfo(alias="longTermDebentures", default=None) + """Debêntures (passivo não circulante).""" + + long_term_debits_from_operations: Optional[float] = FieldInfo(alias="longTermDebitsFromOperations", default=None) + """Débitos de operações (longo prazo).""" + + long_term_debt: Optional[float] = FieldInfo(alias="longTermDebt", default=None) + """Dívida de longo prazo (empréstimos e financiamentos não circulantes).""" + + long_term_deferred_taxes: Optional[float] = FieldInfo(alias="longTermDeferredTaxes", default=None) + """Tributos diferidos (Ativo Não Circulante).""" + + long_term_financial_investments_measured_at_fair_value_through_income: Optional[float] = FieldInfo( + alias="longTermFinancialInvestmentsMeasuredAtFairValueThroughIncome", default=None + ) + """ + Investimentos financeiros de longo prazo mensurados a valor justo por meio do + resultado. + """ + + long_term_insurance_and_reinsurance: Optional[float] = FieldInfo( + alias="longTermInsuranceAndReinsurance", default=None + ) + """Obrigações de seguros e resseguros de longo prazo.""" + + long_term_inventory: Optional[float] = FieldInfo(alias="longTermInventory", default=None) + """Estoques de longo prazo.""" + + long_term_investments: Optional[float] = FieldInfo(alias="longTermInvestments", default=None) + """Investimentos de longo prazo.""" + + long_term_lease_financing: Optional[float] = FieldInfo(alias="longTermLeaseFinancing", default=None) + """Financiamento por arrendamento mercantil (não circulante).""" + + long_term_liabilities: Optional[float] = FieldInfo(alias="longTermLiabilities", default=None) + """Total do passivo de longo prazo.""" + + long_term_loans_and_financing: Optional[float] = FieldInfo(alias="longTermLoansAndFinancing", default=None) + """Empréstimos e financiamentos (não circulante).""" + + long_term_loans_and_financing_in_foreign_currency: Optional[float] = FieldInfo( + alias="longTermLoansAndFinancingInForeignCurrency", default=None + ) + """Empréstimos e financiamentos em moeda estrangeira (não circulante).""" + + long_term_loans_and_financing_in_national_currency: Optional[float] = FieldInfo( + alias="longTermLoansAndFinancingInNationalCurrency", default=None + ) + """Empréstimos e financiamentos em moeda nacional (não circulante).""" + + long_term_prepaid_expenses: Optional[float] = FieldInfo(alias="longTermPrepaidExpenses", default=None) + """Despesas antecipadas de longo prazo.""" + + long_term_provisions: Optional[float] = FieldInfo(alias="longTermProvisions", default=None) + """Provisões (passivo não circulante).""" + + long_term_realizable_assets: Optional[float] = FieldInfo(alias="longTermRealizableAssets", default=None) + """Ativo realizável a longo prazo.""" + + long_term_receivables: Optional[float] = FieldInfo(alias="longTermReceivables", default=None) + """Contas a receber de longo prazo.""" + + long_term_technical_provisions: Optional[float] = FieldInfo(alias="longTermTechnicalProvisions", default=None) + """Provisões técnicas de longo prazo.""" + + minority_interest: Optional[float] = FieldInfo(alias="minorityInterest", default=None) + """Participação de não controladores (no patrimônio líquido).""" + + national_suppliers: Optional[float] = FieldInfo(alias="nationalSuppliers", default=None) + """Fornecedores nacionais.""" + + net_receivables: Optional[float] = FieldInfo(alias="netReceivables", default=None) + """Contas a receber líquidas (clientes).""" + + net_tangible_assets: Optional[float] = FieldInfo(alias="netTangibleAssets", default=None) + """Ativos tangíveis líquidos (Ativo Total - Intangíveis - Passivo Total).""" + + non_controlling_shareholders_equity: Optional[float] = FieldInfo( + alias="nonControllingShareholdersEquity", default=None + ) + """Participação dos não controladores no patrimônio líquido.""" + + non_current_assets: Optional[float] = FieldInfo(alias="nonCurrentAssets", default=None) + """Total do ativo não circulante (sinônimo de `longTermAssets`).""" + + non_current_liabilities: Optional[float] = FieldInfo(alias="nonCurrentLiabilities", default=None) + """Total do passivo não circulante.""" + + other_accounts_receivable: Optional[float] = FieldInfo(alias="otherAccountsReceivable", default=None) + """Outras contas a receber.""" + + other_assets: Optional[float] = FieldInfo(alias="otherAssets", default=None) + """Outros ativos não circulantes.""" + + other_comprehensive_results: Optional[float] = FieldInfo(alias="otherComprehensiveResults", default=None) + """Outros resultados abrangentes.""" + + other_current_assets: Optional[float] = FieldInfo(alias="otherCurrentAssets", default=None) + """Outros ativos circulantes.""" + + other_current_liab: Optional[float] = FieldInfo(alias="otherCurrentLiab", default=None) + """Outros passivos circulantes.""" + + other_current_liabilities: Optional[float] = FieldInfo(alias="otherCurrentLiabilities", default=None) + """Outros passivos circulantes (sinônimo de `otherCurrentLiab`).""" + + other_debits: Optional[float] = FieldInfo(alias="otherDebits", default=None) + """Outros débitos.""" + + other_liab: Optional[float] = FieldInfo(alias="otherLiab", default=None) + """Outros passivos não circulantes.""" + + other_liabilities: Optional[float] = FieldInfo(alias="otherLiabilities", default=None) + """Outros passivos.""" + + other_long_term_obligations: Optional[float] = FieldInfo(alias="otherLongTermObligations", default=None) + """Outras obrigações (passivo não circulante).""" + + other_long_term_provisions: Optional[float] = FieldInfo(alias="otherLongTermProvisions", default=None) + """Outras provisões de longo prazo.""" + + other_long_term_receivables: Optional[float] = FieldInfo(alias="otherLongTermReceivables", default=None) + """Outros créditos/recebíveis de longo prazo.""" + + other_non_current_assets: Optional[float] = FieldInfo(alias="otherNonCurrentAssets", default=None) + """Outros ativos não circulantes (detalhamento).""" + + other_non_current_liabilities: Optional[float] = FieldInfo(alias="otherNonCurrentLiabilities", default=None) + """Outros passivos não circulantes.""" + + other_obligations: Optional[float] = FieldInfo(alias="otherObligations", default=None) + """Outras obrigações (circulante).""" + + other_operations: Optional[float] = FieldInfo(alias="otherOperations", default=None) + """Outras contas operacionais no ativo.""" + + other_provisions: Optional[float] = FieldInfo(alias="otherProvisions", default=None) + """Outras provisões (diversas).""" + + other_stockholder_equity: Optional[float] = FieldInfo(alias="otherStockholderEquity", default=None) + """Outros componentes do patrimônio líquido.""" + + other_values_and_assets: Optional[float] = FieldInfo(alias="otherValuesAndAssets", default=None) + """Outros valores e bens.""" + + prepaid_expenses: Optional[float] = FieldInfo(alias="prepaidExpenses", default=None) + """Despesas antecipadas.""" + + profit_reserves: Optional[float] = FieldInfo(alias="profitReserves", default=None) + """Reservas de lucros.""" + + profits_and_revenues_to_be_appropriated: Optional[float] = FieldInfo( + alias="profitsAndRevenuesToBeAppropriated", default=None + ) + """Lucros e receitas a apropriar.""" + + property_plant_equipment: Optional[float] = FieldInfo(alias="propertyPlantEquipment", default=None) + """Imobilizado (propriedades, instalações e equipamentos).""" + + providers: Optional[float] = None + """Fornecedores (sinônimo de `accountsPayable`).""" + + provisions: Optional[float] = None + """Provisões (passivo).""" + + realized_share_capital: Optional[float] = FieldInfo(alias="realizedShareCapital", default=None) + """Capital social realizado (sinônimo de `commonStock`).""" + + retained_earnings: Optional[float] = FieldInfo(alias="retainedEarnings", default=None) + """Lucros/Prejuízos acumulados.""" + + revaluation_reserves: Optional[float] = FieldInfo(alias="revaluationReserves", default=None) + """Reservas de reavaliação.""" + + securities_and_credits_receivable: Optional[float] = FieldInfo(alias="securitiesAndCreditsReceivable", default=None) + """Títulos e créditos a receber.""" + + shareholders_equity: Optional[float] = FieldInfo(alias="shareholdersEquity", default=None) + """Patrimônio líquido (sinônimo de `totalStockholderEquity`).""" + + shareholdings: Optional[float] = None + """Participações societárias.""" + + short_long_term_debt: Optional[float] = FieldInfo(alias="shortLongTermDebt", default=None) + """Dívida de curto prazo (empréstimos e financiamentos circulantes).""" + + short_term_investments: Optional[float] = FieldInfo(alias="shortTermInvestments", default=None) + """Aplicações financeiras de curto prazo.""" + + social_and_labor_obligations: Optional[float] = FieldInfo(alias="socialAndLaborObligations", default=None) + """Obrigações sociais e trabalhistas.""" + + symbol: Optional[str] = None + """Ticker do ativo ao qual o balanço se refere.""" + + taxes_to_recover: Optional[float] = FieldInfo(alias="taxesToRecover", default=None) + """Impostos a recuperar.""" + + tax_liabilities: Optional[float] = FieldInfo(alias="taxLiabilities", default=None) + """Obrigações fiscais (passivo).""" + + tax_obligations: Optional[float] = FieldInfo(alias="taxObligations", default=None) + """Obrigações fiscais (passivo circulante).""" + + technical_provisions: Optional[float] = FieldInfo(alias="technicalProvisions", default=None) + """Provisões técnicas (seguradoras/previdência).""" + + third_party_deposits: Optional[float] = FieldInfo(alias="thirdPartyDeposits", default=None) + """Depósitos de terceiros.""" + + total_assets: Optional[float] = FieldInfo(alias="totalAssets", default=None) + """Total do ativo.""" + + total_current_assets: Optional[float] = FieldInfo(alias="totalCurrentAssets", default=None) + """Total do ativo circulante.""" + + total_current_liabilities: Optional[float] = FieldInfo(alias="totalCurrentLiabilities", default=None) + """Total do passivo circulante.""" + + total_liab: Optional[float] = FieldInfo(alias="totalLiab", default=None) + """Total do passivo (circulante + não circulante).""" + + total_liabilities: Optional[float] = FieldInfo(alias="totalLiabilities", default=None) + """Total do passivo.""" + + total_stockholder_equity: Optional[float] = FieldInfo(alias="totalStockholderEquity", default=None) + """Total do patrimônio líquido.""" + + treasury_stock: Optional[float] = FieldInfo(alias="treasuryStock", default=None) + """Ações em tesouraria.""" + + type: Optional[Literal["yearly", "quarterly"]] = None + """ + Indica a periodicidade do balanço: `yearly` (anual) ou `quarterly` (trimestral). + """ + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """Data da última atualização deste registro (YYYY-MM-DD).""" diff --git a/src/brapi/types/cashflow_entry.py b/src/brapi/types/cashflow_entry.py new file mode 100644 index 0000000..b8b7fb9 --- /dev/null +++ b/src/brapi/types/cashflow_entry.py @@ -0,0 +1,89 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["CashflowEntry"] + + +class CashflowEntry(BaseModel): + adjustments_to_profit_or_loss: Optional[float] = FieldInfo(alias="adjustmentsToProfitOrLoss", default=None) + """ + Ajustes ao lucro/prejuízo (depreciação, amortização, equivalência patrimonial, + variações não caixa). + """ + + cash_generated_in_operations: Optional[float] = FieldInfo(alias="cashGeneratedInOperations", default=None) + """Caixa gerado nas operações (após variações no capital de giro).""" + + changes_in_assets_and_liabilities: Optional[float] = FieldInfo(alias="changesInAssetsAndLiabilities", default=None) + """ + Variações em Ativos e Passivos Operacionais (Clientes, Estoques, Fornecedores, + etc.). + """ + + end_date: Optional[date] = FieldInfo(alias="endDate", default=None) + """Data de término do período fiscal ao qual a DFC se refere (YYYY-MM-DD).""" + + exchange_variation_without_cash: Optional[float] = FieldInfo(alias="exchangeVariationWithoutCash", default=None) + """Variação cambial sem efeito caixa (ajuste de conversão).""" + + final_cash_balance: Optional[float] = FieldInfo(alias="finalCashBalance", default=None) + """Saldo Final de Caixa e Equivalentes no final do período.""" + + financing_cash_flow: Optional[float] = FieldInfo(alias="financingCashFlow", default=None) + """ + Fluxo de Caixa das Atividades de Financiamento (FCF) (Captação/Pagamento de + Empréstimos, Emissão/Recompra de Ações, Dividendos pagos). + """ + + foreign_exchange_rate_without_cash: Optional[float] = FieldInfo( + alias="foreignExchangeRateWithoutCash", default=None + ) + """Efeito da Variação Cambial sobre o Caixa e Equivalentes.""" + + income_from_operations: Optional[float] = FieldInfo(alias="incomeFromOperations", default=None) + """Caixa Gerado nas Operações (antes das variações de ativos/passivos).""" + + increase_or_decrease_in_cash: Optional[float] = FieldInfo(alias="increaseOrDecreaseInCash", default=None) + """ + Aumento ou Redução Líquida de Caixa e Equivalentes (FCO + FCI + FCF + Variação + Cambial). + """ + + initial_cash_balance: Optional[float] = FieldInfo(alias="initialCashBalance", default=None) + """Saldo Inicial de Caixa e Equivalentes no início do período.""" + + investment_cash_flow: Optional[float] = FieldInfo(alias="investmentCashFlow", default=None) + """ + Fluxo de Caixa das Atividades de Investimento (FCI) (Compra/Venda de + Imobilizado, Investimentos). + """ + + net_income_before_taxes: Optional[float] = FieldInfo(alias="netIncomeBeforeTaxes", default=None) + """ + Lucro líquido antes dos impostos (base para reconciliação pelo método indireto). + """ + + operating_cash_flow: Optional[float] = FieldInfo(alias="operatingCashFlow", default=None) + """Fluxo de Caixa das Atividades Operacionais (FCO).""" + + other_operating_activities: Optional[float] = FieldInfo(alias="otherOperatingActivities", default=None) + """Outras Atividades Operacionais (Juros pagos/recebidos, Impostos pagos, etc.).""" + + symbol: Optional[str] = None + """Ticker do ativo ao qual a DFC se refere.""" + + type: Optional[Literal["yearly", "quarterly"]] = None + """Indica a periodicidade da DFC: `yearly` (anual) ou `quarterly` (trimestral).""" + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """ + Data da última atualização deste registro específico na fonte de dados + (YYYY-MM-DD). + """ diff --git a/src/brapi/types/default_key_statistics_entry.py b/src/brapi/types/default_key_statistics_entry.py new file mode 100644 index 0000000..dc1315b --- /dev/null +++ b/src/brapi/types/default_key_statistics_entry.py @@ -0,0 +1,138 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["DefaultKeyStatisticsEntry"] + + +class DefaultKeyStatisticsEntry(BaseModel): + api_52_week_change: Optional[float] = FieldInfo(alias="52WeekChange", default=None) + """Variação percentual do preço da ação nas últimas 52 semanas.""" + + beta: Optional[float] = None + """Beta da ação (sensibilidade em relação ao mercado).""" + + book_value: Optional[float] = FieldInfo(alias="bookValue", default=None) + """Valor Patrimonial por Ação (VPA): Patrimônio Líquido / Ações em Circulação.""" + + dividend_yield: Optional[float] = FieldInfo(alias="dividendYield", default=None) + """Dividend Yield (provento anualizado sobre o preço atual).""" + + earnings_annual_growth: Optional[float] = FieldInfo(alias="earningsAnnualGrowth", default=None) + """ + Crescimento percentual do lucro líquido no último ano fiscal completo em relação + ao ano anterior. + """ + + earnings_quarterly_growth: Optional[float] = FieldInfo(alias="earningsQuarterlyGrowth", default=None) + """ + Crescimento percentual do lucro líquido no último trimestre em relação ao mesmo + trimestre do ano anterior (YoY). + """ + + enterprise_to_ebitda: Optional[float] = FieldInfo(alias="enterpriseToEbitda", default=None) + """Múltiplo EV/EBITDA (Enterprise Value / EBITDA TTM).""" + + enterprise_to_revenue: Optional[float] = FieldInfo(alias="enterpriseToRevenue", default=None) + """Múltiplo EV/Receita (Enterprise Value / Receita Líquida TTM).""" + + enterprise_value: Optional[float] = FieldInfo(alias="enterpriseValue", default=None) + """Valor da Firma (Enterprise Value - EV): Market Cap + Dívida Total - Caixa.""" + + float_shares: Optional[float] = FieldInfo(alias="floatShares", default=None) + """Ações em livre circulação (free float).""" + + forward_eps: Optional[float] = FieldInfo(alias="forwardEps", default=None) + """Lucro Por Ação projetado (próximo período).""" + + forward_pe: Optional[float] = FieldInfo(alias="forwardPE", default=None) + """ + Preço / Lucro Projetado (Forward P/E): Preço da Ação / LPA estimado para o + próximo período. + """ + + held_percent_insiders: Optional[float] = FieldInfo(alias="heldPercentInsiders", default=None) + """Percentual de ações detidas por insiders (administradores, controladores).""" + + held_percent_institutions: Optional[float] = FieldInfo(alias="heldPercentInstitutions", default=None) + """ + Percentual de ações detidas por instituições (fundos, investidores + institucionais). + """ + + implied_shares_outstanding: Optional[float] = FieldInfo(alias="impliedSharesOutstanding", default=None) + """Ações implícitas em circulação (considerando diluição/derivativos).""" + + last_dividend_date: Optional[date] = FieldInfo(alias="lastDividendDate", default=None) + """Data de pagamento (ou 'Data Com') do último dividendo/JCP (YYYY-MM-DD).""" + + last_dividend_value: Optional[float] = FieldInfo(alias="lastDividendValue", default=None) + """Valor do último dividendo ou JCP pago por ação.""" + + last_fiscal_year_end: Optional[date] = FieldInfo(alias="lastFiscalYearEnd", default=None) + """Data de encerramento do último ano fiscal (YYYY-MM-DD).""" + + last_split_date: Optional[float] = FieldInfo(alias="lastSplitDate", default=None) + """Data do último desdobramento/grupamento (timestamp UNIX em segundos).""" + + last_split_factor: Optional[str] = FieldInfo(alias="lastSplitFactor", default=None) + """Fator do último desdobramento/grupamento (ex.: 2:1, 1:10).""" + + most_recent_quarter: Optional[date] = FieldInfo(alias="mostRecentQuarter", default=None) + """ + Data de término do trimestre mais recente considerado nos cálculos (YYYY-MM-DD). + """ + + net_income_to_common: Optional[float] = FieldInfo(alias="netIncomeToCommon", default=None) + """Lucro Líquido atribuível aos acionistas ordinários (controladores).""" + + next_fiscal_year_end: Optional[date] = FieldInfo(alias="nextFiscalYearEnd", default=None) + """Data de encerramento do próximo ano fiscal (YYYY-MM-DD).""" + + peg_ratio: Optional[float] = FieldInfo(alias="pegRatio", default=None) + """Índice PEG (P/E dividido pelo crescimento esperado dos lucros).""" + + price_to_book: Optional[float] = FieldInfo(alias="priceToBook", default=None) + """Preço sobre Valor Patrimonial (P/VP): Preço da Ação / VPA.""" + + profit_margins: Optional[float] = FieldInfo(alias="profitMargins", default=None) + """Margem de Lucro Líquida (Lucro Líquido / Receita Líquida). + + Geralmente em base TTM ou anual. + """ + + sand_p52_week_change: Optional[float] = FieldInfo(alias="SandP52WeekChange", default=None) + """Variação percentual do índice S&P 500 nas últimas 52 semanas (para referência).""" + + shares_outstanding: Optional[float] = FieldInfo(alias="sharesOutstanding", default=None) + """Número total de ações ordinárias em circulação.""" + + symbol: Optional[str] = None + """Ticker do ativo ao qual as estatísticas se referem.""" + + total_assets: Optional[float] = FieldInfo(alias="totalAssets", default=None) + """Valor total dos ativos registrado no último balanço (anual ou trimestral).""" + + trailing_eps: Optional[float] = FieldInfo(alias="trailingEps", default=None) + """Lucro Por Ação (LPA) dos Últimos 12 Meses (TTM).""" + + type: Optional[Literal["yearly", "quarterly", "ttm"]] = None + """ + Periodicidade dos dados: `yearly` (anual), `quarterly` (trimestral), `ttm` + (Trailing Twelve Months - últimos 12 meses). + """ + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """ + Data da última atualização deste registro específico na fonte de dados + (YYYY-MM-DD). + """ + + ytd_return: Optional[float] = FieldInfo(alias="ytdReturn", default=None) + """Retorno percentual do preço da ação desde o início do ano atual (Year-to-Date).""" diff --git a/src/brapi/types/financial_data_entry.py b/src/brapi/types/financial_data_entry.py new file mode 100644 index 0000000..3db991a --- /dev/null +++ b/src/brapi/types/financial_data_entry.py @@ -0,0 +1,133 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["FinancialDataEntry"] + + +class FinancialDataEntry(BaseModel): + current_price: Optional[float] = FieldInfo(alias="currentPrice", default=None) + """Preço atual da ação (pode ser ligeiramente defasado).""" + + current_ratio: Optional[float] = FieldInfo(alias="currentRatio", default=None) + """Índice de Liquidez Corrente (Ativo Circulante / Passivo Circulante).""" + + debt_to_equity: Optional[float] = FieldInfo(alias="debtToEquity", default=None) + """Índice Dívida Líquida / Patrimônio Líquido.""" + + earnings_growth: Optional[float] = FieldInfo(alias="earningsGrowth", default=None) + """ + Crescimento do Lucro Líquido (geralmente trimestral YoY, como + `earningsQuarterlyGrowth`). + """ + + ebitda: Optional[float] = None + """Lucro Antes de Juros, Impostos, Depreciação e Amortização (LAJIDA ou EBITDA). + + Geralmente TTM. + """ + + ebitda_margins: Optional[float] = FieldInfo(alias="ebitdaMargins", default=None) + """Margem EBITDA (EBITDA TTM / Receita Líquida TTM).""" + + financial_currency: Optional[str] = FieldInfo(alias="financialCurrency", default=None) + """Moeda na qual os dados financeiros são reportados (ex: `BRL`, `USD`).""" + + free_cashflow: Optional[float] = FieldInfo(alias="freeCashflow", default=None) + """Fluxo de Caixa Livre (FCO - CAPEX) - (geralmente TTM).""" + + gross_margins: Optional[float] = FieldInfo(alias="grossMargins", default=None) + """Margem Bruta (Lucro Bruto TTM / Receita Líquida TTM).""" + + gross_profits: Optional[float] = FieldInfo(alias="grossProfits", default=None) + """Lucro Bruto (geralmente TTM).""" + + number_of_analyst_opinions: Optional[float] = FieldInfo(alias="numberOfAnalystOpinions", default=None) + """Número de opiniões de analistas consideradas.""" + + operating_cashflow: Optional[float] = FieldInfo(alias="operatingCashflow", default=None) + """Fluxo de Caixa das Operações (FCO) - (geralmente TTM).""" + + operating_margins: Optional[float] = FieldInfo(alias="operatingMargins", default=None) + """Margem Operacional (EBIT TTM / Receita Líquida TTM).""" + + profit_margins: Optional[float] = FieldInfo(alias="profitMargins", default=None) + """Margem Líquida (Lucro Líquido TTM / Receita Líquida TTM). + + Sinônimo do campo de mesmo nome em `DefaultKeyStatisticsEntry`. + """ + + quick_ratio: Optional[float] = FieldInfo(alias="quickRatio", default=None) + """Índice de Liquidez Seca ((Ativo Circulante - Estoques) / Passivo Circulante).""" + + recommendation_key: Optional[str] = FieldInfo(alias="recommendationKey", default=None) + """Resumo da recomendação (ex.: strong_buy, buy, hold, sell, strong_sell).""" + + recommendation_mean: Optional[float] = FieldInfo(alias="recommendationMean", default=None) + """Média de recomendações dos analistas (1=Compra Forte, 5=Venda Forte).""" + + return_on_assets: Optional[float] = FieldInfo(alias="returnOnAssets", default=None) + """Retorno sobre Ativos (ROA): Lucro Líquido TTM / Ativo Total Médio.""" + + return_on_equity: Optional[float] = FieldInfo(alias="returnOnEquity", default=None) + """ + Retorno sobre Patrimônio Líquido (ROE): Lucro Líquido TTM / Patrimônio Líquido + Médio. + """ + + revenue_growth: Optional[float] = FieldInfo(alias="revenueGrowth", default=None) + """Crescimento da Receita Líquida (geralmente trimestral YoY).""" + + revenue_per_share: Optional[float] = FieldInfo(alias="revenuePerShare", default=None) + """Receita Líquida por Ação (Receita Líquida TTM / Ações em Circulação).""" + + symbol: Optional[str] = None + """Ticker do ativo ao qual os dados se referem.""" + + target_high_price: Optional[float] = FieldInfo(alias="targetHighPrice", default=None) + """Preço-alvo mais alto estimado por analistas.""" + + target_low_price: Optional[float] = FieldInfo(alias="targetLowPrice", default=None) + """Preço-alvo mais baixo estimado por analistas.""" + + target_mean_price: Optional[float] = FieldInfo(alias="targetMeanPrice", default=None) + """Preço-alvo médio estimado por analistas.""" + + target_median_price: Optional[float] = FieldInfo(alias="targetMedianPrice", default=None) + """Preço-alvo mediano estimado por analistas.""" + + total_cash: Optional[float] = FieldInfo(alias="totalCash", default=None) + """ + Caixa e Equivalentes de Caixa + Aplicações Financeiras de Curto Prazo (último + balanço). + """ + + total_cash_per_share: Optional[float] = FieldInfo(alias="totalCashPerShare", default=None) + """Caixa Total por Ação (Caixa Total / Ações em Circulação).""" + + total_debt: Optional[float] = FieldInfo(alias="totalDebt", default=None) + """ + Dívida Bruta Total (Dívida de Curto Prazo + Dívida de Longo Prazo - último + balanço). + """ + + total_revenue: Optional[float] = FieldInfo(alias="totalRevenue", default=None) + """Receita Líquida Total (geralmente TTM).""" + + type: Optional[Literal["yearly", "quarterly", "ttm"]] = None + """ + Periodicidade dos dados: `yearly` (anual), `quarterly` (trimestral), `ttm` + (Trailing Twelve Months). + """ + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """ + Data da última atualização deste registro específico na fonte de dados + (YYYY-MM-DD). + """ diff --git a/src/brapi/types/income_statement_entry.py b/src/brapi/types/income_statement_entry.py new file mode 100644 index 0000000..7429e0a --- /dev/null +++ b/src/brapi/types/income_statement_entry.py @@ -0,0 +1,198 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["IncomeStatementEntry"] + + +class IncomeStatementEntry(BaseModel): + id: Optional[str] = None + """Identificador único deste registro de DRE (interno).""" + + administrative_costs: Optional[float] = FieldInfo(alias="administrativeCosts", default=None) + """Despesas Administrativas (detalhamento, pode estar contido em SG&A).""" + + basic_earnings_per_common_share: Optional[float] = FieldInfo(alias="basicEarningsPerCommonShare", default=None) + """Lucro Básico por Ação Ordinária (ON).""" + + basic_earnings_per_preferred_share: Optional[float] = FieldInfo( + alias="basicEarningsPerPreferredShare", default=None + ) + """Lucro Básico por Ação Preferencial (PN).""" + + basic_earnings_per_share: Optional[float] = FieldInfo(alias="basicEarningsPerShare", default=None) + """Lucro Básico por Ação (LPA Básico) - Geral.""" + + capitalization_operations: Optional[float] = FieldInfo(alias="capitalizationOperations", default=None) + """Resultado de Operações de Capitalização (específico para Seguradoras).""" + + claims_and_operations_costs: Optional[float] = FieldInfo(alias="claimsAndOperationsCosts", default=None) + """Custos com Sinistros e Operações (específico para Seguradoras).""" + + complementary_pension_operations: Optional[float] = FieldInfo(alias="complementaryPensionOperations", default=None) + """ + Resultado de Operações de Previdência Complementar (específico para + Seguradoras/Previdência). + """ + + cost_of_revenue: Optional[float] = FieldInfo(alias="costOfRevenue", default=None) + """Custo dos Produtos Vendidos (CPV) ou Custo dos Serviços Prestados (CSP).""" + + current_taxes: Optional[float] = FieldInfo(alias="currentTaxes", default=None) + """Imposto de Renda e Contribuição Social Correntes.""" + + deferred_taxes: Optional[float] = FieldInfo(alias="deferredTaxes", default=None) + """Imposto de Renda e Contribuição Social Diferidos.""" + + diluted_earnings_per_common_share: Optional[float] = FieldInfo(alias="dilutedEarningsPerCommonShare", default=None) + """Lucro Diluído por Ação Ordinária (ON).""" + + diluted_earnings_per_preferred_share: Optional[float] = FieldInfo( + alias="dilutedEarningsPerPreferredShare", default=None + ) + """Lucro Diluído por Ação Preferencial (PN).""" + + diluted_earnings_per_share: Optional[float] = FieldInfo(alias="dilutedEarningsPerShare", default=None) + """Lucro Diluído por Ação (LPA Diluído) - Geral.""" + + discontinued_operations: Optional[float] = FieldInfo(alias="discontinuedOperations", default=None) + """Resultado Líquido das Operações Descontinuadas.""" + + earnings_per_share: Optional[float] = FieldInfo(alias="earningsPerShare", default=None) + """Lucro por Ação (LPA) - Geral (pode ser básico ou diluído, verificar contexto).""" + + ebit: Optional[float] = None + """Lucro Antes dos Juros e Impostos (LAJIR ou EBIT). + + Geralmente igual a `operatingIncome`. + """ + + effect_of_accounting_charges: Optional[float] = FieldInfo(alias="effectOfAccountingCharges", default=None) + """Efeito de Mudanças Contábeis.""" + + end_date: Optional[date] = FieldInfo(alias="endDate", default=None) + """Data de término do período fiscal ao qual a DRE se refere (YYYY-MM-DD).""" + + equity_income_result: Optional[float] = FieldInfo(alias="equityIncomeResult", default=None) + """Resultado de Equivalência Patrimonial.""" + + extraordinary_items: Optional[float] = FieldInfo(alias="extraordinaryItems", default=None) + """Itens Extraordinários.""" + + financial_expenses: Optional[float] = FieldInfo(alias="financialExpenses", default=None) + """Despesas Financeiras (valor positivo aqui, diferente de `interestExpense`).""" + + financial_income: Optional[float] = FieldInfo(alias="financialIncome", default=None) + """Receitas Financeiras.""" + + financial_result: Optional[float] = FieldInfo(alias="financialResult", default=None) + """Resultado Financeiro Líquido.""" + + gross_profit: Optional[float] = FieldInfo(alias="grossProfit", default=None) + """Lucro Bruto (Receita Líquida - CPV/CSP).""" + + income_before_statutory_participations_and_contributions: Optional[float] = FieldInfo( + alias="incomeBeforeStatutoryParticipationsAndContributions", default=None + ) + """Resultado Antes das Participações Estatutárias.""" + + income_before_tax: Optional[float] = FieldInfo(alias="incomeBeforeTax", default=None) + """Lucro Antes do Imposto de Renda e Contribuição Social (LAIR). + + EBIT + Resultado Financeiro. + """ + + income_tax_expense: Optional[float] = FieldInfo(alias="incomeTaxExpense", default=None) + """Imposto de Renda e Contribuição Social sobre o Lucro.""" + + insurance_operations: Optional[float] = FieldInfo(alias="insuranceOperations", default=None) + """Resultado de Operações de Seguros (específico para Seguradoras).""" + + interest_expense: Optional[float] = FieldInfo(alias="interestExpense", default=None) + """Despesas Financeiras (Juros pagos). Note que este campo é negativo.""" + + losses_due_to_non_recoverability_of_assets: Optional[float] = FieldInfo( + alias="lossesDueToNonRecoverabilityOfAssets", default=None + ) + """Perdas por Não Recuperabilidade de Ativos (Impairment).""" + + minority_interest: Optional[float] = FieldInfo(alias="minorityInterest", default=None) + """Participação de Acionistas Não Controladores (no Lucro Líquido).""" + + net_income: Optional[float] = FieldInfo(alias="netIncome", default=None) + """Lucro Líquido Consolidado do Período.""" + + net_income_applicable_to_common_shares: Optional[float] = FieldInfo( + alias="netIncomeApplicableToCommonShares", default=None + ) + """Lucro Líquido Atribuível aos Acionistas Controladores (Ações Ordinárias).""" + + net_income_from_continuing_ops: Optional[float] = FieldInfo(alias="netIncomeFromContinuingOps", default=None) + """Lucro Líquido das Operações Continuadas.""" + + non_recurring: Optional[float] = FieldInfo(alias="nonRecurring", default=None) + """Itens Não Recorrentes (pode incluir outras despesas/receitas operacionais).""" + + operating_income: Optional[float] = FieldInfo(alias="operatingIncome", default=None) + """Lucro Operacional (EBIT - Earnings Before Interest and Taxes). + + Lucro Bruto - Despesas Operacionais. + """ + + other_items: Optional[float] = FieldInfo(alias="otherItems", default=None) + """Outros Itens.""" + + other_operating_expenses: Optional[float] = FieldInfo(alias="otherOperatingExpenses", default=None) + """Outras Despesas Operacionais.""" + + other_operating_income: Optional[float] = FieldInfo(alias="otherOperatingIncome", default=None) + """Outras Receitas Operacionais (detalhamento).""" + + other_operating_income_and_expenses: Optional[float] = FieldInfo( + alias="otherOperatingIncomeAndExpenses", default=None + ) + """Outras Receitas e Despesas Operacionais (agregado).""" + + profit_sharing_and_statutory_contributions: Optional[float] = FieldInfo( + alias="profitSharingAndStatutoryContributions", default=None + ) + """Participações nos Lucros e Contribuições Estatutárias.""" + + reinsurance_operations: Optional[float] = FieldInfo(alias="reinsuranceOperations", default=None) + """Resultado de Operações de Resseguros (específico para Seguradoras).""" + + research_development: Optional[float] = FieldInfo(alias="researchDevelopment", default=None) + """Despesas com Pesquisa e Desenvolvimento.""" + + sales_expenses: Optional[float] = FieldInfo(alias="salesExpenses", default=None) + """Despesas com Vendas (detalhamento, pode estar contido em SG&A).""" + + selling_general_administrative: Optional[float] = FieldInfo(alias="sellingGeneralAdministrative", default=None) + """Despesas com Vendas, Gerais e Administrativas.""" + + symbol: Optional[str] = None + """Ticker do ativo ao qual a DRE se refere.""" + + total_operating_expenses: Optional[float] = FieldInfo(alias="totalOperatingExpenses", default=None) + """Total das Despesas Operacionais (P&D + SG&A + Outras).""" + + total_other_income_expense_net: Optional[float] = FieldInfo(alias="totalOtherIncomeExpenseNet", default=None) + """Resultado Financeiro Líquido + Outras Receitas/Despesas.""" + + total_revenue: Optional[float] = FieldInfo(alias="totalRevenue", default=None) + """Receita Operacional Líquida.""" + + type: Optional[Literal["yearly", "quarterly"]] = None + """Indica a periodicidade da DRE: `yearly` (anual) ou `quarterly` (trimestral).""" + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """ + Data da última atualização deste registro específico na fonte de dados + (YYYY-MM-DD). + """ diff --git a/src/brapi/types/quote_list_params.py b/src/brapi/types/quote_list_params.py new file mode 100644 index 0000000..9038560 --- /dev/null +++ b/src/brapi/types/quote_list_params.py @@ -0,0 +1,86 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import Literal, Annotated, TypedDict + +from .._utils import PropertyInfo + +__all__ = ["QuoteListParams"] + + +class QuoteListParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + limit: int + """**Opcional.** Número máximo de ativos a serem retornados por página. + + O valor padrão pode variar. + """ + + page: int + """ + **Opcional.** Número da página dos resultados a ser retornada, considerando o + `limit` especificado. Começa em 1. + """ + + search: str + """**Opcional.** Termo para buscar ativos por ticker (correspondência parcial). + + Ex: `PETR` encontrará `PETR4`, `PETR3`. + """ + + sector: Literal[ + "Retail Trade", + "Energy Minerals", + "Health Services", + "Utilities", + "Finance", + "Consumer Services", + "Consumer Non-Durables", + "Non-Energy Minerals", + "Commercial Services", + "Distribution Services", + "Transportation", + "Technology Services", + "Process Industries", + "Communications", + "Producer Manufacturing", + "Miscellaneous", + "Electronic Technology", + "Industrial Services", + "Health Technology", + "Consumer Durables", + ] + """**Opcional.** Filtra os resultados por setor de atuação da empresa. + + Utilize um dos valores retornados em `availableSectors`. + """ + + sort_by: Annotated[ + Literal["name", "close", "change", "change_abs", "volume", "market_cap_basic", "sector"], + PropertyInfo(alias="sortBy"), + ] + """**Opcional.** Campo pelo qual os resultados serão ordenados.""" + + sort_order: Annotated[Literal["asc", "desc"], PropertyInfo(alias="sortOrder")] + """**Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + + Requer que `sortBy` seja especificado. + """ + + type: Literal["stock", "fund", "bdr"] + """**Opcional.** Filtra os resultados por tipo de ativo.""" diff --git a/src/brapi/types/quote_list_response.py b/src/brapi/types/quote_list_response.py new file mode 100644 index 0000000..b1ca5d7 --- /dev/null +++ b/src/brapi/types/quote_list_response.py @@ -0,0 +1,99 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["QuoteListResponse", "Index", "Stock"] + + +class Index(BaseModel): + name: Optional[str] = None + """Nome do índice (ex: `IBOVESPA`).""" + + stock: Optional[str] = None + """Ticker do índice (ex: `^BVSP`).""" + + +class Stock(BaseModel): + change: Optional[float] = None + """Variação percentual do preço em relação ao fechamento anterior.""" + + close: Optional[float] = None + """Preço de fechamento mais recente ou último preço negociado.""" + + logo: Optional[str] = None + """URL para a imagem do logo da empresa/ativo.""" + + market_cap: Optional[float] = None + """Capitalização de mercado (Preço x Quantidade de Ações). + + Pode ser nulo para FIIs ou outros tipos. + """ + + name: Optional[str] = None + """Nome do ativo ou empresa (ex: `PETROBRAS PN`).""" + + sector: Optional[str] = None + """Setor de atuação da empresa (ex: `Energy Minerals`, `Finance`). + + Pode ser nulo ou variar para FIIs. + """ + + stock: Optional[str] = None + """Ticker do ativo (ex: `PETR4`, `MXRF11`).""" + + type: Optional[Literal["stock", "fund", "bdr"]] = None + """ + Tipo do ativo: `stock` (Ação), `fund` (Fundo Imobiliário/FII), `bdr` (Brazilian + Depositary Receipt). + """ + + volume: Optional[int] = None + """Volume financeiro negociado no último pregão ou dia atual.""" + + +class QuoteListResponse(BaseModel): + available_sectors: Optional[List[str]] = FieldInfo(alias="availableSectors", default=None) + """ + Lista de todos os setores disponíveis que podem ser usados no parâmetro de + filtro `sector`. + """ + + available_stock_types: Optional[List[Literal["stock", "fund", "bdr"]]] = FieldInfo( + alias="availableStockTypes", default=None + ) + """ + Lista dos tipos de ativos (`stock`, `fund`, `bdr`) disponíveis que podem ser + usados no parâmetro de filtro `type`. + """ + + current_page: Optional[int] = FieldInfo(alias="currentPage", default=None) + """Número da página atual retornada nos resultados.""" + + has_next_page: Optional[bool] = FieldInfo(alias="hasNextPage", default=None) + """ + Indica se existe uma próxima página de resultados (`true`) ou se esta é a última + página (`false`). + """ + + indexes: Optional[List[Index]] = None + """Lista resumida de índices relevantes (geralmente inclui IBOVESPA).""" + + items_per_page: Optional[int] = FieldInfo(alias="itemsPerPage", default=None) + """Número de itens (ativos) retornados por página (conforme `limit` ou padrão).""" + + stocks: Optional[List[Stock]] = None + """Lista paginada e filtrada dos ativos solicitados.""" + + total_count: Optional[int] = FieldInfo(alias="totalCount", default=None) + """ + Número total de ativos encontrados que correspondem aos filtros aplicados (sem + considerar a paginação). + """ + + total_pages: Optional[int] = FieldInfo(alias="totalPages", default=None) + """Número total de páginas existentes para a consulta/filtros aplicados.""" diff --git a/src/brapi/types/quote_retrieve_params.py b/src/brapi/types/quote_retrieve_params.py new file mode 100644 index 0000000..e219ac0 --- /dev/null +++ b/src/brapi/types/quote_retrieve_params.py @@ -0,0 +1,113 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import List +from typing_extensions import Literal, TypedDict + +__all__ = ["QuoteRetrieveParams"] + + +class QuoteRetrieveParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + dividends: bool + """**Opcional.** Booleano (`true` ou `false`). + + Se `true`, inclui informações sobre dividendos e JCP (Juros sobre Capital + Próprio) pagos historicamente pelo ativo na chave `dividendsData`. + """ + + fundamental: bool + """**Opcional.** Booleano (`true` ou `false`). + + Se `true`, inclui dados fundamentalistas básicos na resposta, como Preço/Lucro + (P/L) e Lucro Por Ação (LPA). + + **Nota:** Para dados fundamentalistas mais completos, utilize o parâmetro + `modules`. + """ + + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + """ + **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. + + **Valores Possíveis:** + + - `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`: Intervalos intraday + (minutos/horas). **Atenção:** Disponibilidade pode variar conforme o `range` e + o ativo. + - `1d`: Diário (padrão se `range` for especificado e `interval` omitido). + - `5d`: 5 dias. + - `1wk`: Semanal. + - `1mo`: Mensal. + - `3mo`: Trimestral. + """ + + modules: List[ + Literal[ + "summaryProfile", + "balanceSheetHistory", + "defaultKeyStatistics", + "balanceSheetHistoryQuarterly", + "incomeStatementHistory", + "incomeStatementHistoryQuarterly", + "financialData", + "financialDataHistory", + "financialDataHistoryQuarterly", + "defaultKeyStatisticsHistory", + "defaultKeyStatisticsHistoryQuarterly", + "valueAddedHistory", + "valueAddedHistoryQuarterly", + "cashflowHistory", + "cashflowHistoryQuarterly", + ] + ] + """ + **Opcional.** Uma lista de módulos de dados adicionais, separados por vírgula + (`,`), para incluir na resposta. Permite buscar dados financeiros detalhados. + + **Exemplos:** + + - `modules=summaryProfile` (retorna perfil da empresa) + - `modules=balanceSheetHistory,incomeStatementHistory` (retorna histórico anual + do BP e DRE) + + Veja a descrição principal do endpoint para a lista completa de módulos e seus + conteúdos. + """ + + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] + """ + **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Se omitido, apenas a cotação mais recente é retornada + (a menos que `interval` seja usado). + + **Valores Possíveis:** + + - `1d`: Último dia de pregão (intraday se `interval` for minutos/horas). + - `5d`: Últimos 5 dias. + - `1mo`: Último mês. + - `3mo`: Últimos 3 meses. + - `6mo`: Últimos 6 meses. + - `1y`: Último ano. + - `2y`: Últimos 2 anos. + - `5y`: Últimos 5 anos. + - `10y`: Últimos 10 anos. + - `ytd`: Desde o início do ano atual (Year-to-Date). + - `max`: Todo o período histórico disponível. + """ diff --git a/src/brapi/types/quote_retrieve_response.py b/src/brapi/types/quote_retrieve_response.py new file mode 100644 index 0000000..8f9e542 --- /dev/null +++ b/src/brapi/types/quote_retrieve_response.py @@ -0,0 +1,476 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional +from datetime import datetime + +from pydantic import Field as FieldInfo + +from .._models import BaseModel +from .cashflow_entry import CashflowEntry +from .value_added_entry import ValueAddedEntry +from .balance_sheet_entry import BalanceSheetEntry +from .financial_data_entry import FinancialDataEntry +from .income_statement_entry import IncomeStatementEntry +from .default_key_statistics_entry import DefaultKeyStatisticsEntry + +__all__ = [ + "QuoteRetrieveResponse", + "Result", + "ResultDividendsData", + "ResultDividendsDataCashDividend", + "ResultDividendsDataStockDividend", + "ResultHistoricalDataPrice", + "ResultSummaryProfile", +] + + +class ResultDividendsDataCashDividend(BaseModel): + approved_on: Optional[datetime] = FieldInfo(alias="approvedOn", default=None) + """Data em que o pagamento do provento foi aprovado pela empresa. + + Pode ser uma estimativa em alguns casos. Formato ISO 8601. + """ + + asset_issued: Optional[str] = FieldInfo(alias="assetIssued", default=None) + """Ticker do ativo que pagou o provento (ex: `ITSA4`). + + Pode incluir sufixos específicos relacionados ao evento. + """ + + isin_code: Optional[str] = FieldInfo(alias="isinCode", default=None) + """ + Código ISIN (International Securities Identification Number) do ativo + relacionado ao provento. + """ + + label: Optional[str] = None + """Tipo do provento em dinheiro. + + Geralmente `DIVIDENDO` ou `JCP` (Juros sobre Capital Próprio). + """ + + last_date_prior: Optional[datetime] = FieldInfo(alias="lastDatePrior", default=None) + """Data Com (Ex-Date). + + Último dia em que era necessário possuir o ativo para ter direito a receber este + provento. Pode ser uma estimativa. Formato ISO 8601. + """ + + payment_date: Optional[datetime] = FieldInfo(alias="paymentDate", default=None) + """Data efetiva em que o pagamento foi realizado (ou está previsto). + + Formato ISO 8601. + """ + + rate: Optional[float] = None + """Valor bruto do provento pago por unidade do ativo (por ação, por cota).""" + + related_to: Optional[str] = FieldInfo(alias="relatedTo", default=None) + """ + Descrição do período ou evento ao qual o provento se refere (ex: + `1º Trimestre/2023`, `Resultado 2022`). + """ + + remarks: Optional[str] = None + """Observações adicionais ou informações relevantes sobre o provento.""" + + +class ResultDividendsDataStockDividend(BaseModel): + approved_on: Optional[datetime] = FieldInfo(alias="approvedOn", default=None) + """Data em que o evento foi aprovado. Formato ISO 8601.""" + + asset_issued: Optional[str] = FieldInfo(alias="assetIssued", default=None) + """Ticker do ativo afetado pelo evento.""" + + complete_factor: Optional[str] = FieldInfo(alias="completeFactor", default=None) + """Descrição textual do fator (ex: `1 / 10`, `10 / 1`).""" + + factor: Optional[float] = None + """Fator numérico do evento. + + - **Bonificação:** Percentual (ex: 0.1 para 10%). + - **Desdobramento/Grupamento:** Fator multiplicativo ou divisor. + """ + + isin_code: Optional[str] = FieldInfo(alias="isinCode", default=None) + """Código ISIN do ativo.""" + + label: Optional[str] = None + """Tipo do evento: `DESDOBRAMENTO`, `GRUPAMENTO`, `BONIFICACAO`.""" + + last_date_prior: Optional[datetime] = FieldInfo(alias="lastDatePrior", default=None) + """Data Com (Ex-Date). + + Último dia para possuir o ativo nas condições antigas. Formato ISO 8601. + """ + + remarks: Optional[str] = None + """Observações adicionais sobre o evento.""" + + +class ResultDividendsData(BaseModel): + cash_dividends: Optional[List[ResultDividendsDataCashDividend]] = FieldInfo(alias="cashDividends", default=None) + """Lista de proventos pagos em dinheiro (Dividendos e JCP).""" + + stock_dividends: Optional[List[ResultDividendsDataStockDividend]] = FieldInfo(alias="stockDividends", default=None) + """Lista de eventos corporativos (Desdobramento, Grupamento, Bonificação).""" + + subscriptions: Optional[List[object]] = None + """Lista de eventos de subscrição de ações (estrutura não detalhada aqui).""" + + +class ResultHistoricalDataPrice(BaseModel): + adjusted_close: Optional[float] = FieldInfo(alias="adjustedClose", default=None) + """ + Preço de fechamento ajustado para proventos (dividendos, JCP, bonificações, + etc.) e desdobramentos/grupamentos. + """ + + close: Optional[float] = None + """Preço de fechamento do ativo no intervalo.""" + + date: Optional[int] = None + """ + Data do pregão ou do ponto de dados, representada como um timestamp UNIX (número + de segundos desde 1970-01-01 UTC). + """ + + high: Optional[float] = None + """Preço máximo atingido pelo ativo no intervalo.""" + + low: Optional[float] = None + """Preço mínimo atingido pelo ativo no intervalo.""" + + open: Optional[float] = None + """Preço de abertura do ativo no intervalo (dia, semana, mês, etc.).""" + + volume: Optional[int] = None + """Volume financeiro negociado no intervalo.""" + + +class ResultSummaryProfile(BaseModel): + address1: Optional[str] = None + """Linha 1 do endereço da sede da empresa.""" + + address2: Optional[str] = None + """Linha 2 do endereço da sede da empresa (complemento).""" + + city: Optional[str] = None + """Cidade da sede da empresa.""" + + company_officers: Optional[List[object]] = FieldInfo(alias="companyOfficers", default=None) + """ + Lista de diretores e executivos principais da empresa (estrutura interna do + objeto não detalhada aqui). + """ + + country: Optional[str] = None + """País da sede da empresa.""" + + full_time_employees: Optional[int] = FieldInfo(alias="fullTimeEmployees", default=None) + """Número estimado de funcionários em tempo integral.""" + + industry: Optional[str] = None + """Nome da indústria em que a empresa atua.""" + + industry_disp: Optional[str] = FieldInfo(alias="industryDisp", default=None) + """Nome de exibição formatado para a indústria.""" + + industry_key: Optional[str] = FieldInfo(alias="industryKey", default=None) + """Chave interna ou código para a indústria.""" + + long_business_summary: Optional[str] = FieldInfo(alias="longBusinessSummary", default=None) + """Descrição longa e detalhada sobre as atividades e o negócio da empresa.""" + + phone: Optional[str] = None + """Número de telefone principal da empresa.""" + + sector: Optional[str] = None + """Nome do setor de atuação da empresa.""" + + sector_disp: Optional[str] = FieldInfo(alias="sectorDisp", default=None) + """Nome de exibição formatado para o setor.""" + + sector_key: Optional[str] = FieldInfo(alias="sectorKey", default=None) + """Chave interna ou código para o setor.""" + + state: Optional[str] = None + """Estado ou província da sede da empresa.""" + + website: Optional[str] = None + """URL do website oficial da empresa.""" + + zip: Optional[str] = None + """Código Postal (CEP) da sede da empresa.""" + + +class Result(BaseModel): + average_daily_volume10_day: Optional[float] = FieldInfo(alias="averageDailyVolume10Day", default=None) + """Média do volume financeiro diário negociado nos últimos 10 dias.""" + + average_daily_volume3_month: Optional[float] = FieldInfo(alias="averageDailyVolume3Month", default=None) + """Média do volume financeiro diário negociado nos últimos 3 meses.""" + + balance_sheet_history: Optional[List[BalanceSheetEntry]] = FieldInfo(alias="balanceSheetHistory", default=None) + """Histórico **anual** do Balanço Patrimonial. + + Retornado apenas se `modules` incluir `balanceSheetHistory`. + """ + + balance_sheet_history_quarterly: Optional[List[BalanceSheetEntry]] = FieldInfo( + alias="balanceSheetHistoryQuarterly", default=None + ) + """Histórico **trimestral** do Balanço Patrimonial. + + Retornado apenas se `modules` incluir `balanceSheetHistoryQuarterly`. + """ + + cashflow_history: Optional[List[CashflowEntry]] = FieldInfo(alias="cashflowHistory", default=None) + """Histórico **anual** da Demonstração do Fluxo de Caixa (DFC). + + Retornado apenas se `modules` incluir `cashflowHistory`. + """ + + cashflow_history_quarterly: Optional[List[CashflowEntry]] = FieldInfo( + alias="cashflowHistoryQuarterly", default=None + ) + """Histórico **trimestral** da Demonstração do Fluxo de Caixa (DFC). + + Retornado apenas se `modules` incluir `cashflowHistoryQuarterly`. + """ + + currency: Optional[str] = None + """Moeda na qual os valores monetários são expressos (geralmente `BRL`).""" + + default_key_statistics: Optional[DefaultKeyStatisticsEntry] = FieldInfo(alias="defaultKeyStatistics", default=None) + """Principais estatísticas financeiras atuais/TTM. + + Retornado apenas se `modules` incluir `defaultKeyStatistics`. + """ + + default_key_statistics_history: Optional[List[DefaultKeyStatisticsEntry]] = FieldInfo( + alias="defaultKeyStatisticsHistory", default=None + ) + """Histórico **anual** das principais estatísticas. + + Retornado apenas se `modules` incluir `defaultKeyStatisticsHistory`. + """ + + default_key_statistics_history_quarterly: Optional[List[DefaultKeyStatisticsEntry]] = FieldInfo( + alias="defaultKeyStatisticsHistoryQuarterly", default=None + ) + """Histórico **trimestral** das principais estatísticas. + + Retornado apenas se `modules` incluir `defaultKeyStatisticsHistoryQuarterly`. + """ + + dividends_data: Optional[ResultDividendsData] = FieldInfo(alias="dividendsData", default=None) + """Objeto contendo informações sobre dividendos, JCP e outros eventos corporativos. + + Retornado apenas se `dividends=true` for especificado na requisição. + """ + + earnings_per_share: Optional[float] = FieldInfo(alias="earningsPerShare", default=None) + """Lucro Por Ação (LPA) dos últimos 12 meses (TTM). + + Retornado se `fundamental=true`. + """ + + fifty_two_week_high: Optional[float] = FieldInfo(alias="fiftyTwoWeekHigh", default=None) + """Preço máximo atingido nas últimas 52 semanas.""" + + fifty_two_week_high_change: Optional[float] = FieldInfo(alias="fiftyTwoWeekHighChange", default=None) + """Variação absoluta entre o preço atual e o preço máximo das últimas 52 semanas.""" + + fifty_two_week_high_change_percent: Optional[float] = FieldInfo(alias="fiftyTwoWeekHighChangePercent", default=None) + """ + Variação percentual entre o preço atual e o preço máximo das últimas 52 semanas. + """ + + fifty_two_week_low: Optional[float] = FieldInfo(alias="fiftyTwoWeekLow", default=None) + """Preço mínimo atingido nas últimas 52 semanas.""" + + fifty_two_week_low_change: Optional[float] = FieldInfo(alias="fiftyTwoWeekLowChange", default=None) + """Variação absoluta entre o preço atual e o preço mínimo das últimas 52 semanas.""" + + fifty_two_week_range: Optional[str] = FieldInfo(alias="fiftyTwoWeekRange", default=None) + """ + String formatada mostrando o intervalo de preço das últimas 52 semanas (Mínimo - + Máximo). + """ + + financial_data: Optional[FinancialDataEntry] = FieldInfo(alias="financialData", default=None) + """Dados financeiros e indicadores TTM. + + Retornado apenas se `modules` incluir `financialData`. + """ + + financial_data_history: Optional[List[FinancialDataEntry]] = FieldInfo(alias="financialDataHistory", default=None) + """Histórico **anual** de dados financeiros e indicadores. + + Retornado apenas se `modules` incluir `financialDataHistory`. + """ + + financial_data_history_quarterly: Optional[List[FinancialDataEntry]] = FieldInfo( + alias="financialDataHistoryQuarterly", default=None + ) + """Histórico **trimestral** de dados financeiros e indicadores. + + Retornado apenas se `modules` incluir `financialDataHistoryQuarterly`. + """ + + historical_data_price: Optional[List[ResultHistoricalDataPrice]] = FieldInfo( + alias="historicalDataPrice", default=None + ) + """ + Array contendo a série histórica de preços, retornado apenas se os parâmetros + `range` e/ou `interval` forem especificados na requisição. + """ + + income_statement_history: Optional[List[IncomeStatementEntry]] = FieldInfo( + alias="incomeStatementHistory", default=None + ) + """Histórico **anual** da Demonstração do Resultado (DRE). + + Retornado apenas se `modules` incluir `incomeStatementHistory`. + """ + + income_statement_history_quarterly: Optional[List[IncomeStatementEntry]] = FieldInfo( + alias="incomeStatementHistoryQuarterly", default=None + ) + """Histórico **trimestral** da Demonstração do Resultado (DRE). + + Retornado apenas se `modules` incluir `incomeStatementHistoryQuarterly`. + """ + + logourl: Optional[str] = None + """URL da imagem do logo do ativo/empresa.""" + + long_name: Optional[str] = FieldInfo(alias="longName", default=None) + """Nome longo ou completo da empresa ou ativo.""" + + market_cap: Optional[float] = FieldInfo(alias="marketCap", default=None) + """Capitalização de mercado total do ativo (Preço Atual x Ações em Circulação).""" + + price_earnings: Optional[float] = FieldInfo(alias="priceEarnings", default=None) + """Indicador Preço/Lucro (P/L): Preço Atual / Lucro Por Ação (LPA) TTM. + + Retornado se `fundamental=true`. + """ + + regular_market_change: Optional[float] = FieldInfo(alias="regularMarketChange", default=None) + """Variação absoluta do preço no dia atual em relação ao fechamento anterior.""" + + regular_market_change_percent: Optional[float] = FieldInfo(alias="regularMarketChangePercent", default=None) + """Variação percentual do preço no dia atual em relação ao fechamento anterior.""" + + regular_market_day_high: Optional[float] = FieldInfo(alias="regularMarketDayHigh", default=None) + """Preço máximo atingido no dia de negociação atual.""" + + regular_market_day_low: Optional[float] = FieldInfo(alias="regularMarketDayLow", default=None) + """Preço mínimo atingido no dia de negociação atual.""" + + regular_market_day_range: Optional[str] = FieldInfo(alias="regularMarketDayRange", default=None) + """String formatada mostrando o intervalo de preço do dia (Mínimo - Máximo).""" + + regular_market_open: Optional[float] = FieldInfo(alias="regularMarketOpen", default=None) + """Preço de abertura no dia de negociação atual.""" + + regular_market_previous_close: Optional[float] = FieldInfo(alias="regularMarketPreviousClose", default=None) + """Preço de fechamento do pregão anterior.""" + + regular_market_price: Optional[float] = FieldInfo(alias="regularMarketPrice", default=None) + """Preço atual ou do último negócio registrado.""" + + regular_market_time: Optional[datetime] = FieldInfo(alias="regularMarketTime", default=None) + """Data e hora da última atualização da cotação (último negócio registrado). + + Formato ISO 8601. + """ + + regular_market_volume: Optional[float] = FieldInfo(alias="regularMarketVolume", default=None) + """Volume financeiro negociado no dia atual.""" + + short_name: Optional[str] = FieldInfo(alias="shortName", default=None) + """Nome curto ou abreviado da empresa ou ativo.""" + + summary_profile: Optional[ResultSummaryProfile] = FieldInfo(alias="summaryProfile", default=None) + """Resumo do perfil da empresa. + + Retornado apenas se `modules` incluir `summaryProfile`. + """ + + symbol: Optional[str] = None + """Ticker (símbolo) do ativo (ex: `PETR4`, `^BVSP`).""" + + two_hundred_day_average: Optional[float] = FieldInfo(alias="twoHundredDayAverage", default=None) + """Média móvel simples dos preços de fechamento dos últimos 200 dias.""" + + two_hundred_day_average_change: Optional[float] = FieldInfo(alias="twoHundredDayAverageChange", default=None) + """Variação absoluta entre o preço atual e a média de 200 dias.""" + + two_hundred_day_average_change_percent: Optional[float] = FieldInfo( + alias="twoHundredDayAverageChangePercent", default=None + ) + """Variação percentual entre o preço atual e a média de 200 dias.""" + + updated_at: Optional[datetime] = FieldInfo(alias="updatedAt", default=None) + """ + Timestamp da última atualização dos dados do índice na fonte (aplicável + principalmente a índices, como `^BVSP`). Formato ISO 8601. + """ + + used_interval: Optional[str] = FieldInfo(alias="usedInterval", default=None) + """ + O intervalo (`interval`) efetivamente utilizado pela API para retornar os dados + históricos, caso solicitado. + """ + + used_range: Optional[str] = FieldInfo(alias="usedRange", default=None) + """ + O período (`range`) efetivamente utilizado pela API para retornar os dados + históricos, caso solicitado. + """ + + valid_intervals: Optional[List[str]] = FieldInfo(alias="validIntervals", default=None) + """ + Lista dos valores válidos que podem ser utilizados no parâmetro `interval` para + este ativo específico. + """ + + valid_ranges: Optional[List[str]] = FieldInfo(alias="validRanges", default=None) + """ + Lista dos valores válidos que podem ser utilizados no parâmetro `range` para + este ativo específico. + """ + + value_added_history: Optional[List[ValueAddedEntry]] = FieldInfo(alias="valueAddedHistory", default=None) + """Histórico **anual** da Demonstração do Valor Adicionado (DVA). + + Retornado apenas se `modules` incluir `valueAddedHistory`. + """ + + value_added_history_quarterly: Optional[List[ValueAddedEntry]] = FieldInfo( + alias="valueAddedHistoryQuarterly", default=None + ) + """Histórico **trimestral** da Demonstração do Valor Adicionado (DVA). + + Retornado apenas se `modules` incluir `valueAddedHistoryQuarterly`. + """ + + +class QuoteRetrieveResponse(BaseModel): + requested_at: Optional[datetime] = FieldInfo(alias="requestedAt", default=None) + """Timestamp indicando quando a requisição foi recebida pelo servidor. + + Formato ISO 8601. + """ + + results: Optional[List[Result]] = None + """Array contendo os resultados detalhados para cada ticker solicitado.""" + + took: Optional[str] = None + """ + Tempo aproximado que o servidor levou para processar a requisição, em formato de + string (ex: `746ms`). + """ diff --git a/src/brapi/types/v2/__init__.py b/src/brapi/types/v2/__init__.py new file mode 100644 index 0000000..febc9e5 --- /dev/null +++ b/src/brapi/types/v2/__init__.py @@ -0,0 +1,20 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from .crypto_retrieve_params import CryptoRetrieveParams as CryptoRetrieveParams +from .crypto_retrieve_response import CryptoRetrieveResponse as CryptoRetrieveResponse +from .currency_retrieve_params import CurrencyRetrieveParams as CurrencyRetrieveParams +from .inflation_retrieve_params import InflationRetrieveParams as InflationRetrieveParams +from .currency_retrieve_response import CurrencyRetrieveResponse as CurrencyRetrieveResponse +from .prime_rate_retrieve_params import PrimeRateRetrieveParams as PrimeRateRetrieveParams +from .inflation_retrieve_response import InflationRetrieveResponse as InflationRetrieveResponse +from .crypto_list_available_params import CryptoListAvailableParams as CryptoListAvailableParams +from .prime_rate_retrieve_response import PrimeRateRetrieveResponse as PrimeRateRetrieveResponse +from .crypto_list_available_response import CryptoListAvailableResponse as CryptoListAvailableResponse +from .currency_list_available_params import CurrencyListAvailableParams as CurrencyListAvailableParams +from .inflation_list_available_params import InflationListAvailableParams as InflationListAvailableParams +from .currency_list_available_response import CurrencyListAvailableResponse as CurrencyListAvailableResponse +from .prime_rate_list_available_params import PrimeRateListAvailableParams as PrimeRateListAvailableParams +from .inflation_list_available_response import InflationListAvailableResponse as InflationListAvailableResponse +from .prime_rate_list_available_response import PrimeRateListAvailableResponse as PrimeRateListAvailableResponse diff --git a/src/brapi/types/v2/crypto_list_available_params.py b/src/brapi/types/v2/crypto_list_available_params.py new file mode 100644 index 0000000..0846d0d --- /dev/null +++ b/src/brapi/types/v2/crypto_list_available_params.py @@ -0,0 +1,31 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import TypedDict + +__all__ = ["CryptoListAvailableParams"] + + +class CryptoListAvailableParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + search: str + """ + **Opcional.** Termo para filtrar a lista de siglas de criptomoedas + (correspondência parcial, case-insensitive). Se omitido, retorna todas as + siglas. + """ diff --git a/src/brapi/types/v2/crypto_list_available_response.py b/src/brapi/types/v2/crypto_list_available_response.py new file mode 100644 index 0000000..5951e9a --- /dev/null +++ b/src/brapi/types/v2/crypto_list_available_response.py @@ -0,0 +1,15 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from ..._models import BaseModel + +__all__ = ["CryptoListAvailableResponse"] + + +class CryptoListAvailableResponse(BaseModel): + coins: Optional[List[str]] = None + """ + Lista de siglas (tickers) das criptomoedas disponíveis (ex: `BTC`, `ETH`, + `LTC`). + """ diff --git a/src/brapi/types/v2/crypto_retrieve_params.py b/src/brapi/types/v2/crypto_retrieve_params.py new file mode 100644 index 0000000..cbd226b --- /dev/null +++ b/src/brapi/types/v2/crypto_retrieve_params.py @@ -0,0 +1,59 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import Literal, Required, TypedDict + +__all__ = ["CryptoRetrieveParams"] + + +class CryptoRetrieveParams(TypedDict, total=False): + coin: Required[str] + """ + **Obrigatório.** Uma ou mais siglas (tickers) de criptomoedas que você deseja + consultar. Separe múltiplas siglas por vírgula (`,`). + + - **Exemplos:** `BTC`, `ETH,ADA`, `SOL`. + """ + + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + currency: str + """ + **Opcional.** A sigla da moeda fiduciária na qual a cotação da(s) criptomoeda(s) + deve ser retornada. Se omitido, o padrão é `BRL` (Real Brasileiro). + """ + + interval: Literal["1m", "2m", "5m", "15m", "30m", "60m", "90m", "1h", "1d", "5d", "1wk", "1mo", "3mo"] + """ + **Opcional.** Define a granularidade (intervalo) dos dados históricos de preço + (`historicalDataPrice`). Requer que `range` também seja especificado. Funciona + de forma análoga ao endpoint de ações. + + - Valores: `1m`, `2m`, `5m`, `15m`, `30m`, `60m`, `90m`, `1h`, `1d`, `5d`, + `1wk`, `1mo`, `3mo`. + """ + + range: Literal["1d", "5d", "1mo", "3mo", "6mo", "1y", "2y", "5y", "10y", "ytd", "max"] + """ + **Opcional.** Define o período para os dados históricos de preço + (`historicalDataPrice`). Funciona de forma análoga ao endpoint de ações. Se + omitido, apenas a cotação mais recente é retornada (a menos que `interval` seja + usado). + + - Valores: `1d`, `5d`, `1mo`, `3mo`, `6mo`, `1y`, `2y`, `5y`, `10y`, `ytd`, + `max`. + """ diff --git a/src/brapi/types/v2/crypto_retrieve_response.py b/src/brapi/types/v2/crypto_retrieve_response.py new file mode 100644 index 0000000..11d50b7 --- /dev/null +++ b/src/brapi/types/v2/crypto_retrieve_response.py @@ -0,0 +1,117 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional +from datetime import datetime + +from pydantic import Field as FieldInfo + +from ..._models import BaseModel + +__all__ = ["CryptoRetrieveResponse", "Coin", "CoinHistoricalDataPrice"] + + +class CoinHistoricalDataPrice(BaseModel): + adjusted_close: Optional[float] = FieldInfo(alias="adjustedClose", default=None) + """Preço de fechamento ajustado (geralmente igual ao `close` para cripto).""" + + close: Optional[float] = None + """Preço de fechamento da criptomoeda no intervalo.""" + + date: Optional[int] = None + """Data do ponto de dados, representada como um timestamp UNIX.""" + + high: Optional[float] = None + """Preço máximo atingido no intervalo.""" + + low: Optional[float] = None + """Preço mínimo atingido no intervalo.""" + + open: Optional[float] = None + """Preço de abertura da criptomoeda no intervalo.""" + + volume: Optional[int] = None + """ + Volume negociado no intervalo (na criptomoeda ou na moeda de referência, + verificar contexto). + """ + + +class Coin(BaseModel): + coin: Optional[str] = None + """Sigla (ticker) da criptomoeda (ex: `BTC`, `ETH`).""" + + coin_image_url: Optional[str] = FieldInfo(alias="coinImageUrl", default=None) + """URL da imagem do logo da criptomoeda.""" + + coin_name: Optional[str] = FieldInfo(alias="coinName", default=None) + """Nome completo da criptomoeda (ex: `Bitcoin`, `Ethereum`).""" + + currency: Optional[str] = None + """Sigla da moeda fiduciária na qual os preços estão cotados (ex: `BRL`, `USD`).""" + + currency_rate_from_usd: Optional[float] = FieldInfo(alias="currencyRateFromUSD", default=None) + """Taxa de câmbio da `currency` em relação ao USD (Dólar Americano). + + `1 USD = X currency`. + """ + + historical_data_price: Optional[List[CoinHistoricalDataPrice]] = FieldInfo( + alias="historicalDataPrice", default=None + ) + """ + Array contendo a série histórica de preços, retornado se `range` ou `interval` + forem especificados. + """ + + market_cap: Optional[int] = FieldInfo(alias="marketCap", default=None) + """Capitalização de mercado da criptomoeda na `currency` especificada.""" + + regular_market_change: Optional[float] = FieldInfo(alias="regularMarketChange", default=None) + """Variação absoluta do preço nas últimas 24 horas (ou período relevante).""" + + regular_market_change_percent: Optional[float] = FieldInfo(alias="regularMarketChangePercent", default=None) + """Variação percentual do preço nas últimas 24 horas (ou período relevante).""" + + regular_market_day_high: Optional[float] = FieldInfo(alias="regularMarketDayHigh", default=None) + """Preço máximo nas últimas 24 horas (ou período relevante).""" + + regular_market_day_low: Optional[float] = FieldInfo(alias="regularMarketDayLow", default=None) + """Preço mínimo nas últimas 24 horas (ou período relevante).""" + + regular_market_day_range: Optional[str] = FieldInfo(alias="regularMarketDayRange", default=None) + """ + String formatada mostrando o intervalo de preço das últimas 24h (Mínimo - + Máximo). + """ + + regular_market_price: Optional[float] = FieldInfo(alias="regularMarketPrice", default=None) + """Preço atual da criptomoeda na `currency` especificada.""" + + regular_market_time: Optional[datetime] = FieldInfo(alias="regularMarketTime", default=None) + """Timestamp da última atualização da cotação. Formato ISO 8601.""" + + regular_market_volume: Optional[int] = FieldInfo(alias="regularMarketVolume", default=None) + """Volume negociado nas últimas 24 horas (na `currency` especificada).""" + + used_interval: Optional[str] = FieldInfo(alias="usedInterval", default=None) + """ + O intervalo (`interval`) efetivamente utilizado para os dados históricos, se + solicitado. + """ + + used_range: Optional[str] = FieldInfo(alias="usedRange", default=None) + """ + O período (`range`) efetivamente utilizado para os dados históricos, se + solicitado. + """ + + valid_intervals: Optional[List[str]] = FieldInfo(alias="validIntervals", default=None) + """Lista dos valores válidos para o parâmetro `interval` nesta criptomoeda.""" + + valid_ranges: Optional[List[str]] = FieldInfo(alias="validRanges", default=None) + """Lista dos valores válidos para o parâmetro `range` nesta criptomoeda.""" + + +class CryptoRetrieveResponse(BaseModel): + coins: Optional[List[Coin]] = None + """Array contendo os resultados detalhados para cada criptomoeda solicitada.""" diff --git a/src/brapi/types/v2/currency_list_available_params.py b/src/brapi/types/v2/currency_list_available_params.py new file mode 100644 index 0000000..8ce9441 --- /dev/null +++ b/src/brapi/types/v2/currency_list_available_params.py @@ -0,0 +1,30 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import TypedDict + +__all__ = ["CurrencyListAvailableParams"] + + +class CurrencyListAvailableParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + search: str + """ + **Opcional.** Termo para filtrar a lista pelo nome da moeda (correspondência + parcial, case-insensitive). + """ diff --git a/src/brapi/types/v2/currency_list_available_response.py b/src/brapi/types/v2/currency_list_available_response.py new file mode 100644 index 0000000..2a515fc --- /dev/null +++ b/src/brapi/types/v2/currency_list_available_response.py @@ -0,0 +1,24 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from ..._models import BaseModel + +__all__ = ["CurrencyListAvailableResponse", "Currency"] + + +class Currency(BaseModel): + currency: Optional[str] = None + """ + Nome da moeda ou par de moedas suportado (ex: `Dólar Americano/Real Brasileiro`, + `Euro/Real Brasileiro`). A sigla pode ser extraída deste nome ou consultada em + documentação adicional. + """ + + +class CurrencyListAvailableResponse(BaseModel): + currencies: Optional[List[Currency]] = None + """ + Lista de objetos, cada um contendo o nome de uma moeda fiduciária ou par + suportado pela API. + """ diff --git a/src/brapi/types/v2/currency_retrieve_params.py b/src/brapi/types/v2/currency_retrieve_params.py new file mode 100644 index 0000000..8ff1825 --- /dev/null +++ b/src/brapi/types/v2/currency_retrieve_params.py @@ -0,0 +1,35 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import Required, TypedDict + +__all__ = ["CurrencyRetrieveParams"] + + +class CurrencyRetrieveParams(TypedDict, total=False): + currency: Required[str] + """ + **Obrigatório.** Uma lista de um ou mais pares de moedas a serem consultados, + separados por vírgula (`,`). + + - **Formato:** `MOEDA_ORIGEM-MOEDA_DESTINO` (ex: `USD-BRL`). + - **Disponibilidade:** Consulte os pares válidos usando o endpoint + [`/api/v2/currency/available`](#/Moedas/getAvailableCurrencies). + - **Exemplo:** `USD-BRL,EUR-BRL,BTC-BRL` + """ + + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ diff --git a/src/brapi/types/v2/currency_retrieve_response.py b/src/brapi/types/v2/currency_retrieve_response.py new file mode 100644 index 0000000..7cc09b5 --- /dev/null +++ b/src/brapi/types/v2/currency_retrieve_response.py @@ -0,0 +1,84 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List + +from pydantic import Field as FieldInfo + +from ..._models import BaseModel + +__all__ = ["CurrencyRetrieveResponse", "Currency"] + + +class Currency(BaseModel): + ask_price: str = FieldInfo(alias="askPrice") + """ + **Preço de Venda (Ask):** Preço atual pelo qual o mercado está disposto a vender + a moeda de origem (`fromCurrency`) recebendo a moeda de destino (`toCurrency`). + Formato String. + """ + + bid_price: str = FieldInfo(alias="bidPrice") + """ + **Preço de Compra (Bid):** Preço atual pelo qual o mercado está disposto a + comprar a moeda de origem (`fromCurrency`) pagando com a moeda de destino + (`toCurrency`). Formato String. + """ + + bid_variation: str = FieldInfo(alias="bidVariation") + """ + **Variação Absoluta (Bid):** Mudança absoluta no preço de compra (bid) desde o + último fechamento ou período de referência. Formato String. + """ + + from_currency: str = FieldInfo(alias="fromCurrency") + """**Moeda de Origem:** Sigla da moeda base do par (ex: `USD` em `USD-BRL`).""" + + high: str + """ + **Máxima:** Preço mais alto atingido pelo par no período recente (geralmente + diário). Formato String. + """ + + low: str + """ + **Mínima:** Preço mais baixo atingido pelo par no período recente (geralmente + diário). Formato String. + """ + + name: str + """ + **Nome do Par:** Nome descritivo do par de moedas (ex: + `Dólar Americano/Real Brasileiro`). + """ + + percentage_change: str = FieldInfo(alias="percentageChange") + """ + **Variação Percentual:** Mudança percentual no preço do par desde o último + fechamento ou período de referência. Formato String. + """ + + to_currency: str = FieldInfo(alias="toCurrency") + """ + **Moeda de Destino:** Sigla da moeda de cotação do par (ex: `BRL` em `USD-BRL`). + """ + + updated_at_date: str = FieldInfo(alias="updatedAtDate") + """ + **Data da Atualização:** Data e hora da última atualização da cotação, formatada + de forma legível (`YYYY-MM-DD HH:MM:SS`). + """ + + updated_at_timestamp: str = FieldInfo(alias="updatedAtTimestamp") + """ + **Timestamp da Atualização:** Data e hora da última atualização da cotação, + representada como um **timestamp UNIX** (string contendo o número de segundos + desde 1970-01-01 UTC). + """ + + +class CurrencyRetrieveResponse(BaseModel): + currency: List[Currency] + """ + Array contendo os objetos `CurrencyQuote`, um para cada par de moeda válido + solicitado no parâmetro `currency`. + """ diff --git a/src/brapi/types/v2/inflation_list_available_params.py b/src/brapi/types/v2/inflation_list_available_params.py new file mode 100644 index 0000000..822dc76 --- /dev/null +++ b/src/brapi/types/v2/inflation_list_available_params.py @@ -0,0 +1,30 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import TypedDict + +__all__ = ["InflationListAvailableParams"] + + +class InflationListAvailableParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + search: str + """ + **Opcional.** Termo para filtrar a lista pelo nome do país (correspondência + parcial, case-insensitive). Se omitido, retorna todos os países. + """ diff --git a/src/brapi/types/v2/inflation_list_available_response.py b/src/brapi/types/v2/inflation_list_available_response.py new file mode 100644 index 0000000..971e902 --- /dev/null +++ b/src/brapi/types/v2/inflation_list_available_response.py @@ -0,0 +1,15 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from ..._models import BaseModel + +__all__ = ["InflationListAvailableResponse"] + + +class InflationListAvailableResponse(BaseModel): + countries: Optional[List[str]] = None + """ + Lista de nomes de países (em minúsculas) para os quais há dados de inflação + disponíveis (ex: `brazil`, `usa`, `argentina`). + """ diff --git a/src/brapi/types/v2/inflation_retrieve_params.py b/src/brapi/types/v2/inflation_retrieve_params.py new file mode 100644 index 0000000..550be1e --- /dev/null +++ b/src/brapi/types/v2/inflation_retrieve_params.py @@ -0,0 +1,63 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import Union +from datetime import date +from typing_extensions import Literal, Annotated, TypedDict + +from ..._utils import PropertyInfo + +__all__ = ["InflationRetrieveParams"] + + +class InflationRetrieveParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + country: str + """**Opcional.** Nome do país para o qual buscar os dados de inflação. + + Use nomes em minúsculas. O padrão é `brazil`. Consulte + `/api/v2/inflation/available` para a lista de países suportados. + """ + + end: Annotated[Union[str, date], PropertyInfo(format="iso8601")] + """ + **Opcional.** Data final do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `start` for especificado. + """ + + historical: bool + """**Opcional.** Booleano (`true` ou `false`). + + Define se dados históricos devem ser incluídos. O comportamento exato em + conjunto com `start`/`end` deve ser verificado. Padrão: `false`. + """ + + sort_by: Annotated[Literal["date", "value"], PropertyInfo(alias="sortBy")] + """**Opcional.** Campo pelo qual os resultados da inflação serão ordenados.""" + + sort_order: Annotated[Literal["asc", "desc"], PropertyInfo(alias="sortOrder")] + """**Opcional.** Direção da ordenação: `asc` (ascendente) ou `desc` (descendente). + + Padrão: `desc`. Requer que `sortBy` seja especificado. + """ + + start: Annotated[Union[str, date], PropertyInfo(format="iso8601")] + """ + **Opcional.** Data de início do período desejado para os dados históricos, no + formato `DD/MM/YYYY`. Requerido se `end` for especificado. + """ diff --git a/src/brapi/types/v2/inflation_retrieve_response.py b/src/brapi/types/v2/inflation_retrieve_response.py new file mode 100644 index 0000000..864488a --- /dev/null +++ b/src/brapi/types/v2/inflation_retrieve_response.py @@ -0,0 +1,34 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from pydantic import Field as FieldInfo + +from ..._models import BaseModel + +__all__ = ["InflationRetrieveResponse", "Inflation"] + + +class Inflation(BaseModel): + date: Optional[str] = None + """Data da medição da inflação, no formato `DD/MM/YYYY`.""" + + epoch_date: Optional[int] = FieldInfo(alias="epochDate", default=None) + """ + Timestamp UNIX (número de segundos desde 1970-01-01 UTC) correspondente à + `date`. + """ + + value: Optional[str] = None + """ + Valor do índice de inflação para a data especificada (formato string, pode + conter `%` ou ser apenas numérico). + """ + + +class InflationRetrieveResponse(BaseModel): + inflation: Optional[List[Inflation]] = None + """ + Array contendo os registros históricos de inflação para o país e período + solicitados. + """ diff --git a/src/brapi/types/v2/prime_rate_list_available_params.py b/src/brapi/types/v2/prime_rate_list_available_params.py new file mode 100644 index 0000000..a13a836 --- /dev/null +++ b/src/brapi/types/v2/prime_rate_list_available_params.py @@ -0,0 +1,30 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing_extensions import TypedDict + +__all__ = ["PrimeRateListAvailableParams"] + + +class PrimeRateListAvailableParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + search: str + """**Opcional.** Termo para filtrar a lista de países por nome. + + Retorna países cujos nomes contenham o termo especificado (case insensitive). + """ diff --git a/src/brapi/types/v2/prime_rate_list_available_response.py b/src/brapi/types/v2/prime_rate_list_available_response.py new file mode 100644 index 0000000..4cfacdd --- /dev/null +++ b/src/brapi/types/v2/prime_rate_list_available_response.py @@ -0,0 +1,15 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from ..._models import BaseModel + +__all__ = ["PrimeRateListAvailableResponse"] + + +class PrimeRateListAvailableResponse(BaseModel): + countries: Optional[List[str]] = None + """ + Lista de países com dados de taxa básica de juros (SELIC) disponíveis para + consulta. + """ diff --git a/src/brapi/types/v2/prime_rate_retrieve_params.py b/src/brapi/types/v2/prime_rate_retrieve_params.py new file mode 100644 index 0000000..0fb3d75 --- /dev/null +++ b/src/brapi/types/v2/prime_rate_retrieve_params.py @@ -0,0 +1,67 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +from typing import Union +from datetime import date +from typing_extensions import Literal, Annotated, TypedDict + +from ..._utils import PropertyInfo + +__all__ = ["PrimeRateRetrieveParams"] + + +class PrimeRateRetrieveParams(TypedDict, total=False): + token: str + """ + **Obrigatório caso não esteja adicionado como header "Authorization".** Seu + token de autenticação pessoal da API Brapi. + + **Formas de Envio:** + + 1. **Query Parameter:** Adicione `?token=SEU_TOKEN` ao final da URL. + 2. **HTTP Header:** Inclua o header `Authorization: Bearer SEU_TOKEN` na sua + requisição. + + Ambos os métodos são aceitos, mas pelo menos um deles deve ser utilizado. + Obtenha seu token em [brapi.dev/dashboard](https://brapi.dev/dashboard). + """ + + country: str + """ + **Opcional.** O país do qual você deseja obter informações sobre a taxa básica + de juros. Por padrão, o país é definido como brazil. Você pode consultar a lista + de países disponíveis através do endpoint `/api/v2/prime-rate/available`. + """ + + end: Annotated[Union[str, date], PropertyInfo(format="iso8601")] + """**Opcional.** Data final do período para busca no formato DD/MM/YYYY. + + Por padrão é a data atual. Útil quando `historical=true` para restringir o + período da série histórica. + """ + + historical: bool + """**Opcional.** Define se os dados históricos serão retornados. + + Se definido como `true`, retorna a série histórica completa. Se `false` (padrão) + ou omitido, retorna apenas o valor mais recente. + """ + + sort_by: Annotated[Literal["date", "value"], PropertyInfo(alias="sortBy")] + """**Opcional.** Campo pelo qual os resultados serão ordenados. + + Por padrão, ordena por `date` (data). + """ + + sort_order: Annotated[Literal["asc", "desc"], PropertyInfo(alias="sortOrder")] + """ + **Opcional.** Define se a ordenação será crescente (`asc`) ou decrescente + (`desc`). Por padrão, é `desc` (decrescente). + """ + + start: Annotated[Union[str, date], PropertyInfo(format="iso8601")] + """**Opcional.** Data inicial do período para busca no formato DD/MM/YYYY. + + Útil quando `historical=true` para restringir o período da série histórica. + """ diff --git a/src/brapi/types/v2/prime_rate_retrieve_response.py b/src/brapi/types/v2/prime_rate_retrieve_response.py new file mode 100644 index 0000000..967860e --- /dev/null +++ b/src/brapi/types/v2/prime_rate_retrieve_response.py @@ -0,0 +1,28 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import List, Optional + +from pydantic import Field as FieldInfo + +from ..._models import BaseModel + +__all__ = ["PrimeRateRetrieveResponse", "PrimeRate"] + + +class PrimeRate(BaseModel): + date: Optional[str] = None + """Data do registro no formato DD/MM/YYYY.""" + + epoch_date: Optional[int] = FieldInfo(alias="epochDate", default=None) + """Timestamp em milissegundos (formato epoch) correspondente à data do registro.""" + + value: Optional[str] = None + """Valor da taxa básica de juros (SELIC) para a data correspondente.""" + + +class PrimeRateRetrieveResponse(BaseModel): + prime_rate: Optional[List[PrimeRate]] = FieldInfo(alias="prime-rate", default=None) + """ + Array contendo os registros históricos de taxa básica de juros (SELIC) para o + país e período solicitados. + """ diff --git a/src/brapi/types/value_added_entry.py b/src/brapi/types/value_added_entry.py new file mode 100644 index 0000000..70c0db2 --- /dev/null +++ b/src/brapi/types/value_added_entry.py @@ -0,0 +1,240 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from typing import Optional +from datetime import date +from typing_extensions import Literal + +from pydantic import Field as FieldInfo + +from .._models import BaseModel + +__all__ = ["ValueAddedEntry"] + + +class ValueAddedEntry(BaseModel): + added_value_received_by_transfer: Optional[float] = FieldInfo(alias="addedValueReceivedByTransfer", default=None) + """ + Valor Adicionado Recebido em Transferência (Resultado de Equivalência + Patrimonial, Receitas Financeiras, etc.). Item 6 da DVA. + """ + + added_value_received_on_transfer: Optional[float] = FieldInfo(alias="addedValueReceivedOnTransfer", default=None) + """ + Valor Adicionado Recebido em Transferência (sinônimo de + `addedValueReceivedByTransfer`). + """ + + added_value_to_distribute: Optional[float] = FieldInfo(alias="addedValueToDistribute", default=None) + """ + Valor Adicionado Total a Distribuir (Líquido Produzido + Recebido em + Transferência). Item 7 da DVA. + """ + + claims_and_benefits: Optional[float] = FieldInfo(alias="claimsAndBenefits", default=None) + """Sinistros Retidos e Benefícios.""" + + complementary_pension_operations_revenue: Optional[float] = FieldInfo( + alias="complementaryPensionOperationsRevenue", default=None + ) + """Receita com Operações de Previdência Complementar.""" + + construction_of_own_assets: Optional[float] = FieldInfo(alias="constructionOfOwnAssets", default=None) + """Construção de Ativos Próprios.""" + + costs_with_products_sold: Optional[float] = FieldInfo(alias="costsWithProductsSold", default=None) + """Custos dos Produtos, Mercadorias e Serviços Vendidos (detalhamento).""" + + depreciation_and_amortization: Optional[float] = FieldInfo(alias="depreciationAndAmortization", default=None) + """Depreciação e Amortização.""" + + distribution_of_added_value: Optional[float] = FieldInfo(alias="distributionOfAddedValue", default=None) + """Distribuição do Valor Adicionado (Soma dos itens seguintes). Item 8 da DVA.""" + + dividends: Optional[float] = None + """Dividendos Distribuídos.""" + + end_date: Optional[date] = FieldInfo(alias="endDate", default=None) + """Data de término do período fiscal ao qual a DVA se refere (YYYY-MM-DD).""" + + equity_income_result: Optional[float] = FieldInfo(alias="equityIncomeResult", default=None) + """Resultado de Equivalência Patrimonial (como receita na DVA).""" + + equity_remuneration: Optional[float] = FieldInfo(alias="equityRemuneration", default=None) + """Remuneração de Capitais Próprios (JCP, Dividendos, Lucros Retidos).""" + + federal_taxes: Optional[float] = FieldInfo(alias="federalTaxes", default=None) + """Impostos Federais (IRPJ, CSLL, PIS, COFINS, IPI).""" + + fees_revenue: Optional[float] = FieldInfo(alias="feesRevenue", default=None) + """Receita com Taxas e Comissões.""" + + financial_income: Optional[float] = FieldInfo(alias="financialIncome", default=None) + """Receitas Financeiras (como valor recebido em transferência).""" + + financial_intermediation_expenses: Optional[float] = FieldInfo( + alias="financialIntermediationExpenses", default=None + ) + """Despesas de Intermediação Financeira (específico para bancos).""" + + financial_intermediation_revenue: Optional[float] = FieldInfo(alias="financialIntermediationRevenue", default=None) + """Receita de Intermediação Financeira (específico para bancos).""" + + gross_added_value: Optional[float] = FieldInfo(alias="grossAddedValue", default=None) + """Valor Adicionado Bruto (Receitas - Insumos). Item 3 da DVA.""" + + insurance_operations_revenue: Optional[float] = FieldInfo(alias="insuranceOperationsRevenue", default=None) + """Receita com Operações de Seguros (específico para Seguradoras).""" + + insurance_operations_variations: Optional[float] = FieldInfo(alias="insuranceOperationsVariations", default=None) + """Variações de Operações de Seguros.""" + + interest_on_own_equity: Optional[float] = FieldInfo(alias="interestOnOwnEquity", default=None) + """Juros sobre o Capital Próprio (JCP).""" + + loss_or_recovery_of_assets: Optional[float] = FieldInfo(alias="lossOrRecoveryOfAssets", default=None) + """Perda/Recuperação de Valores de Ativos (Impairment - como custo/receita).""" + + loss_or_recovery_of_asset_values: Optional[float] = FieldInfo(alias="lossOrRecoveryOfAssetValues", default=None) + """Perda / Recuperação de Valores de Ativos (Impairment).""" + + materials_energy_and_others: Optional[float] = FieldInfo(alias="materialsEnergyAndOthers", default=None) + """Custos com Materiais, Energia, Serviços de Terceiros e Outros.""" + + municipal_taxes: Optional[float] = FieldInfo(alias="municipalTaxes", default=None) + """Impostos Municipais (ISS).""" + + net_added_value: Optional[float] = FieldInfo(alias="netAddedValue", default=None) + """Valor Adicionado Líquido Produzido pela Entidade (Bruto - Retenções). + + Item 5 da DVA. + """ + + net_added_value_produced: Optional[float] = FieldInfo(alias="netAddedValueProduced", default=None) + """Valor Adicionado Líquido Produzido (sinônimo de `netAddedValue`).""" + + net_operating_revenue: Optional[float] = FieldInfo(alias="netOperatingRevenue", default=None) + """Receita Operacional Líquida (detalhamento).""" + + non_controlling_share_of_retained_earnings: Optional[float] = FieldInfo( + alias="nonControllingShareOfRetainedEarnings", default=None + ) + """Participação dos Não Controladores nos Lucros Retidos.""" + + other_distributions: Optional[float] = FieldInfo(alias="otherDistributions", default=None) + """Outras Distribuições.""" + + other_retentions: Optional[float] = FieldInfo(alias="otherRetentions", default=None) + """Outras Retenções (Exaustão, etc.).""" + + other_revenues: Optional[float] = FieldInfo(alias="otherRevenues", default=None) + """Outras Receitas.""" + + other_supplies: Optional[float] = FieldInfo(alias="otherSupplies", default=None) + """Outros Insumos.""" + + other_values_received_by_transfer: Optional[float] = FieldInfo(alias="otherValuesReceivedByTransfer", default=None) + """Outros Valores Recebidos (Receitas Financeiras, Aluguéis, etc.).""" + + other_variations: Optional[float] = FieldInfo(alias="otherVariations", default=None) + """Outras Variações.""" + + own_equity_remuneration: Optional[float] = FieldInfo(alias="ownEquityRemuneration", default=None) + """Remuneração de Capitais Próprios (sinônimo de `equityRemuneration`).""" + + pension_operations_variations: Optional[float] = FieldInfo(alias="pensionOperationsVariations", default=None) + """Variações de Operações de Previdência.""" + + product_sales: Optional[float] = FieldInfo(alias="productSales", default=None) + """Venda de Produtos e Serviços (detalhamento).""" + + provision_or_reversal_of_doubtful_accounts: Optional[float] = FieldInfo( + alias="provisionOrReversalOfDoubtfulAccounts", default=None + ) + """ + Provisão/Reversão para Créditos de Liquidação Duvidosa (PCLD - como + receita/despesa na DVA). + """ + + provision_or_reversal_of_expected_credit_risk_losses: Optional[float] = FieldInfo( + alias="provisionOrReversalOfExpectedCreditRiskLosses", default=None + ) + """Provisão/Reversão de Perdas com Risco de Crédito (PCLD).""" + + remuneration_of_third_party_capitals: Optional[float] = FieldInfo( + alias="remunerationOfThirdPartyCapitals", default=None + ) + """Remuneração de Capitais de Terceiros (Juros, Aluguéis).""" + + result_of_coinsurance_operations_assigned: Optional[float] = FieldInfo( + alias="resultOfCoinsuranceOperationsAssigned", default=None + ) + """Resultado de Operações de Cosseguros Cedidos.""" + + results_of_ceded_reinsurance_operations: Optional[float] = FieldInfo( + alias="resultsOfCededReinsuranceOperations", default=None + ) + """Resultados de Operações de Resseguros Cedidos.""" + + retained_earnings_or_loss: Optional[float] = FieldInfo(alias="retainedEarningsOrLoss", default=None) + """Lucros Retidos ou Prejuízo do Exercício.""" + + retentions: Optional[float] = None + """Retenções (Depreciação, Amortização e Exaustão). Item 4 da DVA.""" + + revenue: Optional[float] = None + """Receitas (Venda de Mercadorias, Produtos e Serviços, etc.). Item 1 da DVA.""" + + revenue_from_the_provision_of_services: Optional[float] = FieldInfo( + alias="revenueFromTheProvisionOfServices", default=None + ) + """Receita da Prestação de Serviços (detalhamento).""" + + services: Optional[float] = None + """Serviços de Terceiros (detalhamento).""" + + state_taxes: Optional[float] = FieldInfo(alias="stateTaxes", default=None) + """Impostos Estaduais (ICMS).""" + + supplies_purchased_from_third_parties: Optional[float] = FieldInfo( + alias="suppliesPurchasedFromThirdParties", default=None + ) + """Insumos Adquiridos de Terceiros (Custo de Mercadorias, Matérias-Primas). + + Item 2 da DVA. + """ + + symbol: Optional[str] = None + """Ticker do ativo ao qual a DVA se refere.""" + + taxes: Optional[float] = None + """Impostos, Taxas e Contribuições (Federais, Estaduais, Municipais).""" + + team_remuneration: Optional[float] = FieldInfo(alias="teamRemuneration", default=None) + """Pessoal e Encargos (Salários, Benefícios, FGTS).""" + + third_party_materials_and_services: Optional[float] = FieldInfo( + alias="thirdPartyMaterialsAndServices", default=None + ) + """Materiais, Energia, Serviços de Terceiros.""" + + total_added_value_to_distribute: Optional[float] = FieldInfo(alias="totalAddedValueToDistribute", default=None) + """Valor Adicionado Total a Distribuir (sinônimo de `addedValueToDistribute`).""" + + type: Optional[Literal["yearly", "quarterly"]] = None + """Indica a periodicidade da DVA: `yearly` (anual) ou `quarterly` (trimestral).""" + + updated_at: Optional[date] = FieldInfo(alias="updatedAt", default=None) + """ + Data da última atualização deste registro específico na fonte de dados + (YYYY-MM-DD). + """ + + variation_in_deferred_selling_expenses: Optional[float] = FieldInfo( + alias="variationInDeferredSellingExpenses", default=None + ) + """Variação nas Despesas de Comercialização Diferidas.""" + + variations_of_technical_provisions: Optional[float] = FieldInfo( + alias="variationsOfTechnicalProvisions", default=None + ) + """Variações das Provisões Técnicas (específico para Seguradoras).""" diff --git a/tests/__init__.py b/tests/__init__.py new file mode 100644 index 0000000..fd8019a --- /dev/null +++ b/tests/__init__.py @@ -0,0 +1 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. diff --git a/tests/api_resources/__init__.py b/tests/api_resources/__init__.py new file mode 100644 index 0000000..fd8019a --- /dev/null +++ b/tests/api_resources/__init__.py @@ -0,0 +1 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. diff --git a/tests/api_resources/test_available.py b/tests/api_resources/test_available.py new file mode 100644 index 0000000..6b366e0 --- /dev/null +++ b/tests/api_resources/test_available.py @@ -0,0 +1,98 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from brapi.types import AvailableListResponse +from tests.utils import assert_matches_type + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestAvailable: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list(self, client: Brapi) -> None: + available = client.available.list() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_with_all_params(self, client: Brapi) -> None: + available = client.available.list( + token="token", + search="search", + ) + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list(self, client: Brapi) -> None: + response = client.available.with_raw_response.list() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + available = response.parse() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list(self, client: Brapi) -> None: + with client.available.with_streaming_response.list() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + available = response.parse() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncAvailable: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list(self, async_client: AsyncBrapi) -> None: + available = await async_client.available.list() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> None: + available = await async_client.available.list( + token="token", + search="search", + ) + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: + response = await async_client.available.with_raw_response.list() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + available = await response.parse() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list(self, async_client: AsyncBrapi) -> None: + async with async_client.available.with_streaming_response.list() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + available = await response.parse() + assert_matches_type(AvailableListResponse, available, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/test_quote.py b/tests/api_resources/test_quote.py new file mode 100644 index 0000000..21a14a6 --- /dev/null +++ b/tests/api_resources/test_quote.py @@ -0,0 +1,222 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from brapi.types import QuoteListResponse, QuoteRetrieveResponse +from tests.utils import assert_matches_type + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestQuote: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve(self, client: Brapi) -> None: + quote = client.quote.retrieve( + tickers="PETR4,MGLU3", + ) + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve_with_all_params(self, client: Brapi) -> None: + quote = client.quote.retrieve( + tickers="PETR4,MGLU3", + token="token", + dividends=True, + fundamental=True, + interval="1d", + modules=["summaryProfile", "balanceSheetHistory", "financialData"], + range="5d", + ) + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_retrieve(self, client: Brapi) -> None: + response = client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + quote = response.parse() + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_retrieve(self, client: Brapi) -> None: + with client.quote.with_streaming_response.retrieve( + tickers="PETR4,MGLU3", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + quote = response.parse() + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_path_params_retrieve(self, client: Brapi) -> None: + with pytest.raises(ValueError, match=r"Expected a non-empty value for `tickers` but received ''"): + client.quote.with_raw_response.retrieve( + tickers="", + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list(self, client: Brapi) -> None: + quote = client.quote.list() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_with_all_params(self, client: Brapi) -> None: + quote = client.quote.list( + token="token", + limit=1, + page=1, + search="search", + sector="Retail Trade", + sort_by="name", + sort_order="asc", + type="stock", + ) + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list(self, client: Brapi) -> None: + response = client.quote.with_raw_response.list() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + quote = response.parse() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list(self, client: Brapi) -> None: + with client.quote.with_streaming_response.list() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + quote = response.parse() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncQuote: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: + quote = await async_client.quote.retrieve( + tickers="PETR4,MGLU3", + ) + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: + quote = await async_client.quote.retrieve( + tickers="PETR4,MGLU3", + token="token", + dividends=True, + fundamental=True, + interval="1d", + modules=["summaryProfile", "balanceSheetHistory", "financialData"], + range="5d", + ) + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: + response = await async_client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + quote = await response.parse() + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: + async with async_client.quote.with_streaming_response.retrieve( + tickers="PETR4,MGLU3", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + quote = await response.parse() + assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_path_params_retrieve(self, async_client: AsyncBrapi) -> None: + with pytest.raises(ValueError, match=r"Expected a non-empty value for `tickers` but received ''"): + await async_client.quote.with_raw_response.retrieve( + tickers="", + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list(self, async_client: AsyncBrapi) -> None: + quote = await async_client.quote.list() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> None: + quote = await async_client.quote.list( + token="token", + limit=1, + page=1, + search="search", + sector="Retail Trade", + sort_by="name", + sort_order="asc", + type="stock", + ) + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: + response = await async_client.quote.with_raw_response.list() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + quote = await response.parse() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list(self, async_client: AsyncBrapi) -> None: + async with async_client.quote.with_streaming_response.list() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + quote = await response.parse() + assert_matches_type(QuoteListResponse, quote, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/v2/__init__.py b/tests/api_resources/v2/__init__.py new file mode 100644 index 0000000..fd8019a --- /dev/null +++ b/tests/api_resources/v2/__init__.py @@ -0,0 +1 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. diff --git a/tests/api_resources/v2/test_crypto.py b/tests/api_resources/v2/test_crypto.py new file mode 100644 index 0000000..957328e --- /dev/null +++ b/tests/api_resources/v2/test_crypto.py @@ -0,0 +1,193 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from tests.utils import assert_matches_type +from brapi.types.v2 import ( + CryptoRetrieveResponse, + CryptoListAvailableResponse, +) + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestCrypto: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve(self, client: Brapi) -> None: + crypto = client.v2.crypto.retrieve( + coin="coin", + ) + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve_with_all_params(self, client: Brapi) -> None: + crypto = client.v2.crypto.retrieve( + coin="coin", + token="token", + currency="currency", + interval="1m", + range="1d", + ) + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_retrieve(self, client: Brapi) -> None: + response = client.v2.crypto.with_raw_response.retrieve( + coin="coin", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + crypto = response.parse() + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_retrieve(self, client: Brapi) -> None: + with client.v2.crypto.with_streaming_response.retrieve( + coin="coin", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + crypto = response.parse() + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available(self, client: Brapi) -> None: + crypto = client.v2.crypto.list_available() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available_with_all_params(self, client: Brapi) -> None: + crypto = client.v2.crypto.list_available( + token="token", + search="search", + ) + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list_available(self, client: Brapi) -> None: + response = client.v2.crypto.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + crypto = response.parse() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list_available(self, client: Brapi) -> None: + with client.v2.crypto.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + crypto = response.parse() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncCrypto: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: + crypto = await async_client.v2.crypto.retrieve( + coin="coin", + ) + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: + crypto = await async_client.v2.crypto.retrieve( + coin="coin", + token="token", + currency="currency", + interval="1m", + range="1d", + ) + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.crypto.with_raw_response.retrieve( + coin="coin", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + crypto = await response.parse() + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.crypto.with_streaming_response.retrieve( + coin="coin", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + crypto = await response.parse() + assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available(self, async_client: AsyncBrapi) -> None: + crypto = await async_client.v2.crypto.list_available() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: + crypto = await async_client.v2.crypto.list_available( + token="token", + search="search", + ) + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.crypto.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + crypto = await response.parse() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.crypto.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + crypto = await response.parse() + assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/v2/test_currency.py b/tests/api_resources/v2/test_currency.py new file mode 100644 index 0000000..0326fc1 --- /dev/null +++ b/tests/api_resources/v2/test_currency.py @@ -0,0 +1,187 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from tests.utils import assert_matches_type +from brapi.types.v2 import ( + CurrencyRetrieveResponse, + CurrencyListAvailableResponse, +) + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestCurrency: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve(self, client: Brapi) -> None: + currency = client.v2.currency.retrieve( + currency="USD-BRL,EUR-USD", + ) + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve_with_all_params(self, client: Brapi) -> None: + currency = client.v2.currency.retrieve( + currency="USD-BRL,EUR-USD", + token="token", + ) + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_retrieve(self, client: Brapi) -> None: + response = client.v2.currency.with_raw_response.retrieve( + currency="USD-BRL,EUR-USD", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + currency = response.parse() + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_retrieve(self, client: Brapi) -> None: + with client.v2.currency.with_streaming_response.retrieve( + currency="USD-BRL,EUR-USD", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + currency = response.parse() + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available(self, client: Brapi) -> None: + currency = client.v2.currency.list_available() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available_with_all_params(self, client: Brapi) -> None: + currency = client.v2.currency.list_available( + token="token", + search="search", + ) + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list_available(self, client: Brapi) -> None: + response = client.v2.currency.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + currency = response.parse() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list_available(self, client: Brapi) -> None: + with client.v2.currency.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + currency = response.parse() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncCurrency: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: + currency = await async_client.v2.currency.retrieve( + currency="USD-BRL,EUR-USD", + ) + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: + currency = await async_client.v2.currency.retrieve( + currency="USD-BRL,EUR-USD", + token="token", + ) + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.currency.with_raw_response.retrieve( + currency="USD-BRL,EUR-USD", + ) + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + currency = await response.parse() + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.currency.with_streaming_response.retrieve( + currency="USD-BRL,EUR-USD", + ) as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + currency = await response.parse() + assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available(self, async_client: AsyncBrapi) -> None: + currency = await async_client.v2.currency.list_available() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: + currency = await async_client.v2.currency.list_available( + token="token", + search="search", + ) + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.currency.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + currency = await response.parse() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.currency.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + currency = await response.parse() + assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/v2/test_inflation.py b/tests/api_resources/v2/test_inflation.py new file mode 100644 index 0000000..4071b4d --- /dev/null +++ b/tests/api_resources/v2/test_inflation.py @@ -0,0 +1,186 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from tests.utils import assert_matches_type +from brapi._utils import parse_date +from brapi.types.v2 import ( + InflationRetrieveResponse, + InflationListAvailableResponse, +) + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestInflation: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve(self, client: Brapi) -> None: + inflation = client.v2.inflation.retrieve() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve_with_all_params(self, client: Brapi) -> None: + inflation = client.v2.inflation.retrieve( + token="token", + country="country", + end=parse_date("2019-12-27"), + historical=True, + sort_by="date", + sort_order="asc", + start=parse_date("2019-12-27"), + ) + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_retrieve(self, client: Brapi) -> None: + response = client.v2.inflation.with_raw_response.retrieve() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + inflation = response.parse() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_retrieve(self, client: Brapi) -> None: + with client.v2.inflation.with_streaming_response.retrieve() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + inflation = response.parse() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available(self, client: Brapi) -> None: + inflation = client.v2.inflation.list_available() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available_with_all_params(self, client: Brapi) -> None: + inflation = client.v2.inflation.list_available( + token="token", + search="search", + ) + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list_available(self, client: Brapi) -> None: + response = client.v2.inflation.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + inflation = response.parse() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list_available(self, client: Brapi) -> None: + with client.v2.inflation.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + inflation = response.parse() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncInflation: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: + inflation = await async_client.v2.inflation.retrieve() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: + inflation = await async_client.v2.inflation.retrieve( + token="token", + country="country", + end=parse_date("2019-12-27"), + historical=True, + sort_by="date", + sort_order="asc", + start=parse_date("2019-12-27"), + ) + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.inflation.with_raw_response.retrieve() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + inflation = await response.parse() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.inflation.with_streaming_response.retrieve() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + inflation = await response.parse() + assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available(self, async_client: AsyncBrapi) -> None: + inflation = await async_client.v2.inflation.list_available() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: + inflation = await async_client.v2.inflation.list_available( + token="token", + search="search", + ) + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.inflation.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + inflation = await response.parse() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.inflation.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + inflation = await response.parse() + assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/api_resources/v2/test_prime_rate.py b/tests/api_resources/v2/test_prime_rate.py new file mode 100644 index 0000000..ecb5738 --- /dev/null +++ b/tests/api_resources/v2/test_prime_rate.py @@ -0,0 +1,186 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +from typing import Any, cast + +import pytest + +from brapi import Brapi, AsyncBrapi +from tests.utils import assert_matches_type +from brapi._utils import parse_date +from brapi.types.v2 import ( + PrimeRateRetrieveResponse, + PrimeRateListAvailableResponse, +) + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + + +class TestPrimeRate: + parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve(self, client: Brapi) -> None: + prime_rate = client.v2.prime_rate.retrieve() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_retrieve_with_all_params(self, client: Brapi) -> None: + prime_rate = client.v2.prime_rate.retrieve( + token="token", + country="country", + end=parse_date("2019-12-27"), + historical=True, + sort_by="date", + sort_order="asc", + start=parse_date("2019-12-27"), + ) + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_retrieve(self, client: Brapi) -> None: + response = client.v2.prime_rate.with_raw_response.retrieve() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + prime_rate = response.parse() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_retrieve(self, client: Brapi) -> None: + with client.v2.prime_rate.with_streaming_response.retrieve() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + prime_rate = response.parse() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available(self, client: Brapi) -> None: + prime_rate = client.v2.prime_rate.list_available() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_method_list_available_with_all_params(self, client: Brapi) -> None: + prime_rate = client.v2.prime_rate.list_available( + token="token", + search="search", + ) + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_raw_response_list_available(self, client: Brapi) -> None: + response = client.v2.prime_rate.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + prime_rate = response.parse() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + def test_streaming_response_list_available(self, client: Brapi) -> None: + with client.v2.prime_rate.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + prime_rate = response.parse() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + assert cast(Any, response.is_closed) is True + + +class TestAsyncPrimeRate: + parametrize = pytest.mark.parametrize( + "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] + ) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: + prime_rate = await async_client.v2.prime_rate.retrieve() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: + prime_rate = await async_client.v2.prime_rate.retrieve( + token="token", + country="country", + end=parse_date("2019-12-27"), + historical=True, + sort_by="date", + sort_order="asc", + start=parse_date("2019-12-27"), + ) + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.prime_rate.with_raw_response.retrieve() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + prime_rate = await response.parse() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.prime_rate.with_streaming_response.retrieve() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + prime_rate = await response.parse() + assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) + + assert cast(Any, response.is_closed) is True + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available(self, async_client: AsyncBrapi) -> None: + prime_rate = await async_client.v2.prime_rate.list_available() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: + prime_rate = await async_client.v2.prime_rate.list_available( + token="token", + search="search", + ) + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: + response = await async_client.v2.prime_rate.with_raw_response.list_available() + + assert response.is_closed is True + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + prime_rate = await response.parse() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + @pytest.mark.skip(reason="Prism tests are disabled") + @parametrize + async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: + async with async_client.v2.prime_rate.with_streaming_response.list_available() as response: + assert not response.is_closed + assert response.http_request.headers.get("X-Stainless-Lang") == "python" + + prime_rate = await response.parse() + assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) + + assert cast(Any, response.is_closed) is True diff --git a/tests/conftest.py b/tests/conftest.py new file mode 100644 index 0000000..a4865b6 --- /dev/null +++ b/tests/conftest.py @@ -0,0 +1,84 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import os +import logging +from typing import TYPE_CHECKING, Iterator, AsyncIterator + +import httpx +import pytest +from pytest_asyncio import is_async_test + +from brapi import Brapi, AsyncBrapi, DefaultAioHttpClient +from brapi._utils import is_dict + +if TYPE_CHECKING: + from _pytest.fixtures import FixtureRequest # pyright: ignore[reportPrivateImportUsage] + +pytest.register_assert_rewrite("tests.utils") + +logging.getLogger("brapi").setLevel(logging.DEBUG) + + +# automatically add `pytest.mark.asyncio()` to all of our async tests +# so we don't have to add that boilerplate everywhere +def pytest_collection_modifyitems(items: list[pytest.Function]) -> None: + pytest_asyncio_tests = (item for item in items if is_async_test(item)) + session_scope_marker = pytest.mark.asyncio(loop_scope="session") + for async_test in pytest_asyncio_tests: + async_test.add_marker(session_scope_marker, append=False) + + # We skip tests that use both the aiohttp client and respx_mock as respx_mock + # doesn't support custom transports. + for item in items: + if "async_client" not in item.fixturenames or "respx_mock" not in item.fixturenames: + continue + + if not hasattr(item, "callspec"): + continue + + async_client_param = item.callspec.params.get("async_client") + if is_dict(async_client_param) and async_client_param.get("http_client") == "aiohttp": + item.add_marker(pytest.mark.skip(reason="aiohttp client is not compatible with respx_mock")) + + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") + +api_key = "My API Key" + + +@pytest.fixture(scope="session") +def client(request: FixtureRequest) -> Iterator[Brapi]: + strict = getattr(request, "param", True) + if not isinstance(strict, bool): + raise TypeError(f"Unexpected fixture parameter type {type(strict)}, expected {bool}") + + with Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=strict) as client: + yield client + + +@pytest.fixture(scope="session") +async def async_client(request: FixtureRequest) -> AsyncIterator[AsyncBrapi]: + param = getattr(request, "param", True) + + # defaults + strict = True + http_client: None | httpx.AsyncClient = None + + if isinstance(param, bool): + strict = param + elif is_dict(param): + strict = param.get("strict", True) + assert isinstance(strict, bool) + + http_client_type = param.get("http_client", "httpx") + if http_client_type == "aiohttp": + http_client = DefaultAioHttpClient() + else: + raise TypeError(f"Unexpected fixture parameter type {type(param)}, expected bool or dict") + + async with AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=strict, http_client=http_client + ) as client: + yield client diff --git a/tests/sample_file.txt b/tests/sample_file.txt new file mode 100644 index 0000000..af5626b --- /dev/null +++ b/tests/sample_file.txt @@ -0,0 +1 @@ +Hello, world! diff --git a/tests/test_client.py b/tests/test_client.py new file mode 100644 index 0000000..0ef232e --- /dev/null +++ b/tests/test_client.py @@ -0,0 +1,1699 @@ +# File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. + +from __future__ import annotations + +import gc +import os +import sys +import json +import asyncio +import inspect +import tracemalloc +from typing import Any, Union, cast +from unittest import mock +from typing_extensions import Literal + +import httpx +import pytest +from respx import MockRouter +from pydantic import ValidationError + +from brapi import Brapi, AsyncBrapi, APIResponseValidationError +from brapi._types import Omit +from brapi._utils import asyncify +from brapi._models import BaseModel, FinalRequestOptions +from brapi._exceptions import BrapiError, APIStatusError, APITimeoutError, APIResponseValidationError +from brapi._base_client import ( + DEFAULT_TIMEOUT, + HTTPX_DEFAULT_TIMEOUT, + BaseClient, + OtherPlatform, + DefaultHttpxClient, + DefaultAsyncHttpxClient, + get_platform, + make_request_options, +) + +from .utils import update_env + +base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") +api_key = "My API Key" + + +def _get_params(client: BaseClient[Any, Any]) -> dict[str, str]: + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + url = httpx.URL(request.url) + return dict(url.params) + + +def _low_retry_timeout(*_args: Any, **_kwargs: Any) -> float: + return 0.1 + + +def _get_open_connections(client: Brapi | AsyncBrapi) -> int: + transport = client._client._transport + assert isinstance(transport, httpx.HTTPTransport) or isinstance(transport, httpx.AsyncHTTPTransport) + + pool = transport._pool + return len(pool._requests) + + +class TestBrapi: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + @pytest.mark.respx(base_url=base_url) + def test_raw_response(self, respx_mock: MockRouter) -> None: + respx_mock.post("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = self.client.post("/foo", cast_to=httpx.Response) + assert response.status_code == 200 + assert isinstance(response, httpx.Response) + assert response.json() == {"foo": "bar"} + + @pytest.mark.respx(base_url=base_url) + def test_raw_response_for_binary(self, respx_mock: MockRouter) -> None: + respx_mock.post("/foo").mock( + return_value=httpx.Response(200, headers={"Content-Type": "application/binary"}, content='{"foo": "bar"}') + ) + + response = self.client.post("/foo", cast_to=httpx.Response) + assert response.status_code == 200 + assert isinstance(response, httpx.Response) + assert response.json() == {"foo": "bar"} + + def test_copy(self) -> None: + copied = self.client.copy() + assert id(copied) != id(self.client) + + copied = self.client.copy(api_key="another My API Key") + assert copied.api_key == "another My API Key" + assert self.client.api_key == "My API Key" + + def test_copy_default_options(self) -> None: + # options that have a default are overridden correctly + copied = self.client.copy(max_retries=7) + assert copied.max_retries == 7 + assert self.client.max_retries == 2 + + copied2 = copied.copy(max_retries=6) + assert copied2.max_retries == 6 + assert copied.max_retries == 7 + + # timeout + assert isinstance(self.client.timeout, httpx.Timeout) + copied = self.client.copy(timeout=None) + assert copied.timeout is None + assert isinstance(self.client.timeout, httpx.Timeout) + + def test_copy_default_headers(self) -> None: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} + ) + assert client.default_headers["X-Foo"] == "bar" + + # does not override the already given value when not specified + copied = client.copy() + assert copied.default_headers["X-Foo"] == "bar" + + # merges already given headers + copied = client.copy(default_headers={"X-Bar": "stainless"}) + assert copied.default_headers["X-Foo"] == "bar" + assert copied.default_headers["X-Bar"] == "stainless" + + # uses new values for any already given headers + copied = client.copy(default_headers={"X-Foo": "stainless"}) + assert copied.default_headers["X-Foo"] == "stainless" + + # set_default_headers + + # completely overrides already set values + copied = client.copy(set_default_headers={}) + assert copied.default_headers.get("X-Foo") is None + + copied = client.copy(set_default_headers={"X-Bar": "Robert"}) + assert copied.default_headers["X-Bar"] == "Robert" + + with pytest.raises( + ValueError, + match="`default_headers` and `set_default_headers` arguments are mutually exclusive", + ): + client.copy(set_default_headers={}, default_headers={"X-Foo": "Bar"}) + + def test_copy_default_query(self) -> None: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"foo": "bar"} + ) + assert _get_params(client)["foo"] == "bar" + + # does not override the already given value when not specified + copied = client.copy() + assert _get_params(copied)["foo"] == "bar" + + # merges already given params + copied = client.copy(default_query={"bar": "stainless"}) + params = _get_params(copied) + assert params["foo"] == "bar" + assert params["bar"] == "stainless" + + # uses new values for any already given headers + copied = client.copy(default_query={"foo": "stainless"}) + assert _get_params(copied)["foo"] == "stainless" + + # set_default_query + + # completely overrides already set values + copied = client.copy(set_default_query={}) + assert _get_params(copied) == {} + + copied = client.copy(set_default_query={"bar": "Robert"}) + assert _get_params(copied)["bar"] == "Robert" + + with pytest.raises( + ValueError, + # TODO: update + match="`default_query` and `set_default_query` arguments are mutually exclusive", + ): + client.copy(set_default_query={}, default_query={"foo": "Bar"}) + + def test_copy_signature(self) -> None: + # ensure the same parameters that can be passed to the client are defined in the `.copy()` method + init_signature = inspect.signature( + # mypy doesn't like that we access the `__init__` property. + self.client.__init__, # type: ignore[misc] + ) + copy_signature = inspect.signature(self.client.copy) + exclude_params = {"transport", "proxies", "_strict_response_validation"} + + for name in init_signature.parameters.keys(): + if name in exclude_params: + continue + + copy_param = copy_signature.parameters.get(name) + assert copy_param is not None, f"copy() signature is missing the {name} param" + + @pytest.mark.skipif(sys.version_info >= (3, 10), reason="fails because of a memory leak that started from 3.12") + def test_copy_build_request(self) -> None: + options = FinalRequestOptions(method="get", url="/foo") + + def build_request(options: FinalRequestOptions) -> None: + client = self.client.copy() + client._build_request(options) + + # ensure that the machinery is warmed up before tracing starts. + build_request(options) + gc.collect() + + tracemalloc.start(1000) + + snapshot_before = tracemalloc.take_snapshot() + + ITERATIONS = 10 + for _ in range(ITERATIONS): + build_request(options) + + gc.collect() + snapshot_after = tracemalloc.take_snapshot() + + tracemalloc.stop() + + def add_leak(leaks: list[tracemalloc.StatisticDiff], diff: tracemalloc.StatisticDiff) -> None: + if diff.count == 0: + # Avoid false positives by considering only leaks (i.e. allocations that persist). + return + + if diff.count % ITERATIONS != 0: + # Avoid false positives by considering only leaks that appear per iteration. + return + + for frame in diff.traceback: + if any( + frame.filename.endswith(fragment) + for fragment in [ + # to_raw_response_wrapper leaks through the @functools.wraps() decorator. + # + # removing the decorator fixes the leak for reasons we don't understand. + "brapi/_legacy_response.py", + "brapi/_response.py", + # pydantic.BaseModel.model_dump || pydantic.BaseModel.dict leak memory for some reason. + "brapi/_compat.py", + # Standard library leaks we don't care about. + "/logging/__init__.py", + ] + ): + return + + leaks.append(diff) + + leaks: list[tracemalloc.StatisticDiff] = [] + for diff in snapshot_after.compare_to(snapshot_before, "traceback"): + add_leak(leaks, diff) + if leaks: + for leak in leaks: + print("MEMORY LEAK:", leak) + for frame in leak.traceback: + print(frame) + raise AssertionError() + + def test_request_timeout(self) -> None: + request = self.client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT + + request = self.client._build_request( + FinalRequestOptions(method="get", url="/foo", timeout=httpx.Timeout(100.0)) + ) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(100.0) + + def test_client_timeout_option(self) -> None: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True, timeout=httpx.Timeout(0)) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(0) + + def test_http_client_timeout_option(self) -> None: + # custom timeout given to the httpx client should be used + with httpx.Client(timeout=None) as http_client: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(None) + + # no timeout given to the httpx client should not use the httpx default + with httpx.Client() as http_client: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT + + # explicitly passing the default timeout currently results in it being ignored + with httpx.Client(timeout=HTTPX_DEFAULT_TIMEOUT) as http_client: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT # our default + + async def test_invalid_http_client(self) -> None: + with pytest.raises(TypeError, match="Invalid `http_client` arg"): + async with httpx.AsyncClient() as http_client: + Brapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + http_client=cast(Any, http_client), + ) + + def test_default_headers_option(self) -> None: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} + ) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("x-foo") == "bar" + assert request.headers.get("x-stainless-lang") == "python" + + client2 = Brapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + default_headers={ + "X-Foo": "stainless", + "X-Stainless-Lang": "my-overriding-header", + }, + ) + request = client2._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("x-foo") == "stainless" + assert request.headers.get("x-stainless-lang") == "my-overriding-header" + + def test_validate_headers(self) -> None: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("Authorization") == f"Bearer {api_key}" + + with pytest.raises(BrapiError): + with update_env(**{"BRAPI_API_KEY": Omit()}): + client2 = Brapi(base_url=base_url, api_key=None, _strict_response_validation=True) + _ = client2 + + def test_default_query_option(self) -> None: + client = Brapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"query_param": "bar"} + ) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + url = httpx.URL(request.url) + assert dict(url.params) == {"query_param": "bar"} + + request = client._build_request( + FinalRequestOptions( + method="get", + url="/foo", + params={"foo": "baz", "query_param": "overridden"}, + ) + ) + url = httpx.URL(request.url) + assert dict(url.params) == {"foo": "baz", "query_param": "overridden"} + + def test_request_extra_json(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + extra_json={"baz": False}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"foo": "bar", "baz": False} + + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + extra_json={"baz": False}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"baz": False} + + # `extra_json` takes priority over `json_data` when keys clash + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar", "baz": True}, + extra_json={"baz": None}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"foo": "bar", "baz": None} + + def test_request_extra_headers(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options(extra_headers={"X-Foo": "Foo"}), + ), + ) + assert request.headers.get("X-Foo") == "Foo" + + # `extra_headers` takes priority over `default_headers` when keys clash + request = self.client.with_options(default_headers={"X-Bar": "true"})._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + extra_headers={"X-Bar": "false"}, + ), + ), + ) + assert request.headers.get("X-Bar") == "false" + + def test_request_extra_query(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + extra_query={"my_query_param": "Foo"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"my_query_param": "Foo"} + + # if both `query` and `extra_query` are given, they are merged + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + query={"bar": "1"}, + extra_query={"foo": "2"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"bar": "1", "foo": "2"} + + # `extra_query` takes priority over `query` when keys clash + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + query={"foo": "1"}, + extra_query={"foo": "2"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"foo": "2"} + + def test_multipart_repeating_array(self, client: Brapi) -> None: + request = client._build_request( + FinalRequestOptions.construct( + method="post", + url="/foo", + headers={"Content-Type": "multipart/form-data; boundary=6b7ba517decee4a450543ea6ae821c82"}, + json_data={"array": ["foo", "bar"]}, + files=[("foo.txt", b"hello world")], + ) + ) + + assert request.read().split(b"\r\n") == [ + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="array[]"', + b"", + b"foo", + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="array[]"', + b"", + b"bar", + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="foo.txt"; filename="upload"', + b"Content-Type: application/octet-stream", + b"", + b"hello world", + b"--6b7ba517decee4a450543ea6ae821c82--", + b"", + ] + + @pytest.mark.respx(base_url=base_url) + def test_basic_union_response(self, respx_mock: MockRouter) -> None: + class Model1(BaseModel): + name: str + + class Model2(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model2) + assert response.foo == "bar" + + @pytest.mark.respx(base_url=base_url) + def test_union_response_different_types(self, respx_mock: MockRouter) -> None: + """Union of objects with the same field name using a different type""" + + class Model1(BaseModel): + foo: int + + class Model2(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model2) + assert response.foo == "bar" + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": 1})) + + response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model1) + assert response.foo == 1 + + @pytest.mark.respx(base_url=base_url) + def test_non_application_json_content_type_for_json_data(self, respx_mock: MockRouter) -> None: + """ + Response that sets Content-Type to something other than application/json but returns json data + """ + + class Model(BaseModel): + foo: int + + respx_mock.get("/foo").mock( + return_value=httpx.Response( + 200, + content=json.dumps({"foo": 2}), + headers={"Content-Type": "application/text"}, + ) + ) + + response = self.client.get("/foo", cast_to=Model) + assert isinstance(response, Model) + assert response.foo == 2 + + def test_base_url_setter(self) -> None: + client = Brapi(base_url="https://example.com/from_init", api_key=api_key, _strict_response_validation=True) + assert client.base_url == "https://example.com/from_init/" + + client.base_url = "https://example.com/from_setter" # type: ignore[assignment] + + assert client.base_url == "https://example.com/from_setter/" + + def test_base_url_env(self) -> None: + with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): + client = Brapi(api_key=api_key, _strict_response_validation=True) + assert client.base_url == "http://localhost:5000/from/env/" + + # explicit environment arg requires explicitness + with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): + with pytest.raises(ValueError, match=r"you must pass base_url=None"): + Brapi(api_key=api_key, _strict_response_validation=True, environment="production") + + client = Brapi(base_url=None, api_key=api_key, _strict_response_validation=True, environment="production") + assert str(client.base_url).startswith("https://brapi.dev") + + @pytest.mark.parametrize( + "client", + [ + Brapi(base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True), + Brapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.Client(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_base_url_trailing_slash(self, client: Brapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "http://localhost:5000/custom/path/foo" + + @pytest.mark.parametrize( + "client", + [ + Brapi(base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True), + Brapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.Client(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_base_url_no_trailing_slash(self, client: Brapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "http://localhost:5000/custom/path/foo" + + @pytest.mark.parametrize( + "client", + [ + Brapi(base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True), + Brapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.Client(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_absolute_request_url(self, client: Brapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="https://myapi.com/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "https://myapi.com/foo" + + def test_copied_client_does_not_close_http(self) -> None: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + assert not client.is_closed() + + copied = client.copy() + assert copied is not client + + del copied + + assert not client.is_closed() + + def test_client_context_manager(self) -> None: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + with client as c2: + assert c2 is client + assert not c2.is_closed() + assert not client.is_closed() + assert client.is_closed() + + @pytest.mark.respx(base_url=base_url) + def test_client_response_validation_error(self, respx_mock: MockRouter) -> None: + class Model(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": {"invalid": True}})) + + with pytest.raises(APIResponseValidationError) as exc: + self.client.get("/foo", cast_to=Model) + + assert isinstance(exc.value.__cause__, ValidationError) + + def test_client_max_retries_validation(self) -> None: + with pytest.raises(TypeError, match=r"max_retries cannot be None"): + Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True, max_retries=cast(Any, None)) + + @pytest.mark.respx(base_url=base_url) + def test_received_text_for_expected_json(self, respx_mock: MockRouter) -> None: + class Model(BaseModel): + name: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, text="my-custom-format")) + + strict_client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + with pytest.raises(APIResponseValidationError): + strict_client.get("/foo", cast_to=Model) + + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) + + response = client.get("/foo", cast_to=Model) + assert isinstance(response, str) # type: ignore[unreachable] + + @pytest.mark.parametrize( + "remaining_retries,retry_after,timeout", + [ + [3, "20", 20], + [3, "0", 0.5], + [3, "-10", 0.5], + [3, "60", 60], + [3, "61", 0.5], + [3, "Fri, 29 Sep 2023 16:26:57 GMT", 20], + [3, "Fri, 29 Sep 2023 16:26:37 GMT", 0.5], + [3, "Fri, 29 Sep 2023 16:26:27 GMT", 0.5], + [3, "Fri, 29 Sep 2023 16:27:37 GMT", 60], + [3, "Fri, 29 Sep 2023 16:27:38 GMT", 0.5], + [3, "99999999999999999999999999999999999", 0.5], + [3, "Zun, 29 Sep 2023 16:26:27 GMT", 0.5], + [3, "", 0.5], + [2, "", 0.5 * 2.0], + [1, "", 0.5 * 4.0], + [-1100, "", 8], # test large number potentially overflowing + ], + ) + @mock.patch("time.time", mock.MagicMock(return_value=1696004797)) + def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str, timeout: float) -> None: + client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + headers = httpx.Headers({"retry-after": retry_after}) + options = FinalRequestOptions(method="get", url="/foo", max_retries=3) + calculated = client._calculate_retry_timeout(remaining_retries, options, headers) + assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType] + + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter, client: Brapi) -> None: + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=httpx.TimeoutException("Test timeout error")) + + with pytest.raises(APITimeoutError): + client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__enter__() + + assert _get_open_connections(self.client) == 0 + + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter, client: Brapi) -> None: + respx_mock.get("/api/quote/PETR4,MGLU3").mock(return_value=httpx.Response(500)) + + with pytest.raises(APIStatusError): + client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__enter__() + assert _get_open_connections(self.client) == 0 + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + @pytest.mark.parametrize("failure_mode", ["status", "exception"]) + def test_retries_taken( + self, + client: Brapi, + failures_before_success: int, + failure_mode: Literal["status", "exception"], + respx_mock: MockRouter, + ) -> None: + client = client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + if failure_mode == "exception": + raise RuntimeError("oops") + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = client.quote.with_raw_response.retrieve(tickers="PETR4,MGLU3") + + assert response.retries_taken == failures_before_success + assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + def test_omit_retry_count_header(self, client: Brapi, failures_before_success: int, respx_mock: MockRouter) -> None: + client = client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", extra_headers={"x-stainless-retry-count": Omit()} + ) + + assert len(response.http_request.headers.get_list("x-stainless-retry-count")) == 0 + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + def test_overwrite_retry_count_header( + self, client: Brapi, failures_before_success: int, respx_mock: MockRouter + ) -> None: + client = client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", extra_headers={"x-stainless-retry-count": "42"} + ) + + assert response.http_request.headers.get("x-stainless-retry-count") == "42" + + def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: + # Test that the proxy environment variables are set correctly + monkeypatch.setenv("HTTPS_PROXY", "https://example.org") + + client = DefaultHttpxClient() + + mounts = tuple(client._mounts.items()) + assert len(mounts) == 1 + assert mounts[0][0].pattern == "https://" + + @pytest.mark.filterwarnings("ignore:.*deprecated.*:DeprecationWarning") + def test_default_client_creation(self) -> None: + # Ensure that the client can be initialized without any exceptions + DefaultHttpxClient( + verify=True, + cert=None, + trust_env=True, + http1=True, + http2=False, + limits=httpx.Limits(max_connections=100, max_keepalive_connections=20), + ) + + @pytest.mark.respx(base_url=base_url) + def test_follow_redirects(self, respx_mock: MockRouter) -> None: + # Test that the default follow_redirects=True allows following redirects + respx_mock.post("/redirect").mock( + return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) + ) + respx_mock.get("/redirected").mock(return_value=httpx.Response(200, json={"status": "ok"})) + + response = self.client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) + assert response.status_code == 200 + assert response.json() == {"status": "ok"} + + @pytest.mark.respx(base_url=base_url) + def test_follow_redirects_disabled(self, respx_mock: MockRouter) -> None: + # Test that follow_redirects=False prevents following redirects + respx_mock.post("/redirect").mock( + return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) + ) + + with pytest.raises(APIStatusError) as exc_info: + self.client.post( + "/redirect", body={"key": "value"}, options={"follow_redirects": False}, cast_to=httpx.Response + ) + + assert exc_info.value.response.status_code == 302 + assert exc_info.value.response.headers["Location"] == f"{base_url}/redirected" + + +class TestAsyncBrapi: + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_raw_response(self, respx_mock: MockRouter) -> None: + respx_mock.post("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = await self.client.post("/foo", cast_to=httpx.Response) + assert response.status_code == 200 + assert isinstance(response, httpx.Response) + assert response.json() == {"foo": "bar"} + + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_raw_response_for_binary(self, respx_mock: MockRouter) -> None: + respx_mock.post("/foo").mock( + return_value=httpx.Response(200, headers={"Content-Type": "application/binary"}, content='{"foo": "bar"}') + ) + + response = await self.client.post("/foo", cast_to=httpx.Response) + assert response.status_code == 200 + assert isinstance(response, httpx.Response) + assert response.json() == {"foo": "bar"} + + def test_copy(self) -> None: + copied = self.client.copy() + assert id(copied) != id(self.client) + + copied = self.client.copy(api_key="another My API Key") + assert copied.api_key == "another My API Key" + assert self.client.api_key == "My API Key" + + def test_copy_default_options(self) -> None: + # options that have a default are overridden correctly + copied = self.client.copy(max_retries=7) + assert copied.max_retries == 7 + assert self.client.max_retries == 2 + + copied2 = copied.copy(max_retries=6) + assert copied2.max_retries == 6 + assert copied.max_retries == 7 + + # timeout + assert isinstance(self.client.timeout, httpx.Timeout) + copied = self.client.copy(timeout=None) + assert copied.timeout is None + assert isinstance(self.client.timeout, httpx.Timeout) + + def test_copy_default_headers(self) -> None: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} + ) + assert client.default_headers["X-Foo"] == "bar" + + # does not override the already given value when not specified + copied = client.copy() + assert copied.default_headers["X-Foo"] == "bar" + + # merges already given headers + copied = client.copy(default_headers={"X-Bar": "stainless"}) + assert copied.default_headers["X-Foo"] == "bar" + assert copied.default_headers["X-Bar"] == "stainless" + + # uses new values for any already given headers + copied = client.copy(default_headers={"X-Foo": "stainless"}) + assert copied.default_headers["X-Foo"] == "stainless" + + # set_default_headers + + # completely overrides already set values + copied = client.copy(set_default_headers={}) + assert copied.default_headers.get("X-Foo") is None + + copied = client.copy(set_default_headers={"X-Bar": "Robert"}) + assert copied.default_headers["X-Bar"] == "Robert" + + with pytest.raises( + ValueError, + match="`default_headers` and `set_default_headers` arguments are mutually exclusive", + ): + client.copy(set_default_headers={}, default_headers={"X-Foo": "Bar"}) + + def test_copy_default_query(self) -> None: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"foo": "bar"} + ) + assert _get_params(client)["foo"] == "bar" + + # does not override the already given value when not specified + copied = client.copy() + assert _get_params(copied)["foo"] == "bar" + + # merges already given params + copied = client.copy(default_query={"bar": "stainless"}) + params = _get_params(copied) + assert params["foo"] == "bar" + assert params["bar"] == "stainless" + + # uses new values for any already given headers + copied = client.copy(default_query={"foo": "stainless"}) + assert _get_params(copied)["foo"] == "stainless" + + # set_default_query + + # completely overrides already set values + copied = client.copy(set_default_query={}) + assert _get_params(copied) == {} + + copied = client.copy(set_default_query={"bar": "Robert"}) + assert _get_params(copied)["bar"] == "Robert" + + with pytest.raises( + ValueError, + # TODO: update + match="`default_query` and `set_default_query` arguments are mutually exclusive", + ): + client.copy(set_default_query={}, default_query={"foo": "Bar"}) + + def test_copy_signature(self) -> None: + # ensure the same parameters that can be passed to the client are defined in the `.copy()` method + init_signature = inspect.signature( + # mypy doesn't like that we access the `__init__` property. + self.client.__init__, # type: ignore[misc] + ) + copy_signature = inspect.signature(self.client.copy) + exclude_params = {"transport", "proxies", "_strict_response_validation"} + + for name in init_signature.parameters.keys(): + if name in exclude_params: + continue + + copy_param = copy_signature.parameters.get(name) + assert copy_param is not None, f"copy() signature is missing the {name} param" + + @pytest.mark.skipif(sys.version_info >= (3, 10), reason="fails because of a memory leak that started from 3.12") + def test_copy_build_request(self) -> None: + options = FinalRequestOptions(method="get", url="/foo") + + def build_request(options: FinalRequestOptions) -> None: + client = self.client.copy() + client._build_request(options) + + # ensure that the machinery is warmed up before tracing starts. + build_request(options) + gc.collect() + + tracemalloc.start(1000) + + snapshot_before = tracemalloc.take_snapshot() + + ITERATIONS = 10 + for _ in range(ITERATIONS): + build_request(options) + + gc.collect() + snapshot_after = tracemalloc.take_snapshot() + + tracemalloc.stop() + + def add_leak(leaks: list[tracemalloc.StatisticDiff], diff: tracemalloc.StatisticDiff) -> None: + if diff.count == 0: + # Avoid false positives by considering only leaks (i.e. allocations that persist). + return + + if diff.count % ITERATIONS != 0: + # Avoid false positives by considering only leaks that appear per iteration. + return + + for frame in diff.traceback: + if any( + frame.filename.endswith(fragment) + for fragment in [ + # to_raw_response_wrapper leaks through the @functools.wraps() decorator. + # + # removing the decorator fixes the leak for reasons we don't understand. + "brapi/_legacy_response.py", + "brapi/_response.py", + # pydantic.BaseModel.model_dump || pydantic.BaseModel.dict leak memory for some reason. + "brapi/_compat.py", + # Standard library leaks we don't care about. + "/logging/__init__.py", + ] + ): + return + + leaks.append(diff) + + leaks: list[tracemalloc.StatisticDiff] = [] + for diff in snapshot_after.compare_to(snapshot_before, "traceback"): + add_leak(leaks, diff) + if leaks: + for leak in leaks: + print("MEMORY LEAK:", leak) + for frame in leak.traceback: + print(frame) + raise AssertionError() + + async def test_request_timeout(self) -> None: + request = self.client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT + + request = self.client._build_request( + FinalRequestOptions(method="get", url="/foo", timeout=httpx.Timeout(100.0)) + ) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(100.0) + + async def test_client_timeout_option(self) -> None: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, timeout=httpx.Timeout(0) + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(0) + + async def test_http_client_timeout_option(self) -> None: + # custom timeout given to the httpx client should be used + async with httpx.AsyncClient(timeout=None) as http_client: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == httpx.Timeout(None) + + # no timeout given to the httpx client should not use the httpx default + async with httpx.AsyncClient() as http_client: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT + + # explicitly passing the default timeout currently results in it being ignored + async with httpx.AsyncClient(timeout=HTTPX_DEFAULT_TIMEOUT) as http_client: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, http_client=http_client + ) + + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore + assert timeout == DEFAULT_TIMEOUT # our default + + def test_invalid_http_client(self) -> None: + with pytest.raises(TypeError, match="Invalid `http_client` arg"): + with httpx.Client() as http_client: + AsyncBrapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + http_client=cast(Any, http_client), + ) + + def test_default_headers_option(self) -> None: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} + ) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("x-foo") == "bar" + assert request.headers.get("x-stainless-lang") == "python" + + client2 = AsyncBrapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + default_headers={ + "X-Foo": "stainless", + "X-Stainless-Lang": "my-overriding-header", + }, + ) + request = client2._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("x-foo") == "stainless" + assert request.headers.get("x-stainless-lang") == "my-overriding-header" + + def test_validate_headers(self) -> None: + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + assert request.headers.get("Authorization") == f"Bearer {api_key}" + + with pytest.raises(BrapiError): + with update_env(**{"BRAPI_API_KEY": Omit()}): + client2 = AsyncBrapi(base_url=base_url, api_key=None, _strict_response_validation=True) + _ = client2 + + def test_default_query_option(self) -> None: + client = AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"query_param": "bar"} + ) + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + url = httpx.URL(request.url) + assert dict(url.params) == {"query_param": "bar"} + + request = client._build_request( + FinalRequestOptions( + method="get", + url="/foo", + params={"foo": "baz", "query_param": "overridden"}, + ) + ) + url = httpx.URL(request.url) + assert dict(url.params) == {"foo": "baz", "query_param": "overridden"} + + def test_request_extra_json(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + extra_json={"baz": False}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"foo": "bar", "baz": False} + + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + extra_json={"baz": False}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"baz": False} + + # `extra_json` takes priority over `json_data` when keys clash + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar", "baz": True}, + extra_json={"baz": None}, + ), + ) + data = json.loads(request.content.decode("utf-8")) + assert data == {"foo": "bar", "baz": None} + + def test_request_extra_headers(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options(extra_headers={"X-Foo": "Foo"}), + ), + ) + assert request.headers.get("X-Foo") == "Foo" + + # `extra_headers` takes priority over `default_headers` when keys clash + request = self.client.with_options(default_headers={"X-Bar": "true"})._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + extra_headers={"X-Bar": "false"}, + ), + ), + ) + assert request.headers.get("X-Bar") == "false" + + def test_request_extra_query(self) -> None: + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + extra_query={"my_query_param": "Foo"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"my_query_param": "Foo"} + + # if both `query` and `extra_query` are given, they are merged + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + query={"bar": "1"}, + extra_query={"foo": "2"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"bar": "1", "foo": "2"} + + # `extra_query` takes priority over `query` when keys clash + request = self.client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + **make_request_options( + query={"foo": "1"}, + extra_query={"foo": "2"}, + ), + ), + ) + params = dict(request.url.params) + assert params == {"foo": "2"} + + def test_multipart_repeating_array(self, async_client: AsyncBrapi) -> None: + request = async_client._build_request( + FinalRequestOptions.construct( + method="post", + url="/foo", + headers={"Content-Type": "multipart/form-data; boundary=6b7ba517decee4a450543ea6ae821c82"}, + json_data={"array": ["foo", "bar"]}, + files=[("foo.txt", b"hello world")], + ) + ) + + assert request.read().split(b"\r\n") == [ + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="array[]"', + b"", + b"foo", + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="array[]"', + b"", + b"bar", + b"--6b7ba517decee4a450543ea6ae821c82", + b'Content-Disposition: form-data; name="foo.txt"; filename="upload"', + b"Content-Type: application/octet-stream", + b"", + b"hello world", + b"--6b7ba517decee4a450543ea6ae821c82--", + b"", + ] + + @pytest.mark.respx(base_url=base_url) + async def test_basic_union_response(self, respx_mock: MockRouter) -> None: + class Model1(BaseModel): + name: str + + class Model2(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model2) + assert response.foo == "bar" + + @pytest.mark.respx(base_url=base_url) + async def test_union_response_different_types(self, respx_mock: MockRouter) -> None: + """Union of objects with the same field name using a different type""" + + class Model1(BaseModel): + foo: int + + class Model2(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) + + response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model2) + assert response.foo == "bar" + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": 1})) + + response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + assert isinstance(response, Model1) + assert response.foo == 1 + + @pytest.mark.respx(base_url=base_url) + async def test_non_application_json_content_type_for_json_data(self, respx_mock: MockRouter) -> None: + """ + Response that sets Content-Type to something other than application/json but returns json data + """ + + class Model(BaseModel): + foo: int + + respx_mock.get("/foo").mock( + return_value=httpx.Response( + 200, + content=json.dumps({"foo": 2}), + headers={"Content-Type": "application/text"}, + ) + ) + + response = await self.client.get("/foo", cast_to=Model) + assert isinstance(response, Model) + assert response.foo == 2 + + def test_base_url_setter(self) -> None: + client = AsyncBrapi(base_url="https://example.com/from_init", api_key=api_key, _strict_response_validation=True) + assert client.base_url == "https://example.com/from_init/" + + client.base_url = "https://example.com/from_setter" # type: ignore[assignment] + + assert client.base_url == "https://example.com/from_setter/" + + def test_base_url_env(self) -> None: + with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): + client = AsyncBrapi(api_key=api_key, _strict_response_validation=True) + assert client.base_url == "http://localhost:5000/from/env/" + + # explicit environment arg requires explicitness + with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): + with pytest.raises(ValueError, match=r"you must pass base_url=None"): + AsyncBrapi(api_key=api_key, _strict_response_validation=True, environment="production") + + client = AsyncBrapi( + base_url=None, api_key=api_key, _strict_response_validation=True, environment="production" + ) + assert str(client.base_url).startswith("https://brapi.dev") + + @pytest.mark.parametrize( + "client", + [ + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True + ), + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.AsyncClient(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_base_url_trailing_slash(self, client: AsyncBrapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "http://localhost:5000/custom/path/foo" + + @pytest.mark.parametrize( + "client", + [ + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True + ), + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.AsyncClient(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_base_url_no_trailing_slash(self, client: AsyncBrapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "http://localhost:5000/custom/path/foo" + + @pytest.mark.parametrize( + "client", + [ + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", api_key=api_key, _strict_response_validation=True + ), + AsyncBrapi( + base_url="http://localhost:5000/custom/path/", + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.AsyncClient(), + ), + ], + ids=["standard", "custom http client"], + ) + def test_absolute_request_url(self, client: AsyncBrapi) -> None: + request = client._build_request( + FinalRequestOptions( + method="post", + url="https://myapi.com/foo", + json_data={"foo": "bar"}, + ), + ) + assert request.url == "https://myapi.com/foo" + + async def test_copied_client_does_not_close_http(self) -> None: + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + assert not client.is_closed() + + copied = client.copy() + assert copied is not client + + del copied + + await asyncio.sleep(0.2) + assert not client.is_closed() + + async def test_client_context_manager(self) -> None: + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + async with client as c2: + assert c2 is client + assert not c2.is_closed() + assert not client.is_closed() + assert client.is_closed() + + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_client_response_validation_error(self, respx_mock: MockRouter) -> None: + class Model(BaseModel): + foo: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": {"invalid": True}})) + + with pytest.raises(APIResponseValidationError) as exc: + await self.client.get("/foo", cast_to=Model) + + assert isinstance(exc.value.__cause__, ValidationError) + + async def test_client_max_retries_validation(self) -> None: + with pytest.raises(TypeError, match=r"max_retries cannot be None"): + AsyncBrapi( + base_url=base_url, api_key=api_key, _strict_response_validation=True, max_retries=cast(Any, None) + ) + + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_received_text_for_expected_json(self, respx_mock: MockRouter) -> None: + class Model(BaseModel): + name: str + + respx_mock.get("/foo").mock(return_value=httpx.Response(200, text="my-custom-format")) + + strict_client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + with pytest.raises(APIResponseValidationError): + await strict_client.get("/foo", cast_to=Model) + + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) + + response = await client.get("/foo", cast_to=Model) + assert isinstance(response, str) # type: ignore[unreachable] + + @pytest.mark.parametrize( + "remaining_retries,retry_after,timeout", + [ + [3, "20", 20], + [3, "0", 0.5], + [3, "-10", 0.5], + [3, "60", 60], + [3, "61", 0.5], + [3, "Fri, 29 Sep 2023 16:26:57 GMT", 20], + [3, "Fri, 29 Sep 2023 16:26:37 GMT", 0.5], + [3, "Fri, 29 Sep 2023 16:26:27 GMT", 0.5], + [3, "Fri, 29 Sep 2023 16:27:37 GMT", 60], + [3, "Fri, 29 Sep 2023 16:27:38 GMT", 0.5], + [3, "99999999999999999999999999999999999", 0.5], + [3, "Zun, 29 Sep 2023 16:26:27 GMT", 0.5], + [3, "", 0.5], + [2, "", 0.5 * 2.0], + [1, "", 0.5 * 4.0], + [-1100, "", 8], # test large number potentially overflowing + ], + ) + @mock.patch("time.time", mock.MagicMock(return_value=1696004797)) + @pytest.mark.asyncio + async def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str, timeout: float) -> None: + client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + + headers = httpx.Headers({"retry-after": retry_after}) + options = FinalRequestOptions(method="get", url="/foo", max_retries=3) + calculated = client._calculate_retry_timeout(remaining_retries, options, headers) + assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType] + + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=httpx.TimeoutException("Test timeout error")) + + with pytest.raises(APITimeoutError): + await async_client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__aenter__() + + assert _get_open_connections(self.client) == 0 + + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + async def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: + respx_mock.get("/api/quote/PETR4,MGLU3").mock(return_value=httpx.Response(500)) + + with pytest.raises(APIStatusError): + await async_client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__aenter__() + assert _get_open_connections(self.client) == 0 + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + @pytest.mark.parametrize("failure_mode", ["status", "exception"]) + async def test_retries_taken( + self, + async_client: AsyncBrapi, + failures_before_success: int, + failure_mode: Literal["status", "exception"], + respx_mock: MockRouter, + ) -> None: + client = async_client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + if failure_mode == "exception": + raise RuntimeError("oops") + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = await client.quote.with_raw_response.retrieve(tickers="PETR4,MGLU3") + + assert response.retries_taken == failures_before_success + assert int(response.http_request.headers.get("x-stainless-retry-count")) == failures_before_success + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_omit_retry_count_header( + self, async_client: AsyncBrapi, failures_before_success: int, respx_mock: MockRouter + ) -> None: + client = async_client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = await client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", extra_headers={"x-stainless-retry-count": Omit()} + ) + + assert len(response.http_request.headers.get_list("x-stainless-retry-count")) == 0 + + @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) + @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) + @pytest.mark.respx(base_url=base_url) + @pytest.mark.asyncio + async def test_overwrite_retry_count_header( + self, async_client: AsyncBrapi, failures_before_success: int, respx_mock: MockRouter + ) -> None: + client = async_client.with_options(max_retries=4) + + nb_retries = 0 + + def retry_handler(_request: httpx.Request) -> httpx.Response: + nonlocal nb_retries + if nb_retries < failures_before_success: + nb_retries += 1 + return httpx.Response(500) + return httpx.Response(200) + + respx_mock.get("/api/quote/PETR4,MGLU3").mock(side_effect=retry_handler) + + response = await client.quote.with_raw_response.retrieve( + tickers="PETR4,MGLU3", extra_headers={"x-stainless-retry-count": "42"} + ) + + assert response.http_request.headers.get("x-stainless-retry-count") == "42" + + async def test_get_platform(self) -> None: + platform = await asyncify(get_platform)() + assert isinstance(platform, (str, OtherPlatform)) + + async def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: + # Test that the proxy environment variables are set correctly + monkeypatch.setenv("HTTPS_PROXY", "https://example.org") + + client = DefaultAsyncHttpxClient() + + mounts = tuple(client._mounts.items()) + assert len(mounts) == 1 + assert mounts[0][0].pattern == "https://" + + @pytest.mark.filterwarnings("ignore:.*deprecated.*:DeprecationWarning") + async def test_default_client_creation(self) -> None: + # Ensure that the client can be initialized without any exceptions + DefaultAsyncHttpxClient( + verify=True, + cert=None, + trust_env=True, + http1=True, + http2=False, + limits=httpx.Limits(max_connections=100, max_keepalive_connections=20), + ) + + @pytest.mark.respx(base_url=base_url) + async def test_follow_redirects(self, respx_mock: MockRouter) -> None: + # Test that the default follow_redirects=True allows following redirects + respx_mock.post("/redirect").mock( + return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) + ) + respx_mock.get("/redirected").mock(return_value=httpx.Response(200, json={"status": "ok"})) + + response = await self.client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) + assert response.status_code == 200 + assert response.json() == {"status": "ok"} + + @pytest.mark.respx(base_url=base_url) + async def test_follow_redirects_disabled(self, respx_mock: MockRouter) -> None: + # Test that follow_redirects=False prevents following redirects + respx_mock.post("/redirect").mock( + return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) + ) + + with pytest.raises(APIStatusError) as exc_info: + await self.client.post( + "/redirect", body={"key": "value"}, options={"follow_redirects": False}, cast_to=httpx.Response + ) + + assert exc_info.value.response.status_code == 302 + assert exc_info.value.response.headers["Location"] == f"{base_url}/redirected" diff --git a/tests/test_deepcopy.py b/tests/test_deepcopy.py new file mode 100644 index 0000000..f5639fb --- /dev/null +++ b/tests/test_deepcopy.py @@ -0,0 +1,58 @@ +from brapi._utils import deepcopy_minimal + + +def assert_different_identities(obj1: object, obj2: object) -> None: + assert obj1 == obj2 + assert id(obj1) != id(obj2) + + +def test_simple_dict() -> None: + obj1 = {"foo": "bar"} + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + + +def test_nested_dict() -> None: + obj1 = {"foo": {"bar": True}} + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + assert_different_identities(obj1["foo"], obj2["foo"]) + + +def test_complex_nested_dict() -> None: + obj1 = {"foo": {"bar": [{"hello": "world"}]}} + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + assert_different_identities(obj1["foo"], obj2["foo"]) + assert_different_identities(obj1["foo"]["bar"], obj2["foo"]["bar"]) + assert_different_identities(obj1["foo"]["bar"][0], obj2["foo"]["bar"][0]) + + +def test_simple_list() -> None: + obj1 = ["a", "b", "c"] + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + + +def test_nested_list() -> None: + obj1 = ["a", [1, 2, 3]] + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + assert_different_identities(obj1[1], obj2[1]) + + +class MyObject: ... + + +def test_ignores_other_types() -> None: + # custom classes + my_obj = MyObject() + obj1 = {"foo": my_obj} + obj2 = deepcopy_minimal(obj1) + assert_different_identities(obj1, obj2) + assert obj1["foo"] is my_obj + + # tuples + obj3 = ("a", "b") + obj4 = deepcopy_minimal(obj3) + assert obj3 is obj4 diff --git a/tests/test_extract_files.py b/tests/test_extract_files.py new file mode 100644 index 0000000..9b6f362 --- /dev/null +++ b/tests/test_extract_files.py @@ -0,0 +1,64 @@ +from __future__ import annotations + +from typing import Sequence + +import pytest + +from brapi._types import FileTypes +from brapi._utils import extract_files + + +def test_removes_files_from_input() -> None: + query = {"foo": "bar"} + assert extract_files(query, paths=[]) == [] + assert query == {"foo": "bar"} + + query2 = {"foo": b"Bar", "hello": "world"} + assert extract_files(query2, paths=[["foo"]]) == [("foo", b"Bar")] + assert query2 == {"hello": "world"} + + query3 = {"foo": {"foo": {"bar": b"Bar"}}, "hello": "world"} + assert extract_files(query3, paths=[["foo", "foo", "bar"]]) == [("foo[foo][bar]", b"Bar")] + assert query3 == {"foo": {"foo": {}}, "hello": "world"} + + query4 = {"foo": {"bar": b"Bar", "baz": "foo"}, "hello": "world"} + assert extract_files(query4, paths=[["foo", "bar"]]) == [("foo[bar]", b"Bar")] + assert query4 == {"hello": "world", "foo": {"baz": "foo"}} + + +def test_multiple_files() -> None: + query = {"documents": [{"file": b"My first file"}, {"file": b"My second file"}]} + assert extract_files(query, paths=[["documents", "", "file"]]) == [ + ("documents[][file]", b"My first file"), + ("documents[][file]", b"My second file"), + ] + assert query == {"documents": [{}, {}]} + + +@pytest.mark.parametrize( + "query,paths,expected", + [ + [ + {"foo": {"bar": "baz"}}, + [["foo", "", "bar"]], + [], + ], + [ + {"foo": ["bar", "baz"]}, + [["foo", "bar"]], + [], + ], + [ + {"foo": {"bar": "baz"}}, + [["foo", "foo"]], + [], + ], + ], + ids=["dict expecting array", "array expecting dict", "unknown keys"], +) +def test_ignores_incorrect_paths( + query: dict[str, object], + paths: Sequence[Sequence[str]], + expected: list[tuple[str, FileTypes]], +) -> None: + assert extract_files(query, paths=paths) == expected diff --git a/tests/test_files.py b/tests/test_files.py new file mode 100644 index 0000000..bdb3673 --- /dev/null +++ b/tests/test_files.py @@ -0,0 +1,51 @@ +from pathlib import Path + +import anyio +import pytest +from dirty_equals import IsDict, IsList, IsBytes, IsTuple + +from brapi._files import to_httpx_files, async_to_httpx_files + +readme_path = Path(__file__).parent.parent.joinpath("README.md") + + +def test_pathlib_includes_file_name() -> None: + result = to_httpx_files({"file": readme_path}) + print(result) + assert result == IsDict({"file": IsTuple("README.md", IsBytes())}) + + +def test_tuple_input() -> None: + result = to_httpx_files([("file", readme_path)]) + print(result) + assert result == IsList(IsTuple("file", IsTuple("README.md", IsBytes()))) + + +@pytest.mark.asyncio +async def test_async_pathlib_includes_file_name() -> None: + result = await async_to_httpx_files({"file": readme_path}) + print(result) + assert result == IsDict({"file": IsTuple("README.md", IsBytes())}) + + +@pytest.mark.asyncio +async def test_async_supports_anyio_path() -> None: + result = await async_to_httpx_files({"file": anyio.Path(readme_path)}) + print(result) + assert result == IsDict({"file": IsTuple("README.md", IsBytes())}) + + +@pytest.mark.asyncio +async def test_async_tuple_input() -> None: + result = await async_to_httpx_files([("file", readme_path)]) + print(result) + assert result == IsList(IsTuple("file", IsTuple("README.md", IsBytes()))) + + +def test_string_not_allowed() -> None: + with pytest.raises(TypeError, match="Expected file types input to be a FileContent type or to be a tuple"): + to_httpx_files( + { + "file": "foo", # type: ignore + } + ) diff --git a/tests/test_models.py b/tests/test_models.py new file mode 100644 index 0000000..cb150da --- /dev/null +++ b/tests/test_models.py @@ -0,0 +1,963 @@ +import json +from typing import TYPE_CHECKING, Any, Dict, List, Union, Optional, cast +from datetime import datetime, timezone +from typing_extensions import Literal, Annotated, TypeAliasType + +import pytest +import pydantic +from pydantic import Field + +from brapi._utils import PropertyInfo +from brapi._compat import PYDANTIC_V1, parse_obj, model_dump, model_json +from brapi._models import BaseModel, construct_type + + +class BasicModel(BaseModel): + foo: str + + +@pytest.mark.parametrize("value", ["hello", 1], ids=["correct type", "mismatched"]) +def test_basic(value: object) -> None: + m = BasicModel.construct(foo=value) + assert m.foo == value + + +def test_directly_nested_model() -> None: + class NestedModel(BaseModel): + nested: BasicModel + + m = NestedModel.construct(nested={"foo": "Foo!"}) + assert m.nested.foo == "Foo!" + + # mismatched types + m = NestedModel.construct(nested="hello!") + assert cast(Any, m.nested) == "hello!" + + +def test_optional_nested_model() -> None: + class NestedModel(BaseModel): + nested: Optional[BasicModel] + + m1 = NestedModel.construct(nested=None) + assert m1.nested is None + + m2 = NestedModel.construct(nested={"foo": "bar"}) + assert m2.nested is not None + assert m2.nested.foo == "bar" + + # mismatched types + m3 = NestedModel.construct(nested={"foo"}) + assert isinstance(cast(Any, m3.nested), set) + assert cast(Any, m3.nested) == {"foo"} + + +def test_list_nested_model() -> None: + class NestedModel(BaseModel): + nested: List[BasicModel] + + m = NestedModel.construct(nested=[{"foo": "bar"}, {"foo": "2"}]) + assert m.nested is not None + assert isinstance(m.nested, list) + assert len(m.nested) == 2 + assert m.nested[0].foo == "bar" + assert m.nested[1].foo == "2" + + # mismatched types + m = NestedModel.construct(nested=True) + assert cast(Any, m.nested) is True + + m = NestedModel.construct(nested=[False]) + assert cast(Any, m.nested) == [False] + + +def test_optional_list_nested_model() -> None: + class NestedModel(BaseModel): + nested: Optional[List[BasicModel]] + + m1 = NestedModel.construct(nested=[{"foo": "bar"}, {"foo": "2"}]) + assert m1.nested is not None + assert isinstance(m1.nested, list) + assert len(m1.nested) == 2 + assert m1.nested[0].foo == "bar" + assert m1.nested[1].foo == "2" + + m2 = NestedModel.construct(nested=None) + assert m2.nested is None + + # mismatched types + m3 = NestedModel.construct(nested={1}) + assert cast(Any, m3.nested) == {1} + + m4 = NestedModel.construct(nested=[False]) + assert cast(Any, m4.nested) == [False] + + +def test_list_optional_items_nested_model() -> None: + class NestedModel(BaseModel): + nested: List[Optional[BasicModel]] + + m = NestedModel.construct(nested=[None, {"foo": "bar"}]) + assert m.nested is not None + assert isinstance(m.nested, list) + assert len(m.nested) == 2 + assert m.nested[0] is None + assert m.nested[1] is not None + assert m.nested[1].foo == "bar" + + # mismatched types + m3 = NestedModel.construct(nested="foo") + assert cast(Any, m3.nested) == "foo" + + m4 = NestedModel.construct(nested=[False]) + assert cast(Any, m4.nested) == [False] + + +def test_list_mismatched_type() -> None: + class NestedModel(BaseModel): + nested: List[str] + + m = NestedModel.construct(nested=False) + assert cast(Any, m.nested) is False + + +def test_raw_dictionary() -> None: + class NestedModel(BaseModel): + nested: Dict[str, str] + + m = NestedModel.construct(nested={"hello": "world"}) + assert m.nested == {"hello": "world"} + + # mismatched types + m = NestedModel.construct(nested=False) + assert cast(Any, m.nested) is False + + +def test_nested_dictionary_model() -> None: + class NestedModel(BaseModel): + nested: Dict[str, BasicModel] + + m = NestedModel.construct(nested={"hello": {"foo": "bar"}}) + assert isinstance(m.nested, dict) + assert m.nested["hello"].foo == "bar" + + # mismatched types + m = NestedModel.construct(nested={"hello": False}) + assert cast(Any, m.nested["hello"]) is False + + +def test_unknown_fields() -> None: + m1 = BasicModel.construct(foo="foo", unknown=1) + assert m1.foo == "foo" + assert cast(Any, m1).unknown == 1 + + m2 = BasicModel.construct(foo="foo", unknown={"foo_bar": True}) + assert m2.foo == "foo" + assert cast(Any, m2).unknown == {"foo_bar": True} + + assert model_dump(m2) == {"foo": "foo", "unknown": {"foo_bar": True}} + + +def test_strict_validation_unknown_fields() -> None: + class Model(BaseModel): + foo: str + + model = parse_obj(Model, dict(foo="hello!", user="Robert")) + assert model.foo == "hello!" + assert cast(Any, model).user == "Robert" + + assert model_dump(model) == {"foo": "hello!", "user": "Robert"} + + +def test_aliases() -> None: + class Model(BaseModel): + my_field: int = Field(alias="myField") + + m = Model.construct(myField=1) + assert m.my_field == 1 + + # mismatched types + m = Model.construct(myField={"hello": False}) + assert cast(Any, m.my_field) == {"hello": False} + + +def test_repr() -> None: + model = BasicModel(foo="bar") + assert str(model) == "BasicModel(foo='bar')" + assert repr(model) == "BasicModel(foo='bar')" + + +def test_repr_nested_model() -> None: + class Child(BaseModel): + name: str + age: int + + class Parent(BaseModel): + name: str + child: Child + + model = Parent(name="Robert", child=Child(name="Foo", age=5)) + assert str(model) == "Parent(name='Robert', child=Child(name='Foo', age=5))" + assert repr(model) == "Parent(name='Robert', child=Child(name='Foo', age=5))" + + +def test_optional_list() -> None: + class Submodel(BaseModel): + name: str + + class Model(BaseModel): + items: Optional[List[Submodel]] + + m = Model.construct(items=None) + assert m.items is None + + m = Model.construct(items=[]) + assert m.items == [] + + m = Model.construct(items=[{"name": "Robert"}]) + assert m.items is not None + assert len(m.items) == 1 + assert m.items[0].name == "Robert" + + +def test_nested_union_of_models() -> None: + class Submodel1(BaseModel): + bar: bool + + class Submodel2(BaseModel): + thing: str + + class Model(BaseModel): + foo: Union[Submodel1, Submodel2] + + m = Model.construct(foo={"thing": "hello"}) + assert isinstance(m.foo, Submodel2) + assert m.foo.thing == "hello" + + +def test_nested_union_of_mixed_types() -> None: + class Submodel1(BaseModel): + bar: bool + + class Model(BaseModel): + foo: Union[Submodel1, Literal[True], Literal["CARD_HOLDER"]] + + m = Model.construct(foo=True) + assert m.foo is True + + m = Model.construct(foo="CARD_HOLDER") + assert m.foo == "CARD_HOLDER" + + m = Model.construct(foo={"bar": False}) + assert isinstance(m.foo, Submodel1) + assert m.foo.bar is False + + +def test_nested_union_multiple_variants() -> None: + class Submodel1(BaseModel): + bar: bool + + class Submodel2(BaseModel): + thing: str + + class Submodel3(BaseModel): + foo: int + + class Model(BaseModel): + foo: Union[Submodel1, Submodel2, None, Submodel3] + + m = Model.construct(foo={"thing": "hello"}) + assert isinstance(m.foo, Submodel2) + assert m.foo.thing == "hello" + + m = Model.construct(foo=None) + assert m.foo is None + + m = Model.construct() + assert m.foo is None + + m = Model.construct(foo={"foo": "1"}) + assert isinstance(m.foo, Submodel3) + assert m.foo.foo == 1 + + +def test_nested_union_invalid_data() -> None: + class Submodel1(BaseModel): + level: int + + class Submodel2(BaseModel): + name: str + + class Model(BaseModel): + foo: Union[Submodel1, Submodel2] + + m = Model.construct(foo=True) + assert cast(bool, m.foo) is True + + m = Model.construct(foo={"name": 3}) + if PYDANTIC_V1: + assert isinstance(m.foo, Submodel2) + assert m.foo.name == "3" + else: + assert isinstance(m.foo, Submodel1) + assert m.foo.name == 3 # type: ignore + + +def test_list_of_unions() -> None: + class Submodel1(BaseModel): + level: int + + class Submodel2(BaseModel): + name: str + + class Model(BaseModel): + items: List[Union[Submodel1, Submodel2]] + + m = Model.construct(items=[{"level": 1}, {"name": "Robert"}]) + assert len(m.items) == 2 + assert isinstance(m.items[0], Submodel1) + assert m.items[0].level == 1 + assert isinstance(m.items[1], Submodel2) + assert m.items[1].name == "Robert" + + m = Model.construct(items=[{"level": -1}, 156]) + assert len(m.items) == 2 + assert isinstance(m.items[0], Submodel1) + assert m.items[0].level == -1 + assert cast(Any, m.items[1]) == 156 + + +def test_union_of_lists() -> None: + class SubModel1(BaseModel): + level: int + + class SubModel2(BaseModel): + name: str + + class Model(BaseModel): + items: Union[List[SubModel1], List[SubModel2]] + + # with one valid entry + m = Model.construct(items=[{"name": "Robert"}]) + assert len(m.items) == 1 + assert isinstance(m.items[0], SubModel2) + assert m.items[0].name == "Robert" + + # with two entries pointing to different types + m = Model.construct(items=[{"level": 1}, {"name": "Robert"}]) + assert len(m.items) == 2 + assert isinstance(m.items[0], SubModel1) + assert m.items[0].level == 1 + assert isinstance(m.items[1], SubModel1) + assert cast(Any, m.items[1]).name == "Robert" + + # with two entries pointing to *completely* different types + m = Model.construct(items=[{"level": -1}, 156]) + assert len(m.items) == 2 + assert isinstance(m.items[0], SubModel1) + assert m.items[0].level == -1 + assert cast(Any, m.items[1]) == 156 + + +def test_dict_of_union() -> None: + class SubModel1(BaseModel): + name: str + + class SubModel2(BaseModel): + foo: str + + class Model(BaseModel): + data: Dict[str, Union[SubModel1, SubModel2]] + + m = Model.construct(data={"hello": {"name": "there"}, "foo": {"foo": "bar"}}) + assert len(list(m.data.keys())) == 2 + assert isinstance(m.data["hello"], SubModel1) + assert m.data["hello"].name == "there" + assert isinstance(m.data["foo"], SubModel2) + assert m.data["foo"].foo == "bar" + + # TODO: test mismatched type + + +def test_double_nested_union() -> None: + class SubModel1(BaseModel): + name: str + + class SubModel2(BaseModel): + bar: str + + class Model(BaseModel): + data: Dict[str, List[Union[SubModel1, SubModel2]]] + + m = Model.construct(data={"foo": [{"bar": "baz"}, {"name": "Robert"}]}) + assert len(m.data["foo"]) == 2 + + entry1 = m.data["foo"][0] + assert isinstance(entry1, SubModel2) + assert entry1.bar == "baz" + + entry2 = m.data["foo"][1] + assert isinstance(entry2, SubModel1) + assert entry2.name == "Robert" + + # TODO: test mismatched type + + +def test_union_of_dict() -> None: + class SubModel1(BaseModel): + name: str + + class SubModel2(BaseModel): + foo: str + + class Model(BaseModel): + data: Union[Dict[str, SubModel1], Dict[str, SubModel2]] + + m = Model.construct(data={"hello": {"name": "there"}, "foo": {"foo": "bar"}}) + assert len(list(m.data.keys())) == 2 + assert isinstance(m.data["hello"], SubModel1) + assert m.data["hello"].name == "there" + assert isinstance(m.data["foo"], SubModel1) + assert cast(Any, m.data["foo"]).foo == "bar" + + +def test_iso8601_datetime() -> None: + class Model(BaseModel): + created_at: datetime + + expected = datetime(2019, 12, 27, 18, 11, 19, 117000, tzinfo=timezone.utc) + + if PYDANTIC_V1: + expected_json = '{"created_at": "2019-12-27T18:11:19.117000+00:00"}' + else: + expected_json = '{"created_at":"2019-12-27T18:11:19.117000Z"}' + + model = Model.construct(created_at="2019-12-27T18:11:19.117Z") + assert model.created_at == expected + assert model_json(model) == expected_json + + model = parse_obj(Model, dict(created_at="2019-12-27T18:11:19.117Z")) + assert model.created_at == expected + assert model_json(model) == expected_json + + +def test_does_not_coerce_int() -> None: + class Model(BaseModel): + bar: int + + assert Model.construct(bar=1).bar == 1 + assert Model.construct(bar=10.9).bar == 10.9 + assert Model.construct(bar="19").bar == "19" # type: ignore[comparison-overlap] + assert Model.construct(bar=False).bar is False + + +def test_int_to_float_safe_conversion() -> None: + class Model(BaseModel): + float_field: float + + m = Model.construct(float_field=10) + assert m.float_field == 10.0 + assert isinstance(m.float_field, float) + + m = Model.construct(float_field=10.12) + assert m.float_field == 10.12 + assert isinstance(m.float_field, float) + + # number too big + m = Model.construct(float_field=2**53 + 1) + assert m.float_field == 2**53 + 1 + assert isinstance(m.float_field, int) + + +def test_deprecated_alias() -> None: + class Model(BaseModel): + resource_id: str = Field(alias="model_id") + + @property + def model_id(self) -> str: + return self.resource_id + + m = Model.construct(model_id="id") + assert m.model_id == "id" + assert m.resource_id == "id" + assert m.resource_id is m.model_id + + m = parse_obj(Model, {"model_id": "id"}) + assert m.model_id == "id" + assert m.resource_id == "id" + assert m.resource_id is m.model_id + + +def test_omitted_fields() -> None: + class Model(BaseModel): + resource_id: Optional[str] = None + + m = Model.construct() + assert m.resource_id is None + assert "resource_id" not in m.model_fields_set + + m = Model.construct(resource_id=None) + assert m.resource_id is None + assert "resource_id" in m.model_fields_set + + m = Model.construct(resource_id="foo") + assert m.resource_id == "foo" + assert "resource_id" in m.model_fields_set + + +def test_to_dict() -> None: + class Model(BaseModel): + foo: Optional[str] = Field(alias="FOO", default=None) + + m = Model(FOO="hello") + assert m.to_dict() == {"FOO": "hello"} + assert m.to_dict(use_api_names=False) == {"foo": "hello"} + + m2 = Model() + assert m2.to_dict() == {} + assert m2.to_dict(exclude_unset=False) == {"FOO": None} + assert m2.to_dict(exclude_unset=False, exclude_none=True) == {} + assert m2.to_dict(exclude_unset=False, exclude_defaults=True) == {} + + m3 = Model(FOO=None) + assert m3.to_dict() == {"FOO": None} + assert m3.to_dict(exclude_none=True) == {} + assert m3.to_dict(exclude_defaults=True) == {} + + class Model2(BaseModel): + created_at: datetime + + time_str = "2024-03-21T11:39:01.275859" + m4 = Model2.construct(created_at=time_str) + assert m4.to_dict(mode="python") == {"created_at": datetime.fromisoformat(time_str)} + assert m4.to_dict(mode="json") == {"created_at": time_str} + + if PYDANTIC_V1: + with pytest.raises(ValueError, match="warnings is only supported in Pydantic v2"): + m.to_dict(warnings=False) + + +def test_forwards_compat_model_dump_method() -> None: + class Model(BaseModel): + foo: Optional[str] = Field(alias="FOO", default=None) + + m = Model(FOO="hello") + assert m.model_dump() == {"foo": "hello"} + assert m.model_dump(include={"bar"}) == {} + assert m.model_dump(exclude={"foo"}) == {} + assert m.model_dump(by_alias=True) == {"FOO": "hello"} + + m2 = Model() + assert m2.model_dump() == {"foo": None} + assert m2.model_dump(exclude_unset=True) == {} + assert m2.model_dump(exclude_none=True) == {} + assert m2.model_dump(exclude_defaults=True) == {} + + m3 = Model(FOO=None) + assert m3.model_dump() == {"foo": None} + assert m3.model_dump(exclude_none=True) == {} + + if PYDANTIC_V1: + with pytest.raises(ValueError, match="round_trip is only supported in Pydantic v2"): + m.model_dump(round_trip=True) + + with pytest.raises(ValueError, match="warnings is only supported in Pydantic v2"): + m.model_dump(warnings=False) + + +def test_compat_method_no_error_for_warnings() -> None: + class Model(BaseModel): + foo: Optional[str] + + m = Model(foo="hello") + assert isinstance(model_dump(m, warnings=False), dict) + + +def test_to_json() -> None: + class Model(BaseModel): + foo: Optional[str] = Field(alias="FOO", default=None) + + m = Model(FOO="hello") + assert json.loads(m.to_json()) == {"FOO": "hello"} + assert json.loads(m.to_json(use_api_names=False)) == {"foo": "hello"} + + if PYDANTIC_V1: + assert m.to_json(indent=None) == '{"FOO": "hello"}' + else: + assert m.to_json(indent=None) == '{"FOO":"hello"}' + + m2 = Model() + assert json.loads(m2.to_json()) == {} + assert json.loads(m2.to_json(exclude_unset=False)) == {"FOO": None} + assert json.loads(m2.to_json(exclude_unset=False, exclude_none=True)) == {} + assert json.loads(m2.to_json(exclude_unset=False, exclude_defaults=True)) == {} + + m3 = Model(FOO=None) + assert json.loads(m3.to_json()) == {"FOO": None} + assert json.loads(m3.to_json(exclude_none=True)) == {} + + if PYDANTIC_V1: + with pytest.raises(ValueError, match="warnings is only supported in Pydantic v2"): + m.to_json(warnings=False) + + +def test_forwards_compat_model_dump_json_method() -> None: + class Model(BaseModel): + foo: Optional[str] = Field(alias="FOO", default=None) + + m = Model(FOO="hello") + assert json.loads(m.model_dump_json()) == {"foo": "hello"} + assert json.loads(m.model_dump_json(include={"bar"})) == {} + assert json.loads(m.model_dump_json(include={"foo"})) == {"foo": "hello"} + assert json.loads(m.model_dump_json(by_alias=True)) == {"FOO": "hello"} + + assert m.model_dump_json(indent=2) == '{\n "foo": "hello"\n}' + + m2 = Model() + assert json.loads(m2.model_dump_json()) == {"foo": None} + assert json.loads(m2.model_dump_json(exclude_unset=True)) == {} + assert json.loads(m2.model_dump_json(exclude_none=True)) == {} + assert json.loads(m2.model_dump_json(exclude_defaults=True)) == {} + + m3 = Model(FOO=None) + assert json.loads(m3.model_dump_json()) == {"foo": None} + assert json.loads(m3.model_dump_json(exclude_none=True)) == {} + + if PYDANTIC_V1: + with pytest.raises(ValueError, match="round_trip is only supported in Pydantic v2"): + m.model_dump_json(round_trip=True) + + with pytest.raises(ValueError, match="warnings is only supported in Pydantic v2"): + m.model_dump_json(warnings=False) + + +def test_type_compat() -> None: + # our model type can be assigned to Pydantic's model type + + def takes_pydantic(model: pydantic.BaseModel) -> None: # noqa: ARG001 + ... + + class OurModel(BaseModel): + foo: Optional[str] = None + + takes_pydantic(OurModel()) + + +def test_annotated_types() -> None: + class Model(BaseModel): + value: str + + m = construct_type( + value={"value": "foo"}, + type_=cast(Any, Annotated[Model, "random metadata"]), + ) + assert isinstance(m, Model) + assert m.value == "foo" + + +def test_discriminated_unions_invalid_data() -> None: + class A(BaseModel): + type: Literal["a"] + + data: str + + class B(BaseModel): + type: Literal["b"] + + data: int + + m = construct_type( + value={"type": "b", "data": "foo"}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="type")]), + ) + assert isinstance(m, B) + assert m.type == "b" + assert m.data == "foo" # type: ignore[comparison-overlap] + + m = construct_type( + value={"type": "a", "data": 100}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="type")]), + ) + assert isinstance(m, A) + assert m.type == "a" + if PYDANTIC_V1: + # pydantic v1 automatically converts inputs to strings + # if the expected type is a str + assert m.data == "100" + else: + assert m.data == 100 # type: ignore[comparison-overlap] + + +def test_discriminated_unions_unknown_variant() -> None: + class A(BaseModel): + type: Literal["a"] + + data: str + + class B(BaseModel): + type: Literal["b"] + + data: int + + m = construct_type( + value={"type": "c", "data": None, "new_thing": "bar"}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="type")]), + ) + + # just chooses the first variant + assert isinstance(m, A) + assert m.type == "c" # type: ignore[comparison-overlap] + assert m.data == None # type: ignore[unreachable] + assert m.new_thing == "bar" + + +def test_discriminated_unions_invalid_data_nested_unions() -> None: + class A(BaseModel): + type: Literal["a"] + + data: str + + class B(BaseModel): + type: Literal["b"] + + data: int + + class C(BaseModel): + type: Literal["c"] + + data: bool + + m = construct_type( + value={"type": "b", "data": "foo"}, + type_=cast(Any, Annotated[Union[Union[A, B], C], PropertyInfo(discriminator="type")]), + ) + assert isinstance(m, B) + assert m.type == "b" + assert m.data == "foo" # type: ignore[comparison-overlap] + + m = construct_type( + value={"type": "c", "data": "foo"}, + type_=cast(Any, Annotated[Union[Union[A, B], C], PropertyInfo(discriminator="type")]), + ) + assert isinstance(m, C) + assert m.type == "c" + assert m.data == "foo" # type: ignore[comparison-overlap] + + +def test_discriminated_unions_with_aliases_invalid_data() -> None: + class A(BaseModel): + foo_type: Literal["a"] = Field(alias="type") + + data: str + + class B(BaseModel): + foo_type: Literal["b"] = Field(alias="type") + + data: int + + m = construct_type( + value={"type": "b", "data": "foo"}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="foo_type")]), + ) + assert isinstance(m, B) + assert m.foo_type == "b" + assert m.data == "foo" # type: ignore[comparison-overlap] + + m = construct_type( + value={"type": "a", "data": 100}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="foo_type")]), + ) + assert isinstance(m, A) + assert m.foo_type == "a" + if PYDANTIC_V1: + # pydantic v1 automatically converts inputs to strings + # if the expected type is a str + assert m.data == "100" + else: + assert m.data == 100 # type: ignore[comparison-overlap] + + +def test_discriminated_unions_overlapping_discriminators_invalid_data() -> None: + class A(BaseModel): + type: Literal["a"] + + data: bool + + class B(BaseModel): + type: Literal["a"] + + data: int + + m = construct_type( + value={"type": "a", "data": "foo"}, + type_=cast(Any, Annotated[Union[A, B], PropertyInfo(discriminator="type")]), + ) + assert isinstance(m, B) + assert m.type == "a" + assert m.data == "foo" # type: ignore[comparison-overlap] + + +def test_discriminated_unions_invalid_data_uses_cache() -> None: + class A(BaseModel): + type: Literal["a"] + + data: str + + class B(BaseModel): + type: Literal["b"] + + data: int + + UnionType = cast(Any, Union[A, B]) + + assert not hasattr(UnionType, "__discriminator__") + + m = construct_type( + value={"type": "b", "data": "foo"}, type_=cast(Any, Annotated[UnionType, PropertyInfo(discriminator="type")]) + ) + assert isinstance(m, B) + assert m.type == "b" + assert m.data == "foo" # type: ignore[comparison-overlap] + + discriminator = UnionType.__discriminator__ + assert discriminator is not None + + m = construct_type( + value={"type": "b", "data": "foo"}, type_=cast(Any, Annotated[UnionType, PropertyInfo(discriminator="type")]) + ) + assert isinstance(m, B) + assert m.type == "b" + assert m.data == "foo" # type: ignore[comparison-overlap] + + # if the discriminator details object stays the same between invocations then + # we hit the cache + assert UnionType.__discriminator__ is discriminator + + +@pytest.mark.skipif(PYDANTIC_V1, reason="TypeAliasType is not supported in Pydantic v1") +def test_type_alias_type() -> None: + Alias = TypeAliasType("Alias", str) # pyright: ignore + + class Model(BaseModel): + alias: Alias + union: Union[int, Alias] + + m = construct_type(value={"alias": "foo", "union": "bar"}, type_=Model) + assert isinstance(m, Model) + assert isinstance(m.alias, str) + assert m.alias == "foo" + assert isinstance(m.union, str) + assert m.union == "bar" + + +@pytest.mark.skipif(PYDANTIC_V1, reason="TypeAliasType is not supported in Pydantic v1") +def test_field_named_cls() -> None: + class Model(BaseModel): + cls: str + + m = construct_type(value={"cls": "foo"}, type_=Model) + assert isinstance(m, Model) + assert isinstance(m.cls, str) + + +def test_discriminated_union_case() -> None: + class A(BaseModel): + type: Literal["a"] + + data: bool + + class B(BaseModel): + type: Literal["b"] + + data: List[Union[A, object]] + + class ModelA(BaseModel): + type: Literal["modelA"] + + data: int + + class ModelB(BaseModel): + type: Literal["modelB"] + + required: str + + data: Union[A, B] + + # when constructing ModelA | ModelB, value data doesn't match ModelB exactly - missing `required` + m = construct_type( + value={"type": "modelB", "data": {"type": "a", "data": True}}, + type_=cast(Any, Annotated[Union[ModelA, ModelB], PropertyInfo(discriminator="type")]), + ) + + assert isinstance(m, ModelB) + + +def test_nested_discriminated_union() -> None: + class InnerType1(BaseModel): + type: Literal["type_1"] + + class InnerModel(BaseModel): + inner_value: str + + class InnerType2(BaseModel): + type: Literal["type_2"] + some_inner_model: InnerModel + + class Type1(BaseModel): + base_type: Literal["base_type_1"] + value: Annotated[ + Union[ + InnerType1, + InnerType2, + ], + PropertyInfo(discriminator="type"), + ] + + class Type2(BaseModel): + base_type: Literal["base_type_2"] + + T = Annotated[ + Union[ + Type1, + Type2, + ], + PropertyInfo(discriminator="base_type"), + ] + + model = construct_type( + type_=T, + value={ + "base_type": "base_type_1", + "value": { + "type": "type_2", + }, + }, + ) + assert isinstance(model, Type1) + assert isinstance(model.value, InnerType2) + + +@pytest.mark.skipif(PYDANTIC_V1, reason="this is only supported in pydantic v2 for now") +def test_extra_properties() -> None: + class Item(BaseModel): + prop: int + + class Model(BaseModel): + __pydantic_extra__: Dict[str, Item] = Field(init=False) # pyright: ignore[reportIncompatibleVariableOverride] + + other: str + + if TYPE_CHECKING: + + def __getattr__(self, attr: str) -> Item: ... + + model = construct_type( + type_=Model, + value={ + "a": {"prop": 1}, + "other": "foo", + }, + ) + assert isinstance(model, Model) + assert model.a.prop == 1 + assert isinstance(model.a, Item) + assert model.other == "foo" diff --git a/tests/test_qs.py b/tests/test_qs.py new file mode 100644 index 0000000..99accfc --- /dev/null +++ b/tests/test_qs.py @@ -0,0 +1,78 @@ +from typing import Any, cast +from functools import partial +from urllib.parse import unquote + +import pytest + +from brapi._qs import Querystring, stringify + + +def test_empty() -> None: + assert stringify({}) == "" + assert stringify({"a": {}}) == "" + assert stringify({"a": {"b": {"c": {}}}}) == "" + + +def test_basic() -> None: + assert stringify({"a": 1}) == "a=1" + assert stringify({"a": "b"}) == "a=b" + assert stringify({"a": True}) == "a=true" + assert stringify({"a": False}) == "a=false" + assert stringify({"a": 1.23456}) == "a=1.23456" + assert stringify({"a": None}) == "" + + +@pytest.mark.parametrize("method", ["class", "function"]) +def test_nested_dotted(method: str) -> None: + if method == "class": + serialise = Querystring(nested_format="dots").stringify + else: + serialise = partial(stringify, nested_format="dots") + + assert unquote(serialise({"a": {"b": "c"}})) == "a.b=c" + assert unquote(serialise({"a": {"b": "c", "d": "e", "f": "g"}})) == "a.b=c&a.d=e&a.f=g" + assert unquote(serialise({"a": {"b": {"c": {"d": "e"}}}})) == "a.b.c.d=e" + assert unquote(serialise({"a": {"b": True}})) == "a.b=true" + + +def test_nested_brackets() -> None: + assert unquote(stringify({"a": {"b": "c"}})) == "a[b]=c" + assert unquote(stringify({"a": {"b": "c", "d": "e", "f": "g"}})) == "a[b]=c&a[d]=e&a[f]=g" + assert unquote(stringify({"a": {"b": {"c": {"d": "e"}}}})) == "a[b][c][d]=e" + assert unquote(stringify({"a": {"b": True}})) == "a[b]=true" + + +@pytest.mark.parametrize("method", ["class", "function"]) +def test_array_comma(method: str) -> None: + if method == "class": + serialise = Querystring(array_format="comma").stringify + else: + serialise = partial(stringify, array_format="comma") + + assert unquote(serialise({"in": ["foo", "bar"]})) == "in=foo,bar" + assert unquote(serialise({"a": {"b": [True, False]}})) == "a[b]=true,false" + assert unquote(serialise({"a": {"b": [True, False, None, True]}})) == "a[b]=true,false,true" + + +def test_array_repeat() -> None: + assert unquote(stringify({"in": ["foo", "bar"]})) == "in=foo&in=bar" + assert unquote(stringify({"a": {"b": [True, False]}})) == "a[b]=true&a[b]=false" + assert unquote(stringify({"a": {"b": [True, False, None, True]}})) == "a[b]=true&a[b]=false&a[b]=true" + assert unquote(stringify({"in": ["foo", {"b": {"c": ["d", "e"]}}]})) == "in=foo&in[b][c]=d&in[b][c]=e" + + +@pytest.mark.parametrize("method", ["class", "function"]) +def test_array_brackets(method: str) -> None: + if method == "class": + serialise = Querystring(array_format="brackets").stringify + else: + serialise = partial(stringify, array_format="brackets") + + assert unquote(serialise({"in": ["foo", "bar"]})) == "in[]=foo&in[]=bar" + assert unquote(serialise({"a": {"b": [True, False]}})) == "a[b][]=true&a[b][]=false" + assert unquote(serialise({"a": {"b": [True, False, None, True]}})) == "a[b][]=true&a[b][]=false&a[b][]=true" + + +def test_unknown_array_format() -> None: + with pytest.raises(NotImplementedError, match="Unknown array_format value: foo, choose from comma, repeat"): + stringify({"a": ["foo", "bar"]}, array_format=cast(Any, "foo")) diff --git a/tests/test_required_args.py b/tests/test_required_args.py new file mode 100644 index 0000000..958fd4f --- /dev/null +++ b/tests/test_required_args.py @@ -0,0 +1,111 @@ +from __future__ import annotations + +import pytest + +from brapi._utils import required_args + + +def test_too_many_positional_params() -> None: + @required_args(["a"]) + def foo(a: str | None = None) -> str | None: + return a + + with pytest.raises(TypeError, match=r"foo\(\) takes 1 argument\(s\) but 2 were given"): + foo("a", "b") # type: ignore + + +def test_positional_param() -> None: + @required_args(["a"]) + def foo(a: str | None = None) -> str | None: + return a + + assert foo("a") == "a" + assert foo(None) is None + assert foo(a="b") == "b" + + with pytest.raises(TypeError, match="Missing required argument: 'a'"): + foo() + + +def test_keyword_only_param() -> None: + @required_args(["a"]) + def foo(*, a: str | None = None) -> str | None: + return a + + assert foo(a="a") == "a" + assert foo(a=None) is None + assert foo(a="b") == "b" + + with pytest.raises(TypeError, match="Missing required argument: 'a'"): + foo() + + +def test_multiple_params() -> None: + @required_args(["a", "b", "c"]) + def foo(a: str = "", *, b: str = "", c: str = "") -> str | None: + return f"{a} {b} {c}" + + assert foo(a="a", b="b", c="c") == "a b c" + + error_message = r"Missing required arguments.*" + + with pytest.raises(TypeError, match=error_message): + foo() + + with pytest.raises(TypeError, match=error_message): + foo(a="a") + + with pytest.raises(TypeError, match=error_message): + foo(b="b") + + with pytest.raises(TypeError, match=error_message): + foo(c="c") + + with pytest.raises(TypeError, match=r"Missing required argument: 'a'"): + foo(b="a", c="c") + + with pytest.raises(TypeError, match=r"Missing required argument: 'b'"): + foo("a", c="c") + + +def test_multiple_variants() -> None: + @required_args(["a"], ["b"]) + def foo(*, a: str | None = None, b: str | None = None) -> str | None: + return a if a is not None else b + + assert foo(a="foo") == "foo" + assert foo(b="bar") == "bar" + assert foo(a=None) is None + assert foo(b=None) is None + + # TODO: this error message could probably be improved + with pytest.raises( + TypeError, + match=r"Missing required arguments; Expected either \('a'\) or \('b'\) arguments to be given", + ): + foo() + + +def test_multiple_params_multiple_variants() -> None: + @required_args(["a", "b"], ["c"]) + def foo(*, a: str | None = None, b: str | None = None, c: str | None = None) -> str | None: + if a is not None: + return a + if b is not None: + return b + return c + + error_message = r"Missing required arguments; Expected either \('a' and 'b'\) or \('c'\) arguments to be given" + + with pytest.raises(TypeError, match=error_message): + foo(a="foo") + + with pytest.raises(TypeError, match=error_message): + foo(b="bar") + + with pytest.raises(TypeError, match=error_message): + foo() + + assert foo(a=None, b="bar") == "bar" + assert foo(c=None) is None + assert foo(c="foo") == "foo" diff --git a/tests/test_response.py b/tests/test_response.py new file mode 100644 index 0000000..5bdfd2e --- /dev/null +++ b/tests/test_response.py @@ -0,0 +1,277 @@ +import json +from typing import Any, List, Union, cast +from typing_extensions import Annotated + +import httpx +import pytest +import pydantic + +from brapi import Brapi, BaseModel, AsyncBrapi +from brapi._response import ( + APIResponse, + BaseAPIResponse, + AsyncAPIResponse, + BinaryAPIResponse, + AsyncBinaryAPIResponse, + extract_response_type, +) +from brapi._streaming import Stream +from brapi._base_client import FinalRequestOptions + + +class ConcreteBaseAPIResponse(APIResponse[bytes]): ... + + +class ConcreteAPIResponse(APIResponse[List[str]]): ... + + +class ConcreteAsyncAPIResponse(APIResponse[httpx.Response]): ... + + +def test_extract_response_type_direct_classes() -> None: + assert extract_response_type(BaseAPIResponse[str]) == str + assert extract_response_type(APIResponse[str]) == str + assert extract_response_type(AsyncAPIResponse[str]) == str + + +def test_extract_response_type_direct_class_missing_type_arg() -> None: + with pytest.raises( + RuntimeError, + match="Expected type to have a type argument at index 0 but it did not", + ): + extract_response_type(AsyncAPIResponse) + + +def test_extract_response_type_concrete_subclasses() -> None: + assert extract_response_type(ConcreteBaseAPIResponse) == bytes + assert extract_response_type(ConcreteAPIResponse) == List[str] + assert extract_response_type(ConcreteAsyncAPIResponse) == httpx.Response + + +def test_extract_response_type_binary_response() -> None: + assert extract_response_type(BinaryAPIResponse) == bytes + assert extract_response_type(AsyncBinaryAPIResponse) == bytes + + +class PydanticModel(pydantic.BaseModel): ... + + +def test_response_parse_mismatched_basemodel(client: Brapi) -> None: + response = APIResponse( + raw=httpx.Response(200, content=b"foo"), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + with pytest.raises( + TypeError, + match="Pydantic models must subclass our base model type, e.g. `from brapi import BaseModel`", + ): + response.parse(to=PydanticModel) + + +@pytest.mark.asyncio +async def test_async_response_parse_mismatched_basemodel(async_client: AsyncBrapi) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=b"foo"), + client=async_client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + with pytest.raises( + TypeError, + match="Pydantic models must subclass our base model type, e.g. `from brapi import BaseModel`", + ): + await response.parse(to=PydanticModel) + + +def test_response_parse_custom_stream(client: Brapi) -> None: + response = APIResponse( + raw=httpx.Response(200, content=b"foo"), + client=client, + stream=True, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + stream = response.parse(to=Stream[int]) + assert stream._cast_to == int + + +@pytest.mark.asyncio +async def test_async_response_parse_custom_stream(async_client: AsyncBrapi) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=b"foo"), + client=async_client, + stream=True, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + stream = await response.parse(to=Stream[int]) + assert stream._cast_to == int + + +class CustomModel(BaseModel): + foo: str + bar: int + + +def test_response_parse_custom_model(client: Brapi) -> None: + response = APIResponse( + raw=httpx.Response(200, content=json.dumps({"foo": "hello!", "bar": 2})), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = response.parse(to=CustomModel) + assert obj.foo == "hello!" + assert obj.bar == 2 + + +@pytest.mark.asyncio +async def test_async_response_parse_custom_model(async_client: AsyncBrapi) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=json.dumps({"foo": "hello!", "bar": 2})), + client=async_client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = await response.parse(to=CustomModel) + assert obj.foo == "hello!" + assert obj.bar == 2 + + +def test_response_parse_annotated_type(client: Brapi) -> None: + response = APIResponse( + raw=httpx.Response(200, content=json.dumps({"foo": "hello!", "bar": 2})), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = response.parse( + to=cast("type[CustomModel]", Annotated[CustomModel, "random metadata"]), + ) + assert obj.foo == "hello!" + assert obj.bar == 2 + + +async def test_async_response_parse_annotated_type(async_client: AsyncBrapi) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=json.dumps({"foo": "hello!", "bar": 2})), + client=async_client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = await response.parse( + to=cast("type[CustomModel]", Annotated[CustomModel, "random metadata"]), + ) + assert obj.foo == "hello!" + assert obj.bar == 2 + + +@pytest.mark.parametrize( + "content, expected", + [ + ("false", False), + ("true", True), + ("False", False), + ("True", True), + ("TrUe", True), + ("FalSe", False), + ], +) +def test_response_parse_bool(client: Brapi, content: str, expected: bool) -> None: + response = APIResponse( + raw=httpx.Response(200, content=content), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + result = response.parse(to=bool) + assert result is expected + + +@pytest.mark.parametrize( + "content, expected", + [ + ("false", False), + ("true", True), + ("False", False), + ("True", True), + ("TrUe", True), + ("FalSe", False), + ], +) +async def test_async_response_parse_bool(client: AsyncBrapi, content: str, expected: bool) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=content), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + result = await response.parse(to=bool) + assert result is expected + + +class OtherModel(BaseModel): + a: str + + +@pytest.mark.parametrize("client", [False], indirect=True) # loose validation +def test_response_parse_expect_model_union_non_json_content(client: Brapi) -> None: + response = APIResponse( + raw=httpx.Response(200, content=b"foo", headers={"Content-Type": "application/text"}), + client=client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = response.parse(to=cast(Any, Union[CustomModel, OtherModel])) + assert isinstance(obj, str) + assert obj == "foo" + + +@pytest.mark.asyncio +@pytest.mark.parametrize("async_client", [False], indirect=True) # loose validation +async def test_async_response_parse_expect_model_union_non_json_content(async_client: AsyncBrapi) -> None: + response = AsyncAPIResponse( + raw=httpx.Response(200, content=b"foo", headers={"Content-Type": "application/text"}), + client=async_client, + stream=False, + stream_cls=None, + cast_to=str, + options=FinalRequestOptions.construct(method="get", url="/foo"), + ) + + obj = await response.parse(to=cast(Any, Union[CustomModel, OtherModel])) + assert isinstance(obj, str) + assert obj == "foo" diff --git a/tests/test_streaming.py b/tests/test_streaming.py new file mode 100644 index 0000000..ab4a852 --- /dev/null +++ b/tests/test_streaming.py @@ -0,0 +1,248 @@ +from __future__ import annotations + +from typing import Iterator, AsyncIterator + +import httpx +import pytest + +from brapi import Brapi, AsyncBrapi +from brapi._streaming import Stream, AsyncStream, ServerSentEvent + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_basic(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: completion\n" + yield b'data: {"foo":true}\n' + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "completion" + assert sse.json() == {"foo": True} + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_data_missing_event(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b'data: {"foo":true}\n' + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event is None + assert sse.json() == {"foo": True} + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_event_missing_data(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.data == "" + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_multiple_events(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b"\n" + yield b"event: completion\n" + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.data == "" + + sse = await iter_next(iterator) + assert sse.event == "completion" + assert sse.data == "" + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_multiple_events_with_data(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b'data: {"foo":true}\n' + yield b"\n" + yield b"event: completion\n" + yield b'data: {"bar":false}\n' + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.json() == {"foo": True} + + sse = await iter_next(iterator) + assert sse.event == "completion" + assert sse.json() == {"bar": False} + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_multiple_data_lines_with_empty_line(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b"data: {\n" + yield b'data: "foo":\n' + yield b"data: \n" + yield b"data:\n" + yield b"data: true}\n" + yield b"\n\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.json() == {"foo": True} + assert sse.data == '{\n"foo":\n\n\ntrue}' + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_data_json_escaped_double_new_line(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b'data: {"foo": "my long\\n\\ncontent"}' + yield b"\n\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.json() == {"foo": "my long\n\ncontent"} + + await assert_empty_iter(iterator) + + +@pytest.mark.asyncio +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_multiple_data_lines(sync: bool, client: Brapi, async_client: AsyncBrapi) -> None: + def body() -> Iterator[bytes]: + yield b"event: ping\n" + yield b"data: {\n" + yield b'data: "foo":\n' + yield b"data: true}\n" + yield b"\n\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event == "ping" + assert sse.json() == {"foo": True} + + await assert_empty_iter(iterator) + + +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_special_new_line_character( + sync: bool, + client: Brapi, + async_client: AsyncBrapi, +) -> None: + def body() -> Iterator[bytes]: + yield b'data: {"content":" culpa"}\n' + yield b"\n" + yield b'data: {"content":" \xe2\x80\xa8"}\n' + yield b"\n" + yield b'data: {"content":"foo"}\n' + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event is None + assert sse.json() == {"content": " culpa"} + + sse = await iter_next(iterator) + assert sse.event is None + assert sse.json() == {"content": " 
"} + + sse = await iter_next(iterator) + assert sse.event is None + assert sse.json() == {"content": "foo"} + + await assert_empty_iter(iterator) + + +@pytest.mark.parametrize("sync", [True, False], ids=["sync", "async"]) +async def test_multi_byte_character_multiple_chunks( + sync: bool, + client: Brapi, + async_client: AsyncBrapi, +) -> None: + def body() -> Iterator[bytes]: + yield b'data: {"content":"' + # bytes taken from the string 'известни' and arbitrarily split + # so that some multi-byte characters span multiple chunks + yield b"\xd0" + yield b"\xb8\xd0\xb7\xd0" + yield b"\xb2\xd0\xb5\xd1\x81\xd1\x82\xd0\xbd\xd0\xb8" + yield b'"}\n' + yield b"\n" + + iterator = make_event_iterator(content=body(), sync=sync, client=client, async_client=async_client) + + sse = await iter_next(iterator) + assert sse.event is None + assert sse.json() == {"content": "известни"} + + +async def to_aiter(iter: Iterator[bytes]) -> AsyncIterator[bytes]: + for chunk in iter: + yield chunk + + +async def iter_next(iter: Iterator[ServerSentEvent] | AsyncIterator[ServerSentEvent]) -> ServerSentEvent: + if isinstance(iter, AsyncIterator): + return await iter.__anext__() + + return next(iter) + + +async def assert_empty_iter(iter: Iterator[ServerSentEvent] | AsyncIterator[ServerSentEvent]) -> None: + with pytest.raises((StopAsyncIteration, RuntimeError)): + await iter_next(iter) + + +def make_event_iterator( + content: Iterator[bytes], + *, + sync: bool, + client: Brapi, + async_client: AsyncBrapi, +) -> Iterator[ServerSentEvent] | AsyncIterator[ServerSentEvent]: + if sync: + return Stream(cast_to=object, client=client, response=httpx.Response(200, content=content))._iter_events() + + return AsyncStream( + cast_to=object, client=async_client, response=httpx.Response(200, content=to_aiter(content)) + )._iter_events() diff --git a/tests/test_transform.py b/tests/test_transform.py new file mode 100644 index 0000000..90f69e9 --- /dev/null +++ b/tests/test_transform.py @@ -0,0 +1,460 @@ +from __future__ import annotations + +import io +import pathlib +from typing import Any, Dict, List, Union, TypeVar, Iterable, Optional, cast +from datetime import date, datetime +from typing_extensions import Required, Annotated, TypedDict + +import pytest + +from brapi._types import Base64FileInput, omit, not_given +from brapi._utils import ( + PropertyInfo, + transform as _transform, + parse_datetime, + async_transform as _async_transform, +) +from brapi._compat import PYDANTIC_V1 +from brapi._models import BaseModel + +_T = TypeVar("_T") + +SAMPLE_FILE_PATH = pathlib.Path(__file__).parent.joinpath("sample_file.txt") + + +async def transform( + data: _T, + expected_type: object, + use_async: bool, +) -> _T: + if use_async: + return await _async_transform(data, expected_type=expected_type) + + return _transform(data, expected_type=expected_type) + + +parametrize = pytest.mark.parametrize("use_async", [False, True], ids=["sync", "async"]) + + +class Foo1(TypedDict): + foo_bar: Annotated[str, PropertyInfo(alias="fooBar")] + + +@parametrize +@pytest.mark.asyncio +async def test_top_level_alias(use_async: bool) -> None: + assert await transform({"foo_bar": "hello"}, expected_type=Foo1, use_async=use_async) == {"fooBar": "hello"} + + +class Foo2(TypedDict): + bar: Bar2 + + +class Bar2(TypedDict): + this_thing: Annotated[int, PropertyInfo(alias="this__thing")] + baz: Annotated[Baz2, PropertyInfo(alias="Baz")] + + +class Baz2(TypedDict): + my_baz: Annotated[str, PropertyInfo(alias="myBaz")] + + +@parametrize +@pytest.mark.asyncio +async def test_recursive_typeddict(use_async: bool) -> None: + assert await transform({"bar": {"this_thing": 1}}, Foo2, use_async) == {"bar": {"this__thing": 1}} + assert await transform({"bar": {"baz": {"my_baz": "foo"}}}, Foo2, use_async) == {"bar": {"Baz": {"myBaz": "foo"}}} + + +class Foo3(TypedDict): + things: List[Bar3] + + +class Bar3(TypedDict): + my_field: Annotated[str, PropertyInfo(alias="myField")] + + +@parametrize +@pytest.mark.asyncio +async def test_list_of_typeddict(use_async: bool) -> None: + result = await transform({"things": [{"my_field": "foo"}, {"my_field": "foo2"}]}, Foo3, use_async) + assert result == {"things": [{"myField": "foo"}, {"myField": "foo2"}]} + + +class Foo4(TypedDict): + foo: Union[Bar4, Baz4] + + +class Bar4(TypedDict): + foo_bar: Annotated[str, PropertyInfo(alias="fooBar")] + + +class Baz4(TypedDict): + foo_baz: Annotated[str, PropertyInfo(alias="fooBaz")] + + +@parametrize +@pytest.mark.asyncio +async def test_union_of_typeddict(use_async: bool) -> None: + assert await transform({"foo": {"foo_bar": "bar"}}, Foo4, use_async) == {"foo": {"fooBar": "bar"}} + assert await transform({"foo": {"foo_baz": "baz"}}, Foo4, use_async) == {"foo": {"fooBaz": "baz"}} + assert await transform({"foo": {"foo_baz": "baz", "foo_bar": "bar"}}, Foo4, use_async) == { + "foo": {"fooBaz": "baz", "fooBar": "bar"} + } + + +class Foo5(TypedDict): + foo: Annotated[Union[Bar4, List[Baz4]], PropertyInfo(alias="FOO")] + + +class Bar5(TypedDict): + foo_bar: Annotated[str, PropertyInfo(alias="fooBar")] + + +class Baz5(TypedDict): + foo_baz: Annotated[str, PropertyInfo(alias="fooBaz")] + + +@parametrize +@pytest.mark.asyncio +async def test_union_of_list(use_async: bool) -> None: + assert await transform({"foo": {"foo_bar": "bar"}}, Foo5, use_async) == {"FOO": {"fooBar": "bar"}} + assert await transform( + { + "foo": [ + {"foo_baz": "baz"}, + {"foo_baz": "baz"}, + ] + }, + Foo5, + use_async, + ) == {"FOO": [{"fooBaz": "baz"}, {"fooBaz": "baz"}]} + + +class Foo6(TypedDict): + bar: Annotated[str, PropertyInfo(alias="Bar")] + + +@parametrize +@pytest.mark.asyncio +async def test_includes_unknown_keys(use_async: bool) -> None: + assert await transform({"bar": "bar", "baz_": {"FOO": 1}}, Foo6, use_async) == { + "Bar": "bar", + "baz_": {"FOO": 1}, + } + + +class Foo7(TypedDict): + bar: Annotated[List[Bar7], PropertyInfo(alias="bAr")] + foo: Bar7 + + +class Bar7(TypedDict): + foo: str + + +@parametrize +@pytest.mark.asyncio +async def test_ignores_invalid_input(use_async: bool) -> None: + assert await transform({"bar": ""}, Foo7, use_async) == {"bAr": ""} + assert await transform({"foo": ""}, Foo7, use_async) == {"foo": ""} + + +class DatetimeDict(TypedDict, total=False): + foo: Annotated[datetime, PropertyInfo(format="iso8601")] + + bar: Annotated[Optional[datetime], PropertyInfo(format="iso8601")] + + required: Required[Annotated[Optional[datetime], PropertyInfo(format="iso8601")]] + + list_: Required[Annotated[Optional[List[datetime]], PropertyInfo(format="iso8601")]] + + union: Annotated[Union[int, datetime], PropertyInfo(format="iso8601")] + + +class DateDict(TypedDict, total=False): + foo: Annotated[date, PropertyInfo(format="iso8601")] + + +class DatetimeModel(BaseModel): + foo: datetime + + +class DateModel(BaseModel): + foo: Optional[date] + + +@parametrize +@pytest.mark.asyncio +async def test_iso8601_format(use_async: bool) -> None: + dt = datetime.fromisoformat("2023-02-23T14:16:36.337692+00:00") + tz = "+00:00" if PYDANTIC_V1 else "Z" + assert await transform({"foo": dt}, DatetimeDict, use_async) == {"foo": "2023-02-23T14:16:36.337692+00:00"} # type: ignore[comparison-overlap] + assert await transform(DatetimeModel(foo=dt), Any, use_async) == {"foo": "2023-02-23T14:16:36.337692" + tz} # type: ignore[comparison-overlap] + + dt = dt.replace(tzinfo=None) + assert await transform({"foo": dt}, DatetimeDict, use_async) == {"foo": "2023-02-23T14:16:36.337692"} # type: ignore[comparison-overlap] + assert await transform(DatetimeModel(foo=dt), Any, use_async) == {"foo": "2023-02-23T14:16:36.337692"} # type: ignore[comparison-overlap] + + assert await transform({"foo": None}, DateDict, use_async) == {"foo": None} # type: ignore[comparison-overlap] + assert await transform(DateModel(foo=None), Any, use_async) == {"foo": None} # type: ignore + assert await transform({"foo": date.fromisoformat("2023-02-23")}, DateDict, use_async) == {"foo": "2023-02-23"} # type: ignore[comparison-overlap] + assert await transform(DateModel(foo=date.fromisoformat("2023-02-23")), DateDict, use_async) == { + "foo": "2023-02-23" + } # type: ignore[comparison-overlap] + + +@parametrize +@pytest.mark.asyncio +async def test_optional_iso8601_format(use_async: bool) -> None: + dt = datetime.fromisoformat("2023-02-23T14:16:36.337692+00:00") + assert await transform({"bar": dt}, DatetimeDict, use_async) == {"bar": "2023-02-23T14:16:36.337692+00:00"} # type: ignore[comparison-overlap] + + assert await transform({"bar": None}, DatetimeDict, use_async) == {"bar": None} + + +@parametrize +@pytest.mark.asyncio +async def test_required_iso8601_format(use_async: bool) -> None: + dt = datetime.fromisoformat("2023-02-23T14:16:36.337692+00:00") + assert await transform({"required": dt}, DatetimeDict, use_async) == { + "required": "2023-02-23T14:16:36.337692+00:00" + } # type: ignore[comparison-overlap] + + assert await transform({"required": None}, DatetimeDict, use_async) == {"required": None} + + +@parametrize +@pytest.mark.asyncio +async def test_union_datetime(use_async: bool) -> None: + dt = datetime.fromisoformat("2023-02-23T14:16:36.337692+00:00") + assert await transform({"union": dt}, DatetimeDict, use_async) == { # type: ignore[comparison-overlap] + "union": "2023-02-23T14:16:36.337692+00:00" + } + + assert await transform({"union": "foo"}, DatetimeDict, use_async) == {"union": "foo"} + + +@parametrize +@pytest.mark.asyncio +async def test_nested_list_iso6801_format(use_async: bool) -> None: + dt1 = datetime.fromisoformat("2023-02-23T14:16:36.337692+00:00") + dt2 = parse_datetime("2022-01-15T06:34:23Z") + assert await transform({"list_": [dt1, dt2]}, DatetimeDict, use_async) == { # type: ignore[comparison-overlap] + "list_": ["2023-02-23T14:16:36.337692+00:00", "2022-01-15T06:34:23+00:00"] + } + + +@parametrize +@pytest.mark.asyncio +async def test_datetime_custom_format(use_async: bool) -> None: + dt = parse_datetime("2022-01-15T06:34:23Z") + + result = await transform(dt, Annotated[datetime, PropertyInfo(format="custom", format_template="%H")], use_async) + assert result == "06" # type: ignore[comparison-overlap] + + +class DateDictWithRequiredAlias(TypedDict, total=False): + required_prop: Required[Annotated[date, PropertyInfo(format="iso8601", alias="prop")]] + + +@parametrize +@pytest.mark.asyncio +async def test_datetime_with_alias(use_async: bool) -> None: + assert await transform({"required_prop": None}, DateDictWithRequiredAlias, use_async) == {"prop": None} # type: ignore[comparison-overlap] + assert await transform( + {"required_prop": date.fromisoformat("2023-02-23")}, DateDictWithRequiredAlias, use_async + ) == {"prop": "2023-02-23"} # type: ignore[comparison-overlap] + + +class MyModel(BaseModel): + foo: str + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_model_to_dictionary(use_async: bool) -> None: + assert cast(Any, await transform(MyModel(foo="hi!"), Any, use_async)) == {"foo": "hi!"} + assert cast(Any, await transform(MyModel.construct(foo="hi!"), Any, use_async)) == {"foo": "hi!"} + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_empty_model(use_async: bool) -> None: + assert cast(Any, await transform(MyModel.construct(), Any, use_async)) == {} + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_unknown_field(use_async: bool) -> None: + assert cast(Any, await transform(MyModel.construct(my_untyped_field=True), Any, use_async)) == { + "my_untyped_field": True + } + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_mismatched_types(use_async: bool) -> None: + model = MyModel.construct(foo=True) + if PYDANTIC_V1: + params = await transform(model, Any, use_async) + else: + with pytest.warns(UserWarning): + params = await transform(model, Any, use_async) + assert cast(Any, params) == {"foo": True} + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_mismatched_object_type(use_async: bool) -> None: + model = MyModel.construct(foo=MyModel.construct(hello="world")) + if PYDANTIC_V1: + params = await transform(model, Any, use_async) + else: + with pytest.warns(UserWarning): + params = await transform(model, Any, use_async) + assert cast(Any, params) == {"foo": {"hello": "world"}} + + +class ModelNestedObjects(BaseModel): + nested: MyModel + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_nested_objects(use_async: bool) -> None: + model = ModelNestedObjects.construct(nested={"foo": "stainless"}) + assert isinstance(model.nested, MyModel) + assert cast(Any, await transform(model, Any, use_async)) == {"nested": {"foo": "stainless"}} + + +class ModelWithDefaultField(BaseModel): + foo: str + with_none_default: Union[str, None] = None + with_str_default: str = "foo" + + +@parametrize +@pytest.mark.asyncio +async def test_pydantic_default_field(use_async: bool) -> None: + # should be excluded when defaults are used + model = ModelWithDefaultField.construct() + assert model.with_none_default is None + assert model.with_str_default == "foo" + assert cast(Any, await transform(model, Any, use_async)) == {} + + # should be included when the default value is explicitly given + model = ModelWithDefaultField.construct(with_none_default=None, with_str_default="foo") + assert model.with_none_default is None + assert model.with_str_default == "foo" + assert cast(Any, await transform(model, Any, use_async)) == {"with_none_default": None, "with_str_default": "foo"} + + # should be included when a non-default value is explicitly given + model = ModelWithDefaultField.construct(with_none_default="bar", with_str_default="baz") + assert model.with_none_default == "bar" + assert model.with_str_default == "baz" + assert cast(Any, await transform(model, Any, use_async)) == {"with_none_default": "bar", "with_str_default": "baz"} + + +class TypedDictIterableUnion(TypedDict): + foo: Annotated[Union[Bar8, Iterable[Baz8]], PropertyInfo(alias="FOO")] + + +class Bar8(TypedDict): + foo_bar: Annotated[str, PropertyInfo(alias="fooBar")] + + +class Baz8(TypedDict): + foo_baz: Annotated[str, PropertyInfo(alias="fooBaz")] + + +@parametrize +@pytest.mark.asyncio +async def test_iterable_of_dictionaries(use_async: bool) -> None: + assert await transform({"foo": [{"foo_baz": "bar"}]}, TypedDictIterableUnion, use_async) == { + "FOO": [{"fooBaz": "bar"}] + } + assert cast(Any, await transform({"foo": ({"foo_baz": "bar"},)}, TypedDictIterableUnion, use_async)) == { + "FOO": [{"fooBaz": "bar"}] + } + + def my_iter() -> Iterable[Baz8]: + yield {"foo_baz": "hello"} + yield {"foo_baz": "world"} + + assert await transform({"foo": my_iter()}, TypedDictIterableUnion, use_async) == { + "FOO": [{"fooBaz": "hello"}, {"fooBaz": "world"}] + } + + +@parametrize +@pytest.mark.asyncio +async def test_dictionary_items(use_async: bool) -> None: + class DictItems(TypedDict): + foo_baz: Annotated[str, PropertyInfo(alias="fooBaz")] + + assert await transform({"foo": {"foo_baz": "bar"}}, Dict[str, DictItems], use_async) == {"foo": {"fooBaz": "bar"}} + + +class TypedDictIterableUnionStr(TypedDict): + foo: Annotated[Union[str, Iterable[Baz8]], PropertyInfo(alias="FOO")] + + +@parametrize +@pytest.mark.asyncio +async def test_iterable_union_str(use_async: bool) -> None: + assert await transform({"foo": "bar"}, TypedDictIterableUnionStr, use_async) == {"FOO": "bar"} + assert cast(Any, await transform(iter([{"foo_baz": "bar"}]), Union[str, Iterable[Baz8]], use_async)) == [ + {"fooBaz": "bar"} + ] + + +class TypedDictBase64Input(TypedDict): + foo: Annotated[Union[str, Base64FileInput], PropertyInfo(format="base64")] + + +@parametrize +@pytest.mark.asyncio +async def test_base64_file_input(use_async: bool) -> None: + # strings are left as-is + assert await transform({"foo": "bar"}, TypedDictBase64Input, use_async) == {"foo": "bar"} + + # pathlib.Path is automatically converted to base64 + assert await transform({"foo": SAMPLE_FILE_PATH}, TypedDictBase64Input, use_async) == { + "foo": "SGVsbG8sIHdvcmxkIQo=" + } # type: ignore[comparison-overlap] + + # io instances are automatically converted to base64 + assert await transform({"foo": io.StringIO("Hello, world!")}, TypedDictBase64Input, use_async) == { + "foo": "SGVsbG8sIHdvcmxkIQ==" + } # type: ignore[comparison-overlap] + assert await transform({"foo": io.BytesIO(b"Hello, world!")}, TypedDictBase64Input, use_async) == { + "foo": "SGVsbG8sIHdvcmxkIQ==" + } # type: ignore[comparison-overlap] + + +@parametrize +@pytest.mark.asyncio +async def test_transform_skipping(use_async: bool) -> None: + # lists of ints are left as-is + data = [1, 2, 3] + assert await transform(data, List[int], use_async) is data + + # iterables of ints are converted to a list + data = iter([1, 2, 3]) + assert await transform(data, Iterable[int], use_async) == [1, 2, 3] + + +@parametrize +@pytest.mark.asyncio +async def test_strips_notgiven(use_async: bool) -> None: + assert await transform({"foo_bar": "bar"}, Foo1, use_async) == {"fooBar": "bar"} + assert await transform({"foo_bar": not_given}, Foo1, use_async) == {} + + +@parametrize +@pytest.mark.asyncio +async def test_strips_omit(use_async: bool) -> None: + assert await transform({"foo_bar": "bar"}, Foo1, use_async) == {"fooBar": "bar"} + assert await transform({"foo_bar": omit}, Foo1, use_async) == {} diff --git a/tests/test_utils/test_datetime_parse.py b/tests/test_utils/test_datetime_parse.py new file mode 100644 index 0000000..0017d67 --- /dev/null +++ b/tests/test_utils/test_datetime_parse.py @@ -0,0 +1,110 @@ +""" +Copied from https://github.com/pydantic/pydantic/blob/v1.10.22/tests/test_datetime_parse.py +with modifications so it works without pydantic v1 imports. +""" + +from typing import Type, Union +from datetime import date, datetime, timezone, timedelta + +import pytest + +from brapi._utils import parse_date, parse_datetime + + +def create_tz(minutes: int) -> timezone: + return timezone(timedelta(minutes=minutes)) + + +@pytest.mark.parametrize( + "value,result", + [ + # Valid inputs + ("1494012444.883309", date(2017, 5, 5)), + (b"1494012444.883309", date(2017, 5, 5)), + (1_494_012_444.883_309, date(2017, 5, 5)), + ("1494012444", date(2017, 5, 5)), + (1_494_012_444, date(2017, 5, 5)), + (0, date(1970, 1, 1)), + ("2012-04-23", date(2012, 4, 23)), + (b"2012-04-23", date(2012, 4, 23)), + ("2012-4-9", date(2012, 4, 9)), + (date(2012, 4, 9), date(2012, 4, 9)), + (datetime(2012, 4, 9, 12, 15), date(2012, 4, 9)), + # Invalid inputs + ("x20120423", ValueError), + ("2012-04-56", ValueError), + (19_999_999_999, date(2603, 10, 11)), # just before watershed + (20_000_000_001, date(1970, 8, 20)), # just after watershed + (1_549_316_052, date(2019, 2, 4)), # nowish in s + (1_549_316_052_104, date(2019, 2, 4)), # nowish in ms + (1_549_316_052_104_324, date(2019, 2, 4)), # nowish in μs + (1_549_316_052_104_324_096, date(2019, 2, 4)), # nowish in ns + ("infinity", date(9999, 12, 31)), + ("inf", date(9999, 12, 31)), + (float("inf"), date(9999, 12, 31)), + ("infinity ", date(9999, 12, 31)), + (int("1" + "0" * 100), date(9999, 12, 31)), + (1e1000, date(9999, 12, 31)), + ("-infinity", date(1, 1, 1)), + ("-inf", date(1, 1, 1)), + ("nan", ValueError), + ], +) +def test_date_parsing(value: Union[str, bytes, int, float], result: Union[date, Type[Exception]]) -> None: + if type(result) == type and issubclass(result, Exception): # pyright: ignore[reportUnnecessaryIsInstance] + with pytest.raises(result): + parse_date(value) + else: + assert parse_date(value) == result + + +@pytest.mark.parametrize( + "value,result", + [ + # Valid inputs + # values in seconds + ("1494012444.883309", datetime(2017, 5, 5, 19, 27, 24, 883_309, tzinfo=timezone.utc)), + (1_494_012_444.883_309, datetime(2017, 5, 5, 19, 27, 24, 883_309, tzinfo=timezone.utc)), + ("1494012444", datetime(2017, 5, 5, 19, 27, 24, tzinfo=timezone.utc)), + (b"1494012444", datetime(2017, 5, 5, 19, 27, 24, tzinfo=timezone.utc)), + (1_494_012_444, datetime(2017, 5, 5, 19, 27, 24, tzinfo=timezone.utc)), + # values in ms + ("1494012444000.883309", datetime(2017, 5, 5, 19, 27, 24, 883, tzinfo=timezone.utc)), + ("-1494012444000.883309", datetime(1922, 8, 29, 4, 32, 35, 999117, tzinfo=timezone.utc)), + (1_494_012_444_000, datetime(2017, 5, 5, 19, 27, 24, tzinfo=timezone.utc)), + ("2012-04-23T09:15:00", datetime(2012, 4, 23, 9, 15)), + ("2012-4-9 4:8:16", datetime(2012, 4, 9, 4, 8, 16)), + ("2012-04-23T09:15:00Z", datetime(2012, 4, 23, 9, 15, 0, 0, timezone.utc)), + ("2012-4-9 4:8:16-0320", datetime(2012, 4, 9, 4, 8, 16, 0, create_tz(-200))), + ("2012-04-23T10:20:30.400+02:30", datetime(2012, 4, 23, 10, 20, 30, 400_000, create_tz(150))), + ("2012-04-23T10:20:30.400+02", datetime(2012, 4, 23, 10, 20, 30, 400_000, create_tz(120))), + ("2012-04-23T10:20:30.400-02", datetime(2012, 4, 23, 10, 20, 30, 400_000, create_tz(-120))), + (b"2012-04-23T10:20:30.400-02", datetime(2012, 4, 23, 10, 20, 30, 400_000, create_tz(-120))), + (datetime(2017, 5, 5), datetime(2017, 5, 5)), + (0, datetime(1970, 1, 1, 0, 0, 0, tzinfo=timezone.utc)), + # Invalid inputs + ("x20120423091500", ValueError), + ("2012-04-56T09:15:90", ValueError), + ("2012-04-23T11:05:00-25:00", ValueError), + (19_999_999_999, datetime(2603, 10, 11, 11, 33, 19, tzinfo=timezone.utc)), # just before watershed + (20_000_000_001, datetime(1970, 8, 20, 11, 33, 20, 1000, tzinfo=timezone.utc)), # just after watershed + (1_549_316_052, datetime(2019, 2, 4, 21, 34, 12, 0, tzinfo=timezone.utc)), # nowish in s + (1_549_316_052_104, datetime(2019, 2, 4, 21, 34, 12, 104_000, tzinfo=timezone.utc)), # nowish in ms + (1_549_316_052_104_324, datetime(2019, 2, 4, 21, 34, 12, 104_324, tzinfo=timezone.utc)), # nowish in μs + (1_549_316_052_104_324_096, datetime(2019, 2, 4, 21, 34, 12, 104_324, tzinfo=timezone.utc)), # nowish in ns + ("infinity", datetime(9999, 12, 31, 23, 59, 59, 999999)), + ("inf", datetime(9999, 12, 31, 23, 59, 59, 999999)), + ("inf ", datetime(9999, 12, 31, 23, 59, 59, 999999)), + (1e50, datetime(9999, 12, 31, 23, 59, 59, 999999)), + (float("inf"), datetime(9999, 12, 31, 23, 59, 59, 999999)), + ("-infinity", datetime(1, 1, 1, 0, 0)), + ("-inf", datetime(1, 1, 1, 0, 0)), + ("nan", ValueError), + ], +) +def test_datetime_parsing(value: Union[str, bytes, int, float], result: Union[datetime, Type[Exception]]) -> None: + if type(result) == type and issubclass(result, Exception): # pyright: ignore[reportUnnecessaryIsInstance] + with pytest.raises(result): + parse_datetime(value) + else: + assert parse_datetime(value) == result diff --git a/tests/test_utils/test_proxy.py b/tests/test_utils/test_proxy.py new file mode 100644 index 0000000..eae01e5 --- /dev/null +++ b/tests/test_utils/test_proxy.py @@ -0,0 +1,34 @@ +import operator +from typing import Any +from typing_extensions import override + +from brapi._utils import LazyProxy + + +class RecursiveLazyProxy(LazyProxy[Any]): + @override + def __load__(self) -> Any: + return self + + def __call__(self, *_args: Any, **_kwds: Any) -> Any: + raise RuntimeError("This should never be called!") + + +def test_recursive_proxy() -> None: + proxy = RecursiveLazyProxy() + assert repr(proxy) == "RecursiveLazyProxy" + assert str(proxy) == "RecursiveLazyProxy" + assert dir(proxy) == [] + assert type(proxy).__name__ == "RecursiveLazyProxy" + assert type(operator.attrgetter("name.foo.bar.baz")(proxy)).__name__ == "RecursiveLazyProxy" + + +def test_isinstance_does_not_error() -> None: + class AlwaysErrorProxy(LazyProxy[Any]): + @override + def __load__(self) -> Any: + raise RuntimeError("Mocking missing dependency") + + proxy = AlwaysErrorProxy() + assert not isinstance(proxy, dict) + assert isinstance(proxy, LazyProxy) diff --git a/tests/test_utils/test_typing.py b/tests/test_utils/test_typing.py new file mode 100644 index 0000000..c2aca49 --- /dev/null +++ b/tests/test_utils/test_typing.py @@ -0,0 +1,73 @@ +from __future__ import annotations + +from typing import Generic, TypeVar, cast + +from brapi._utils import extract_type_var_from_base + +_T = TypeVar("_T") +_T2 = TypeVar("_T2") +_T3 = TypeVar("_T3") + + +class BaseGeneric(Generic[_T]): ... + + +class SubclassGeneric(BaseGeneric[_T]): ... + + +class BaseGenericMultipleTypeArgs(Generic[_T, _T2, _T3]): ... + + +class SubclassGenericMultipleTypeArgs(BaseGenericMultipleTypeArgs[_T, _T2, _T3]): ... + + +class SubclassDifferentOrderGenericMultipleTypeArgs(BaseGenericMultipleTypeArgs[_T2, _T, _T3]): ... + + +def test_extract_type_var() -> None: + assert ( + extract_type_var_from_base( + BaseGeneric[int], + index=0, + generic_bases=cast("tuple[type, ...]", (BaseGeneric,)), + ) + == int + ) + + +def test_extract_type_var_generic_subclass() -> None: + assert ( + extract_type_var_from_base( + SubclassGeneric[int], + index=0, + generic_bases=cast("tuple[type, ...]", (BaseGeneric,)), + ) + == int + ) + + +def test_extract_type_var_multiple() -> None: + typ = BaseGenericMultipleTypeArgs[int, str, None] + + generic_bases = cast("tuple[type, ...]", (BaseGenericMultipleTypeArgs,)) + assert extract_type_var_from_base(typ, index=0, generic_bases=generic_bases) == int + assert extract_type_var_from_base(typ, index=1, generic_bases=generic_bases) == str + assert extract_type_var_from_base(typ, index=2, generic_bases=generic_bases) == type(None) + + +def test_extract_type_var_generic_subclass_multiple() -> None: + typ = SubclassGenericMultipleTypeArgs[int, str, None] + + generic_bases = cast("tuple[type, ...]", (BaseGenericMultipleTypeArgs,)) + assert extract_type_var_from_base(typ, index=0, generic_bases=generic_bases) == int + assert extract_type_var_from_base(typ, index=1, generic_bases=generic_bases) == str + assert extract_type_var_from_base(typ, index=2, generic_bases=generic_bases) == type(None) + + +def test_extract_type_var_generic_subclass_different_ordering_multiple() -> None: + typ = SubclassDifferentOrderGenericMultipleTypeArgs[int, str, None] + + generic_bases = cast("tuple[type, ...]", (BaseGenericMultipleTypeArgs,)) + assert extract_type_var_from_base(typ, index=0, generic_bases=generic_bases) == int + assert extract_type_var_from_base(typ, index=1, generic_bases=generic_bases) == str + assert extract_type_var_from_base(typ, index=2, generic_bases=generic_bases) == type(None) diff --git a/tests/utils.py b/tests/utils.py new file mode 100644 index 0000000..008bc85 --- /dev/null +++ b/tests/utils.py @@ -0,0 +1,167 @@ +from __future__ import annotations + +import os +import inspect +import traceback +import contextlib +from typing import Any, TypeVar, Iterator, Sequence, cast +from datetime import date, datetime +from typing_extensions import Literal, get_args, get_origin, assert_type + +from brapi._types import Omit, NoneType +from brapi._utils import ( + is_dict, + is_list, + is_list_type, + is_union_type, + extract_type_arg, + is_sequence_type, + is_annotated_type, + is_type_alias_type, +) +from brapi._compat import PYDANTIC_V1, field_outer_type, get_model_fields +from brapi._models import BaseModel + +BaseModelT = TypeVar("BaseModelT", bound=BaseModel) + + +def assert_matches_model(model: type[BaseModelT], value: BaseModelT, *, path: list[str]) -> bool: + for name, field in get_model_fields(model).items(): + field_value = getattr(value, name) + if PYDANTIC_V1: + # in v1 nullability was structured differently + # https://docs.pydantic.dev/2.0/migration/#required-optional-and-nullable-fields + allow_none = getattr(field, "allow_none", False) + else: + allow_none = False + + assert_matches_type( + field_outer_type(field), + field_value, + path=[*path, name], + allow_none=allow_none, + ) + + return True + + +# Note: the `path` argument is only used to improve error messages when `--showlocals` is used +def assert_matches_type( + type_: Any, + value: object, + *, + path: list[str], + allow_none: bool = False, +) -> None: + if is_type_alias_type(type_): + type_ = type_.__value__ + + # unwrap `Annotated[T, ...]` -> `T` + if is_annotated_type(type_): + type_ = extract_type_arg(type_, 0) + + if allow_none and value is None: + return + + if type_ is None or type_ is NoneType: + assert value is None + return + + origin = get_origin(type_) or type_ + + if is_list_type(type_): + return _assert_list_type(type_, value) + + if is_sequence_type(type_): + assert isinstance(value, Sequence) + inner_type = get_args(type_)[0] + for entry in value: # type: ignore + assert_type(inner_type, entry) # type: ignore + return + + if origin == str: + assert isinstance(value, str) + elif origin == int: + assert isinstance(value, int) + elif origin == bool: + assert isinstance(value, bool) + elif origin == float: + assert isinstance(value, float) + elif origin == bytes: + assert isinstance(value, bytes) + elif origin == datetime: + assert isinstance(value, datetime) + elif origin == date: + assert isinstance(value, date) + elif origin == object: + # nothing to do here, the expected type is unknown + pass + elif origin == Literal: + assert value in get_args(type_) + elif origin == dict: + assert is_dict(value) + + args = get_args(type_) + key_type = args[0] + items_type = args[1] + + for key, item in value.items(): + assert_matches_type(key_type, key, path=[*path, ""]) + assert_matches_type(items_type, item, path=[*path, ""]) + elif is_union_type(type_): + variants = get_args(type_) + + try: + none_index = variants.index(type(None)) + except ValueError: + pass + else: + # special case Optional[T] for better error messages + if len(variants) == 2: + if value is None: + # valid + return + + return assert_matches_type(type_=variants[not none_index], value=value, path=path) + + for i, variant in enumerate(variants): + try: + assert_matches_type(variant, value, path=[*path, f"variant {i}"]) + return + except AssertionError: + traceback.print_exc() + continue + + raise AssertionError("Did not match any variants") + elif issubclass(origin, BaseModel): + assert isinstance(value, type_) + assert assert_matches_model(type_, cast(Any, value), path=path) + elif inspect.isclass(origin) and origin.__name__ == "HttpxBinaryResponseContent": + assert value.__class__.__name__ == "HttpxBinaryResponseContent" + else: + assert None, f"Unhandled field type: {type_}" + + +def _assert_list_type(type_: type[object], value: object) -> None: + assert is_list(value) + + inner_type = get_args(type_)[0] + for entry in value: + assert_type(inner_type, entry) # type: ignore + + +@contextlib.contextmanager +def update_env(**new_env: str | Omit) -> Iterator[None]: + old = os.environ.copy() + + try: + for name, value in new_env.items(): + if isinstance(value, Omit): + os.environ.pop(name, None) + else: + os.environ[name] = value + + yield None + finally: + os.environ.clear() + os.environ.update(old) From 3d4efc9a676a07aa511725ff992312db9f03633a Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:26:00 +0000 Subject: [PATCH 03/47] codegen metadata --- .stats.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.stats.yml b/.stats.yml index ec152aa..9e32c91 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 -config_hash: 6de3ea33802724ce95047bc7f2a45358 +config_hash: c8b63d688f18091e728375c9dc421428 From da1aa34b51255e3b4800f0cce60c7463ae0089a3 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:26:18 +0000 Subject: [PATCH 04/47] codegen metadata --- .stats.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.stats.yml b/.stats.yml index 9e32c91..ec152aa 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 -config_hash: c8b63d688f18091e728375c9dc421428 +config_hash: 6de3ea33802724ce95047bc7f2a45358 From 58a2faf9dc8ad7f3c7464afcaa41a5efbf58aa6f Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:26:41 +0000 Subject: [PATCH 05/47] chore: update SDK settings --- .github/workflows/publish-pypi.yml | 31 +++++++++++++ .github/workflows/release-doctor.yml | 21 +++++++++ .release-please-manifest.json | 3 ++ .stats.yml | 2 +- CONTRIBUTING.md | 4 +- README.md | 14 +++--- bin/check-release-environment | 21 +++++++++ pyproject.toml | 6 +-- release-please-config.json | 66 ++++++++++++++++++++++++++++ src/brapi/_version.py | 2 +- src/brapi/resources/available.py | 8 ++-- src/brapi/resources/quote.py | 8 ++-- src/brapi/resources/v2/crypto.py | 8 ++-- src/brapi/resources/v2/currency.py | 8 ++-- src/brapi/resources/v2/inflation.py | 8 ++-- src/brapi/resources/v2/prime_rate.py | 8 ++-- src/brapi/resources/v2/v2.py | 8 ++-- 17 files changed, 184 insertions(+), 42 deletions(-) create mode 100644 .github/workflows/publish-pypi.yml create mode 100644 .github/workflows/release-doctor.yml create mode 100644 .release-please-manifest.json create mode 100644 bin/check-release-environment create mode 100644 release-please-config.json diff --git a/.github/workflows/publish-pypi.yml b/.github/workflows/publish-pypi.yml new file mode 100644 index 0000000..8181d42 --- /dev/null +++ b/.github/workflows/publish-pypi.yml @@ -0,0 +1,31 @@ +# This workflow is triggered when a GitHub release is created. +# It can also be run manually to re-publish to PyPI in case it failed for some reason. +# You can run this workflow by navigating to https://www.github.com/brapi-dev/brapi-python/actions/workflows/publish-pypi.yml +name: Publish PyPI +on: + workflow_dispatch: + + release: + types: [published] + +jobs: + publish: + name: publish + runs-on: ubuntu-latest + + steps: + - uses: actions/checkout@v4 + + - name: Install Rye + run: | + curl -sSf https://rye.astral.sh/get | bash + echo "$HOME/.rye/shims" >> $GITHUB_PATH + env: + RYE_VERSION: '0.44.0' + RYE_INSTALL_OPTION: '--yes' + + - name: Publish to PyPI + run: | + bash ./bin/publish-pypi + env: + PYPI_TOKEN: ${{ secrets.BRAPI_PYPI_TOKEN || secrets.PYPI_TOKEN }} diff --git a/.github/workflows/release-doctor.yml b/.github/workflows/release-doctor.yml new file mode 100644 index 0000000..790ebde --- /dev/null +++ b/.github/workflows/release-doctor.yml @@ -0,0 +1,21 @@ +name: Release Doctor +on: + pull_request: + branches: + - main + workflow_dispatch: + +jobs: + release_doctor: + name: release doctor + runs-on: ubuntu-latest + if: github.repository == 'brapi-dev/brapi-python' && (github.event_name == 'push' || github.event_name == 'workflow_dispatch' || startsWith(github.head_ref, 'release-please') || github.head_ref == 'next') + + steps: + - uses: actions/checkout@v4 + + - name: Check release environment + run: | + bash ./bin/check-release-environment + env: + PYPI_TOKEN: ${{ secrets.BRAPI_PYPI_TOKEN || secrets.PYPI_TOKEN }} diff --git a/.release-please-manifest.json b/.release-please-manifest.json new file mode 100644 index 0000000..1332969 --- /dev/null +++ b/.release-please-manifest.json @@ -0,0 +1,3 @@ +{ + ".": "0.0.1" +} \ No newline at end of file diff --git a/.stats.yml b/.stats.yml index ec152aa..9ec36be 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 -config_hash: 6de3ea33802724ce95047bc7f2a45358 +config_hash: c64d73bbc698643552178b1152a09f8e diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 8266ffe..5068a3b 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -62,7 +62,7 @@ If you’d like to use the repository from source, you can either install from g To install via git: ```sh -$ pip install git+ssh://git@github.com/stainless-sdks/brapi-python.git +$ pip install git+ssh://git@github.com/brapi-dev/brapi-python.git ``` Alternatively, you can build from source and install the wheel file: @@ -120,7 +120,7 @@ the changes aren't made through the automated pipeline, you may want to make rel ### Publish with a GitHub workflow -You can release to package managers by using [the `Publish PyPI` GitHub action](https://www.github.com/stainless-sdks/brapi-python/actions/workflows/publish-pypi.yml). This requires a setup organization or repository secret to be set up. +You can release to package managers by using [the `Publish PyPI` GitHub action](https://www.github.com/brapi-dev/brapi-python/actions/workflows/publish-pypi.yml). This requires a setup organization or repository secret to be set up. ### Publish manually diff --git a/README.md b/README.md index b0480b6..77a8a02 100644 --- a/README.md +++ b/README.md @@ -16,8 +16,8 @@ The REST API documentation can be found on [brapi.dev](https://brapi.dev). The f ## Installation ```sh -# install from this staging repo -pip install git+ssh://git@github.com/stainless-sdks/brapi-python.git +# install from the production repo +pip install git+ssh://git@github.com/brapi-dev/brapi-python.git ``` > [!NOTE] @@ -83,8 +83,8 @@ By default, the async client uses `httpx` for HTTP requests. However, for improv You can enable this by installing `aiohttp`: ```sh -# install from this staging repo -pip install 'brapi[aiohttp] @ git+ssh://git@github.com/stainless-sdks/brapi-python.git' +# install from the production repo +pip install 'brapi[aiohttp] @ git+ssh://git@github.com/brapi-dev/brapi-python.git' ``` Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: @@ -256,9 +256,9 @@ quote = response.parse() # get the object that `quote.retrieve()` would have re print(quote.requested_at) ``` -These methods return an [`APIResponse`](https://github.com/stainless-sdks/brapi-python/tree/main/src/brapi/_response.py) object. +These methods return an [`APIResponse`](https://github.com/brapi-dev/brapi-python/tree/main/src/brapi/_response.py) object. -The async client returns an [`AsyncAPIResponse`](https://github.com/stainless-sdks/brapi-python/tree/main/src/brapi/_response.py) with the same structure, the only difference being `await`able methods for reading the response content. +The async client returns an [`AsyncAPIResponse`](https://github.com/brapi-dev/brapi-python/tree/main/src/brapi/_response.py) with the same structure, the only difference being `await`able methods for reading the response content. #### `.with_streaming_response` @@ -364,7 +364,7 @@ This package generally follows [SemVer](https://semver.org/spec/v2.0.0.html) con We take backwards-compatibility seriously and work hard to ensure you can rely on a smooth upgrade experience. -We are keen for your feedback; please open an [issue](https://www.github.com/stainless-sdks/brapi-python/issues) with questions, bugs, or suggestions. +We are keen for your feedback; please open an [issue](https://www.github.com/brapi-dev/brapi-python/issues) with questions, bugs, or suggestions. ### Determining the installed version diff --git a/bin/check-release-environment b/bin/check-release-environment new file mode 100644 index 0000000..b845b0f --- /dev/null +++ b/bin/check-release-environment @@ -0,0 +1,21 @@ +#!/usr/bin/env bash + +errors=() + +if [ -z "${PYPI_TOKEN}" ]; then + errors+=("The PYPI_TOKEN secret has not been set. Please set it in either this repository's secrets or your organization secrets.") +fi + +lenErrors=${#errors[@]} + +if [[ lenErrors -gt 0 ]]; then + echo -e "Found the following errors in the release environment:\n" + + for error in "${errors[@]}"; do + echo -e "- $error\n" + done + + exit 1 +fi + +echo "The environment is ready to push releases!" diff --git a/pyproject.toml b/pyproject.toml index 968a4bf..99fd66f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -35,8 +35,8 @@ classifiers = [ ] [project.urls] -Homepage = "https://github.com/stainless-sdks/brapi-python" -Repository = "https://github.com/stainless-sdks/brapi-python" +Homepage = "https://github.com/brapi-dev/brapi-python" +Repository = "https://github.com/brapi-dev/brapi-python" [project.optional-dependencies] aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.8"] @@ -124,7 +124,7 @@ path = "README.md" [[tool.hatch.metadata.hooks.fancy-pypi-readme.substitutions]] # replace relative links with absolute links pattern = '\[(.+?)\]\(((?!https?://)\S+?)\)' -replacement = '[\1](https://github.com/stainless-sdks/brapi-python/tree/main/\g<2>)' +replacement = '[\1](https://github.com/brapi-dev/brapi-python/tree/main/\g<2>)' [tool.pytest.ini_options] testpaths = ["tests"] diff --git a/release-please-config.json b/release-please-config.json new file mode 100644 index 0000000..f365cd0 --- /dev/null +++ b/release-please-config.json @@ -0,0 +1,66 @@ +{ + "packages": { + ".": {} + }, + "$schema": "https://raw.githubusercontent.com/stainless-api/release-please/main/schemas/config.json", + "include-v-in-tag": true, + "include-component-in-tag": false, + "versioning": "prerelease", + "prerelease": true, + "bump-minor-pre-major": true, + "bump-patch-for-minor-pre-major": false, + "pull-request-header": "Automated Release PR", + "pull-request-title-pattern": "release: ${version}", + "changelog-sections": [ + { + "type": "feat", + "section": "Features" + }, + { + "type": "fix", + "section": "Bug Fixes" + }, + { + "type": "perf", + "section": "Performance Improvements" + }, + { + "type": "revert", + "section": "Reverts" + }, + { + "type": "chore", + "section": "Chores" + }, + { + "type": "docs", + "section": "Documentation" + }, + { + "type": "style", + "section": "Styles" + }, + { + "type": "refactor", + "section": "Refactors" + }, + { + "type": "test", + "section": "Tests", + "hidden": true + }, + { + "type": "build", + "section": "Build System" + }, + { + "type": "ci", + "section": "Continuous Integration", + "hidden": true + } + ], + "release-type": "python", + "extra-files": [ + "src/brapi/_version.py" + ] +} \ No newline at end of file diff --git a/src/brapi/_version.py b/src/brapi/_version.py index 905eebf..a7dcbe3 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "0.0.1" +__version__ = "0.0.1" # x-release-please-version diff --git a/src/brapi/resources/available.py b/src/brapi/resources/available.py index 7c65696..26e04de 100644 --- a/src/brapi/resources/available.py +++ b/src/brapi/resources/available.py @@ -28,7 +28,7 @@ def with_raw_response(self) -> AvailableResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AvailableResourceWithRawResponse(self) @@ -37,7 +37,7 @@ def with_streaming_response(self) -> AvailableResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AvailableResourceWithStreamingResponse(self) @@ -141,7 +141,7 @@ def with_raw_response(self) -> AsyncAvailableResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncAvailableResourceWithRawResponse(self) @@ -150,7 +150,7 @@ def with_streaming_response(self) -> AsyncAvailableResourceWithStreamingResponse """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncAvailableResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/quote.py b/src/brapi/resources/quote.py index 48da3f8..ba428d5 100644 --- a/src/brapi/resources/quote.py +++ b/src/brapi/resources/quote.py @@ -32,7 +32,7 @@ def with_raw_response(self) -> QuoteResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return QuoteResourceWithRawResponse(self) @@ -41,7 +41,7 @@ def with_streaming_response(self) -> QuoteResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return QuoteResourceWithStreamingResponse(self) @@ -475,7 +475,7 @@ def with_raw_response(self) -> AsyncQuoteResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncQuoteResourceWithRawResponse(self) @@ -484,7 +484,7 @@ def with_streaming_response(self) -> AsyncQuoteResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncQuoteResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/v2/crypto.py b/src/brapi/resources/v2/crypto.py index c4db8df..317e8d9 100644 --- a/src/brapi/resources/v2/crypto.py +++ b/src/brapi/resources/v2/crypto.py @@ -31,7 +31,7 @@ def with_raw_response(self) -> CryptoResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return CryptoResourceWithRawResponse(self) @@ -40,7 +40,7 @@ def with_streaming_response(self) -> CryptoResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return CryptoResourceWithStreamingResponse(self) @@ -257,7 +257,7 @@ def with_raw_response(self) -> AsyncCryptoResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncCryptoResourceWithRawResponse(self) @@ -266,7 +266,7 @@ def with_streaming_response(self) -> AsyncCryptoResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncCryptoResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/v2/currency.py b/src/brapi/resources/v2/currency.py index e082a6d..2195980 100644 --- a/src/brapi/resources/v2/currency.py +++ b/src/brapi/resources/v2/currency.py @@ -29,7 +29,7 @@ def with_raw_response(self) -> CurrencyResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return CurrencyResourceWithRawResponse(self) @@ -38,7 +38,7 @@ def with_streaming_response(self) -> CurrencyResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return CurrencyResourceWithStreamingResponse(self) @@ -222,7 +222,7 @@ def with_raw_response(self) -> AsyncCurrencyResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncCurrencyResourceWithRawResponse(self) @@ -231,7 +231,7 @@ def with_streaming_response(self) -> AsyncCurrencyResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncCurrencyResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/v2/inflation.py b/src/brapi/resources/v2/inflation.py index b0e1138..7098b22 100644 --- a/src/brapi/resources/v2/inflation.py +++ b/src/brapi/resources/v2/inflation.py @@ -33,7 +33,7 @@ def with_raw_response(self) -> InflationResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return InflationResourceWithRawResponse(self) @@ -42,7 +42,7 @@ def with_streaming_response(self) -> InflationResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return InflationResourceWithStreamingResponse(self) @@ -261,7 +261,7 @@ def with_raw_response(self) -> AsyncInflationResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncInflationResourceWithRawResponse(self) @@ -270,7 +270,7 @@ def with_streaming_response(self) -> AsyncInflationResourceWithStreamingResponse """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncInflationResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/v2/prime_rate.py b/src/brapi/resources/v2/prime_rate.py index 21d9e81..0dedc9b 100644 --- a/src/brapi/resources/v2/prime_rate.py +++ b/src/brapi/resources/v2/prime_rate.py @@ -33,7 +33,7 @@ def with_raw_response(self) -> PrimeRateResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return PrimeRateResourceWithRawResponse(self) @@ -42,7 +42,7 @@ def with_streaming_response(self) -> PrimeRateResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return PrimeRateResourceWithStreamingResponse(self) @@ -241,7 +241,7 @@ def with_raw_response(self) -> AsyncPrimeRateResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncPrimeRateResourceWithRawResponse(self) @@ -250,7 +250,7 @@ def with_streaming_response(self) -> AsyncPrimeRateResourceWithStreamingResponse """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncPrimeRateResourceWithStreamingResponse(self) diff --git a/src/brapi/resources/v2/v2.py b/src/brapi/resources/v2/v2.py index 54e46f1..3fa73c9 100644 --- a/src/brapi/resources/v2/v2.py +++ b/src/brapi/resources/v2/v2.py @@ -63,7 +63,7 @@ def with_raw_response(self) -> V2ResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return V2ResourceWithRawResponse(self) @@ -72,7 +72,7 @@ def with_streaming_response(self) -> V2ResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return V2ResourceWithStreamingResponse(self) @@ -100,7 +100,7 @@ def with_raw_response(self) -> AsyncV2ResourceWithRawResponse: This property can be used as a prefix for any HTTP method call to return the raw response object instead of the parsed content. - For more information, see https://www.github.com/stainless-sdks/brapi-python#accessing-raw-response-data-eg-headers + For more information, see https://www.github.com/brapi-dev/brapi-python#accessing-raw-response-data-eg-headers """ return AsyncV2ResourceWithRawResponse(self) @@ -109,7 +109,7 @@ def with_streaming_response(self) -> AsyncV2ResourceWithStreamingResponse: """ An alternative to `.with_raw_response` that doesn't eagerly read the response body. - For more information, see https://www.github.com/stainless-sdks/brapi-python#with_streaming_response + For more information, see https://www.github.com/brapi-dev/brapi-python#with_streaming_response """ return AsyncV2ResourceWithStreamingResponse(self) From 029d08bd9eae5f37f91c921823b297c1d7ffbf7a Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:27:01 +0000 Subject: [PATCH 06/47] chore: update SDK settings --- .stats.yml | 2 +- README.md | 11 ++++------- 2 files changed, 5 insertions(+), 8 deletions(-) diff --git a/.stats.yml b/.stats.yml index 9ec36be..00e0370 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 -config_hash: c64d73bbc698643552178b1152a09f8e +config_hash: 6f10a67950f65bf850612b59838ad03b diff --git a/README.md b/README.md index 77a8a02..4d5b6c5 100644 --- a/README.md +++ b/README.md @@ -16,13 +16,10 @@ The REST API documentation can be found on [brapi.dev](https://brapi.dev). The f ## Installation ```sh -# install from the production repo -pip install git+ssh://git@github.com/brapi-dev/brapi-python.git +# install from PyPI +pip install brapi ``` -> [!NOTE] -> Once this package is [published to PyPI](https://www.stainless.com/docs/guides/publish), this will become: `pip install brapi` - ## Usage The full API of this library can be found in [api.md](api.md). @@ -83,8 +80,8 @@ By default, the async client uses `httpx` for HTTP requests. However, for improv You can enable this by installing `aiohttp`: ```sh -# install from the production repo -pip install 'brapi[aiohttp] @ git+ssh://git@github.com/brapi-dev/brapi-python.git' +# install from PyPI +pip install brapi[aiohttp] ``` Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: From 930f75be9238606243217a2df7112a76bbacda63 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sun, 12 Oct 2025 18:34:33 +0000 Subject: [PATCH 07/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 1332969..fea3454 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "0.0.1" + ".": "1.0.0" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index 99fd66f..86d0287 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "0.0.1" +version = "1.0.0" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index a7dcbe3..d163e19 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "0.0.1" # x-release-please-version +__version__ = "1.0.0" # x-release-please-version From e1ad501c7cdb003db65125ae3d401d3a66a18db6 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sat, 18 Oct 2025 02:17:59 +0000 Subject: [PATCH 08/47] chore: bump `httpx-aiohttp` version to 0.1.9 --- pyproject.toml | 2 +- requirements-dev.lock | 2 +- requirements.lock | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 86d0287..b50829e 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -39,7 +39,7 @@ Homepage = "https://github.com/brapi-dev/brapi-python" Repository = "https://github.com/brapi-dev/brapi-python" [project.optional-dependencies] -aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.8"] +aiohttp = ["aiohttp", "httpx_aiohttp>=0.1.9"] [tool.rye] managed = true diff --git a/requirements-dev.lock b/requirements-dev.lock index 6f06451..3050a50 100644 --- a/requirements-dev.lock +++ b/requirements-dev.lock @@ -56,7 +56,7 @@ httpx==0.28.1 # via brapi # via httpx-aiohttp # via respx -httpx-aiohttp==0.1.8 +httpx-aiohttp==0.1.9 # via brapi idna==3.4 # via anyio diff --git a/requirements.lock b/requirements.lock index 6018e55..cf79912 100644 --- a/requirements.lock +++ b/requirements.lock @@ -43,7 +43,7 @@ httpcore==1.0.9 httpx==0.28.1 # via brapi # via httpx-aiohttp -httpx-aiohttp==0.1.8 +httpx-aiohttp==0.1.9 # via brapi idna==3.4 # via anyio From 543c21737ba9fbe790e9865a37135a16e69ea333 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 28 Oct 2025 19:55:19 +0000 Subject: [PATCH 09/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index fea3454..a8f7122 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.0" + ".": "1.0.1" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index b50829e..8fe119f 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.0" +version = "1.0.1" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index d163e19..9269671 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.0" # x-release-please-version +__version__ = "1.0.1" # x-release-please-version From b3444f2ba543983a76a2e63697f3e3224394ac79 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 30 Oct 2025 02:32:26 +0000 Subject: [PATCH 10/47] fix(client): close streams without requiring full consumption --- src/brapi/_streaming.py | 10 ++++------ 1 file changed, 4 insertions(+), 6 deletions(-) diff --git a/src/brapi/_streaming.py b/src/brapi/_streaming.py index b9d6d1c..d0ff08c 100644 --- a/src/brapi/_streaming.py +++ b/src/brapi/_streaming.py @@ -57,9 +57,8 @@ def __stream__(self) -> Iterator[_T]: for sse in iterator: yield process_data(data=sse.json(), cast_to=cast_to, response=response) - # Ensure the entire stream is consumed - for _sse in iterator: - ... + # As we might not fully consume the response stream, we need to close it explicitly + response.close() def __enter__(self) -> Self: return self @@ -121,9 +120,8 @@ async def __stream__(self) -> AsyncIterator[_T]: async for sse in iterator: yield process_data(data=sse.json(), cast_to=cast_to, response=response) - # Ensure the entire stream is consumed - async for _sse in iterator: - ... + # As we might not fully consume the response stream, we need to close it explicitly + await response.aclose() async def __aenter__(self) -> Self: return self From e4decabae07cc0f2a8b18149b1de284f31708b81 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 31 Oct 2025 03:23:28 +0000 Subject: [PATCH 11/47] chore(internal/tests): avoid race condition with implicit client cleanup --- tests/test_client.py | 366 ++++++++++++++++++++++++------------------- 1 file changed, 202 insertions(+), 164 deletions(-) diff --git a/tests/test_client.py b/tests/test_client.py index 0ef232e..f57ce48 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -59,51 +59,49 @@ def _get_open_connections(client: Brapi | AsyncBrapi) -> int: class TestBrapi: - client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - @pytest.mark.respx(base_url=base_url) - def test_raw_response(self, respx_mock: MockRouter) -> None: + def test_raw_response(self, respx_mock: MockRouter, client: Brapi) -> None: respx_mock.post("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = self.client.post("/foo", cast_to=httpx.Response) + response = client.post("/foo", cast_to=httpx.Response) assert response.status_code == 200 assert isinstance(response, httpx.Response) assert response.json() == {"foo": "bar"} @pytest.mark.respx(base_url=base_url) - def test_raw_response_for_binary(self, respx_mock: MockRouter) -> None: + def test_raw_response_for_binary(self, respx_mock: MockRouter, client: Brapi) -> None: respx_mock.post("/foo").mock( return_value=httpx.Response(200, headers={"Content-Type": "application/binary"}, content='{"foo": "bar"}') ) - response = self.client.post("/foo", cast_to=httpx.Response) + response = client.post("/foo", cast_to=httpx.Response) assert response.status_code == 200 assert isinstance(response, httpx.Response) assert response.json() == {"foo": "bar"} - def test_copy(self) -> None: - copied = self.client.copy() - assert id(copied) != id(self.client) + def test_copy(self, client: Brapi) -> None: + copied = client.copy() + assert id(copied) != id(client) - copied = self.client.copy(api_key="another My API Key") + copied = client.copy(api_key="another My API Key") assert copied.api_key == "another My API Key" - assert self.client.api_key == "My API Key" + assert client.api_key == "My API Key" - def test_copy_default_options(self) -> None: + def test_copy_default_options(self, client: Brapi) -> None: # options that have a default are overridden correctly - copied = self.client.copy(max_retries=7) + copied = client.copy(max_retries=7) assert copied.max_retries == 7 - assert self.client.max_retries == 2 + assert client.max_retries == 2 copied2 = copied.copy(max_retries=6) assert copied2.max_retries == 6 assert copied.max_retries == 7 # timeout - assert isinstance(self.client.timeout, httpx.Timeout) - copied = self.client.copy(timeout=None) + assert isinstance(client.timeout, httpx.Timeout) + copied = client.copy(timeout=None) assert copied.timeout is None - assert isinstance(self.client.timeout, httpx.Timeout) + assert isinstance(client.timeout, httpx.Timeout) def test_copy_default_headers(self) -> None: client = Brapi( @@ -138,6 +136,7 @@ def test_copy_default_headers(self) -> None: match="`default_headers` and `set_default_headers` arguments are mutually exclusive", ): client.copy(set_default_headers={}, default_headers={"X-Foo": "Bar"}) + client.close() def test_copy_default_query(self) -> None: client = Brapi( @@ -175,13 +174,15 @@ def test_copy_default_query(self) -> None: ): client.copy(set_default_query={}, default_query={"foo": "Bar"}) - def test_copy_signature(self) -> None: + client.close() + + def test_copy_signature(self, client: Brapi) -> None: # ensure the same parameters that can be passed to the client are defined in the `.copy()` method init_signature = inspect.signature( # mypy doesn't like that we access the `__init__` property. - self.client.__init__, # type: ignore[misc] + client.__init__, # type: ignore[misc] ) - copy_signature = inspect.signature(self.client.copy) + copy_signature = inspect.signature(client.copy) exclude_params = {"transport", "proxies", "_strict_response_validation"} for name in init_signature.parameters.keys(): @@ -192,12 +193,12 @@ def test_copy_signature(self) -> None: assert copy_param is not None, f"copy() signature is missing the {name} param" @pytest.mark.skipif(sys.version_info >= (3, 10), reason="fails because of a memory leak that started from 3.12") - def test_copy_build_request(self) -> None: + def test_copy_build_request(self, client: Brapi) -> None: options = FinalRequestOptions(method="get", url="/foo") def build_request(options: FinalRequestOptions) -> None: - client = self.client.copy() - client._build_request(options) + client_copy = client.copy() + client_copy._build_request(options) # ensure that the machinery is warmed up before tracing starts. build_request(options) @@ -254,14 +255,12 @@ def add_leak(leaks: list[tracemalloc.StatisticDiff], diff: tracemalloc.Statistic print(frame) raise AssertionError() - def test_request_timeout(self) -> None: - request = self.client._build_request(FinalRequestOptions(method="get", url="/foo")) + def test_request_timeout(self, client: Brapi) -> None: + request = client._build_request(FinalRequestOptions(method="get", url="/foo")) timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT - request = self.client._build_request( - FinalRequestOptions(method="get", url="/foo", timeout=httpx.Timeout(100.0)) - ) + request = client._build_request(FinalRequestOptions(method="get", url="/foo", timeout=httpx.Timeout(100.0))) timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == httpx.Timeout(100.0) @@ -272,6 +271,8 @@ def test_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == httpx.Timeout(0) + client.close() + def test_http_client_timeout_option(self) -> None: # custom timeout given to the httpx client should be used with httpx.Client(timeout=None) as http_client: @@ -283,6 +284,8 @@ def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == httpx.Timeout(None) + client.close() + # no timeout given to the httpx client should not use the httpx default with httpx.Client() as http_client: client = Brapi( @@ -293,6 +296,8 @@ def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT + client.close() + # explicitly passing the default timeout currently results in it being ignored with httpx.Client(timeout=HTTPX_DEFAULT_TIMEOUT) as http_client: client = Brapi( @@ -303,6 +308,8 @@ def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT # our default + client.close() + async def test_invalid_http_client(self) -> None: with pytest.raises(TypeError, match="Invalid `http_client` arg"): async with httpx.AsyncClient() as http_client: @@ -314,14 +321,14 @@ async def test_invalid_http_client(self) -> None: ) def test_default_headers_option(self) -> None: - client = Brapi( + test_client = Brapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} ) - request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + request = test_client._build_request(FinalRequestOptions(method="get", url="/foo")) assert request.headers.get("x-foo") == "bar" assert request.headers.get("x-stainless-lang") == "python" - client2 = Brapi( + test_client2 = Brapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, @@ -330,10 +337,13 @@ def test_default_headers_option(self) -> None: "X-Stainless-Lang": "my-overriding-header", }, ) - request = client2._build_request(FinalRequestOptions(method="get", url="/foo")) + request = test_client2._build_request(FinalRequestOptions(method="get", url="/foo")) assert request.headers.get("x-foo") == "stainless" assert request.headers.get("x-stainless-lang") == "my-overriding-header" + test_client.close() + test_client2.close() + def test_validate_headers(self) -> None: client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) request = client._build_request(FinalRequestOptions(method="get", url="/foo")) @@ -362,8 +372,10 @@ def test_default_query_option(self) -> None: url = httpx.URL(request.url) assert dict(url.params) == {"foo": "baz", "query_param": "overridden"} - def test_request_extra_json(self) -> None: - request = self.client._build_request( + client.close() + + def test_request_extra_json(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -374,7 +386,7 @@ def test_request_extra_json(self) -> None: data = json.loads(request.content.decode("utf-8")) assert data == {"foo": "bar", "baz": False} - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -385,7 +397,7 @@ def test_request_extra_json(self) -> None: assert data == {"baz": False} # `extra_json` takes priority over `json_data` when keys clash - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -396,8 +408,8 @@ def test_request_extra_json(self) -> None: data = json.loads(request.content.decode("utf-8")) assert data == {"foo": "bar", "baz": None} - def test_request_extra_headers(self) -> None: - request = self.client._build_request( + def test_request_extra_headers(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -407,7 +419,7 @@ def test_request_extra_headers(self) -> None: assert request.headers.get("X-Foo") == "Foo" # `extra_headers` takes priority over `default_headers` when keys clash - request = self.client.with_options(default_headers={"X-Bar": "true"})._build_request( + request = client.with_options(default_headers={"X-Bar": "true"})._build_request( FinalRequestOptions( method="post", url="/foo", @@ -418,8 +430,8 @@ def test_request_extra_headers(self) -> None: ) assert request.headers.get("X-Bar") == "false" - def test_request_extra_query(self) -> None: - request = self.client._build_request( + def test_request_extra_query(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -432,7 +444,7 @@ def test_request_extra_query(self) -> None: assert params == {"my_query_param": "Foo"} # if both `query` and `extra_query` are given, they are merged - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -446,7 +458,7 @@ def test_request_extra_query(self) -> None: assert params == {"bar": "1", "foo": "2"} # `extra_query` takes priority over `query` when keys clash - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -489,7 +501,7 @@ def test_multipart_repeating_array(self, client: Brapi) -> None: ] @pytest.mark.respx(base_url=base_url) - def test_basic_union_response(self, respx_mock: MockRouter) -> None: + def test_basic_union_response(self, respx_mock: MockRouter, client: Brapi) -> None: class Model1(BaseModel): name: str @@ -498,12 +510,12 @@ class Model2(BaseModel): respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model2) assert response.foo == "bar" @pytest.mark.respx(base_url=base_url) - def test_union_response_different_types(self, respx_mock: MockRouter) -> None: + def test_union_response_different_types(self, respx_mock: MockRouter, client: Brapi) -> None: """Union of objects with the same field name using a different type""" class Model1(BaseModel): @@ -514,18 +526,18 @@ class Model2(BaseModel): respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model2) assert response.foo == "bar" respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": 1})) - response = self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model1) assert response.foo == 1 @pytest.mark.respx(base_url=base_url) - def test_non_application_json_content_type_for_json_data(self, respx_mock: MockRouter) -> None: + def test_non_application_json_content_type_for_json_data(self, respx_mock: MockRouter, client: Brapi) -> None: """ Response that sets Content-Type to something other than application/json but returns json data """ @@ -541,7 +553,7 @@ class Model(BaseModel): ) ) - response = self.client.get("/foo", cast_to=Model) + response = client.get("/foo", cast_to=Model) assert isinstance(response, Model) assert response.foo == 2 @@ -553,6 +565,8 @@ def test_base_url_setter(self) -> None: assert client.base_url == "https://example.com/from_setter/" + client.close() + def test_base_url_env(self) -> None: with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): client = Brapi(api_key=api_key, _strict_response_validation=True) @@ -566,6 +580,8 @@ def test_base_url_env(self) -> None: client = Brapi(base_url=None, api_key=api_key, _strict_response_validation=True, environment="production") assert str(client.base_url).startswith("https://brapi.dev") + client.close() + @pytest.mark.parametrize( "client", [ @@ -588,6 +604,7 @@ def test_base_url_trailing_slash(self, client: Brapi) -> None: ), ) assert request.url == "http://localhost:5000/custom/path/foo" + client.close() @pytest.mark.parametrize( "client", @@ -611,6 +628,7 @@ def test_base_url_no_trailing_slash(self, client: Brapi) -> None: ), ) assert request.url == "http://localhost:5000/custom/path/foo" + client.close() @pytest.mark.parametrize( "client", @@ -634,35 +652,36 @@ def test_absolute_request_url(self, client: Brapi) -> None: ), ) assert request.url == "https://myapi.com/foo" + client.close() def test_copied_client_does_not_close_http(self) -> None: - client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - assert not client.is_closed() + test_client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + assert not test_client.is_closed() - copied = client.copy() - assert copied is not client + copied = test_client.copy() + assert copied is not test_client del copied - assert not client.is_closed() + assert not test_client.is_closed() def test_client_context_manager(self) -> None: - client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - with client as c2: - assert c2 is client + test_client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + with test_client as c2: + assert c2 is test_client assert not c2.is_closed() - assert not client.is_closed() - assert client.is_closed() + assert not test_client.is_closed() + assert test_client.is_closed() @pytest.mark.respx(base_url=base_url) - def test_client_response_validation_error(self, respx_mock: MockRouter) -> None: + def test_client_response_validation_error(self, respx_mock: MockRouter, client: Brapi) -> None: class Model(BaseModel): foo: str respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": {"invalid": True}})) with pytest.raises(APIResponseValidationError) as exc: - self.client.get("/foo", cast_to=Model) + client.get("/foo", cast_to=Model) assert isinstance(exc.value.__cause__, ValidationError) @@ -682,11 +701,14 @@ class Model(BaseModel): with pytest.raises(APIResponseValidationError): strict_client.get("/foo", cast_to=Model) - client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) + non_strict_client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) - response = client.get("/foo", cast_to=Model) + response = non_strict_client.get("/foo", cast_to=Model) assert isinstance(response, str) # type: ignore[unreachable] + strict_client.close() + non_strict_client.close() + @pytest.mark.parametrize( "remaining_retries,retry_after,timeout", [ @@ -709,9 +731,9 @@ class Model(BaseModel): ], ) @mock.patch("time.time", mock.MagicMock(return_value=1696004797)) - def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str, timeout: float) -> None: - client = Brapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - + def test_parse_retry_after_header( + self, remaining_retries: int, retry_after: str, timeout: float, client: Brapi + ) -> None: headers = httpx.Headers({"retry-after": retry_after}) options = FinalRequestOptions(method="get", url="/foo", max_retries=3) calculated = client._calculate_retry_timeout(remaining_retries, options, headers) @@ -725,7 +747,7 @@ def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter, clien with pytest.raises(APITimeoutError): client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__enter__() - assert _get_open_connections(self.client) == 0 + assert _get_open_connections(client) == 0 @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @pytest.mark.respx(base_url=base_url) @@ -734,7 +756,7 @@ def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter, client with pytest.raises(APIStatusError): client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__enter__() - assert _get_open_connections(self.client) == 0 + assert _get_open_connections(client) == 0 @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @@ -838,83 +860,77 @@ def test_default_client_creation(self) -> None: ) @pytest.mark.respx(base_url=base_url) - def test_follow_redirects(self, respx_mock: MockRouter) -> None: + def test_follow_redirects(self, respx_mock: MockRouter, client: Brapi) -> None: # Test that the default follow_redirects=True allows following redirects respx_mock.post("/redirect").mock( return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) ) respx_mock.get("/redirected").mock(return_value=httpx.Response(200, json={"status": "ok"})) - response = self.client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) + response = client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) assert response.status_code == 200 assert response.json() == {"status": "ok"} @pytest.mark.respx(base_url=base_url) - def test_follow_redirects_disabled(self, respx_mock: MockRouter) -> None: + def test_follow_redirects_disabled(self, respx_mock: MockRouter, client: Brapi) -> None: # Test that follow_redirects=False prevents following redirects respx_mock.post("/redirect").mock( return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) ) with pytest.raises(APIStatusError) as exc_info: - self.client.post( - "/redirect", body={"key": "value"}, options={"follow_redirects": False}, cast_to=httpx.Response - ) + client.post("/redirect", body={"key": "value"}, options={"follow_redirects": False}, cast_to=httpx.Response) assert exc_info.value.response.status_code == 302 assert exc_info.value.response.headers["Location"] == f"{base_url}/redirected" class TestAsyncBrapi: - client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio - async def test_raw_response(self, respx_mock: MockRouter) -> None: + async def test_raw_response(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: respx_mock.post("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = await self.client.post("/foo", cast_to=httpx.Response) + response = await async_client.post("/foo", cast_to=httpx.Response) assert response.status_code == 200 assert isinstance(response, httpx.Response) assert response.json() == {"foo": "bar"} @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio - async def test_raw_response_for_binary(self, respx_mock: MockRouter) -> None: + async def test_raw_response_for_binary(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: respx_mock.post("/foo").mock( return_value=httpx.Response(200, headers={"Content-Type": "application/binary"}, content='{"foo": "bar"}') ) - response = await self.client.post("/foo", cast_to=httpx.Response) + response = await async_client.post("/foo", cast_to=httpx.Response) assert response.status_code == 200 assert isinstance(response, httpx.Response) assert response.json() == {"foo": "bar"} - def test_copy(self) -> None: - copied = self.client.copy() - assert id(copied) != id(self.client) + def test_copy(self, async_client: AsyncBrapi) -> None: + copied = async_client.copy() + assert id(copied) != id(async_client) - copied = self.client.copy(api_key="another My API Key") + copied = async_client.copy(api_key="another My API Key") assert copied.api_key == "another My API Key" - assert self.client.api_key == "My API Key" + assert async_client.api_key == "My API Key" - def test_copy_default_options(self) -> None: + def test_copy_default_options(self, async_client: AsyncBrapi) -> None: # options that have a default are overridden correctly - copied = self.client.copy(max_retries=7) + copied = async_client.copy(max_retries=7) assert copied.max_retries == 7 - assert self.client.max_retries == 2 + assert async_client.max_retries == 2 copied2 = copied.copy(max_retries=6) assert copied2.max_retries == 6 assert copied.max_retries == 7 # timeout - assert isinstance(self.client.timeout, httpx.Timeout) - copied = self.client.copy(timeout=None) + assert isinstance(async_client.timeout, httpx.Timeout) + copied = async_client.copy(timeout=None) assert copied.timeout is None - assert isinstance(self.client.timeout, httpx.Timeout) + assert isinstance(async_client.timeout, httpx.Timeout) - def test_copy_default_headers(self) -> None: + async def test_copy_default_headers(self) -> None: client = AsyncBrapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} ) @@ -947,8 +963,9 @@ def test_copy_default_headers(self) -> None: match="`default_headers` and `set_default_headers` arguments are mutually exclusive", ): client.copy(set_default_headers={}, default_headers={"X-Foo": "Bar"}) + await client.close() - def test_copy_default_query(self) -> None: + async def test_copy_default_query(self) -> None: client = AsyncBrapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"foo": "bar"} ) @@ -984,13 +1001,15 @@ def test_copy_default_query(self) -> None: ): client.copy(set_default_query={}, default_query={"foo": "Bar"}) - def test_copy_signature(self) -> None: + await client.close() + + def test_copy_signature(self, async_client: AsyncBrapi) -> None: # ensure the same parameters that can be passed to the client are defined in the `.copy()` method init_signature = inspect.signature( # mypy doesn't like that we access the `__init__` property. - self.client.__init__, # type: ignore[misc] + async_client.__init__, # type: ignore[misc] ) - copy_signature = inspect.signature(self.client.copy) + copy_signature = inspect.signature(async_client.copy) exclude_params = {"transport", "proxies", "_strict_response_validation"} for name in init_signature.parameters.keys(): @@ -1001,12 +1020,12 @@ def test_copy_signature(self) -> None: assert copy_param is not None, f"copy() signature is missing the {name} param" @pytest.mark.skipif(sys.version_info >= (3, 10), reason="fails because of a memory leak that started from 3.12") - def test_copy_build_request(self) -> None: + def test_copy_build_request(self, async_client: AsyncBrapi) -> None: options = FinalRequestOptions(method="get", url="/foo") def build_request(options: FinalRequestOptions) -> None: - client = self.client.copy() - client._build_request(options) + client_copy = async_client.copy() + client_copy._build_request(options) # ensure that the machinery is warmed up before tracing starts. build_request(options) @@ -1063,12 +1082,12 @@ def add_leak(leaks: list[tracemalloc.StatisticDiff], diff: tracemalloc.Statistic print(frame) raise AssertionError() - async def test_request_timeout(self) -> None: - request = self.client._build_request(FinalRequestOptions(method="get", url="/foo")) + async def test_request_timeout(self, async_client: AsyncBrapi) -> None: + request = async_client._build_request(FinalRequestOptions(method="get", url="/foo")) timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT - request = self.client._build_request( + request = async_client._build_request( FinalRequestOptions(method="get", url="/foo", timeout=httpx.Timeout(100.0)) ) timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore @@ -1083,6 +1102,8 @@ async def test_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == httpx.Timeout(0) + await client.close() + async def test_http_client_timeout_option(self) -> None: # custom timeout given to the httpx client should be used async with httpx.AsyncClient(timeout=None) as http_client: @@ -1094,6 +1115,8 @@ async def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == httpx.Timeout(None) + await client.close() + # no timeout given to the httpx client should not use the httpx default async with httpx.AsyncClient() as http_client: client = AsyncBrapi( @@ -1104,6 +1127,8 @@ async def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT + await client.close() + # explicitly passing the default timeout currently results in it being ignored async with httpx.AsyncClient(timeout=HTTPX_DEFAULT_TIMEOUT) as http_client: client = AsyncBrapi( @@ -1114,6 +1139,8 @@ async def test_http_client_timeout_option(self) -> None: timeout = httpx.Timeout(**request.extensions["timeout"]) # type: ignore assert timeout == DEFAULT_TIMEOUT # our default + await client.close() + def test_invalid_http_client(self) -> None: with pytest.raises(TypeError, match="Invalid `http_client` arg"): with httpx.Client() as http_client: @@ -1124,15 +1151,15 @@ def test_invalid_http_client(self) -> None: http_client=cast(Any, http_client), ) - def test_default_headers_option(self) -> None: - client = AsyncBrapi( + async def test_default_headers_option(self) -> None: + test_client = AsyncBrapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, default_headers={"X-Foo": "bar"} ) - request = client._build_request(FinalRequestOptions(method="get", url="/foo")) + request = test_client._build_request(FinalRequestOptions(method="get", url="/foo")) assert request.headers.get("x-foo") == "bar" assert request.headers.get("x-stainless-lang") == "python" - client2 = AsyncBrapi( + test_client2 = AsyncBrapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, @@ -1141,10 +1168,13 @@ def test_default_headers_option(self) -> None: "X-Stainless-Lang": "my-overriding-header", }, ) - request = client2._build_request(FinalRequestOptions(method="get", url="/foo")) + request = test_client2._build_request(FinalRequestOptions(method="get", url="/foo")) assert request.headers.get("x-foo") == "stainless" assert request.headers.get("x-stainless-lang") == "my-overriding-header" + await test_client.close() + await test_client2.close() + def test_validate_headers(self) -> None: client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) request = client._build_request(FinalRequestOptions(method="get", url="/foo")) @@ -1155,7 +1185,7 @@ def test_validate_headers(self) -> None: client2 = AsyncBrapi(base_url=base_url, api_key=None, _strict_response_validation=True) _ = client2 - def test_default_query_option(self) -> None: + async def test_default_query_option(self) -> None: client = AsyncBrapi( base_url=base_url, api_key=api_key, _strict_response_validation=True, default_query={"query_param": "bar"} ) @@ -1173,8 +1203,10 @@ def test_default_query_option(self) -> None: url = httpx.URL(request.url) assert dict(url.params) == {"foo": "baz", "query_param": "overridden"} - def test_request_extra_json(self) -> None: - request = self.client._build_request( + await client.close() + + def test_request_extra_json(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1185,7 +1217,7 @@ def test_request_extra_json(self) -> None: data = json.loads(request.content.decode("utf-8")) assert data == {"foo": "bar", "baz": False} - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1196,7 +1228,7 @@ def test_request_extra_json(self) -> None: assert data == {"baz": False} # `extra_json` takes priority over `json_data` when keys clash - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1207,8 +1239,8 @@ def test_request_extra_json(self) -> None: data = json.loads(request.content.decode("utf-8")) assert data == {"foo": "bar", "baz": None} - def test_request_extra_headers(self) -> None: - request = self.client._build_request( + def test_request_extra_headers(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1218,7 +1250,7 @@ def test_request_extra_headers(self) -> None: assert request.headers.get("X-Foo") == "Foo" # `extra_headers` takes priority over `default_headers` when keys clash - request = self.client.with_options(default_headers={"X-Bar": "true"})._build_request( + request = client.with_options(default_headers={"X-Bar": "true"})._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1229,8 +1261,8 @@ def test_request_extra_headers(self) -> None: ) assert request.headers.get("X-Bar") == "false" - def test_request_extra_query(self) -> None: - request = self.client._build_request( + def test_request_extra_query(self, client: Brapi) -> None: + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1243,7 +1275,7 @@ def test_request_extra_query(self) -> None: assert params == {"my_query_param": "Foo"} # if both `query` and `extra_query` are given, they are merged - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1257,7 +1289,7 @@ def test_request_extra_query(self) -> None: assert params == {"bar": "1", "foo": "2"} # `extra_query` takes priority over `query` when keys clash - request = self.client._build_request( + request = client._build_request( FinalRequestOptions( method="post", url="/foo", @@ -1300,7 +1332,7 @@ def test_multipart_repeating_array(self, async_client: AsyncBrapi) -> None: ] @pytest.mark.respx(base_url=base_url) - async def test_basic_union_response(self, respx_mock: MockRouter) -> None: + async def test_basic_union_response(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: class Model1(BaseModel): name: str @@ -1309,12 +1341,12 @@ class Model2(BaseModel): respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = await async_client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model2) assert response.foo == "bar" @pytest.mark.respx(base_url=base_url) - async def test_union_response_different_types(self, respx_mock: MockRouter) -> None: + async def test_union_response_different_types(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: """Union of objects with the same field name using a different type""" class Model1(BaseModel): @@ -1325,18 +1357,20 @@ class Model2(BaseModel): respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": "bar"})) - response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = await async_client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model2) assert response.foo == "bar" respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": 1})) - response = await self.client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) + response = await async_client.get("/foo", cast_to=cast(Any, Union[Model1, Model2])) assert isinstance(response, Model1) assert response.foo == 1 @pytest.mark.respx(base_url=base_url) - async def test_non_application_json_content_type_for_json_data(self, respx_mock: MockRouter) -> None: + async def test_non_application_json_content_type_for_json_data( + self, respx_mock: MockRouter, async_client: AsyncBrapi + ) -> None: """ Response that sets Content-Type to something other than application/json but returns json data """ @@ -1352,11 +1386,11 @@ class Model(BaseModel): ) ) - response = await self.client.get("/foo", cast_to=Model) + response = await async_client.get("/foo", cast_to=Model) assert isinstance(response, Model) assert response.foo == 2 - def test_base_url_setter(self) -> None: + async def test_base_url_setter(self) -> None: client = AsyncBrapi(base_url="https://example.com/from_init", api_key=api_key, _strict_response_validation=True) assert client.base_url == "https://example.com/from_init/" @@ -1364,7 +1398,9 @@ def test_base_url_setter(self) -> None: assert client.base_url == "https://example.com/from_setter/" - def test_base_url_env(self) -> None: + await client.close() + + async def test_base_url_env(self) -> None: with update_env(BRAPI_BASE_URL="http://localhost:5000/from/env"): client = AsyncBrapi(api_key=api_key, _strict_response_validation=True) assert client.base_url == "http://localhost:5000/from/env/" @@ -1379,6 +1415,8 @@ def test_base_url_env(self) -> None: ) assert str(client.base_url).startswith("https://brapi.dev") + await client.close() + @pytest.mark.parametrize( "client", [ @@ -1394,7 +1432,7 @@ def test_base_url_env(self) -> None: ], ids=["standard", "custom http client"], ) - def test_base_url_trailing_slash(self, client: AsyncBrapi) -> None: + async def test_base_url_trailing_slash(self, client: AsyncBrapi) -> None: request = client._build_request( FinalRequestOptions( method="post", @@ -1403,6 +1441,7 @@ def test_base_url_trailing_slash(self, client: AsyncBrapi) -> None: ), ) assert request.url == "http://localhost:5000/custom/path/foo" + await client.close() @pytest.mark.parametrize( "client", @@ -1419,7 +1458,7 @@ def test_base_url_trailing_slash(self, client: AsyncBrapi) -> None: ], ids=["standard", "custom http client"], ) - def test_base_url_no_trailing_slash(self, client: AsyncBrapi) -> None: + async def test_base_url_no_trailing_slash(self, client: AsyncBrapi) -> None: request = client._build_request( FinalRequestOptions( method="post", @@ -1428,6 +1467,7 @@ def test_base_url_no_trailing_slash(self, client: AsyncBrapi) -> None: ), ) assert request.url == "http://localhost:5000/custom/path/foo" + await client.close() @pytest.mark.parametrize( "client", @@ -1444,7 +1484,7 @@ def test_base_url_no_trailing_slash(self, client: AsyncBrapi) -> None: ], ids=["standard", "custom http client"], ) - def test_absolute_request_url(self, client: AsyncBrapi) -> None: + async def test_absolute_request_url(self, client: AsyncBrapi) -> None: request = client._build_request( FinalRequestOptions( method="post", @@ -1453,37 +1493,37 @@ def test_absolute_request_url(self, client: AsyncBrapi) -> None: ), ) assert request.url == "https://myapi.com/foo" + await client.close() async def test_copied_client_does_not_close_http(self) -> None: - client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - assert not client.is_closed() + test_client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + assert not test_client.is_closed() - copied = client.copy() - assert copied is not client + copied = test_client.copy() + assert copied is not test_client del copied await asyncio.sleep(0.2) - assert not client.is_closed() + assert not test_client.is_closed() async def test_client_context_manager(self) -> None: - client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - async with client as c2: - assert c2 is client + test_client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) + async with test_client as c2: + assert c2 is test_client assert not c2.is_closed() - assert not client.is_closed() - assert client.is_closed() + assert not test_client.is_closed() + assert test_client.is_closed() @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio - async def test_client_response_validation_error(self, respx_mock: MockRouter) -> None: + async def test_client_response_validation_error(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: class Model(BaseModel): foo: str respx_mock.get("/foo").mock(return_value=httpx.Response(200, json={"foo": {"invalid": True}})) with pytest.raises(APIResponseValidationError) as exc: - await self.client.get("/foo", cast_to=Model) + await async_client.get("/foo", cast_to=Model) assert isinstance(exc.value.__cause__, ValidationError) @@ -1494,7 +1534,6 @@ async def test_client_max_retries_validation(self) -> None: ) @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio async def test_received_text_for_expected_json(self, respx_mock: MockRouter) -> None: class Model(BaseModel): name: str @@ -1506,11 +1545,14 @@ class Model(BaseModel): with pytest.raises(APIResponseValidationError): await strict_client.get("/foo", cast_to=Model) - client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) + non_strict_client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=False) - response = await client.get("/foo", cast_to=Model) + response = await non_strict_client.get("/foo", cast_to=Model) assert isinstance(response, str) # type: ignore[unreachable] + await strict_client.close() + await non_strict_client.close() + @pytest.mark.parametrize( "remaining_retries,retry_after,timeout", [ @@ -1533,13 +1575,12 @@ class Model(BaseModel): ], ) @mock.patch("time.time", mock.MagicMock(return_value=1696004797)) - @pytest.mark.asyncio - async def test_parse_retry_after_header(self, remaining_retries: int, retry_after: str, timeout: float) -> None: - client = AsyncBrapi(base_url=base_url, api_key=api_key, _strict_response_validation=True) - + async def test_parse_retry_after_header( + self, remaining_retries: int, retry_after: str, timeout: float, async_client: AsyncBrapi + ) -> None: headers = httpx.Headers({"retry-after": retry_after}) options = FinalRequestOptions(method="get", url="/foo", max_retries=3) - calculated = client._calculate_retry_timeout(remaining_retries, options, headers) + calculated = async_client._calculate_retry_timeout(remaining_retries, options, headers) assert calculated == pytest.approx(timeout, 0.5 * 0.875) # pyright: ignore[reportUnknownMemberType] @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @@ -1550,7 +1591,7 @@ async def test_retrying_timeout_errors_doesnt_leak(self, respx_mock: MockRouter, with pytest.raises(APITimeoutError): await async_client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__aenter__() - assert _get_open_connections(self.client) == 0 + assert _get_open_connections(async_client) == 0 @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @pytest.mark.respx(base_url=base_url) @@ -1559,12 +1600,11 @@ async def test_retrying_status_errors_doesnt_leak(self, respx_mock: MockRouter, with pytest.raises(APIStatusError): await async_client.quote.with_streaming_response.retrieve(tickers="PETR4,MGLU3").__aenter__() - assert _get_open_connections(self.client) == 0 + assert _get_open_connections(async_client) == 0 @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio @pytest.mark.parametrize("failure_mode", ["status", "exception"]) async def test_retries_taken( self, @@ -1596,7 +1636,6 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio async def test_omit_retry_count_header( self, async_client: AsyncBrapi, failures_before_success: int, respx_mock: MockRouter ) -> None: @@ -1622,7 +1661,6 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: @pytest.mark.parametrize("failures_before_success", [0, 2, 4]) @mock.patch("brapi._base_client.BaseClient._calculate_retry_timeout", _low_retry_timeout) @pytest.mark.respx(base_url=base_url) - @pytest.mark.asyncio async def test_overwrite_retry_count_header( self, async_client: AsyncBrapi, failures_before_success: int, respx_mock: MockRouter ) -> None: @@ -1672,26 +1710,26 @@ async def test_default_client_creation(self) -> None: ) @pytest.mark.respx(base_url=base_url) - async def test_follow_redirects(self, respx_mock: MockRouter) -> None: + async def test_follow_redirects(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: # Test that the default follow_redirects=True allows following redirects respx_mock.post("/redirect").mock( return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) ) respx_mock.get("/redirected").mock(return_value=httpx.Response(200, json={"status": "ok"})) - response = await self.client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) + response = await async_client.post("/redirect", body={"key": "value"}, cast_to=httpx.Response) assert response.status_code == 200 assert response.json() == {"status": "ok"} @pytest.mark.respx(base_url=base_url) - async def test_follow_redirects_disabled(self, respx_mock: MockRouter) -> None: + async def test_follow_redirects_disabled(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: # Test that follow_redirects=False prevents following redirects respx_mock.post("/redirect").mock( return_value=httpx.Response(302, headers={"Location": f"{base_url}/redirected"}) ) with pytest.raises(APIStatusError) as exc_info: - await self.client.post( + await async_client.post( "/redirect", body={"key": "value"}, options={"follow_redirects": False}, cast_to=httpx.Response ) From 6f83c5adcac181ccc5590f8225cb32554656b3e8 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 4 Nov 2025 04:58:22 +0000 Subject: [PATCH 12/47] chore(internal): grammar fix (it's -> its) --- src/brapi/_utils/_utils.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/brapi/_utils/_utils.py b/src/brapi/_utils/_utils.py index 50d5926..eec7f4a 100644 --- a/src/brapi/_utils/_utils.py +++ b/src/brapi/_utils/_utils.py @@ -133,7 +133,7 @@ def is_given(obj: _T | NotGiven | Omit) -> TypeGuard[_T]: # Type safe methods for narrowing types with TypeVars. # The default narrowing for isinstance(obj, dict) is dict[unknown, unknown], # however this cause Pyright to rightfully report errors. As we know we don't -# care about the contained types we can safely use `object` in it's place. +# care about the contained types we can safely use `object` in its place. # # There are two separate functions defined, `is_*` and `is_*_t` for different use cases. # `is_*` is for when you're dealing with an unknown input From df5d0628de5dada0f3aa2c6d1e394e092cc47502 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 11 Nov 2025 04:36:31 +0000 Subject: [PATCH 13/47] chore(package): drop Python 3.8 support --- README.md | 4 ++-- pyproject.toml | 5 ++--- src/brapi/_utils/_sync.py | 34 +++------------------------------- 3 files changed, 7 insertions(+), 36 deletions(-) diff --git a/README.md b/README.md index 4d5b6c5..ccd626c 100644 --- a/README.md +++ b/README.md @@ -3,7 +3,7 @@ [![PyPI version](https://img.shields.io/pypi/v/brapi.svg?label=pypi%20(stable))](https://pypi.org/project/brapi/) -The Brapi Python library provides convenient access to the Brapi REST API from any Python 3.8+ +The Brapi Python library provides convenient access to the Brapi REST API from any Python 3.9+ application. The library includes type definitions for all request params and response fields, and offers both synchronous and asynchronous clients powered by [httpx](https://github.com/encode/httpx). @@ -376,7 +376,7 @@ print(brapi.__version__) ## Requirements -Python 3.8 or higher. +Python 3.9 or higher. ## Contributing diff --git a/pyproject.toml b/pyproject.toml index 8fe119f..58efa79 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -15,11 +15,10 @@ dependencies = [ "distro>=1.7.0, <2", "sniffio", ] -requires-python = ">= 3.8" +requires-python = ">= 3.9" classifiers = [ "Typing :: Typed", "Intended Audience :: Developers", - "Programming Language :: Python :: 3.8", "Programming Language :: Python :: 3.9", "Programming Language :: Python :: 3.10", "Programming Language :: Python :: 3.11", @@ -141,7 +140,7 @@ filterwarnings = [ # there are a couple of flags that are still disabled by # default in strict mode as they are experimental and niche. typeCheckingMode = "strict" -pythonVersion = "3.8" +pythonVersion = "3.9" exclude = [ "_dev", diff --git a/src/brapi/_utils/_sync.py b/src/brapi/_utils/_sync.py index ad7ec71..f6027c1 100644 --- a/src/brapi/_utils/_sync.py +++ b/src/brapi/_utils/_sync.py @@ -1,10 +1,8 @@ from __future__ import annotations -import sys import asyncio import functools -import contextvars -from typing import Any, TypeVar, Callable, Awaitable +from typing import TypeVar, Callable, Awaitable from typing_extensions import ParamSpec import anyio @@ -15,34 +13,11 @@ T_ParamSpec = ParamSpec("T_ParamSpec") -if sys.version_info >= (3, 9): - _asyncio_to_thread = asyncio.to_thread -else: - # backport of https://docs.python.org/3/library/asyncio-task.html#asyncio.to_thread - # for Python 3.8 support - async def _asyncio_to_thread( - func: Callable[T_ParamSpec, T_Retval], /, *args: T_ParamSpec.args, **kwargs: T_ParamSpec.kwargs - ) -> Any: - """Asynchronously run function *func* in a separate thread. - - Any *args and **kwargs supplied for this function are directly passed - to *func*. Also, the current :class:`contextvars.Context` is propagated, - allowing context variables from the main thread to be accessed in the - separate thread. - - Returns a coroutine that can be awaited to get the eventual result of *func*. - """ - loop = asyncio.events.get_running_loop() - ctx = contextvars.copy_context() - func_call = functools.partial(ctx.run, func, *args, **kwargs) - return await loop.run_in_executor(None, func_call) - - async def to_thread( func: Callable[T_ParamSpec, T_Retval], /, *args: T_ParamSpec.args, **kwargs: T_ParamSpec.kwargs ) -> T_Retval: if sniffio.current_async_library() == "asyncio": - return await _asyncio_to_thread(func, *args, **kwargs) + return await asyncio.to_thread(func, *args, **kwargs) return await anyio.to_thread.run_sync( functools.partial(func, *args, **kwargs), @@ -53,10 +28,7 @@ async def to_thread( def asyncify(function: Callable[T_ParamSpec, T_Retval]) -> Callable[T_ParamSpec, Awaitable[T_Retval]]: """ Take a blocking function and create an async one that receives the same - positional and keyword arguments. For python version 3.9 and above, it uses - asyncio.to_thread to run the function in a separate thread. For python version - 3.8, it uses locally defined copy of the asyncio.to_thread function which was - introduced in python 3.9. + positional and keyword arguments. Usage: From 03df33b17fbced0feb1c1cd0fb0aa38dac33b28d Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 11 Nov 2025 04:37:09 +0000 Subject: [PATCH 14/47] fix: compat with Python 3.14 --- src/brapi/_models.py | 11 ++++++++--- tests/test_models.py | 8 ++++---- 2 files changed, 12 insertions(+), 7 deletions(-) diff --git a/src/brapi/_models.py b/src/brapi/_models.py index 6a3cd1d..fcec2cf 100644 --- a/src/brapi/_models.py +++ b/src/brapi/_models.py @@ -2,6 +2,7 @@ import os import inspect +import weakref from typing import TYPE_CHECKING, Any, Type, Union, Generic, TypeVar, Callable, Optional, cast from datetime import date, datetime from typing_extensions import ( @@ -573,6 +574,9 @@ class CachedDiscriminatorType(Protocol): __discriminator__: DiscriminatorDetails +DISCRIMINATOR_CACHE: weakref.WeakKeyDictionary[type, DiscriminatorDetails] = weakref.WeakKeyDictionary() + + class DiscriminatorDetails: field_name: str """The name of the discriminator field in the variant class, e.g. @@ -615,8 +619,9 @@ def __init__( def _build_discriminated_union_meta(*, union: type, meta_annotations: tuple[Any, ...]) -> DiscriminatorDetails | None: - if isinstance(union, CachedDiscriminatorType): - return union.__discriminator__ + cached = DISCRIMINATOR_CACHE.get(union) + if cached is not None: + return cached discriminator_field_name: str | None = None @@ -669,7 +674,7 @@ def _build_discriminated_union_meta(*, union: type, meta_annotations: tuple[Any, discriminator_field=discriminator_field_name, discriminator_alias=discriminator_alias, ) - cast(CachedDiscriminatorType, union).__discriminator__ = details + DISCRIMINATOR_CACHE.setdefault(union, details) return details diff --git a/tests/test_models.py b/tests/test_models.py index cb150da..a8c46c0 100644 --- a/tests/test_models.py +++ b/tests/test_models.py @@ -9,7 +9,7 @@ from brapi._utils import PropertyInfo from brapi._compat import PYDANTIC_V1, parse_obj, model_dump, model_json -from brapi._models import BaseModel, construct_type +from brapi._models import DISCRIMINATOR_CACHE, BaseModel, construct_type class BasicModel(BaseModel): @@ -809,7 +809,7 @@ class B(BaseModel): UnionType = cast(Any, Union[A, B]) - assert not hasattr(UnionType, "__discriminator__") + assert not DISCRIMINATOR_CACHE.get(UnionType) m = construct_type( value={"type": "b", "data": "foo"}, type_=cast(Any, Annotated[UnionType, PropertyInfo(discriminator="type")]) @@ -818,7 +818,7 @@ class B(BaseModel): assert m.type == "b" assert m.data == "foo" # type: ignore[comparison-overlap] - discriminator = UnionType.__discriminator__ + discriminator = DISCRIMINATOR_CACHE.get(UnionType) assert discriminator is not None m = construct_type( @@ -830,7 +830,7 @@ class B(BaseModel): # if the discriminator details object stays the same between invocations then # we hit the cache - assert UnionType.__discriminator__ is discriminator + assert DISCRIMINATOR_CACHE.get(UnionType) is discriminator @pytest.mark.skipif(PYDANTIC_V1, reason="TypeAliasType is not supported in Pydantic v1") From e45731b617db3ac23fb211f94542ceb8f3fe09de Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 12 Nov 2025 04:22:58 +0000 Subject: [PATCH 15/47] fix(compat): update signatures of `model_dump` and `model_dump_json` for Pydantic v1 --- src/brapi/_models.py | 41 +++++++++++++++++++++++++++++------------ 1 file changed, 29 insertions(+), 12 deletions(-) diff --git a/src/brapi/_models.py b/src/brapi/_models.py index fcec2cf..ca9500b 100644 --- a/src/brapi/_models.py +++ b/src/brapi/_models.py @@ -257,15 +257,16 @@ def model_dump( mode: Literal["json", "python"] | str = "python", include: IncEx | None = None, exclude: IncEx | None = None, + context: Any | None = None, by_alias: bool | None = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, + exclude_computed_fields: bool = False, round_trip: bool = False, warnings: bool | Literal["none", "warn", "error"] = True, - context: dict[str, Any] | None = None, - serialize_as_any: bool = False, fallback: Callable[[Any], Any] | None = None, + serialize_as_any: bool = False, ) -> dict[str, Any]: """Usage docs: https://docs.pydantic.dev/2.4/concepts/serialization/#modelmodel_dump @@ -273,16 +274,24 @@ def model_dump( Args: mode: The mode in which `to_python` should run. - If mode is 'json', the dictionary will only contain JSON serializable types. - If mode is 'python', the dictionary may contain any Python objects. - include: A list of fields to include in the output. - exclude: A list of fields to exclude from the output. + If mode is 'json', the output will only contain JSON serializable types. + If mode is 'python', the output may contain non-JSON-serializable Python objects. + include: A set of fields to include in the output. + exclude: A set of fields to exclude from the output. + context: Additional context to pass to the serializer. by_alias: Whether to use the field's alias in the dictionary key if defined. - exclude_unset: Whether to exclude fields that are unset or None from the output. - exclude_defaults: Whether to exclude fields that are set to their default value from the output. - exclude_none: Whether to exclude fields that have a value of `None` from the output. - round_trip: Whether to enable serialization and deserialization round-trip support. - warnings: Whether to log warnings when invalid fields are encountered. + exclude_unset: Whether to exclude fields that have not been explicitly set. + exclude_defaults: Whether to exclude fields that are set to their default value. + exclude_none: Whether to exclude fields that have a value of `None`. + exclude_computed_fields: Whether to exclude computed fields. + While this can be useful for round-tripping, it is usually recommended to use the dedicated + `round_trip` parameter instead. + round_trip: If True, dumped values should be valid as input for non-idempotent types such as Json[T]. + warnings: How to handle serialization errors. False/"none" ignores them, True/"warn" logs errors, + "error" raises a [`PydanticSerializationError`][pydantic_core.PydanticSerializationError]. + fallback: A function to call when an unknown value is encountered. If not provided, + a [`PydanticSerializationError`][pydantic_core.PydanticSerializationError] error is raised. + serialize_as_any: Whether to serialize fields with duck-typing serialization behavior. Returns: A dictionary representation of the model. @@ -299,6 +308,8 @@ def model_dump( raise ValueError("serialize_as_any is only supported in Pydantic v2") if fallback is not None: raise ValueError("fallback is only supported in Pydantic v2") + if exclude_computed_fields != False: + raise ValueError("exclude_computed_fields is only supported in Pydantic v2") dumped = super().dict( # pyright: ignore[reportDeprecated] include=include, exclude=exclude, @@ -315,15 +326,17 @@ def model_dump_json( self, *, indent: int | None = None, + ensure_ascii: bool = False, include: IncEx | None = None, exclude: IncEx | None = None, + context: Any | None = None, by_alias: bool | None = None, exclude_unset: bool = False, exclude_defaults: bool = False, exclude_none: bool = False, + exclude_computed_fields: bool = False, round_trip: bool = False, warnings: bool | Literal["none", "warn", "error"] = True, - context: dict[str, Any] | None = None, fallback: Callable[[Any], Any] | None = None, serialize_as_any: bool = False, ) -> str: @@ -355,6 +368,10 @@ def model_dump_json( raise ValueError("serialize_as_any is only supported in Pydantic v2") if fallback is not None: raise ValueError("fallback is only supported in Pydantic v2") + if ensure_ascii != False: + raise ValueError("ensure_ascii is only supported in Pydantic v2") + if exclude_computed_fields != False: + raise ValueError("exclude_computed_fields is only supported in Pydantic v2") return super().json( # type: ignore[reportDeprecated] indent=indent, include=include, From e9ba21da5758717a2d83b11704415c448f3dcf51 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 14 Nov 2025 19:06:19 +0000 Subject: [PATCH 16/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index a8f7122..06d6df2 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.1" + ".": "1.0.2" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index 58efa79..a166ff1 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.1" +version = "1.0.2" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index 9269671..26061e0 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.1" # x-release-please-version +__version__ = "1.0.2" # x-release-please-version From 18c4f73dd065d17f643352f952112b0d213e5947 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sat, 22 Nov 2025 04:05:16 +0000 Subject: [PATCH 17/47] chore: add Python 3.14 classifier and testing --- pyproject.toml | 1 + 1 file changed, 1 insertion(+) diff --git a/pyproject.toml b/pyproject.toml index a166ff1..b12397d 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -24,6 +24,7 @@ classifiers = [ "Programming Language :: Python :: 3.11", "Programming Language :: Python :: 3.12", "Programming Language :: Python :: 3.13", + "Programming Language :: Python :: 3.14", "Operating System :: OS Independent", "Operating System :: POSIX", "Operating System :: MacOS", From 50ac94565393053f0c58e183049e18f7ee6e7af5 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Mon, 24 Nov 2025 20:00:38 +0000 Subject: [PATCH 18/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 06d6df2..b7634f9 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.2" + ".": "1.0.3" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index b12397d..eb53185 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.2" +version = "1.0.3" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index 26061e0..59676ab 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.2" # x-release-please-version +__version__ = "1.0.3" # x-release-please-version From 8756a972fcb2f5beadd5889f4ba686f107be9069 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 28 Nov 2025 03:22:36 +0000 Subject: [PATCH 19/47] codegen metadata --- .stats.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.stats.yml b/.stats.yml index 00e0370..3d5ec16 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-bf7b0065e4057ae80522a943caa4967f1fe0aa0a6989122f5687788f39dfbdea.yml -openapi_spec_hash: 7ac81061bb9f3cb0c180b82b5ea83258 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-11630ce70d2f2c187912989dd9004b10828eaab889f617ba39d5a7a8e4b03b62.yml +openapi_spec_hash: 4198e5f7a76f3002723c113663465c00 config_hash: 6f10a67950f65bf850612b59838ad03b From 72472efd8e5755817e7b6c7b16d21276b586f78d Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 28 Nov 2025 03:25:11 +0000 Subject: [PATCH 20/47] fix: ensure streams are always closed --- src/brapi/_streaming.py | 22 ++++++++++++---------- 1 file changed, 12 insertions(+), 10 deletions(-) diff --git a/src/brapi/_streaming.py b/src/brapi/_streaming.py index d0ff08c..e5f7242 100644 --- a/src/brapi/_streaming.py +++ b/src/brapi/_streaming.py @@ -54,11 +54,12 @@ def __stream__(self) -> Iterator[_T]: process_data = self._client._process_response_data iterator = self._iter_events() - for sse in iterator: - yield process_data(data=sse.json(), cast_to=cast_to, response=response) - - # As we might not fully consume the response stream, we need to close it explicitly - response.close() + try: + for sse in iterator: + yield process_data(data=sse.json(), cast_to=cast_to, response=response) + finally: + # Ensure the response is closed even if the consumer doesn't read all data + response.close() def __enter__(self) -> Self: return self @@ -117,11 +118,12 @@ async def __stream__(self) -> AsyncIterator[_T]: process_data = self._client._process_response_data iterator = self._iter_events() - async for sse in iterator: - yield process_data(data=sse.json(), cast_to=cast_to, response=response) - - # As we might not fully consume the response stream, we need to close it explicitly - await response.aclose() + try: + async for sse in iterator: + yield process_data(data=sse.json(), cast_to=cast_to, response=response) + finally: + # Ensure the response is closed even if the consumer doesn't read all data + await response.aclose() async def __aenter__(self) -> Self: return self From c3c4439bfa87043475927473219e12556ea91c8d Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 28 Nov 2025 03:26:09 +0000 Subject: [PATCH 21/47] chore(deps): mypy 1.18.1 has a regression, pin to 1.17 --- pyproject.toml | 2 +- requirements-dev.lock | 4 +++- requirements.lock | 8 ++++---- 3 files changed, 8 insertions(+), 6 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index eb53185..5f8a1bf 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -46,7 +46,7 @@ managed = true # version pins are in requirements-dev.lock dev-dependencies = [ "pyright==1.1.399", - "mypy", + "mypy==1.17", "respx", "pytest", "pytest-asyncio", diff --git a/requirements-dev.lock b/requirements-dev.lock index 3050a50..6025905 100644 --- a/requirements-dev.lock +++ b/requirements-dev.lock @@ -72,7 +72,7 @@ mdurl==0.1.2 multidict==6.4.4 # via aiohttp # via yarl -mypy==1.14.1 +mypy==1.17.0 mypy-extensions==1.0.0 # via mypy nodeenv==1.8.0 @@ -81,6 +81,8 @@ nox==2023.4.22 packaging==23.2 # via nox # via pytest +pathspec==0.12.1 + # via mypy platformdirs==3.11.0 # via virtualenv pluggy==1.5.0 diff --git a/requirements.lock b/requirements.lock index cf79912..6aa2116 100644 --- a/requirements.lock +++ b/requirements.lock @@ -55,21 +55,21 @@ multidict==6.4.4 propcache==0.3.1 # via aiohttp # via yarl -pydantic==2.11.9 +pydantic==2.12.5 # via brapi -pydantic-core==2.33.2 +pydantic-core==2.41.5 # via pydantic sniffio==1.3.0 # via anyio # via brapi -typing-extensions==4.12.2 +typing-extensions==4.15.0 # via anyio # via brapi # via multidict # via pydantic # via pydantic-core # via typing-inspection -typing-inspection==0.4.1 +typing-inspection==0.4.2 # via pydantic yarl==1.20.0 # via aiohttp From 11f75c794edc7ec0d9493bcdea4cdbeff0d093e1 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 3 Dec 2025 05:27:03 +0000 Subject: [PATCH 22/47] chore: update lockfile --- pyproject.toml | 14 +++--- requirements-dev.lock | 108 +++++++++++++++++++++++------------------- requirements.lock | 31 ++++++------ 3 files changed, 83 insertions(+), 70 deletions(-) diff --git a/pyproject.toml b/pyproject.toml index 5f8a1bf..6c5e7ad 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -7,14 +7,16 @@ license = "Apache-2.0" authors = [ { name = "Brapi", email = "contact@brapi.dev" }, ] + dependencies = [ - "httpx>=0.23.0, <1", - "pydantic>=1.9.0, <3", - "typing-extensions>=4.10, <5", - "anyio>=3.5.0, <5", - "distro>=1.7.0, <2", - "sniffio", + "httpx>=0.23.0, <1", + "pydantic>=1.9.0, <3", + "typing-extensions>=4.10, <5", + "anyio>=3.5.0, <5", + "distro>=1.7.0, <2", + "sniffio", ] + requires-python = ">= 3.9" classifiers = [ "Typing :: Typed", diff --git a/requirements-dev.lock b/requirements-dev.lock index 6025905..659c238 100644 --- a/requirements-dev.lock +++ b/requirements-dev.lock @@ -12,40 +12,45 @@ -e file:. aiohappyeyeballs==2.6.1 # via aiohttp -aiohttp==3.12.8 +aiohttp==3.13.2 # via brapi # via httpx-aiohttp -aiosignal==1.3.2 +aiosignal==1.4.0 # via aiohttp -annotated-types==0.6.0 +annotated-types==0.7.0 # via pydantic -anyio==4.4.0 +anyio==4.12.0 # via brapi # via httpx -argcomplete==3.1.2 +argcomplete==3.6.3 # via nox async-timeout==5.0.1 # via aiohttp -attrs==25.3.0 +attrs==25.4.0 # via aiohttp -certifi==2023.7.22 + # via nox +backports-asyncio-runner==1.2.0 + # via pytest-asyncio +certifi==2025.11.12 # via httpcore # via httpx -colorlog==6.7.0 +colorlog==6.10.1 + # via nox +dependency-groups==1.3.1 # via nox -dirty-equals==0.6.0 -distlib==0.3.7 +dirty-equals==0.11 +distlib==0.4.0 # via virtualenv -distro==1.8.0 +distro==1.9.0 # via brapi -exceptiongroup==1.2.2 +exceptiongroup==1.3.1 # via anyio # via pytest -execnet==2.1.1 +execnet==2.1.2 # via pytest-xdist -filelock==3.12.4 +filelock==3.19.1 # via virtualenv -frozenlist==1.6.2 +frozenlist==1.8.0 # via aiohttp # via aiosignal h11==0.16.0 @@ -58,82 +63,87 @@ httpx==0.28.1 # via respx httpx-aiohttp==0.1.9 # via brapi -idna==3.4 +humanize==4.13.0 + # via nox +idna==3.11 # via anyio # via httpx # via yarl -importlib-metadata==7.0.0 -iniconfig==2.0.0 +importlib-metadata==8.7.0 +iniconfig==2.1.0 # via pytest markdown-it-py==3.0.0 # via rich mdurl==0.1.2 # via markdown-it-py -multidict==6.4.4 +multidict==6.7.0 # via aiohttp # via yarl mypy==1.17.0 -mypy-extensions==1.0.0 +mypy-extensions==1.1.0 # via mypy -nodeenv==1.8.0 +nodeenv==1.9.1 # via pyright -nox==2023.4.22 -packaging==23.2 +nox==2025.11.12 +packaging==25.0 + # via dependency-groups # via nox # via pytest pathspec==0.12.1 # via mypy -platformdirs==3.11.0 +platformdirs==4.4.0 # via virtualenv -pluggy==1.5.0 +pluggy==1.6.0 # via pytest -propcache==0.3.1 +propcache==0.4.1 # via aiohttp # via yarl -pydantic==2.11.9 +pydantic==2.12.5 # via brapi -pydantic-core==2.33.2 +pydantic-core==2.41.5 # via pydantic -pygments==2.18.0 +pygments==2.19.2 + # via pytest # via rich pyright==1.1.399 -pytest==8.3.3 +pytest==8.4.2 # via pytest-asyncio # via pytest-xdist -pytest-asyncio==0.24.0 -pytest-xdist==3.7.0 -python-dateutil==2.8.2 +pytest-asyncio==1.2.0 +pytest-xdist==3.8.0 +python-dateutil==2.9.0.post0 # via time-machine -pytz==2023.3.post1 - # via dirty-equals respx==0.22.0 -rich==13.7.1 -ruff==0.9.4 -setuptools==68.2.2 - # via nodeenv -six==1.16.0 +rich==14.2.0 +ruff==0.14.7 +six==1.17.0 # via python-dateutil -sniffio==1.3.0 - # via anyio +sniffio==1.3.1 # via brapi -time-machine==2.9.0 -tomli==2.0.2 +time-machine==2.19.0 +tomli==2.3.0 + # via dependency-groups # via mypy + # via nox # via pytest -typing-extensions==4.12.2 +typing-extensions==4.15.0 + # via aiosignal # via anyio # via brapi + # via exceptiongroup # via multidict # via mypy # via pydantic # via pydantic-core # via pyright + # via pytest-asyncio # via typing-inspection -typing-inspection==0.4.1 + # via virtualenv +typing-inspection==0.4.2 # via pydantic -virtualenv==20.24.5 +virtualenv==20.35.4 # via nox -yarl==1.20.0 +yarl==1.22.0 # via aiohttp -zipp==3.17.0 +zipp==3.23.0 # via importlib-metadata diff --git a/requirements.lock b/requirements.lock index 6aa2116..f030e74 100644 --- a/requirements.lock +++ b/requirements.lock @@ -12,28 +12,28 @@ -e file:. aiohappyeyeballs==2.6.1 # via aiohttp -aiohttp==3.12.8 +aiohttp==3.13.2 # via brapi # via httpx-aiohttp -aiosignal==1.3.2 +aiosignal==1.4.0 # via aiohttp -annotated-types==0.6.0 +annotated-types==0.7.0 # via pydantic -anyio==4.4.0 +anyio==4.12.0 # via brapi # via httpx async-timeout==5.0.1 # via aiohttp -attrs==25.3.0 +attrs==25.4.0 # via aiohttp -certifi==2023.7.22 +certifi==2025.11.12 # via httpcore # via httpx -distro==1.8.0 +distro==1.9.0 # via brapi -exceptiongroup==1.2.2 +exceptiongroup==1.3.1 # via anyio -frozenlist==1.6.2 +frozenlist==1.8.0 # via aiohttp # via aiosignal h11==0.16.0 @@ -45,31 +45,32 @@ httpx==0.28.1 # via httpx-aiohttp httpx-aiohttp==0.1.9 # via brapi -idna==3.4 +idna==3.11 # via anyio # via httpx # via yarl -multidict==6.4.4 +multidict==6.7.0 # via aiohttp # via yarl -propcache==0.3.1 +propcache==0.4.1 # via aiohttp # via yarl pydantic==2.12.5 # via brapi pydantic-core==2.41.5 # via pydantic -sniffio==1.3.0 - # via anyio +sniffio==1.3.1 # via brapi typing-extensions==4.15.0 + # via aiosignal # via anyio # via brapi + # via exceptiongroup # via multidict # via pydantic # via pydantic-core # via typing-inspection typing-inspection==0.4.2 # via pydantic -yarl==1.20.0 +yarl==1.22.0 # via aiohttp From 0ce0d94f774d87977320d25927f15a8306e9f9e9 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 3 Dec 2025 05:35:59 +0000 Subject: [PATCH 23/47] chore(docs): use environment variables for authentication in code snippets --- README.md | 3 ++- 1 file changed, 2 insertions(+), 1 deletion(-) diff --git a/README.md b/README.md index ccd626c..0c5bf22 100644 --- a/README.md +++ b/README.md @@ -87,6 +87,7 @@ pip install brapi[aiohttp] Then you can enable it by instantiating the client with `http_client=DefaultAioHttpClient()`: ```python +import os import asyncio from brapi import DefaultAioHttpClient from brapi import AsyncBrapi @@ -94,7 +95,7 @@ from brapi import AsyncBrapi async def main() -> None: async with AsyncBrapi( - api_key="My API Key", + api_key=os.environ.get("BRAPI_API_KEY"), # This is the default and can be omitted http_client=DefaultAioHttpClient(), ) as client: quote = await client.quote.retrieve( From 60f4cd027ada6179049c83cf324830458adcd126 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 9 Dec 2025 04:30:31 +0000 Subject: [PATCH 24/47] fix(types): allow pyright to infer TypedDict types within SequenceNotStr --- src/brapi/_types.py | 5 +++-- 1 file changed, 3 insertions(+), 2 deletions(-) diff --git a/src/brapi/_types.py b/src/brapi/_types.py index a3e695f..c98ff51 100644 --- a/src/brapi/_types.py +++ b/src/brapi/_types.py @@ -243,6 +243,9 @@ class HttpxSendArgs(TypedDict, total=False): if TYPE_CHECKING: # This works because str.__contains__ does not accept object (either in typeshed or at runtime) # https://github.com/hauntsaninja/useful_types/blob/5e9710f3875107d068e7679fd7fec9cfab0eff3b/useful_types/__init__.py#L285 + # + # Note: index() and count() methods are intentionally omitted to allow pyright to properly + # infer TypedDict types when dict literals are used in lists assigned to SequenceNotStr. class SequenceNotStr(Protocol[_T_co]): @overload def __getitem__(self, index: SupportsIndex, /) -> _T_co: ... @@ -251,8 +254,6 @@ def __getitem__(self, index: slice, /) -> Sequence[_T_co]: ... def __contains__(self, value: object, /) -> bool: ... def __len__(self) -> int: ... def __iter__(self) -> Iterator[_T_co]: ... - def index(self, value: Any, start: int = 0, stop: int = ..., /) -> int: ... - def count(self, value: Any, /) -> int: ... def __reversed__(self) -> Iterator[_T_co]: ... else: # just point this to a normal `Sequence` at runtime to avoid having to special case From f07fc6803cf1726432f7a9f27a446abb13e13f7b Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 9 Dec 2025 04:32:46 +0000 Subject: [PATCH 25/47] chore: add missing docstrings --- src/brapi/types/available_list_response.py | 2 ++ src/brapi/types/balance_sheet_entry.py | 4 ++++ src/brapi/types/cashflow_entry.py | 4 ++++ .../types/default_key_statistics_entry.py | 4 ++++ src/brapi/types/financial_data_entry.py | 4 ++++ src/brapi/types/income_statement_entry.py | 4 ++++ src/brapi/types/quote_list_response.py | 8 +++++++ src/brapi/types/quote_retrieve_response.py | 24 +++++++++++++++++++ .../v2/crypto_list_available_response.py | 2 ++ .../types/v2/crypto_retrieve_response.py | 8 +++++++ .../v2/currency_list_available_response.py | 2 ++ .../types/v2/currency_retrieve_response.py | 6 +++++ .../v2/inflation_list_available_response.py | 2 ++ .../types/v2/inflation_retrieve_response.py | 4 ++++ .../v2/prime_rate_list_available_response.py | 4 ++++ .../types/v2/prime_rate_retrieve_response.py | 6 +++++ src/brapi/types/value_added_entry.py | 4 ++++ 17 files changed, 92 insertions(+) diff --git a/src/brapi/types/available_list_response.py b/src/brapi/types/available_list_response.py index 20f59dd..07e0719 100644 --- a/src/brapi/types/available_list_response.py +++ b/src/brapi/types/available_list_response.py @@ -8,6 +8,8 @@ class AvailableListResponse(BaseModel): + """Resposta do endpoint que lista todos os tickers disponíveis.""" + indexes: List[str] """Lista de tickers de **índices** disponíveis (ex: `^BVSP`, `^IFIX`).""" diff --git a/src/brapi/types/balance_sheet_entry.py b/src/brapi/types/balance_sheet_entry.py index f40cd15..38a4aa7 100644 --- a/src/brapi/types/balance_sheet_entry.py +++ b/src/brapi/types/balance_sheet_entry.py @@ -12,6 +12,10 @@ class BalanceSheetEntry(BaseModel): + """ + Representa os dados de um Balanço Patrimonial para um período específico (anual ou trimestral). + """ + accounts_payable: Optional[float] = FieldInfo(alias="accountsPayable", default=None) """Contas a pagar (fornecedores).""" diff --git a/src/brapi/types/cashflow_entry.py b/src/brapi/types/cashflow_entry.py index b8b7fb9..a17f213 100644 --- a/src/brapi/types/cashflow_entry.py +++ b/src/brapi/types/cashflow_entry.py @@ -12,6 +12,10 @@ class CashflowEntry(BaseModel): + """ + Representa os dados de uma Demonstração do Fluxo de Caixa (DFC) para um período específico (anual ou trimestral). + """ + adjustments_to_profit_or_loss: Optional[float] = FieldInfo(alias="adjustmentsToProfitOrLoss", default=None) """ Ajustes ao lucro/prejuízo (depreciação, amortização, equivalência patrimonial, diff --git a/src/brapi/types/default_key_statistics_entry.py b/src/brapi/types/default_key_statistics_entry.py index dc1315b..44b1660 100644 --- a/src/brapi/types/default_key_statistics_entry.py +++ b/src/brapi/types/default_key_statistics_entry.py @@ -12,6 +12,10 @@ class DefaultKeyStatisticsEntry(BaseModel): + """ + Representa um conjunto de principais indicadores e estatísticas financeiras para um período (TTM, anual ou trimestral). + """ + api_52_week_change: Optional[float] = FieldInfo(alias="52WeekChange", default=None) """Variação percentual do preço da ação nas últimas 52 semanas.""" diff --git a/src/brapi/types/financial_data_entry.py b/src/brapi/types/financial_data_entry.py index 3db991a..3b6a05b 100644 --- a/src/brapi/types/financial_data_entry.py +++ b/src/brapi/types/financial_data_entry.py @@ -12,6 +12,10 @@ class FinancialDataEntry(BaseModel): + """ + Representa um conjunto de dados e indicadores financeiros calculados para um período (TTM, anual ou trimestral). + """ + current_price: Optional[float] = FieldInfo(alias="currentPrice", default=None) """Preço atual da ação (pode ser ligeiramente defasado).""" diff --git a/src/brapi/types/income_statement_entry.py b/src/brapi/types/income_statement_entry.py index 7429e0a..34bfc0b 100644 --- a/src/brapi/types/income_statement_entry.py +++ b/src/brapi/types/income_statement_entry.py @@ -12,6 +12,10 @@ class IncomeStatementEntry(BaseModel): + """ + Representa os dados de uma Demonstração do Resultado do Exercício (DRE) para um período específico (anual ou trimestral). + """ + id: Optional[str] = None """Identificador único deste registro de DRE (interno).""" diff --git a/src/brapi/types/quote_list_response.py b/src/brapi/types/quote_list_response.py index b1ca5d7..194ef67 100644 --- a/src/brapi/types/quote_list_response.py +++ b/src/brapi/types/quote_list_response.py @@ -11,6 +11,8 @@ class Index(BaseModel): + """Resumo de informações de um índice, geralmente retornado em listas.""" + name: Optional[str] = None """Nome do índice (ex: `IBOVESPA`).""" @@ -19,6 +21,10 @@ class Index(BaseModel): class Stock(BaseModel): + """ + Resumo de informações de um ativo (ação, FII, BDR), geralmente retornado em listas. + """ + change: Optional[float] = None """Variação percentual do preço em relação ao fechamento anterior.""" @@ -57,6 +63,8 @@ class Stock(BaseModel): class QuoteListResponse(BaseModel): + """Resposta do endpoint de listagem de cotações (`/api/quote/list`).""" + available_sectors: Optional[List[str]] = FieldInfo(alias="availableSectors", default=None) """ Lista de todos os setores disponíveis que podem ser usados no parâmetro de diff --git a/src/brapi/types/quote_retrieve_response.py b/src/brapi/types/quote_retrieve_response.py index 8f9e542..9deefa1 100644 --- a/src/brapi/types/quote_retrieve_response.py +++ b/src/brapi/types/quote_retrieve_response.py @@ -25,6 +25,8 @@ class ResultDividendsDataCashDividend(BaseModel): + """Detalhes sobre um pagamento de provento em dinheiro (Dividendo ou JCP).""" + approved_on: Optional[datetime] = FieldInfo(alias="approvedOn", default=None) """Data em que o pagamento do provento foi aprovado pela empresa. @@ -76,6 +78,10 @@ class ResultDividendsDataCashDividend(BaseModel): class ResultDividendsDataStockDividend(BaseModel): + """ + Detalhes sobre um evento corporativo que afeta a quantidade de ações (Desdobramento/Split, Grupamento/Inplit, Bonificação). + """ + approved_on: Optional[datetime] = FieldInfo(alias="approvedOn", default=None) """Data em que o evento foi aprovado. Formato ISO 8601.""" @@ -109,6 +115,11 @@ class ResultDividendsDataStockDividend(BaseModel): class ResultDividendsData(BaseModel): + """Objeto contendo informações sobre dividendos, JCP e outros eventos corporativos. + + Retornado apenas se `dividends=true` for especificado na requisição. + """ + cash_dividends: Optional[List[ResultDividendsDataCashDividend]] = FieldInfo(alias="cashDividends", default=None) """Lista de proventos pagos em dinheiro (Dividendos e JCP).""" @@ -120,6 +131,8 @@ class ResultDividendsData(BaseModel): class ResultHistoricalDataPrice(BaseModel): + """Representa um ponto na série histórica de preços de um ativo.""" + adjusted_close: Optional[float] = FieldInfo(alias="adjustedClose", default=None) """ Preço de fechamento ajustado para proventos (dividendos, JCP, bonificações, @@ -149,6 +162,11 @@ class ResultHistoricalDataPrice(BaseModel): class ResultSummaryProfile(BaseModel): + """Resumo do perfil da empresa. + + Retornado apenas se `modules` incluir `summaryProfile`. + """ + address1: Optional[str] = None """Linha 1 do endereço da sede da empresa.""" @@ -205,6 +223,10 @@ class ResultSummaryProfile(BaseModel): class Result(BaseModel): + """ + Contém os dados detalhados de um ativo específico retornado pelo endpoint `/api/quote/{tickers}`. + """ + average_daily_volume10_day: Optional[float] = FieldInfo(alias="averageDailyVolume10Day", default=None) """Média do volume financeiro diário negociado nos últimos 10 dias.""" @@ -460,6 +482,8 @@ class Result(BaseModel): class QuoteRetrieveResponse(BaseModel): + """Resposta principal do endpoint `/api/quote/{tickers}`.""" + requested_at: Optional[datetime] = FieldInfo(alias="requestedAt", default=None) """Timestamp indicando quando a requisição foi recebida pelo servidor. diff --git a/src/brapi/types/v2/crypto_list_available_response.py b/src/brapi/types/v2/crypto_list_available_response.py index 5951e9a..d4dab7f 100644 --- a/src/brapi/types/v2/crypto_list_available_response.py +++ b/src/brapi/types/v2/crypto_list_available_response.py @@ -8,6 +8,8 @@ class CryptoListAvailableResponse(BaseModel): + """Resposta do endpoint que lista todas as criptomoedas disponíveis.""" + coins: Optional[List[str]] = None """ Lista de siglas (tickers) das criptomoedas disponíveis (ex: `BTC`, `ETH`, diff --git a/src/brapi/types/v2/crypto_retrieve_response.py b/src/brapi/types/v2/crypto_retrieve_response.py index 11d50b7..2b4ffa8 100644 --- a/src/brapi/types/v2/crypto_retrieve_response.py +++ b/src/brapi/types/v2/crypto_retrieve_response.py @@ -11,6 +11,8 @@ class CoinHistoricalDataPrice(BaseModel): + """Representa um ponto na série histórica de preços de uma criptomoeda.""" + adjusted_close: Optional[float] = FieldInfo(alias="adjustedClose", default=None) """Preço de fechamento ajustado (geralmente igual ao `close` para cripto).""" @@ -37,6 +39,10 @@ class CoinHistoricalDataPrice(BaseModel): class Coin(BaseModel): + """ + Contém os dados detalhados de uma criptomoeda específica retornada pelo endpoint `/api/v2/crypto`. + """ + coin: Optional[str] = None """Sigla (ticker) da criptomoeda (ex: `BTC`, `ETH`).""" @@ -113,5 +119,7 @@ class Coin(BaseModel): class CryptoRetrieveResponse(BaseModel): + """Resposta principal do endpoint `/api/v2/crypto`.""" + coins: Optional[List[Coin]] = None """Array contendo os resultados detalhados para cada criptomoeda solicitada.""" diff --git a/src/brapi/types/v2/currency_list_available_response.py b/src/brapi/types/v2/currency_list_available_response.py index 2a515fc..0ceda29 100644 --- a/src/brapi/types/v2/currency_list_available_response.py +++ b/src/brapi/types/v2/currency_list_available_response.py @@ -17,6 +17,8 @@ class Currency(BaseModel): class CurrencyListAvailableResponse(BaseModel): + """Resposta do endpoint que lista todas as moedas fiduciárias disponíveis.""" + currencies: Optional[List[Currency]] = None """ Lista de objetos, cada um contendo o nome de uma moeda fiduciária ou par diff --git a/src/brapi/types/v2/currency_retrieve_response.py b/src/brapi/types/v2/currency_retrieve_response.py index 7cc09b5..fdf3b33 100644 --- a/src/brapi/types/v2/currency_retrieve_response.py +++ b/src/brapi/types/v2/currency_retrieve_response.py @@ -10,6 +10,10 @@ class Currency(BaseModel): + """ + Contém os dados detalhados da cotação de um **par de moedas fiduciárias específico**, retornado como um elemento do array `currency` no endpoint `/api/v2/currency`. + """ + ask_price: str = FieldInfo(alias="askPrice") """ **Preço de Venda (Ask):** Preço atual pelo qual o mercado está disposto a vender @@ -77,6 +81,8 @@ class Currency(BaseModel): class CurrencyRetrieveResponse(BaseModel): + """Estrutura da **resposta principal** do endpoint `GET /api/v2/currency`.""" + currency: List[Currency] """ Array contendo os objetos `CurrencyQuote`, um para cada par de moeda válido diff --git a/src/brapi/types/v2/inflation_list_available_response.py b/src/brapi/types/v2/inflation_list_available_response.py index 971e902..964adf4 100644 --- a/src/brapi/types/v2/inflation_list_available_response.py +++ b/src/brapi/types/v2/inflation_list_available_response.py @@ -8,6 +8,8 @@ class InflationListAvailableResponse(BaseModel): + """Resposta do endpoint que lista os países com dados de inflação disponíveis.""" + countries: Optional[List[str]] = None """ Lista de nomes de países (em minúsculas) para os quais há dados de inflação diff --git a/src/brapi/types/v2/inflation_retrieve_response.py b/src/brapi/types/v2/inflation_retrieve_response.py index 864488a..a7daeaa 100644 --- a/src/brapi/types/v2/inflation_retrieve_response.py +++ b/src/brapi/types/v2/inflation_retrieve_response.py @@ -10,6 +10,8 @@ class Inflation(BaseModel): + """Representa um ponto de dado histórico de inflação para um país.""" + date: Optional[str] = None """Data da medição da inflação, no formato `DD/MM/YYYY`.""" @@ -27,6 +29,8 @@ class Inflation(BaseModel): class InflationRetrieveResponse(BaseModel): + """Resposta principal do endpoint `/api/v2/inflation`.""" + inflation: Optional[List[Inflation]] = None """ Array contendo os registros históricos de inflação para o país e período diff --git a/src/brapi/types/v2/prime_rate_list_available_response.py b/src/brapi/types/v2/prime_rate_list_available_response.py index 4cfacdd..67e901a 100644 --- a/src/brapi/types/v2/prime_rate_list_available_response.py +++ b/src/brapi/types/v2/prime_rate_list_available_response.py @@ -8,6 +8,10 @@ class PrimeRateListAvailableResponse(BaseModel): + """ + Resposta do endpoint `/api/v2/prime-rate/available` que lista os países disponíveis para consulta de taxa básica de juros (SELIC). + """ + countries: Optional[List[str]] = None """ Lista de países com dados de taxa básica de juros (SELIC) disponíveis para diff --git a/src/brapi/types/v2/prime_rate_retrieve_response.py b/src/brapi/types/v2/prime_rate_retrieve_response.py index 967860e..b4e9a78 100644 --- a/src/brapi/types/v2/prime_rate_retrieve_response.py +++ b/src/brapi/types/v2/prime_rate_retrieve_response.py @@ -10,6 +10,10 @@ class PrimeRate(BaseModel): + """ + Representa um registro individual de taxa básica de juros (SELIC) para uma data específica. + """ + date: Optional[str] = None """Data do registro no formato DD/MM/YYYY.""" @@ -21,6 +25,8 @@ class PrimeRate(BaseModel): class PrimeRateRetrieveResponse(BaseModel): + """Resposta principal do endpoint `/api/v2/prime-rate`.""" + prime_rate: Optional[List[PrimeRate]] = FieldInfo(alias="prime-rate", default=None) """ Array contendo os registros históricos de taxa básica de juros (SELIC) para o diff --git a/src/brapi/types/value_added_entry.py b/src/brapi/types/value_added_entry.py index 70c0db2..b553683 100644 --- a/src/brapi/types/value_added_entry.py +++ b/src/brapi/types/value_added_entry.py @@ -12,6 +12,10 @@ class ValueAddedEntry(BaseModel): + """ + Representa os dados de uma Demonstração do Valor Adicionado (DVA) para um período específico (anual ou trimestral). A DVA mostra como a riqueza gerada pela empresa foi distribuída. + """ + added_value_received_by_transfer: Optional[float] = FieldInfo(alias="addedValueReceivedByTransfer", default=None) """ Valor Adicionado Recebido em Transferência (Resultado de Equivalência From fa4150eb3a6509a2077720d19fa99730bc40cc02 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 10 Dec 2025 13:12:38 +0000 Subject: [PATCH 26/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index b7634f9..80d368a 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.3" + ".": "1.0.4" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index 6c5e7ad..0ffae3a 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.3" +version = "1.0.4" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index 59676ab..aa6a0a5 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.3" # x-release-please-version +__version__ = "1.0.4" # x-release-please-version From a455c1c876e7d44b05c728e52a2108252d25cc0e Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 16 Dec 2025 04:26:37 +0000 Subject: [PATCH 27/47] chore(internal): add missing files argument to base client --- src/brapi/_base_client.py | 10 ++++++++-- 1 file changed, 8 insertions(+), 2 deletions(-) diff --git a/src/brapi/_base_client.py b/src/brapi/_base_client.py index 2c643d3..ac40494 100644 --- a/src/brapi/_base_client.py +++ b/src/brapi/_base_client.py @@ -1247,9 +1247,12 @@ def patch( *, cast_to: Type[ResponseT], body: Body | None = None, + files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: - opts = FinalRequestOptions.construct(method="patch", url=path, json_data=body, **options) + opts = FinalRequestOptions.construct( + method="patch", url=path, json_data=body, files=to_httpx_files(files), **options + ) return self.request(cast_to, opts) def put( @@ -1767,9 +1770,12 @@ async def patch( *, cast_to: Type[ResponseT], body: Body | None = None, + files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: - opts = FinalRequestOptions.construct(method="patch", url=path, json_data=body, **options) + opts = FinalRequestOptions.construct( + method="patch", url=path, json_data=body, files=to_httpx_files(files), **options + ) return await self.request(cast_to, opts) async def put( From 3d087715d08619f08601f9ad7641989d6e8b6324 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 17 Dec 2025 05:59:26 +0000 Subject: [PATCH 28/47] chore: speedup initial import --- src/brapi/_client.py | 179 ++++++++++++++++++++++++++++++++++--------- 1 file changed, 142 insertions(+), 37 deletions(-) diff --git a/src/brapi/_client.py b/src/brapi/_client.py index 992de0c..80b62ff 100644 --- a/src/brapi/_client.py +++ b/src/brapi/_client.py @@ -3,7 +3,7 @@ from __future__ import annotations import os -from typing import Any, Dict, Mapping, cast +from typing import TYPE_CHECKING, Any, Dict, Mapping, cast from typing_extensions import Self, Literal, override import httpx @@ -20,8 +20,8 @@ not_given, ) from ._utils import is_given, get_async_library +from ._compat import cached_property from ._version import __version__ -from .resources import quote, available from ._streaming import Stream as Stream, AsyncStream as AsyncStream from ._exceptions import BrapiError, APIStatusError from ._base_client import ( @@ -29,7 +29,12 @@ SyncAPIClient, AsyncAPIClient, ) -from .resources.v2 import v2 + +if TYPE_CHECKING: + from .resources import v2, quote, available + from .resources.quote import QuoteResource, AsyncQuoteResource + from .resources.v2.v2 import V2Resource, AsyncV2Resource + from .resources.available import AvailableResource, AsyncAvailableResource __all__ = [ "ENVIRONMENTS", @@ -50,12 +55,6 @@ class Brapi(SyncAPIClient): - quote: quote.QuoteResource - available: available.AvailableResource - v2: v2.V2Resource - with_raw_response: BrapiWithRawResponse - with_streaming_response: BrapiWithStreamedResponse - # client options api_key: str @@ -134,11 +133,31 @@ def __init__( _strict_response_validation=_strict_response_validation, ) - self.quote = quote.QuoteResource(self) - self.available = available.AvailableResource(self) - self.v2 = v2.V2Resource(self) - self.with_raw_response = BrapiWithRawResponse(self) - self.with_streaming_response = BrapiWithStreamedResponse(self) + @cached_property + def quote(self) -> QuoteResource: + from .resources.quote import QuoteResource + + return QuoteResource(self) + + @cached_property + def available(self) -> AvailableResource: + from .resources.available import AvailableResource + + return AvailableResource(self) + + @cached_property + def v2(self) -> V2Resource: + from .resources.v2 import V2Resource + + return V2Resource(self) + + @cached_property + def with_raw_response(self) -> BrapiWithRawResponse: + return BrapiWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> BrapiWithStreamedResponse: + return BrapiWithStreamedResponse(self) @property @override @@ -248,12 +267,6 @@ def _make_status_error( class AsyncBrapi(AsyncAPIClient): - quote: quote.AsyncQuoteResource - available: available.AsyncAvailableResource - v2: v2.AsyncV2Resource - with_raw_response: AsyncBrapiWithRawResponse - with_streaming_response: AsyncBrapiWithStreamedResponse - # client options api_key: str @@ -332,11 +345,31 @@ def __init__( _strict_response_validation=_strict_response_validation, ) - self.quote = quote.AsyncQuoteResource(self) - self.available = available.AsyncAvailableResource(self) - self.v2 = v2.AsyncV2Resource(self) - self.with_raw_response = AsyncBrapiWithRawResponse(self) - self.with_streaming_response = AsyncBrapiWithStreamedResponse(self) + @cached_property + def quote(self) -> AsyncQuoteResource: + from .resources.quote import AsyncQuoteResource + + return AsyncQuoteResource(self) + + @cached_property + def available(self) -> AsyncAvailableResource: + from .resources.available import AsyncAvailableResource + + return AsyncAvailableResource(self) + + @cached_property + def v2(self) -> AsyncV2Resource: + from .resources.v2 import AsyncV2Resource + + return AsyncV2Resource(self) + + @cached_property + def with_raw_response(self) -> AsyncBrapiWithRawResponse: + return AsyncBrapiWithRawResponse(self) + + @cached_property + def with_streaming_response(self) -> AsyncBrapiWithStreamedResponse: + return AsyncBrapiWithStreamedResponse(self) @property @override @@ -446,31 +479,103 @@ def _make_status_error( class BrapiWithRawResponse: + _client: Brapi + def __init__(self, client: Brapi) -> None: - self.quote = quote.QuoteResourceWithRawResponse(client.quote) - self.available = available.AvailableResourceWithRawResponse(client.available) - self.v2 = v2.V2ResourceWithRawResponse(client.v2) + self._client = client + + @cached_property + def quote(self) -> quote.QuoteResourceWithRawResponse: + from .resources.quote import QuoteResourceWithRawResponse + + return QuoteResourceWithRawResponse(self._client.quote) + + @cached_property + def available(self) -> available.AvailableResourceWithRawResponse: + from .resources.available import AvailableResourceWithRawResponse + + return AvailableResourceWithRawResponse(self._client.available) + + @cached_property + def v2(self) -> v2.V2ResourceWithRawResponse: + from .resources.v2 import V2ResourceWithRawResponse + + return V2ResourceWithRawResponse(self._client.v2) class AsyncBrapiWithRawResponse: + _client: AsyncBrapi + def __init__(self, client: AsyncBrapi) -> None: - self.quote = quote.AsyncQuoteResourceWithRawResponse(client.quote) - self.available = available.AsyncAvailableResourceWithRawResponse(client.available) - self.v2 = v2.AsyncV2ResourceWithRawResponse(client.v2) + self._client = client + + @cached_property + def quote(self) -> quote.AsyncQuoteResourceWithRawResponse: + from .resources.quote import AsyncQuoteResourceWithRawResponse + + return AsyncQuoteResourceWithRawResponse(self._client.quote) + + @cached_property + def available(self) -> available.AsyncAvailableResourceWithRawResponse: + from .resources.available import AsyncAvailableResourceWithRawResponse + + return AsyncAvailableResourceWithRawResponse(self._client.available) + + @cached_property + def v2(self) -> v2.AsyncV2ResourceWithRawResponse: + from .resources.v2 import AsyncV2ResourceWithRawResponse + + return AsyncV2ResourceWithRawResponse(self._client.v2) class BrapiWithStreamedResponse: + _client: Brapi + def __init__(self, client: Brapi) -> None: - self.quote = quote.QuoteResourceWithStreamingResponse(client.quote) - self.available = available.AvailableResourceWithStreamingResponse(client.available) - self.v2 = v2.V2ResourceWithStreamingResponse(client.v2) + self._client = client + + @cached_property + def quote(self) -> quote.QuoteResourceWithStreamingResponse: + from .resources.quote import QuoteResourceWithStreamingResponse + + return QuoteResourceWithStreamingResponse(self._client.quote) + + @cached_property + def available(self) -> available.AvailableResourceWithStreamingResponse: + from .resources.available import AvailableResourceWithStreamingResponse + + return AvailableResourceWithStreamingResponse(self._client.available) + + @cached_property + def v2(self) -> v2.V2ResourceWithStreamingResponse: + from .resources.v2 import V2ResourceWithStreamingResponse + + return V2ResourceWithStreamingResponse(self._client.v2) class AsyncBrapiWithStreamedResponse: + _client: AsyncBrapi + def __init__(self, client: AsyncBrapi) -> None: - self.quote = quote.AsyncQuoteResourceWithStreamingResponse(client.quote) - self.available = available.AsyncAvailableResourceWithStreamingResponse(client.available) - self.v2 = v2.AsyncV2ResourceWithStreamingResponse(client.v2) + self._client = client + + @cached_property + def quote(self) -> quote.AsyncQuoteResourceWithStreamingResponse: + from .resources.quote import AsyncQuoteResourceWithStreamingResponse + + return AsyncQuoteResourceWithStreamingResponse(self._client.quote) + + @cached_property + def available(self) -> available.AsyncAvailableResourceWithStreamingResponse: + from .resources.available import AsyncAvailableResourceWithStreamingResponse + + return AsyncAvailableResourceWithStreamingResponse(self._client.available) + + @cached_property + def v2(self) -> v2.AsyncV2ResourceWithStreamingResponse: + from .resources.v2 import AsyncV2ResourceWithStreamingResponse + + return AsyncV2ResourceWithStreamingResponse(self._client.v2) Client = Brapi From 6b0b1b6c67410fcf5b31d269cd7d6323e6bae9e5 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 18 Dec 2025 06:49:08 +0000 Subject: [PATCH 29/47] fix: use async_to_httpx_files in patch method --- src/brapi/_base_client.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/brapi/_base_client.py b/src/brapi/_base_client.py index ac40494..f4e99c6 100644 --- a/src/brapi/_base_client.py +++ b/src/brapi/_base_client.py @@ -1774,7 +1774,7 @@ async def patch( options: RequestOptions = {}, ) -> ResponseT: opts = FinalRequestOptions.construct( - method="patch", url=path, json_data=body, files=to_httpx_files(files), **options + method="patch", url=path, json_data=body, files=await async_to_httpx_files(files), **options ) return await self.request(cast_to, opts) From 7ad61720e72bb3e1b9a45ffd5ce8237f8230b734 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 19 Dec 2025 06:00:30 +0000 Subject: [PATCH 30/47] chore(internal): add `--fix` argument to lint script --- scripts/lint | 9 +++++++-- 1 file changed, 7 insertions(+), 2 deletions(-) diff --git a/scripts/lint b/scripts/lint index e8935a5..5e07b8d 100755 --- a/scripts/lint +++ b/scripts/lint @@ -4,8 +4,13 @@ set -e cd "$(dirname "$0")/.." -echo "==> Running lints" -rye run lint +if [ "$1" = "--fix" ]; then + echo "==> Running lints with --fix" + rye run fix:ruff +else + echo "==> Running lints" + rye run lint +fi echo "==> Making sure it imports" rye run python -c 'import brapi' From b3ba109742f8637530fac60938f3e143a5228b27 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Mon, 29 Dec 2025 15:42:38 +0000 Subject: [PATCH 31/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 80d368a..1214610 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.4" + ".": "1.0.5" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index 0ffae3a..cd620ad 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.4" +version = "1.0.5" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index aa6a0a5..7a778d3 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.4" # x-release-please-version +__version__ = "1.0.5" # x-release-please-version From d6557f4032a636f8da158b265aa3f4aed6d45004 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 6 Jan 2026 05:21:33 +0000 Subject: [PATCH 32/47] chore(internal): codegen related update --- LICENSE | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/LICENSE b/LICENSE index 00b551f..74e4dd7 100644 --- a/LICENSE +++ b/LICENSE @@ -186,7 +186,7 @@ same "printed page" as the copyright notice for easier identification within third-party archives. - Copyright 2025 Brapi + Copyright 2026 Brapi Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. From f38347712a1c1f699650783bfbce306978d85d6c Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 9 Jan 2026 21:21:59 +0000 Subject: [PATCH 33/47] codegen metadata --- .stats.yml | 4 ++-- 1 file changed, 2 insertions(+), 2 deletions(-) diff --git a/.stats.yml b/.stats.yml index 3d5ec16..7d05cc5 100644 --- a/.stats.yml +++ b/.stats.yml @@ -1,4 +1,4 @@ configured_endpoints: 11 -openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-11630ce70d2f2c187912989dd9004b10828eaab889f617ba39d5a7a8e4b03b62.yml -openapi_spec_hash: 4198e5f7a76f3002723c113663465c00 +openapi_spec_url: https://storage.googleapis.com/stainless-sdk-openapi-specs/alisson%2Fbrapi-76a60a630b8eaac37bdec27ffec5cbdf6640fb884186adb08211f1b81832c075.yml +openapi_spec_hash: 51fab4b9fd59ce7421f3fdf03644c987 config_hash: 6f10a67950f65bf850612b59838ad03b From 56b001dc47dec692c1a76efca68e7f7fa035134d Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 14 Jan 2026 07:33:28 +0000 Subject: [PATCH 34/47] feat(client): add support for binary request streaming --- src/brapi/_base_client.py | 145 ++++++++++++++++++++++++++--- src/brapi/_models.py | 17 +++- src/brapi/_types.py | 9 ++ tests/test_client.py | 187 +++++++++++++++++++++++++++++++++++++- 4 files changed, 344 insertions(+), 14 deletions(-) diff --git a/src/brapi/_base_client.py b/src/brapi/_base_client.py index f4e99c6..585efb7 100644 --- a/src/brapi/_base_client.py +++ b/src/brapi/_base_client.py @@ -9,6 +9,7 @@ import inspect import logging import platform +import warnings import email.utils from types import TracebackType from random import random @@ -51,9 +52,11 @@ ResponseT, AnyMapping, PostParser, + BinaryTypes, RequestFiles, HttpxSendArgs, RequestOptions, + AsyncBinaryTypes, HttpxRequestFiles, ModelBuilderProtocol, not_given, @@ -477,8 +480,19 @@ def _build_request( retries_taken: int = 0, ) -> httpx.Request: if log.isEnabledFor(logging.DEBUG): - log.debug("Request options: %s", model_dump(options, exclude_unset=True)) - + log.debug( + "Request options: %s", + model_dump( + options, + exclude_unset=True, + # Pydantic v1 can't dump every type we support in content, so we exclude it for now. + exclude={ + "content", + } + if PYDANTIC_V1 + else {}, + ), + ) kwargs: dict[str, Any] = {} json_data = options.json_data @@ -532,7 +546,13 @@ def _build_request( is_body_allowed = options.method.lower() != "get" if is_body_allowed: - if isinstance(json_data, bytes): + if options.content is not None and json_data is not None: + raise TypeError("Passing both `content` and `json_data` is not supported") + if options.content is not None and files is not None: + raise TypeError("Passing both `content` and `files` is not supported") + if options.content is not None: + kwargs["content"] = options.content + elif isinstance(json_data, bytes): kwargs["content"] = json_data else: kwargs["json"] = json_data if is_given(json_data) else None @@ -1194,6 +1214,7 @@ def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, options: RequestOptions = {}, files: RequestFiles | None = None, stream: Literal[False] = False, @@ -1206,6 +1227,7 @@ def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, options: RequestOptions = {}, files: RequestFiles | None = None, stream: Literal[True], @@ -1219,6 +1241,7 @@ def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, options: RequestOptions = {}, files: RequestFiles | None = None, stream: bool, @@ -1231,13 +1254,25 @@ def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, options: RequestOptions = {}, files: RequestFiles | None = None, stream: bool = False, stream_cls: type[_StreamT] | None = None, ) -> ResponseT | _StreamT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="post", url=path, json_data=body, files=to_httpx_files(files), **options + method="post", url=path, json_data=body, content=content, files=to_httpx_files(files), **options ) return cast(ResponseT, self.request(cast_to, opts, stream=stream, stream_cls=stream_cls)) @@ -1247,11 +1282,23 @@ def patch( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="patch", url=path, json_data=body, files=to_httpx_files(files), **options + method="patch", url=path, json_data=body, content=content, files=to_httpx_files(files), **options ) return self.request(cast_to, opts) @@ -1261,11 +1308,23 @@ def put( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="put", url=path, json_data=body, files=to_httpx_files(files), **options + method="put", url=path, json_data=body, content=content, files=to_httpx_files(files), **options ) return self.request(cast_to, opts) @@ -1275,9 +1334,19 @@ def delete( *, cast_to: Type[ResponseT], body: Body | None = None, + content: BinaryTypes | None = None, options: RequestOptions = {}, ) -> ResponseT: - opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, **options) + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) + opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, content=content, **options) return self.request(cast_to, opts) def get_api_list( @@ -1717,6 +1786,7 @@ async def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, stream: Literal[False] = False, @@ -1729,6 +1799,7 @@ async def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, stream: Literal[True], @@ -1742,6 +1813,7 @@ async def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, stream: bool, @@ -1754,13 +1826,25 @@ async def post( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, stream: bool = False, stream_cls: type[_AsyncStreamT] | None = None, ) -> ResponseT | _AsyncStreamT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="post", url=path, json_data=body, files=await async_to_httpx_files(files), **options + method="post", url=path, json_data=body, content=content, files=await async_to_httpx_files(files), **options ) return await self.request(cast_to, opts, stream=stream, stream_cls=stream_cls) @@ -1770,11 +1854,28 @@ async def patch( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="patch", url=path, json_data=body, files=await async_to_httpx_files(files), **options + method="patch", + url=path, + json_data=body, + content=content, + files=await async_to_httpx_files(files), + **options, ) return await self.request(cast_to, opts) @@ -1784,11 +1885,23 @@ async def put( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, files: RequestFiles | None = None, options: RequestOptions = {}, ) -> ResponseT: + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if files is not None and content is not None: + raise TypeError("Passing both `files` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) opts = FinalRequestOptions.construct( - method="put", url=path, json_data=body, files=await async_to_httpx_files(files), **options + method="put", url=path, json_data=body, content=content, files=await async_to_httpx_files(files), **options ) return await self.request(cast_to, opts) @@ -1798,9 +1911,19 @@ async def delete( *, cast_to: Type[ResponseT], body: Body | None = None, + content: AsyncBinaryTypes | None = None, options: RequestOptions = {}, ) -> ResponseT: - opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, **options) + if body is not None and content is not None: + raise TypeError("Passing both `body` and `content` is not supported") + if isinstance(body, bytes): + warnings.warn( + "Passing raw bytes as `body` is deprecated and will be removed in a future version. " + "Please pass raw bytes via the `content` parameter instead.", + DeprecationWarning, + stacklevel=2, + ) + opts = FinalRequestOptions.construct(method="delete", url=path, json_data=body, content=content, **options) return await self.request(cast_to, opts) def get_api_list( diff --git a/src/brapi/_models.py b/src/brapi/_models.py index ca9500b..29070e0 100644 --- a/src/brapi/_models.py +++ b/src/brapi/_models.py @@ -3,7 +3,20 @@ import os import inspect import weakref -from typing import TYPE_CHECKING, Any, Type, Union, Generic, TypeVar, Callable, Optional, cast +from typing import ( + IO, + TYPE_CHECKING, + Any, + Type, + Union, + Generic, + TypeVar, + Callable, + Iterable, + Optional, + AsyncIterable, + cast, +) from datetime import date, datetime from typing_extensions import ( List, @@ -787,6 +800,7 @@ class FinalRequestOptionsInput(TypedDict, total=False): timeout: float | Timeout | None files: HttpxRequestFiles | None idempotency_key: str + content: Union[bytes, bytearray, IO[bytes], Iterable[bytes], AsyncIterable[bytes], None] json_data: Body extra_json: AnyMapping follow_redirects: bool @@ -805,6 +819,7 @@ class FinalRequestOptions(pydantic.BaseModel): post_parser: Union[Callable[[Any], Any], NotGiven] = NotGiven() follow_redirects: Union[bool, None] = None + content: Union[bytes, bytearray, IO[bytes], Iterable[bytes], AsyncIterable[bytes], None] = None # It should be noted that we cannot use `json` here as that would override # a BaseModel method in an incompatible fashion. json_data: Union[Body, None] = None diff --git a/src/brapi/_types.py b/src/brapi/_types.py index c98ff51..60c1ea8 100644 --- a/src/brapi/_types.py +++ b/src/brapi/_types.py @@ -13,9 +13,11 @@ Mapping, TypeVar, Callable, + Iterable, Iterator, Optional, Sequence, + AsyncIterable, ) from typing_extensions import ( Set, @@ -56,6 +58,13 @@ else: Base64FileInput = Union[IO[bytes], PathLike] FileContent = Union[IO[bytes], bytes, PathLike] # PathLike is not subscriptable in Python 3.8. + + +# Used for sending raw binary data / streaming data in request bodies +# e.g. for file uploads without multipart encoding +BinaryTypes = Union[bytes, bytearray, IO[bytes], Iterable[bytes]] +AsyncBinaryTypes = Union[bytes, bytearray, IO[bytes], AsyncIterable[bytes]] + FileTypes = Union[ # file (or bytes) FileContent, diff --git a/tests/test_client.py b/tests/test_client.py index f57ce48..ad052f5 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -8,10 +8,11 @@ import json import asyncio import inspect +import dataclasses import tracemalloc -from typing import Any, Union, cast +from typing import Any, Union, TypeVar, Callable, Iterable, Iterator, Optional, Coroutine, cast from unittest import mock -from typing_extensions import Literal +from typing_extensions import Literal, AsyncIterator, override import httpx import pytest @@ -36,6 +37,7 @@ from .utils import update_env +T = TypeVar("T") base_url = os.environ.get("TEST_API_BASE_URL", "http://127.0.0.1:4010") api_key = "My API Key" @@ -50,6 +52,57 @@ def _low_retry_timeout(*_args: Any, **_kwargs: Any) -> float: return 0.1 +def mirror_request_content(request: httpx.Request) -> httpx.Response: + return httpx.Response(200, content=request.content) + + +# note: we can't use the httpx.MockTransport class as it consumes the request +# body itself, which means we can't test that the body is read lazily +class MockTransport(httpx.BaseTransport, httpx.AsyncBaseTransport): + def __init__( + self, + handler: Callable[[httpx.Request], httpx.Response] + | Callable[[httpx.Request], Coroutine[Any, Any, httpx.Response]], + ) -> None: + self.handler = handler + + @override + def handle_request( + self, + request: httpx.Request, + ) -> httpx.Response: + assert not inspect.iscoroutinefunction(self.handler), "handler must not be a coroutine function" + assert inspect.isfunction(self.handler), "handler must be a function" + return self.handler(request) + + @override + async def handle_async_request( + self, + request: httpx.Request, + ) -> httpx.Response: + assert inspect.iscoroutinefunction(self.handler), "handler must be a coroutine function" + return await self.handler(request) + + +@dataclasses.dataclass +class Counter: + value: int = 0 + + +def _make_sync_iterator(iterable: Iterable[T], counter: Optional[Counter] = None) -> Iterator[T]: + for item in iterable: + if counter: + counter.value += 1 + yield item + + +async def _make_async_iterator(iterable: Iterable[T], counter: Optional[Counter] = None) -> AsyncIterator[T]: + for item in iterable: + if counter: + counter.value += 1 + yield item + + def _get_open_connections(client: Brapi | AsyncBrapi) -> int: transport = client._client._transport assert isinstance(transport, httpx.HTTPTransport) or isinstance(transport, httpx.AsyncHTTPTransport) @@ -500,6 +553,70 @@ def test_multipart_repeating_array(self, client: Brapi) -> None: b"", ] + @pytest.mark.respx(base_url=base_url) + def test_binary_content_upload(self, respx_mock: MockRouter, client: Brapi) -> None: + respx_mock.post("/upload").mock(side_effect=mirror_request_content) + + file_content = b"Hello, this is a test file." + + response = client.post( + "/upload", + content=file_content, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + + def test_binary_content_upload_with_iterator(self) -> None: + file_content = b"Hello, this is a test file." + counter = Counter() + iterator = _make_sync_iterator([file_content], counter=counter) + + def mock_handler(request: httpx.Request) -> httpx.Response: + assert counter.value == 0, "the request body should not have been read" + return httpx.Response(200, content=request.read()) + + with Brapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.Client(transport=MockTransport(handler=mock_handler)), + ) as client: + response = client.post( + "/upload", + content=iterator, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + assert counter.value == 1 + + @pytest.mark.respx(base_url=base_url) + def test_binary_content_upload_with_body_is_deprecated(self, respx_mock: MockRouter, client: Brapi) -> None: + respx_mock.post("/upload").mock(side_effect=mirror_request_content) + + file_content = b"Hello, this is a test file." + + with pytest.deprecated_call( + match="Passing raw bytes as `body` is deprecated and will be removed in a future version. Please pass raw bytes via the `content` parameter instead." + ): + response = client.post( + "/upload", + body=file_content, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + @pytest.mark.respx(base_url=base_url) def test_basic_union_response(self, respx_mock: MockRouter, client: Brapi) -> None: class Model1(BaseModel): @@ -1331,6 +1448,72 @@ def test_multipart_repeating_array(self, async_client: AsyncBrapi) -> None: b"", ] + @pytest.mark.respx(base_url=base_url) + async def test_binary_content_upload(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: + respx_mock.post("/upload").mock(side_effect=mirror_request_content) + + file_content = b"Hello, this is a test file." + + response = await async_client.post( + "/upload", + content=file_content, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + + async def test_binary_content_upload_with_asynciterator(self) -> None: + file_content = b"Hello, this is a test file." + counter = Counter() + iterator = _make_async_iterator([file_content], counter=counter) + + async def mock_handler(request: httpx.Request) -> httpx.Response: + assert counter.value == 0, "the request body should not have been read" + return httpx.Response(200, content=await request.aread()) + + async with AsyncBrapi( + base_url=base_url, + api_key=api_key, + _strict_response_validation=True, + http_client=httpx.AsyncClient(transport=MockTransport(handler=mock_handler)), + ) as client: + response = await client.post( + "/upload", + content=iterator, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + assert counter.value == 1 + + @pytest.mark.respx(base_url=base_url) + async def test_binary_content_upload_with_body_is_deprecated( + self, respx_mock: MockRouter, async_client: AsyncBrapi + ) -> None: + respx_mock.post("/upload").mock(side_effect=mirror_request_content) + + file_content = b"Hello, this is a test file." + + with pytest.deprecated_call( + match="Passing raw bytes as `body` is deprecated and will be removed in a future version. Please pass raw bytes via the `content` parameter instead." + ): + response = await async_client.post( + "/upload", + body=file_content, + cast_to=httpx.Response, + options={"headers": {"Content-Type": "application/octet-stream"}}, + ) + + assert response.status_code == 200 + assert response.request.headers["Content-Type"] == "application/octet-stream" + assert response.content == file_content + @pytest.mark.respx(base_url=base_url) async def test_basic_union_response(self, respx_mock: MockRouter, async_client: AsyncBrapi) -> None: class Model1(BaseModel): From f0704ba2bcec7344e1cc610b95367911223f2224 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 14 Jan 2026 15:20:40 +0000 Subject: [PATCH 35/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 1214610..2601677 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.0.5" + ".": "1.1.0" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index cd620ad..f9e2cb9 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.0.5" +version = "1.1.0" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index 7a778d3..d3062e3 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.0.5" # x-release-please-version +__version__ = "1.1.0" # x-release-please-version From 87d08ca28e2b185dfdee5f854195c9b32763ea6f Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sat, 17 Jan 2026 05:54:33 +0000 Subject: [PATCH 36/47] chore(internal): update `actions/checkout` version --- .github/workflows/ci.yml | 6 +++--- .github/workflows/publish-pypi.yml | 2 +- .github/workflows/release-doctor.yml | 2 +- 3 files changed, 5 insertions(+), 5 deletions(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 78bafd2..654a199 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -19,7 +19,7 @@ jobs: runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} if: github.event_name == 'push' || github.event.pull_request.head.repo.fork steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Install Rye run: | @@ -44,7 +44,7 @@ jobs: id-token: write runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Install Rye run: | @@ -81,7 +81,7 @@ jobs: runs-on: ${{ github.repository == 'stainless-sdks/brapi-python' && 'depot-ubuntu-24.04' || 'ubuntu-latest' }} if: github.event_name == 'push' || github.event.pull_request.head.repo.fork steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Install Rye run: | diff --git a/.github/workflows/publish-pypi.yml b/.github/workflows/publish-pypi.yml index 8181d42..b339e1f 100644 --- a/.github/workflows/publish-pypi.yml +++ b/.github/workflows/publish-pypi.yml @@ -14,7 +14,7 @@ jobs: runs-on: ubuntu-latest steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Install Rye run: | diff --git a/.github/workflows/release-doctor.yml b/.github/workflows/release-doctor.yml index 790ebde..1739d31 100644 --- a/.github/workflows/release-doctor.yml +++ b/.github/workflows/release-doctor.yml @@ -12,7 +12,7 @@ jobs: if: github.repository == 'brapi-dev/brapi-python' && (github.event_name == 'push' || github.event_name == 'workflow_dispatch' || startsWith(github.head_ref, 'release-please') || github.head_ref == 'next') steps: - - uses: actions/checkout@v4 + - uses: actions/checkout@v6 - name: Check release environment run: | From d202c2c0a45986dfab6c9a274a08b5ca545f9e6f Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Sat, 24 Jan 2026 05:23:44 +0000 Subject: [PATCH 37/47] chore(ci): upgrade `actions/github-script` --- .github/workflows/ci.yml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/.github/workflows/ci.yml b/.github/workflows/ci.yml index 654a199..120d183 100644 --- a/.github/workflows/ci.yml +++ b/.github/workflows/ci.yml @@ -63,7 +63,7 @@ jobs: - name: Get GitHub OIDC Token if: github.repository == 'stainless-sdks/brapi-python' id: github-oidc - uses: actions/github-script@v6 + uses: actions/github-script@v8 with: script: core.setOutput('github_token', await core.getIDToken()); From 2879eac2c3e63a7e5b12b71313240e6d097ed7e1 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 30 Jan 2026 05:10:55 +0000 Subject: [PATCH 38/47] feat(client): add custom JSON encoder for extended type support --- src/brapi/_base_client.py | 7 +- src/brapi/_compat.py | 6 +- src/brapi/_utils/_json.py | 35 ++++++++++ tests/test_utils/test_json.py | 126 ++++++++++++++++++++++++++++++++++ 4 files changed, 169 insertions(+), 5 deletions(-) create mode 100644 src/brapi/_utils/_json.py create mode 100644 tests/test_utils/test_json.py diff --git a/src/brapi/_base_client.py b/src/brapi/_base_client.py index 585efb7..9adac9a 100644 --- a/src/brapi/_base_client.py +++ b/src/brapi/_base_client.py @@ -86,6 +86,7 @@ APIConnectionError, APIResponseValidationError, ) +from ._utils._json import openapi_dumps log: logging.Logger = logging.getLogger(__name__) @@ -554,8 +555,10 @@ def _build_request( kwargs["content"] = options.content elif isinstance(json_data, bytes): kwargs["content"] = json_data - else: - kwargs["json"] = json_data if is_given(json_data) else None + elif not files: + # Don't set content when JSON is sent as multipart/form-data, + # since httpx's content param overrides other body arguments + kwargs["content"] = openapi_dumps(json_data) if is_given(json_data) and json_data is not None else None kwargs["files"] = files else: headers.pop("Content-Type", None) diff --git a/src/brapi/_compat.py b/src/brapi/_compat.py index bdef67f..786ff42 100644 --- a/src/brapi/_compat.py +++ b/src/brapi/_compat.py @@ -139,6 +139,7 @@ def model_dump( exclude_defaults: bool = False, warnings: bool = True, mode: Literal["json", "python"] = "python", + by_alias: bool | None = None, ) -> dict[str, Any]: if (not PYDANTIC_V1) or hasattr(model, "model_dump"): return model.model_dump( @@ -148,13 +149,12 @@ def model_dump( exclude_defaults=exclude_defaults, # warnings are not supported in Pydantic v1 warnings=True if PYDANTIC_V1 else warnings, + by_alias=by_alias, ) return cast( "dict[str, Any]", model.dict( # pyright: ignore[reportDeprecated, reportUnnecessaryCast] - exclude=exclude, - exclude_unset=exclude_unset, - exclude_defaults=exclude_defaults, + exclude=exclude, exclude_unset=exclude_unset, exclude_defaults=exclude_defaults, by_alias=bool(by_alias) ), ) diff --git a/src/brapi/_utils/_json.py b/src/brapi/_utils/_json.py new file mode 100644 index 0000000..6058421 --- /dev/null +++ b/src/brapi/_utils/_json.py @@ -0,0 +1,35 @@ +import json +from typing import Any +from datetime import datetime +from typing_extensions import override + +import pydantic + +from .._compat import model_dump + + +def openapi_dumps(obj: Any) -> bytes: + """ + Serialize an object to UTF-8 encoded JSON bytes. + + Extends the standard json.dumps with support for additional types + commonly used in the SDK, such as `datetime`, `pydantic.BaseModel`, etc. + """ + return json.dumps( + obj, + cls=_CustomEncoder, + # Uses the same defaults as httpx's JSON serialization + ensure_ascii=False, + separators=(",", ":"), + allow_nan=False, + ).encode() + + +class _CustomEncoder(json.JSONEncoder): + @override + def default(self, o: Any) -> Any: + if isinstance(o, datetime): + return o.isoformat() + if isinstance(o, pydantic.BaseModel): + return model_dump(o, exclude_unset=True, mode="json", by_alias=True) + return super().default(o) diff --git a/tests/test_utils/test_json.py b/tests/test_utils/test_json.py new file mode 100644 index 0000000..ca6ea27 --- /dev/null +++ b/tests/test_utils/test_json.py @@ -0,0 +1,126 @@ +from __future__ import annotations + +import datetime +from typing import Union + +import pydantic + +from brapi import _compat +from brapi._utils._json import openapi_dumps + + +class TestOpenapiDumps: + def test_basic(self) -> None: + data = {"key": "value", "number": 42} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"key":"value","number":42}' + + def test_datetime_serialization(self) -> None: + dt = datetime.datetime(2023, 1, 1, 12, 0, 0) + data = {"datetime": dt} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"datetime":"2023-01-01T12:00:00"}' + + def test_pydantic_model_serialization(self) -> None: + class User(pydantic.BaseModel): + first_name: str + last_name: str + age: int + + model_instance = User(first_name="John", last_name="Kramer", age=83) + data = {"model": model_instance} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"first_name":"John","last_name":"Kramer","age":83}}' + + def test_pydantic_model_with_default_values(self) -> None: + class User(pydantic.BaseModel): + name: str + role: str = "user" + active: bool = True + score: int = 0 + + model_instance = User(name="Alice") + data = {"model": model_instance} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"name":"Alice"}}' + + def test_pydantic_model_with_default_values_overridden(self) -> None: + class User(pydantic.BaseModel): + name: str + role: str = "user" + active: bool = True + + model_instance = User(name="Bob", role="admin", active=False) + data = {"model": model_instance} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"name":"Bob","role":"admin","active":false}}' + + def test_pydantic_model_with_alias(self) -> None: + class User(pydantic.BaseModel): + first_name: str = pydantic.Field(alias="firstName") + last_name: str = pydantic.Field(alias="lastName") + + model_instance = User(firstName="John", lastName="Doe") + data = {"model": model_instance} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"firstName":"John","lastName":"Doe"}}' + + def test_pydantic_model_with_alias_and_default(self) -> None: + class User(pydantic.BaseModel): + user_name: str = pydantic.Field(alias="userName") + user_role: str = pydantic.Field(default="member", alias="userRole") + is_active: bool = pydantic.Field(default=True, alias="isActive") + + model_instance = User(userName="charlie") + data = {"model": model_instance} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"userName":"charlie"}}' + + model_with_overrides = User(userName="diana", userRole="admin", isActive=False) + data = {"model": model_with_overrides} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"userName":"diana","userRole":"admin","isActive":false}}' + + def test_pydantic_model_with_nested_models_and_defaults(self) -> None: + class Address(pydantic.BaseModel): + street: str + city: str = "Unknown" + + class User(pydantic.BaseModel): + name: str + address: Address + verified: bool = False + + if _compat.PYDANTIC_V1: + # to handle forward references in Pydantic v1 + User.update_forward_refs(**locals()) # type: ignore[reportDeprecated] + + address = Address(street="123 Main St") + user = User(name="Diana", address=address) + data = {"user": user} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"user":{"name":"Diana","address":{"street":"123 Main St"}}}' + + address_with_city = Address(street="456 Oak Ave", city="Boston") + user_verified = User(name="Eve", address=address_with_city, verified=True) + data = {"user": user_verified} + json_bytes = openapi_dumps(data) + assert ( + json_bytes == b'{"user":{"name":"Eve","address":{"street":"456 Oak Ave","city":"Boston"},"verified":true}}' + ) + + def test_pydantic_model_with_optional_fields(self) -> None: + class User(pydantic.BaseModel): + name: str + email: Union[str, None] + phone: Union[str, None] + + model_with_none = User(name="Eve", email=None, phone=None) + data = {"model": model_with_none} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"name":"Eve","email":null,"phone":null}}' + + model_with_values = User(name="Frank", email="frank@example.com", phone=None) + data = {"model": model_with_values} + json_bytes = openapi_dumps(data) + assert json_bytes == b'{"model":{"name":"Frank","email":"frank@example.com","phone":null}}' From 1664399499932467f7ec4912bed0574497b29772 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Mon, 2 Feb 2026 01:35:51 +0000 Subject: [PATCH 39/47] chore(internal): version bump --- .release-please-manifest.json | 2 +- pyproject.toml | 2 +- src/brapi/_version.py | 2 +- 3 files changed, 3 insertions(+), 3 deletions(-) diff --git a/.release-please-manifest.json b/.release-please-manifest.json index 2601677..d0ab664 100644 --- a/.release-please-manifest.json +++ b/.release-please-manifest.json @@ -1,3 +1,3 @@ { - ".": "1.1.0" + ".": "1.2.0" } \ No newline at end of file diff --git a/pyproject.toml b/pyproject.toml index f9e2cb9..ecae5b3 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -1,6 +1,6 @@ [project] name = "brapi" -version = "1.1.0" +version = "1.2.0" description = "The official Python library for the brapi API" dynamic = ["readme"] license = "Apache-2.0" diff --git a/src/brapi/_version.py b/src/brapi/_version.py index d3062e3..165bd5b 100644 --- a/src/brapi/_version.py +++ b/src/brapi/_version.py @@ -1,4 +1,4 @@ # File generated from our OpenAPI spec by Stainless. See CONTRIBUTING.md for details. __title__ = "brapi" -__version__ = "1.1.0" # x-release-please-version +__version__ = "1.2.0" # x-release-please-version From 71a5b0a7d491a989beb1bd21309586cbdecde252 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 10 Feb 2026 05:21:49 +0000 Subject: [PATCH 40/47] chore(internal): bump dependencies --- requirements-dev.lock | 20 ++++++++++---------- requirements.lock | 8 ++++---- 2 files changed, 14 insertions(+), 14 deletions(-) diff --git a/requirements-dev.lock b/requirements-dev.lock index 659c238..3581988 100644 --- a/requirements-dev.lock +++ b/requirements-dev.lock @@ -12,14 +12,14 @@ -e file:. aiohappyeyeballs==2.6.1 # via aiohttp -aiohttp==3.13.2 +aiohttp==3.13.3 # via brapi # via httpx-aiohttp aiosignal==1.4.0 # via aiohttp annotated-types==0.7.0 # via pydantic -anyio==4.12.0 +anyio==4.12.1 # via brapi # via httpx argcomplete==3.6.3 @@ -31,7 +31,7 @@ attrs==25.4.0 # via nox backports-asyncio-runner==1.2.0 # via pytest-asyncio -certifi==2025.11.12 +certifi==2026.1.4 # via httpcore # via httpx colorlog==6.10.1 @@ -61,7 +61,7 @@ httpx==0.28.1 # via brapi # via httpx-aiohttp # via respx -httpx-aiohttp==0.1.9 +httpx-aiohttp==0.1.12 # via brapi humanize==4.13.0 # via nox @@ -69,7 +69,7 @@ idna==3.11 # via anyio # via httpx # via yarl -importlib-metadata==8.7.0 +importlib-metadata==8.7.1 iniconfig==2.1.0 # via pytest markdown-it-py==3.0.0 @@ -82,14 +82,14 @@ multidict==6.7.0 mypy==1.17.0 mypy-extensions==1.1.0 # via mypy -nodeenv==1.9.1 +nodeenv==1.10.0 # via pyright nox==2025.11.12 packaging==25.0 # via dependency-groups # via nox # via pytest -pathspec==0.12.1 +pathspec==1.0.3 # via mypy platformdirs==4.4.0 # via virtualenv @@ -115,13 +115,13 @@ python-dateutil==2.9.0.post0 # via time-machine respx==0.22.0 rich==14.2.0 -ruff==0.14.7 +ruff==0.14.13 six==1.17.0 # via python-dateutil sniffio==1.3.1 # via brapi time-machine==2.19.0 -tomli==2.3.0 +tomli==2.4.0 # via dependency-groups # via mypy # via nox @@ -141,7 +141,7 @@ typing-extensions==4.15.0 # via virtualenv typing-inspection==0.4.2 # via pydantic -virtualenv==20.35.4 +virtualenv==20.36.1 # via nox yarl==1.22.0 # via aiohttp diff --git a/requirements.lock b/requirements.lock index f030e74..b885522 100644 --- a/requirements.lock +++ b/requirements.lock @@ -12,21 +12,21 @@ -e file:. aiohappyeyeballs==2.6.1 # via aiohttp -aiohttp==3.13.2 +aiohttp==3.13.3 # via brapi # via httpx-aiohttp aiosignal==1.4.0 # via aiohttp annotated-types==0.7.0 # via pydantic -anyio==4.12.0 +anyio==4.12.1 # via brapi # via httpx async-timeout==5.0.1 # via aiohttp attrs==25.4.0 # via aiohttp -certifi==2025.11.12 +certifi==2026.1.4 # via httpcore # via httpx distro==1.9.0 @@ -43,7 +43,7 @@ httpcore==1.0.9 httpx==0.28.1 # via brapi # via httpx-aiohttp -httpx-aiohttp==0.1.9 +httpx-aiohttp==0.1.12 # via brapi idna==3.11 # via anyio From 1c353194205e2adb3671225b8381be553c90f827 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Thu, 12 Feb 2026 06:59:56 +0000 Subject: [PATCH 41/47] chore(internal): fix lint error on Python 3.14 --- src/brapi/_utils/_compat.py | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/src/brapi/_utils/_compat.py b/src/brapi/_utils/_compat.py index dd70323..2c70b29 100644 --- a/src/brapi/_utils/_compat.py +++ b/src/brapi/_utils/_compat.py @@ -26,7 +26,7 @@ def is_union(tp: Optional[Type[Any]]) -> bool: else: import types - return tp is Union or tp is types.UnionType + return tp is Union or tp is types.UnionType # type: ignore[comparison-overlap] def is_typeddict(tp: Type[Any]) -> bool: From 83cbbbda4f39b47922d82fb0cfb76a74363b9e17 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 13 Feb 2026 04:46:11 +0000 Subject: [PATCH 42/47] chore: format all `api.md` files --- pyproject.toml | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/pyproject.toml b/pyproject.toml index ecae5b3..b7861d4 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -69,7 +69,7 @@ format = { chain = [ # run formatting again to fix any inconsistencies when imports are stripped "format:ruff", ]} -"format:docs" = "python scripts/utils/ruffen-docs.py README.md api.md" +"format:docs" = "bash -c 'python scripts/utils/ruffen-docs.py README.md $(find . -type f -name api.md)'" "format:ruff" = "ruff format" "lint" = { chain = [ From 60eee7b843567c18cd3c3d73dd15e0e9d1f97569 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 20 Feb 2026 06:26:51 +0000 Subject: [PATCH 43/47] chore(internal): remove mock server code --- scripts/mock | 41 ----------------------------------------- scripts/test | 46 ---------------------------------------------- 2 files changed, 87 deletions(-) delete mode 100755 scripts/mock diff --git a/scripts/mock b/scripts/mock deleted file mode 100755 index 0b28f6e..0000000 --- a/scripts/mock +++ /dev/null @@ -1,41 +0,0 @@ -#!/usr/bin/env bash - -set -e - -cd "$(dirname "$0")/.." - -if [[ -n "$1" && "$1" != '--'* ]]; then - URL="$1" - shift -else - URL="$(grep 'openapi_spec_url' .stats.yml | cut -d' ' -f2)" -fi - -# Check if the URL is empty -if [ -z "$URL" ]; then - echo "Error: No OpenAPI spec path/url provided or found in .stats.yml" - exit 1 -fi - -echo "==> Starting mock server with URL ${URL}" - -# Run prism mock on the given spec -if [ "$1" == "--daemon" ]; then - npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock "$URL" &> .prism.log & - - # Wait for server to come online - echo -n "Waiting for server" - while ! grep -q "✖ fatal\|Prism is listening" ".prism.log" ; do - echo -n "." - sleep 0.1 - done - - if grep -q "✖ fatal" ".prism.log"; then - cat .prism.log - exit 1 - fi - - echo -else - npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock "$URL" -fi diff --git a/scripts/test b/scripts/test index dbeda2d..39729d0 100755 --- a/scripts/test +++ b/scripts/test @@ -4,53 +4,7 @@ set -e cd "$(dirname "$0")/.." -RED='\033[0;31m' -GREEN='\033[0;32m' -YELLOW='\033[0;33m' -NC='\033[0m' # No Color -function prism_is_running() { - curl --silent "http://localhost:4010" >/dev/null 2>&1 -} - -kill_server_on_port() { - pids=$(lsof -t -i tcp:"$1" || echo "") - if [ "$pids" != "" ]; then - kill "$pids" - echo "Stopped $pids." - fi -} - -function is_overriding_api_base_url() { - [ -n "$TEST_API_BASE_URL" ] -} - -if ! is_overriding_api_base_url && ! prism_is_running ; then - # When we exit this script, make sure to kill the background mock server process - trap 'kill_server_on_port 4010' EXIT - - # Start the dev server - ./scripts/mock --daemon -fi - -if is_overriding_api_base_url ; then - echo -e "${GREEN}✔ Running tests against ${TEST_API_BASE_URL}${NC}" - echo -elif ! prism_is_running ; then - echo -e "${RED}ERROR:${NC} The test suite will not run without a mock Prism server" - echo -e "running against your OpenAPI spec." - echo - echo -e "To run the server, pass in the path or url of your OpenAPI" - echo -e "spec to the prism command:" - echo - echo -e " \$ ${YELLOW}npm exec --package=@stainless-api/prism-cli@5.15.0 -- prism mock path/to/your.openapi.yml${NC}" - echo - - exit 1 -else - echo -e "${GREEN}✔ Mock prism server is running with your OpenAPI spec${NC}" - echo -fi export DEFER_PYDANTIC_BUILD=false From 8ed2a1860140a9b393c28fb312570c3f3476c3af Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Fri, 20 Feb 2026 06:28:01 +0000 Subject: [PATCH 44/47] chore: update mock server docs --- CONTRIBUTING.md | 7 ----- tests/api_resources/test_available.py | 16 +++++----- tests/api_resources/test_quote.py | 36 +++++++++++------------ tests/api_resources/v2/test_crypto.py | 32 ++++++++++---------- tests/api_resources/v2/test_currency.py | 32 ++++++++++---------- tests/api_resources/v2/test_inflation.py | 32 ++++++++++---------- tests/api_resources/v2/test_prime_rate.py | 32 ++++++++++---------- 7 files changed, 90 insertions(+), 97 deletions(-) diff --git a/CONTRIBUTING.md b/CONTRIBUTING.md index 5068a3b..9d51ff5 100644 --- a/CONTRIBUTING.md +++ b/CONTRIBUTING.md @@ -85,13 +85,6 @@ $ pip install ./path-to-wheel-file.whl ## Running tests -Most tests require you to [set up a mock server](https://github.com/stoplightio/prism) against the OpenAPI spec to run the tests. - -```sh -# you will need npm installed -$ npx prism mock path/to/your/openapi.yml -``` - ```sh $ ./scripts/test ``` diff --git a/tests/api_resources/test_available.py b/tests/api_resources/test_available.py index 6b366e0..59ececf 100644 --- a/tests/api_resources/test_available.py +++ b/tests/api_resources/test_available.py @@ -17,13 +17,13 @@ class TestAvailable: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list(self, client: Brapi) -> None: available = client.available.list() assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_with_all_params(self, client: Brapi) -> None: available = client.available.list( @@ -32,7 +32,7 @@ def test_method_list_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list(self, client: Brapi) -> None: response = client.available.with_raw_response.list() @@ -42,7 +42,7 @@ def test_raw_response_list(self, client: Brapi) -> None: available = response.parse() assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list(self, client: Brapi) -> None: with client.available.with_streaming_response.list() as response: @@ -60,13 +60,13 @@ class TestAsyncAvailable: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list(self, async_client: AsyncBrapi) -> None: available = await async_client.available.list() assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> None: available = await async_client.available.list( @@ -75,7 +75,7 @@ async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> No ) assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: response = await async_client.available.with_raw_response.list() @@ -85,7 +85,7 @@ async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: available = await response.parse() assert_matches_type(AvailableListResponse, available, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list(self, async_client: AsyncBrapi) -> None: async with async_client.available.with_streaming_response.list() as response: diff --git a/tests/api_resources/test_quote.py b/tests/api_resources/test_quote.py index 21a14a6..a80bd69 100644 --- a/tests/api_resources/test_quote.py +++ b/tests/api_resources/test_quote.py @@ -17,7 +17,7 @@ class TestQuote: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve(self, client: Brapi) -> None: quote = client.quote.retrieve( @@ -25,7 +25,7 @@ def test_method_retrieve(self, client: Brapi) -> None: ) assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve_with_all_params(self, client: Brapi) -> None: quote = client.quote.retrieve( @@ -39,7 +39,7 @@ def test_method_retrieve_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_retrieve(self, client: Brapi) -> None: response = client.quote.with_raw_response.retrieve( @@ -51,7 +51,7 @@ def test_raw_response_retrieve(self, client: Brapi) -> None: quote = response.parse() assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_retrieve(self, client: Brapi) -> None: with client.quote.with_streaming_response.retrieve( @@ -65,7 +65,7 @@ def test_streaming_response_retrieve(self, client: Brapi) -> None: assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_path_params_retrieve(self, client: Brapi) -> None: with pytest.raises(ValueError, match=r"Expected a non-empty value for `tickers` but received ''"): @@ -73,13 +73,13 @@ def test_path_params_retrieve(self, client: Brapi) -> None: tickers="", ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list(self, client: Brapi) -> None: quote = client.quote.list() assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_with_all_params(self, client: Brapi) -> None: quote = client.quote.list( @@ -94,7 +94,7 @@ def test_method_list_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list(self, client: Brapi) -> None: response = client.quote.with_raw_response.list() @@ -104,7 +104,7 @@ def test_raw_response_list(self, client: Brapi) -> None: quote = response.parse() assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list(self, client: Brapi) -> None: with client.quote.with_streaming_response.list() as response: @@ -122,7 +122,7 @@ class TestAsyncQuote: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: quote = await async_client.quote.retrieve( @@ -130,7 +130,7 @@ async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: ) assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: quote = await async_client.quote.retrieve( @@ -144,7 +144,7 @@ async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) - ) assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: response = await async_client.quote.with_raw_response.retrieve( @@ -156,7 +156,7 @@ async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: quote = await response.parse() assert_matches_type(QuoteRetrieveResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: async with async_client.quote.with_streaming_response.retrieve( @@ -170,7 +170,7 @@ async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> No assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_path_params_retrieve(self, async_client: AsyncBrapi) -> None: with pytest.raises(ValueError, match=r"Expected a non-empty value for `tickers` but received ''"): @@ -178,13 +178,13 @@ async def test_path_params_retrieve(self, async_client: AsyncBrapi) -> None: tickers="", ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list(self, async_client: AsyncBrapi) -> None: quote = await async_client.quote.list() assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> None: quote = await async_client.quote.list( @@ -199,7 +199,7 @@ async def test_method_list_with_all_params(self, async_client: AsyncBrapi) -> No ) assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: response = await async_client.quote.with_raw_response.list() @@ -209,7 +209,7 @@ async def test_raw_response_list(self, async_client: AsyncBrapi) -> None: quote = await response.parse() assert_matches_type(QuoteListResponse, quote, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list(self, async_client: AsyncBrapi) -> None: async with async_client.quote.with_streaming_response.list() as response: diff --git a/tests/api_resources/v2/test_crypto.py b/tests/api_resources/v2/test_crypto.py index 957328e..5379377 100644 --- a/tests/api_resources/v2/test_crypto.py +++ b/tests/api_resources/v2/test_crypto.py @@ -20,7 +20,7 @@ class TestCrypto: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve(self, client: Brapi) -> None: crypto = client.v2.crypto.retrieve( @@ -28,7 +28,7 @@ def test_method_retrieve(self, client: Brapi) -> None: ) assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve_with_all_params(self, client: Brapi) -> None: crypto = client.v2.crypto.retrieve( @@ -40,7 +40,7 @@ def test_method_retrieve_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_retrieve(self, client: Brapi) -> None: response = client.v2.crypto.with_raw_response.retrieve( @@ -52,7 +52,7 @@ def test_raw_response_retrieve(self, client: Brapi) -> None: crypto = response.parse() assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_retrieve(self, client: Brapi) -> None: with client.v2.crypto.with_streaming_response.retrieve( @@ -66,13 +66,13 @@ def test_streaming_response_retrieve(self, client: Brapi) -> None: assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available(self, client: Brapi) -> None: crypto = client.v2.crypto.list_available() assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available_with_all_params(self, client: Brapi) -> None: crypto = client.v2.crypto.list_available( @@ -81,7 +81,7 @@ def test_method_list_available_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list_available(self, client: Brapi) -> None: response = client.v2.crypto.with_raw_response.list_available() @@ -91,7 +91,7 @@ def test_raw_response_list_available(self, client: Brapi) -> None: crypto = response.parse() assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list_available(self, client: Brapi) -> None: with client.v2.crypto.with_streaming_response.list_available() as response: @@ -109,7 +109,7 @@ class TestAsyncCrypto: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: crypto = await async_client.v2.crypto.retrieve( @@ -117,7 +117,7 @@ async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: ) assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: crypto = await async_client.v2.crypto.retrieve( @@ -129,7 +129,7 @@ async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) - ) assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.crypto.with_raw_response.retrieve( @@ -141,7 +141,7 @@ async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: crypto = await response.parse() assert_matches_type(CryptoRetrieveResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: async with async_client.v2.crypto.with_streaming_response.retrieve( @@ -155,13 +155,13 @@ async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> No assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available(self, async_client: AsyncBrapi) -> None: crypto = await async_client.v2.crypto.list_available() assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: crypto = await async_client.v2.crypto.list_available( @@ -170,7 +170,7 @@ async def test_method_list_available_with_all_params(self, async_client: AsyncBr ) assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.crypto.with_raw_response.list_available() @@ -180,7 +180,7 @@ async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> No crypto = await response.parse() assert_matches_type(CryptoListAvailableResponse, crypto, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: async with async_client.v2.crypto.with_streaming_response.list_available() as response: diff --git a/tests/api_resources/v2/test_currency.py b/tests/api_resources/v2/test_currency.py index 0326fc1..3af6cfd 100644 --- a/tests/api_resources/v2/test_currency.py +++ b/tests/api_resources/v2/test_currency.py @@ -20,7 +20,7 @@ class TestCurrency: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve(self, client: Brapi) -> None: currency = client.v2.currency.retrieve( @@ -28,7 +28,7 @@ def test_method_retrieve(self, client: Brapi) -> None: ) assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve_with_all_params(self, client: Brapi) -> None: currency = client.v2.currency.retrieve( @@ -37,7 +37,7 @@ def test_method_retrieve_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_retrieve(self, client: Brapi) -> None: response = client.v2.currency.with_raw_response.retrieve( @@ -49,7 +49,7 @@ def test_raw_response_retrieve(self, client: Brapi) -> None: currency = response.parse() assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_retrieve(self, client: Brapi) -> None: with client.v2.currency.with_streaming_response.retrieve( @@ -63,13 +63,13 @@ def test_streaming_response_retrieve(self, client: Brapi) -> None: assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available(self, client: Brapi) -> None: currency = client.v2.currency.list_available() assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available_with_all_params(self, client: Brapi) -> None: currency = client.v2.currency.list_available( @@ -78,7 +78,7 @@ def test_method_list_available_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list_available(self, client: Brapi) -> None: response = client.v2.currency.with_raw_response.list_available() @@ -88,7 +88,7 @@ def test_raw_response_list_available(self, client: Brapi) -> None: currency = response.parse() assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list_available(self, client: Brapi) -> None: with client.v2.currency.with_streaming_response.list_available() as response: @@ -106,7 +106,7 @@ class TestAsyncCurrency: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: currency = await async_client.v2.currency.retrieve( @@ -114,7 +114,7 @@ async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: ) assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: currency = await async_client.v2.currency.retrieve( @@ -123,7 +123,7 @@ async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) - ) assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.currency.with_raw_response.retrieve( @@ -135,7 +135,7 @@ async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: currency = await response.parse() assert_matches_type(CurrencyRetrieveResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: async with async_client.v2.currency.with_streaming_response.retrieve( @@ -149,13 +149,13 @@ async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> No assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available(self, async_client: AsyncBrapi) -> None: currency = await async_client.v2.currency.list_available() assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: currency = await async_client.v2.currency.list_available( @@ -164,7 +164,7 @@ async def test_method_list_available_with_all_params(self, async_client: AsyncBr ) assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.currency.with_raw_response.list_available() @@ -174,7 +174,7 @@ async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> No currency = await response.parse() assert_matches_type(CurrencyListAvailableResponse, currency, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: async with async_client.v2.currency.with_streaming_response.list_available() as response: diff --git a/tests/api_resources/v2/test_inflation.py b/tests/api_resources/v2/test_inflation.py index 4071b4d..ac67b5e 100644 --- a/tests/api_resources/v2/test_inflation.py +++ b/tests/api_resources/v2/test_inflation.py @@ -21,13 +21,13 @@ class TestInflation: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve(self, client: Brapi) -> None: inflation = client.v2.inflation.retrieve() assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve_with_all_params(self, client: Brapi) -> None: inflation = client.v2.inflation.retrieve( @@ -41,7 +41,7 @@ def test_method_retrieve_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_retrieve(self, client: Brapi) -> None: response = client.v2.inflation.with_raw_response.retrieve() @@ -51,7 +51,7 @@ def test_raw_response_retrieve(self, client: Brapi) -> None: inflation = response.parse() assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_retrieve(self, client: Brapi) -> None: with client.v2.inflation.with_streaming_response.retrieve() as response: @@ -63,13 +63,13 @@ def test_streaming_response_retrieve(self, client: Brapi) -> None: assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available(self, client: Brapi) -> None: inflation = client.v2.inflation.list_available() assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available_with_all_params(self, client: Brapi) -> None: inflation = client.v2.inflation.list_available( @@ -78,7 +78,7 @@ def test_method_list_available_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list_available(self, client: Brapi) -> None: response = client.v2.inflation.with_raw_response.list_available() @@ -88,7 +88,7 @@ def test_raw_response_list_available(self, client: Brapi) -> None: inflation = response.parse() assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list_available(self, client: Brapi) -> None: with client.v2.inflation.with_streaming_response.list_available() as response: @@ -106,13 +106,13 @@ class TestAsyncInflation: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: inflation = await async_client.v2.inflation.retrieve() assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: inflation = await async_client.v2.inflation.retrieve( @@ -126,7 +126,7 @@ async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) - ) assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.inflation.with_raw_response.retrieve() @@ -136,7 +136,7 @@ async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: inflation = await response.parse() assert_matches_type(InflationRetrieveResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: async with async_client.v2.inflation.with_streaming_response.retrieve() as response: @@ -148,13 +148,13 @@ async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> No assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available(self, async_client: AsyncBrapi) -> None: inflation = await async_client.v2.inflation.list_available() assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: inflation = await async_client.v2.inflation.list_available( @@ -163,7 +163,7 @@ async def test_method_list_available_with_all_params(self, async_client: AsyncBr ) assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.inflation.with_raw_response.list_available() @@ -173,7 +173,7 @@ async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> No inflation = await response.parse() assert_matches_type(InflationListAvailableResponse, inflation, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: async with async_client.v2.inflation.with_streaming_response.list_available() as response: diff --git a/tests/api_resources/v2/test_prime_rate.py b/tests/api_resources/v2/test_prime_rate.py index ecb5738..ec59675 100644 --- a/tests/api_resources/v2/test_prime_rate.py +++ b/tests/api_resources/v2/test_prime_rate.py @@ -21,13 +21,13 @@ class TestPrimeRate: parametrize = pytest.mark.parametrize("client", [False, True], indirect=True, ids=["loose", "strict"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve(self, client: Brapi) -> None: prime_rate = client.v2.prime_rate.retrieve() assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_retrieve_with_all_params(self, client: Brapi) -> None: prime_rate = client.v2.prime_rate.retrieve( @@ -41,7 +41,7 @@ def test_method_retrieve_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_retrieve(self, client: Brapi) -> None: response = client.v2.prime_rate.with_raw_response.retrieve() @@ -51,7 +51,7 @@ def test_raw_response_retrieve(self, client: Brapi) -> None: prime_rate = response.parse() assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_retrieve(self, client: Brapi) -> None: with client.v2.prime_rate.with_streaming_response.retrieve() as response: @@ -63,13 +63,13 @@ def test_streaming_response_retrieve(self, client: Brapi) -> None: assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available(self, client: Brapi) -> None: prime_rate = client.v2.prime_rate.list_available() assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_method_list_available_with_all_params(self, client: Brapi) -> None: prime_rate = client.v2.prime_rate.list_available( @@ -78,7 +78,7 @@ def test_method_list_available_with_all_params(self, client: Brapi) -> None: ) assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_raw_response_list_available(self, client: Brapi) -> None: response = client.v2.prime_rate.with_raw_response.list_available() @@ -88,7 +88,7 @@ def test_raw_response_list_available(self, client: Brapi) -> None: prime_rate = response.parse() assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize def test_streaming_response_list_available(self, client: Brapi) -> None: with client.v2.prime_rate.with_streaming_response.list_available() as response: @@ -106,13 +106,13 @@ class TestAsyncPrimeRate: "async_client", [False, True, {"http_client": "aiohttp"}], indirect=True, ids=["loose", "strict", "aiohttp"] ) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve(self, async_client: AsyncBrapi) -> None: prime_rate = await async_client.v2.prime_rate.retrieve() assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) -> None: prime_rate = await async_client.v2.prime_rate.retrieve( @@ -126,7 +126,7 @@ async def test_method_retrieve_with_all_params(self, async_client: AsyncBrapi) - ) assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.prime_rate.with_raw_response.retrieve() @@ -136,7 +136,7 @@ async def test_raw_response_retrieve(self, async_client: AsyncBrapi) -> None: prime_rate = await response.parse() assert_matches_type(PrimeRateRetrieveResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> None: async with async_client.v2.prime_rate.with_streaming_response.retrieve() as response: @@ -148,13 +148,13 @@ async def test_streaming_response_retrieve(self, async_client: AsyncBrapi) -> No assert cast(Any, response.is_closed) is True - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available(self, async_client: AsyncBrapi) -> None: prime_rate = await async_client.v2.prime_rate.list_available() assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_method_list_available_with_all_params(self, async_client: AsyncBrapi) -> None: prime_rate = await async_client.v2.prime_rate.list_available( @@ -163,7 +163,7 @@ async def test_method_list_available_with_all_params(self, async_client: AsyncBr ) assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> None: response = await async_client.v2.prime_rate.with_raw_response.list_available() @@ -173,7 +173,7 @@ async def test_raw_response_list_available(self, async_client: AsyncBrapi) -> No prime_rate = await response.parse() assert_matches_type(PrimeRateListAvailableResponse, prime_rate, path=["response"]) - @pytest.mark.skip(reason="Prism tests are disabled") + @pytest.mark.skip(reason="Mock server tests are disabled") @parametrize async def test_streaming_response_list_available(self, async_client: AsyncBrapi) -> None: async with async_client.v2.prime_rate.with_streaming_response.list_available() as response: From f495dbdf764a98956d32f2a9c16839e16c8adb68 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 24 Feb 2026 06:41:04 +0000 Subject: [PATCH 45/47] chore(internal): add request options to SSE classes --- src/brapi/_response.py | 3 +++ src/brapi/_streaming.py | 11 ++++++++--- 2 files changed, 11 insertions(+), 3 deletions(-) diff --git a/src/brapi/_response.py b/src/brapi/_response.py index 0f7b71f..b2aad2b 100644 --- a/src/brapi/_response.py +++ b/src/brapi/_response.py @@ -152,6 +152,7 @@ def _parse(self, *, to: type[_T] | None = None) -> R | _T: ), response=self.http_response, client=cast(Any, self._client), + options=self._options, ), ) @@ -162,6 +163,7 @@ def _parse(self, *, to: type[_T] | None = None) -> R | _T: cast_to=extract_stream_chunk_type(self._stream_cls), response=self.http_response, client=cast(Any, self._client), + options=self._options, ), ) @@ -175,6 +177,7 @@ def _parse(self, *, to: type[_T] | None = None) -> R | _T: cast_to=cast_to, response=self.http_response, client=cast(Any, self._client), + options=self._options, ), ) diff --git a/src/brapi/_streaming.py b/src/brapi/_streaming.py index e5f7242..258e79d 100644 --- a/src/brapi/_streaming.py +++ b/src/brapi/_streaming.py @@ -4,7 +4,7 @@ import json import inspect from types import TracebackType -from typing import TYPE_CHECKING, Any, Generic, TypeVar, Iterator, AsyncIterator, cast +from typing import TYPE_CHECKING, Any, Generic, TypeVar, Iterator, Optional, AsyncIterator, cast from typing_extensions import Self, Protocol, TypeGuard, override, get_origin, runtime_checkable import httpx @@ -13,6 +13,7 @@ if TYPE_CHECKING: from ._client import Brapi, AsyncBrapi + from ._models import FinalRequestOptions _T = TypeVar("_T") @@ -22,7 +23,7 @@ class Stream(Generic[_T]): """Provides the core interface to iterate over a synchronous stream response.""" response: httpx.Response - + _options: Optional[FinalRequestOptions] = None _decoder: SSEBytesDecoder def __init__( @@ -31,10 +32,12 @@ def __init__( cast_to: type[_T], response: httpx.Response, client: Brapi, + options: Optional[FinalRequestOptions] = None, ) -> None: self.response = response self._cast_to = cast_to self._client = client + self._options = options self._decoder = client._make_sse_decoder() self._iterator = self.__stream__() @@ -85,7 +88,7 @@ class AsyncStream(Generic[_T]): """Provides the core interface to iterate over an asynchronous stream response.""" response: httpx.Response - + _options: Optional[FinalRequestOptions] = None _decoder: SSEDecoder | SSEBytesDecoder def __init__( @@ -94,10 +97,12 @@ def __init__( cast_to: type[_T], response: httpx.Response, client: AsyncBrapi, + options: Optional[FinalRequestOptions] = None, ) -> None: self.response = response self._cast_to = cast_to self._client = client + self._options = options self._decoder = client._make_sse_decoder() self._iterator = self.__stream__() From 8bd0ee9078838832014e9569a4acedcb484fb31d Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Tue, 24 Feb 2026 06:51:14 +0000 Subject: [PATCH 46/47] chore(internal): make `test_proxy_environment_variables` more resilient --- tests/test_client.py | 4 ++++ 1 file changed, 4 insertions(+) diff --git a/tests/test_client.py b/tests/test_client.py index ad052f5..a170d3d 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -957,6 +957,8 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: # Test that the proxy environment variables are set correctly monkeypatch.setenv("HTTPS_PROXY", "https://example.org") + # Delete in case our environment has this set + monkeypatch.delenv("HTTP_PROXY", raising=False) client = DefaultHttpxClient() @@ -1873,6 +1875,8 @@ async def test_get_platform(self) -> None: async def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: # Test that the proxy environment variables are set correctly monkeypatch.setenv("HTTPS_PROXY", "https://example.org") + # Delete in case our environment has this set + monkeypatch.delenv("HTTP_PROXY", raising=False) client = DefaultAsyncHttpxClient() From f6fbe4d9575625cb289ff165ae2f2ac7e6878ed3 Mon Sep 17 00:00:00 2001 From: "stainless-app[bot]" <142633134+stainless-app[bot]@users.noreply.github.com> Date: Wed, 25 Feb 2026 06:31:02 +0000 Subject: [PATCH 47/47] chore(internal): make `test_proxy_environment_variables` more resilient to env --- tests/test_client.py | 16 ++++++++++++++-- 1 file changed, 14 insertions(+), 2 deletions(-) diff --git a/tests/test_client.py b/tests/test_client.py index a170d3d..62158e8 100644 --- a/tests/test_client.py +++ b/tests/test_client.py @@ -957,8 +957,14 @@ def retry_handler(_request: httpx.Request) -> httpx.Response: def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: # Test that the proxy environment variables are set correctly monkeypatch.setenv("HTTPS_PROXY", "https://example.org") - # Delete in case our environment has this set + # Delete in case our environment has any proxy env vars set monkeypatch.delenv("HTTP_PROXY", raising=False) + monkeypatch.delenv("ALL_PROXY", raising=False) + monkeypatch.delenv("NO_PROXY", raising=False) + monkeypatch.delenv("http_proxy", raising=False) + monkeypatch.delenv("https_proxy", raising=False) + monkeypatch.delenv("all_proxy", raising=False) + monkeypatch.delenv("no_proxy", raising=False) client = DefaultHttpxClient() @@ -1875,8 +1881,14 @@ async def test_get_platform(self) -> None: async def test_proxy_environment_variables(self, monkeypatch: pytest.MonkeyPatch) -> None: # Test that the proxy environment variables are set correctly monkeypatch.setenv("HTTPS_PROXY", "https://example.org") - # Delete in case our environment has this set + # Delete in case our environment has any proxy env vars set monkeypatch.delenv("HTTP_PROXY", raising=False) + monkeypatch.delenv("ALL_PROXY", raising=False) + monkeypatch.delenv("NO_PROXY", raising=False) + monkeypatch.delenv("http_proxy", raising=False) + monkeypatch.delenv("https_proxy", raising=False) + monkeypatch.delenv("all_proxy", raising=False) + monkeypatch.delenv("no_proxy", raising=False) client = DefaultAsyncHttpxClient()