- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 33 for level (0.17 sec)
-
tests/test_include_router_defaults_overrides.py
url = "" content_type_level = "0" if override1: url += "/level1" content_type_level = "1" if override2: url += "/level2" content_type_level = "2" if override3: url += "/level3" content_type_level = "3" if override4: url += "/level4" content_type_level = "4" if override5: url += "/override5"
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Fri Jun 30 18:25:16 GMT 2023 - 358.6K bytes - Viewed (0) -
docs/en/docs/advanced/sub-applications.md
"Mounting" means adding a completely "independent" application in a specific path, that then takes care of handling everything under that path, with the _path operations_ declared in that sub-application. ### Top-level application First, create the main, top-level, **FastAPI** application, and its *path operations*: ```Python hl_lines="3 6-8" {!../../../docs_src/sub_applications/tutorial001.py!} ``` ### Sub-application
Plain Text - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Thu Apr 18 19:53:19 GMT 2024 - 2.9K bytes - Viewed (0) -
tests/test_response_model_data_filter_no_inheritance.py
return [pet1, pet2] client = TestClient(app) def test_filter_top_level_model(): response = client.post( "/users", json={"email": "******@****.***", "password": "secret"} ) assert response.json() == {"email": "******@****.***"} def test_filter_second_level_model(): response = client.get("/pets/1") assert response.json() == { "name": "Nibbler",
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Fri Jul 07 17:12:13 GMT 2023 - 1.7K bytes - Viewed (0) -
docs/en/docs/deployment/https.md
And that's what **HTTPS** is, it's just plain **HTTP** inside a **secure TLS connection** instead of a pure (unencrypted) TCP connection. !!! tip Notice that the encryption of the communication happens at the **TCP level**, not at the HTTP level. ### HTTPS Request Now that the client and server (specifically the browser and the TLS Termination Proxy) have an **encrypted TCP connection**, they can start the **HTTP communication**.
Plain Text - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Thu Jan 11 16:31:18 GMT 2024 - 12K bytes - Viewed (0) -
tests/test_response_model_data_filter.py
return [pet1, pet2] client = TestClient(app) def test_filter_top_level_model(): response = client.post( "/users", json={"email": "******@****.***", "password": "secret"} ) assert response.json() == {"email": "******@****.***"} def test_filter_second_level_model(): response = client.get("/pets/1") assert response.json() == { "name": "Nibbler",
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Fri Jul 07 17:12:13 GMT 2023 - 1.7K bytes - Viewed (0) -
.github/actions/notify-translations/app/main.py
return response.data.updateDiscussionComment.comment if __name__ == "__main__": settings = Settings() if settings.input_debug: logging.basicConfig(level=logging.DEBUG) else: logging.basicConfig(level=logging.INFO) logging.debug(f"Using config: {settings.json()}") g = Github(settings.input_token.get_secret_value()) repo = g.get_repo(settings.github_repository)
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Wed Sep 27 23:01:46 GMT 2023 - 12.4K bytes - Viewed (0) -
tests/test_custom_schema_fields.py
name: str if PYDANTIC_V2: model_config = { "json_schema_extra": { "x-something-internal": {"level": 4}, } } else: class Config: schema_extra = { "x-something-internal": {"level": 4}, } @app.get("/foo", response_model=Item) def foo(): return {"name": "Foo item"}
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Fri Jul 07 17:12:13 GMT 2023 - 1.2K bytes - Viewed (0) -
.github/actions/people/app/main.py
"avatarUrl": author.avatarUrl, "url": author.url, } ) return users if __name__ == "__main__": logging.basicConfig(level=logging.INFO) settings = Settings() logging.info(f"Using config: {settings.model_dump_json()}") g = Github(settings.input_token.get_secret_value()) repo = g.get_repo(settings.github_repository)
Python - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Tue Mar 26 17:38:21 GMT 2024 - 19.2K bytes - Viewed (1) -
docs/en/docs/advanced/events.md
Let's imagine that loading the model can **take quite some time**, because it has to read a lot of **data from disk**. So you don't want to do it for every request.
Plain Text - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Thu Apr 18 19:53:19 GMT 2024 - 7.8K bytes - Viewed (0) -
docs/en/docs/tutorial/bigger-applications.md
Plain Text - Registered: Sun Apr 28 07:19:10 GMT 2024 - Last Modified: Thu Apr 18 19:53:19 GMT 2024 - 18.6K bytes - Viewed (0)