- Sort Score
- Num 10 results
- Language All
Results 41 - 50 of 6,836 for to (0.02 seconds)
-
docs/sts/wso2.md
Once WSO2 is up and running, configure WSO2 to generate Self contained id_tokens. In OAuth 2.0 specification there are primarily two ways to provide id_tokens 1. The id_token is an identifier that is hard to guess. For example, a randomly generated string of sufficient length, that the server handling the protected resource can use to lookup the associated authorization information.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 8.7K bytes - Click Count (0) -
docs/en/docs/tutorial/sql-databases.md
A SQLModel `engine` (underneath it's actually a SQLAlchemy `engine`) is what **holds the connections** to the database. You would have **one single `engine` object** for all your code to connect to the same database. {* ../../docs_src/sql_databases/tutorial001_an_py310.py ln[14:18] hl[14:15,17:18] *}
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Tue Dec 02 05:06:56 GMT 2025 - 15.8K bytes - Click Count (0) -
fastapi/encoders.py
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Sat Dec 27 12:54:56 GMT 2025 - 10.7K bytes - Click Count (0) -
PULL_REQUESTS_ETIQUETTE.md
## Tips for Success - **Small PRs**: Easier to review, faster to merge. Split large changes logically. - **Clear Commits**: Use `git rebase -i` to refine history before submitting. - **Engage Early**: Discuss complex changes in issues or Slack (https://slack.min.io) before coding. - **Be Responsive**: Address reviewer feedback promptly to keep PRs moving. - **Learn from Reviews**: Use feedback to improve future contributions. ## Resources
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Sun May 25 16:32:03 GMT 2025 - 4.7K bytes - Click Count (0) -
cmd/server_test.go
c.Assert(err, nil) // execute the request to complete object upload. response, err = s.client.Do(request) c.Assert(err, nil) // assert the status code of the response. c.Assert(response.StatusCode, http.StatusOK) // constructing http request to delete the bucket. // making an attempt to delete an non-empty bucket. // expected to fail.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 118.1K bytes - Click Count (0) -
docs/metrics/prometheus/list.md
| `minio_notify_events_errors_total` | Events that were failed to be sent to the targets (deprecated, please use `minio_notify_target_failed_events` instead) |
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 43.4K bytes - Click Count (0) -
compat/maven-artifact/src/main/java/org/apache/maven/artifact/repository/Authentication.java
} /** * Get the absolute path to the private key file. * * @return absolute path to private key */ public String getPrivateKey() { return privateKey; } /** * Set the absolute path to private key file. * * @param privateKey path to private key in local file system */Created: Sun Dec 28 03:35:09 GMT 2025 - Last Modified: Fri Oct 25 12:31:46 GMT 2024 - 3.1K bytes - Click Count (0) -
tests/migrate_test.go
{name: "unique to uniqueIndex", from: &UniqueStruct3{}, to: &UniqueStruct5{}, checkFunc: checkUniqueIndex}, {name: "unique to uniqueAndUniqueIndex", from: &UniqueStruct3{}, to: &UniqueStruct8{}, checkFunc: checkMyIndex}, {name: "uniqueIndex to uniqueIndex", from: &UniqueStruct5{}, to: &UniqueStruct6{}, checkFunc: checkUniqueIndex},Created: Sun Dec 28 09:35:17 GMT 2025 - Last Modified: Wed Aug 20 04:51:17 GMT 2025 - 65.2K bytes - Click Count (0) -
docs/extensions/s3zip/README.md
## How to enable S3 ZIP behavior ? Ensure to set the following header `x-minio-extract` to `true` in your S3 requests. ## How to access to files inside a ZIP archive Accessing to contents inside an archive can be done using regular S3 API with a modified request path. You just need to append the path of the content inside the archive to the path of the archive itself. e.g.:
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed Apr 10 16:28:27 GMT 2024 - 3K bytes - Click Count (0) -
docs/bucket/versioning/README.md
Spark/Hadoop workloads which use Hadoop MR Committer v1/v2 algorithm upload objects to a temporary prefix in a bucket. These objects are 'renamed' to a different prefix on Job commit. Object storage admins are forced to configure separate ILM policies to expire these objects and their versions to reclaim space. ### Solution
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 12K bytes - Click Count (0)