- Sort Score
- Num 10 results
- Language All
Results 31 - 40 of 61 for overheads (0.27 seconds)
The search processing time has exceeded the limit. The displayed results may be partial.
-
tensorflow/c/c_api_experimental.h
// in tests only.) TF_CAPI_EXPORT int TF_PickUnusedPortOrDie(void); // Fast path method that makes constructing a single scalar tensor require less // overhead and copies. TF_CAPI_EXPORT extern TFE_TensorHandle* TFE_NewTensorHandleFromScalar( TF_DataType data_type, void* data, size_t len, TF_Status* status); // Specify the server_def that enables collective ops.
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 15.1K bytes - Click Count (0) -
android/guava/src/com/google/common/collect/Ordering.java
* serializable. * * <p><b>Java 8+ users:</b> Use {@code thisComparator.thenComparing(secondaryComparator)} instead. * Depending on what {@code secondaryComparator} is, one of the other overloads of {@code * thenComparing} may be even more useful. */ public <U extends T> Ordering<U> compound(Comparator<? super U> secondaryComparator) {Created: Fri Dec 26 12:43:10 GMT 2025 - Last Modified: Tue Sep 23 17:50:58 GMT 2025 - 39.5K bytes - Click Count (0) -
docs/en/docs/advanced/custom-response.md
But if you are certain that the content that you are returning is **serializable with JSON**, you can pass it directly to the response class and avoid the extra overhead that FastAPI would have by passing your return content through the `jsonable_encoder` before passing it to the response class. {* ../../docs_src/custom_response/tutorial001b_py39.py hl[2,7] *} /// infoCreated: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 20:41:43 GMT 2025 - 12.4K bytes - Click Count (0) -
docs/security/README.md
The *Secure Channel* splits the object content into chunks of a fixed size of `65536` bytes. The last chunk may be smaller to avoid adding additional overhead and is treated specially to prevent truncation attacks. The nonce value is 96 bits long and generated randomly per object / multi-part part. The *Secure Channel* supports plaintexts up to `65536 * 2^32 = 256 TiB`. #### Randomness
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Wed Feb 26 09:25:50 GMT 2025 - 13.8K bytes - Click Count (0) -
build-logic/kotlin-dsl-shared-runtime/src/main/kotlin/org/gradle/kotlin/dsl/internal/sharedruntime/codegen/ApiExtensionGeneratorFacade.kt
// 2. all type parameters must all be reifeid or values (TypeOf, KClass or Class) // 3. when overloading, prefer TypeOf over Class // 4. in case the policy forbids your overloads, discuss private fun kotlinExtensionFunctionsFor(type: ApiType, sinceSupplier: (String) -> String?): Sequence<KotlinExtensionFunction> = candidatesForExtensionFrom(type) .sortedWithTypeOfTakingFunctionsFirst()
Created: Wed Dec 31 11:36:14 GMT 2025 - Last Modified: Wed Mar 12 15:56:18 GMT 2025 - 17.8K bytes - Click Count (0) -
android/guava/src/com/google/common/math/Quantiles.java
* form}), the memory requirement is 8*N bytes for the copy of the dataset plus an overhead which is * independent of N (but depends on the quantiles being computed). When calling {@link * ScaleAndIndex#computeInPlace computeInPlace} (in {@linkplain ScaleAndIndexes#computeInPlace * either form}), only the overhead is required. The number of object allocations is independent of * N in both cases. *Created: Fri Dec 26 12:43:10 GMT 2025 - Last Modified: Mon Mar 17 20:26:29 GMT 2025 - 30.1K bytes - Click Count (0) -
docs/bucket/replication/DESIGN.md
### Multi destination replication The replication design for multiple sites works in a similar manner as described above for two site scenario. However there are some important exceptions.
Created: Sun Dec 28 19:28:13 GMT 2025 - Last Modified: Tue Aug 12 18:20:36 GMT 2025 - 14.7K bytes - Click Count (0) -
RELEASE.md
issue found, _BigtableXYZDataset (of type DatasetV2) does not implement the _as_variant_tensor() of DatasetV1, when moving contrib.bigtable to tensorflow_io. Converting into DatasetV2 removes the overheads to maintain V1 while we are moving into TF 2.0. * Add dataset ops to the graph (or create kernels in Eager execution) during the python Dataset object creation instead doing it during
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Tue Oct 28 22:27:41 GMT 2025 - 740.4K bytes - Click Count (3) -
guava/src/com/google/common/util/concurrent/ExecutionSequencer.java
* non-static, and is initialized/removed frequently - this causes churn in the Thread specific * hashmaps. Using a static ThreadLocal to avoid that overhead would mean that different * ExecutionSequencer objects interfere with each other, which would be undesirable, in addition * to increasing the memory footprint of every thread that interacted with it. In order to release
Created: Fri Dec 26 12:43:10 GMT 2025 - Last Modified: Tue Sep 23 01:35:55 GMT 2025 - 22.1K bytes - Click Count (0) -
okhttp/src/commonJvmAndroid/kotlin/okhttp3/internal/http2/Http2Stream.kt
) : Sink { /** * Buffer of outgoing data. This batches writes of small writes into this sink as larges frames * written to the outgoing connection. Batching saves the (small) framing overhead. */ private val sendBuffer = Buffer() /** Trailers to send at the end of the stream. */ var trailers: Headers? = null var closed: Boolean = falseCreated: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Mon Jul 07 18:57:05 GMT 2025 - 22.4K bytes - Click Count (0)