- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 11 for Scattered (0.43 sec)
-
guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
android/guava/src/com/google/common/cache/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (2) -
android/guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
guava/src/com/google/common/hash/Striped64.java
* * Table entries are of class Cell; a variant of AtomicLong padded * to reduce cache contention on most processors. Padding is * overkill for most Atomics because they are usually irregularly * scattered in memory and thus don't interfere much with each * other. But Atomic objects residing in arrays will tend to be * placed adjacent to each other, and so will most often share
Registered: Wed Jun 12 16:38:11 UTC 2024 - Last Modified: Fri Jun 07 22:25:23 UTC 2024 - 11.5K bytes - Viewed (0) -
staging/src/k8s.io/apiserver/pkg/features/kube_features.go
// // Feature gates should be listed in alphabetical, case-sensitive // (upper before any lower case character) order. This reduces the risk // of code conflicts because changes are more likely to be scattered // across the file. // owner: @ivelichkovich, @tallclair // alpha: v1.27 // beta: v1.28 // stable: v1.30 // kep: https://kep.k8s.io/3716 //
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Tue Jun 04 08:36:46 UTC 2024 - 13.4K bytes - Viewed (0) -
pkg/features/kube_features.go
// // Feature gates should be listed in alphabetical, case-sensitive // (upper before any lower case character) order. This reduces the risk // of code conflicts because changes are more likely to be scattered // across the file. // owner: @ttakahashi21 @mkimuram // kep: https://kep.k8s.io/3294 // alpha: v1.26 // // Enable usage of Provision of PVCs from snapshots in other namespaces
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 22:51:23 UTC 2024 - 45.2K bytes - Viewed (0) -
src/cmd/cgo/internal/test/test.go
} xxpthread_mutex_t; struct ibv_async_event { union { int x; } element; }; struct ibv_context { xxpthread_mutex_t mutex; }; // issue 1635 // Mac OS X's gcc will generate scattered relocation 2/1 for // this function on Darwin/386, and 8l couldn't handle it. // this example is in issue 1635 void scatter() { void *p = scatter; printf("scatter = %p\n", p); }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon May 20 15:50:06 UTC 2024 - 48.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td
Arg<TF_I32OrI64Tensor, [{Array containing the starting indices of the slices that must be scattered to.}]>:$scatter_indices,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Tue Jun 11 23:24:08 UTC 2024 - 793K bytes - Viewed (0) -
pkg/apis/core/validation/validation.go
// been performed. // It doesn't return an error for rootscoped resources with namespace, because namespace should already be cleared before. // TODO: Remove calls to this method scattered in validations of specific resources, e.g., ValidatePodUpdate. func ValidateObjectMeta(meta *metav1.ObjectMeta, requiresNamespace bool, nameFn ValidateNameFunc, fldPath *field.Path) field.ErrorList {
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed May 29 22:40:29 UTC 2024 - 349.5K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/legalize-tf.mlir
} func.func @scatterNdVectorIndices(%arg0: tensor<5x1xi32>, %arg1: tensor<5x3x2xf32>) -> tensor<10x3x2xf32> { %cst = "tf.Const"() { value = dense<[10, 3, 2]> : tensor<3xi32> } : () -> tensor<3xi32> %1 = "tf.ScatterNd"(%arg0, %arg1, %cst) : (tensor<5x1xi32>, tensor<5x3x2xf32>, tensor<3xi32>) -> tensor<10x3x2xf32> func.return %1 : tensor<10x3x2xf32> // CHECK-LABEL:scatterNdVectorIndices
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jun 05 01:54:33 UTC 2024 - 153.4K bytes - Viewed (0)