- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 27 for TILE (0.04 sec)
-
src/cmd/vendor/golang.org/x/mod/sumdb/tlog/tile.go
package tlog import ( "fmt" "strconv" "strings" ) // A Tile is a description of a transparency log tile. // A tile of height H at level L offset N lists W consecutive hashes // at level H*L of the tree starting at offset N*(2**H). // A complete tile lists 2**H hashes; a partial tile lists fewer. // Note that a tile represents the entire subtree of height H // with those hashes as the leaves. The levels above H*L
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Fri May 03 02:38:00 UTC 2024 - 13K bytes - Viewed (0) -
src/cmd/vendor/golang.org/x/mod/sumdb/client.go
return "/" + tile.Path() } // readTile reads a single tile, either from the on-disk cache or the server. func (c *Client) readTile(tile tlog.Tile) ([]byte, error) { type cached struct { data []byte err error } result := c.tileCache.Do(tile, func() interface{} { // Try the requested tile in on-disk cache. data, err := c.ops.ReadCache(c.tileCacheKey(tile))
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Thu Feb 08 17:50:49 UTC 2024 - 19.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.cc
} } else { output_type = input_type; } return builder->create<mlir::TF::ConcatOp>( location, output_type, concat_dimension_op.getOutput(), inputs); } // For tile sharded inputs to TPU computation, inject split op between the // input values and TPU computation so that tiled input values are passed in // as inputs to TPU computations. If more than one dimension is sharded, then
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 22 21:28:13 UTC 2024 - 34K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/tests/fuse-tftext.mlir
%177 = "tf.Reshape"(%176, %9) {device = ""} : (tensor<i32>, tensor<0xi32>) -> tensor<i32> %178 = "tf.Pack"(%7, %177) {axis = 0 : i64, device = ""} : (tensor<i32>, tensor<i32>) -> tensor<2xi32> %179 = "tf.Tile"(%115, %178) {device = ""} : (tensor<?x1xi64>, tensor<2xi32>) -> tensor<?x?xi64> %180 = "tf.Mul"(%177, %118) {device = ""} : (tensor<i32>, tensor<i32>) -> tensor<i32>
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 02 09:41:17 UTC 2024 - 460.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/attributes.mlir
%multiples = "tf.Const"() { device = "/device:CPU:0", value = dense<[7,3]> : tensor<2xi32> } : () -> tensor<2xi32> // CHECK: T = !corert.string %output = "tf.Tile"(%arg, %multiples) { device = "/device:CPU:0" } : (tensor<1x2x!tf_type.string>, tensor<2xi32>) -> tensor<?x6x!tf_type.string> func.return %output : tensor<?x6x!tf_type.string> } // CHECK-LABEL: func @shape
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed May 08 00:18:59 UTC 2024 - 4.8K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/xla_sharding_util.h
// A sharding with OTHER type may be REPLICATED if: // 'replicate_on_last_tile_dim' is true OR // 'last_tile_dims' is not empty // AND // other than replicated last tile dims, all other dims are not sharded. bool IsOtherReplicatedSharding(const xla::OpSharding& xla_sharding); // Returns whether the sharding is split sharding. i.e. A sharding with OTHER // type but not replicated.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 28 22:18:34 UTC 2024 - 6K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/tests/constant-fold.mlir
%2 = "tf.ExpandDims"(%1, %const_0) : (tensor<10000xi32>, tensor<i32>) -> tensor<1x10000xi32> %3 = "tf.Tile"(%2, %const_64_1) : (tensor<1x10000xi32>, tensor<2xi32>) -> tensor<64x10000xi32> // CHECK-NOT: tf.Range // CHECK-NOT: tf.ExpandDims // CHECK: [[TILE:%.*]] = "tf.Tile" // CHECK: return [[TILE]] func.return %3 : tensor<64x10000xi32> }
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jan 31 23:22:24 UTC 2024 - 36.7K bytes - Viewed (0) -
src/vendor/golang.org/x/sys/cpu/cpu.go
HasAVX512BITALG bool // Advanced vector extension 512 Bit Algorithms HasAVX512BF16 bool // Advanced vector extension 512 BFloat16 Instructions HasAMXTile bool // Advanced Matrix Extension Tile instructions HasAMXInt8 bool // Advanced Matrix Extension Int8 instructions HasAMXBF16 bool // Advanced Matrix Extension BFloat16 instructions HasBMI1 bool // Bit manipulation instruction set 1
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Wed May 08 16:12:58 UTC 2024 - 12.1K bytes - Viewed (0) -
tensorflow/compiler/mlir/lite/transforms/optimize_patterns.td
"TFL::HasOneUseOrUsedByOnlyBinaryOps($0)">>; // Pattern for skipping Tile if it is mainly for broadcasting and the // Op is already supporting broadcasting. multiclass FuseTileBroadcastIntoFollowingBinary<Op BinaryOp> { def FuseTileBroadcastToBinaryOp1#BinaryOp : Pat< (BinaryOp:$result (TFL_TileOp $input, (Arith_ConstantOp $tile)), $operand, $act_func), (BinaryOp $input, $operand, $act_func),
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu May 16 20:31:41 UTC 2024 - 66.4K bytes - Viewed (0) -
cmd/update.go
if vsphereVersion != "" { uaAppend(" vsphere-plugin-", vsphereVersion) } } if IsPCFTile() { pcfTileVersion := env.Get("MINIO_PCF_TILE_VERSION", "") if pcfTileVersion != "" { uaAppend(" pcf-tile-", pcfTileVersion) } } uaAppend("; ", "") if cpus, err := gopsutilcpu.Info(); err == nil && len(cpus) > 0 { cpuMap := make(map[string]struct{}, len(cpus))
Registered: Sun Jun 16 00:44:34 UTC 2024 - Last Modified: Fri May 24 23:05:23 UTC 2024 - 18.7K bytes - Viewed (0)