Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 48 for relu (0.24 sec)

  1. tensorflow/compiler/mlir/tensorflow/utils/cluster_util_test.cc

    func.func @main(%arg0: tensor<?xi32>) -> (tensor<?xi32>, tensor<?xi32>) {
        %0 = "tf.Relu"(%arg0) : (tensor<?xi32>) -> tensor<?xi32>
        %1 = "tf.Relu"(%0) {device = "tpu0"} : (tensor<?xi32>) -> tensor<?xi32>
        %2 = "tf.Add"(%0, %1) {device = "tpu0"} : (tensor<?xi32>, tensor<?xi32>) -> tensor<?xi32>
        %3 = "tf.Relu"(%2) : (tensor<?xi32>) -> tensor<?xi32>
        %4 = "tf.Relu"(%1) {device = "tpu0"} : (tensor<?xi32>) -> tensor<?xi32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri Apr 26 09:37:10 UTC 2024
    - 7.8K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/basic_lstm.mlir

    // CHECK-LABEL: @main
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 1.1K bytes
    - Viewed (0)
  3. tensorflow/c/experimental/gradients/nn_grad_test.cc

    using tensorflow::TF_StatusPtr;
    
    Status ReluModel(AbstractContext* ctx,
                     absl::Span<AbstractTensorHandle* const> inputs,
                     absl::Span<AbstractTensorHandle*> outputs) {
      return ops::Relu(ctx, inputs[0], &outputs[0], "Relu");
    }
    
    Status SparseSoftmaxCrossEntropyWithLogitsModel(
        AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs,
        absl::Span<AbstractTensorHandle*> outputs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Feb 28 13:53:47 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/quantization/common/quantization_lib/quantization_driver_test.cc

          %0 = "tfl.conv_2d"(%arg0, %arg1, %arg2) {dilation_h_factor = 1 : i32, dilation_w_factor = 1 : i32, fused_activation_function = "RELU", padding = "VALID", stride_h = 1 : i32, stride_w = 1 : i32} : (tensor<1x4x4x3xf32>, tensor<3x1x1x3xf32>, tensor<3xf32>) -> tensor<1x4x4x3xf32>
          return %0 : tensor<1x4x4x3xf32>
        }
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 7.9K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/lite/tests/optimize_no_verify.mlir

      %cst = arith.constant dense<0.0> : tensor<2x3xbf16>
      %0 = "tfl.maximum"(%arg0, %cst) : (tensor<2x3xbf16>, tensor<2x3xbf16>) -> tensor<2x3xbf16>
      func.return %0 : tensor<2x3xbf16>
    
      // CHECK: %[[RESULT:.*]] = "tfl.relu"(%arg0)
      // CHECK: return %[[RESULT]]
    }
    
    // CHECK-LABEL: fuseScalarAddIntoConv2dBf16
    func.func @fuseScalarAddIntoConv2dBf16(%arg0: tensor<256x32x32x3xbf16>, %arg1: tensor<16x3x3x3xbf16>) -> tensor<256x8x7x16xbf16> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 5.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/stablehlo/passes/merge_fusion_with_dequantize.cc

        func_op.eraseResult(0);
        func_op.insertResult(0, new_call_op.getResult(0).getType(),
                             /*resultAttrs=*/nullptr);
    
        // Modify the quantized fused function to do dequantize+relu(6).
        rewriter.setInsertionPoint(req_op);
        Value new_result = rewriter.create<mlir::stablehlo::UniformDequantizeOp>(
            req_op.getLoc(), func_op.getResultTypes()[0], req_op.getOperand());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.9K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/tfrt/tests/tf_to_corert/tf_to_corert_pipeline.mlir

          %outputs_12, %control_13 = tf_executor.island wraps "tf.Relu"(%outputs_10) {device = ""} : (tensor<16x112x112x?xf32>) -> tensor<16x112x112x?xf32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 00:18:59 UTC 2024
    - 7.7K bytes
    - Viewed (0)
  8. src/cmd/link/internal/loong64/asm.go

    	//				0: R_LARCH_PCALA_HI20	local.moduledata
    	o(0x1a000004)
    	rel, _ := initfunc.AddRel(objabi.R_LOONG64_ADDR_HI)
    	rel.SetOff(0)
    	rel.SetSiz(4)
    	rel.SetSym(ctxt.Moduledata)
    
    	//	4:	02c00084	addi.d	$a0, $a0, 0
    	//				4: R_LARCH_PCALA_LO12	local.moduledata
    	o(0x02c00084)
    	rel2, _ := initfunc.AddRel(objabi.R_LOONG64_ADDR_LO)
    	rel2.SetOff(4)
    	rel2.SetSiz(4)
    	rel2.SetSym(ctxt.Moduledata)
    
    	//	8:	50000000	b	0
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Tue Feb 27 17:26:07 UTC 2024
    - 7.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/tests/flatbuffer2mlir/simple.mlir

      // CHECK:          %{{.*}} = "tfl.pseudo_const"() <{value = dense<{{\[\[1, 2\], \[3, 4\], \[5, 6\]\]}}> : tensor<3x2xi32>}>
      // CHECK-NEXT:     [[SUB:%.*]] = tfl.sub %{{.*}}, %{{.*}} {fused_activation_function = "RELU6"} : tensor<3x2xi32>
      // CHECK-NEXT:     [[SCALAR:%.*]] = "tfl.pseudo_const"() <{value = dense<10> : tensor<i32>}> : () -> tensor<i32>
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 1.6K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/stablehlo/odml_converter/transforms/outline_composites.cc

    namespace mlir {
    namespace odml {
    namespace {
    
    // TODO - b/330337238: Surface these to other files when needed.
    constexpr llvm::StringLiteral kCompositeNamespace = "odml.internal";
    constexpr llvm::StringLiteral kGelu = "gelu";
    
    std::string MakeCompositeName(llvm::StringRef op_name) {
      return (kCompositeNamespace + "." + op_name).str();
    }
    
    #define GEN_PASS_DEF_OUTLINECOMPOSITESPASS
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 9.6K bytes
    - Viewed (0)
Back to top