Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 81 for conv3d (0.13 sec)

  1. tensorflow/compiler/mlir/lite/experimental/tac/tests/device-transform-gpu.mlir

    // -----
    
    func.func @ensureBiasForConv2d(%arg0: tensor<128x32x32x3xf32>, %arg1: tensor<32x1x1x3xf32>) -> tensor<128x32x32x32xf32> {
      %cst = "tfl.no_value"() {value = unit} : () -> none
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 15.6K bytes
    - Viewed (0)
  2. pkg/kubelet/cm/devicemanager/pod_devices_test.go

    	expectResourceDeviceInstances(t, resp, expected)
    
    	cond = map[string]sets.Set[string]{
    		"foo": sets.New[string]("dev-foo1", "dev-foo2"),
    		"bar": sets.New[string]("dev-bar1"),
    	}
    	resp = devs.Filter(cond)
    	expected = `{"bar":{"dev-bar1":{"ID":"bar1"}},"foo":{"dev-foo1":{"ID":"foo1"},"dev-foo2":{"ID":"foo2"}}}`
    	expectResourceDeviceInstances(t, resp, expected)
    
    	cond = map[string]sets.Set[string]{
    Registered: Sat Jun 15 01:39:40 UTC 2024
    - Last Modified: Wed Apr 24 18:25:29 UTC 2024
    - 10.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/experimental/tac/tests/fold-constants-to-subgraph.mlir

    }
    
    // ALL-LABEL: @fold_all_test
    func.func @fold_all_test(%arg0: tensor<256x32x32x3xf32>, %arg1: tensor<16x3x3x3xf32>, %arg2: tensor<16xf32>) -> tensor<256x30x30x16xf32> {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 02 09:41:17 UTC 2024
    - 10.5K bytes
    - Viewed (0)
  4. tensorflow/cc/gradients/nn_grad_test.cc

    #include "tensorflow/core/lib/core/status_test_util.h"
    #include "tensorflow/core/lib/random/random.h"
    
    namespace tensorflow {
    namespace {
    
    using ops::AvgPool;
    using ops::AvgPool3D;
    using ops::BiasAdd;
    using ops::Conv2D;
    using ops::Conv2DBackpropInput;
    using ops::DepthwiseConv2dNative;
    using ops::Elu;
    using ops::FractionalAvgPool;
    using ops::FractionalMaxPool;
    using ops::FusedBatchNormV3;
    using ops::L2Loss;
    using ops::LogSoftmax;
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Mar 22 20:45:22 UTC 2022
    - 15K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tf2xla/tests/legalize-tf-prefer-tf2xla.mlir

      %input_scale = "tf.Const"() {value = dense<1.0> : tensor<f32>} : () -> tensor<f32>
      %side_input_scale = "tf.Const"() {value = dense<2.0> : tensor<f32>} : () -> tensor<f32>
      %conv2d = "tf._FusedConv2D"(%input, %filter, %bias, %act, %input_scale, %side_input_scale) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Sat Apr 06 15:32:52 UTC 2024
    - 15.8K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/passes/prepare_lifting.cc

    // - Conv2DOp
    // - Conv3DOp
    // - DepthwiseConv2dNativeOp
    // - MatMulOp
    // - BatchMatMulV2Op
    LogicalResult MatchSupportedAffineOp(Operation* op, Value& binding_output,
                                         Value& binding_input,
                                         Value& binding_weight) {
      bool is_supported_affine_op = false;
      if (llvm::isa<TF::Conv2DOp, TF::Conv3DOp, TF::DepthwiseConv2dNativeOp>(op)) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 17:58:54 UTC 2024
    - 13.3K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/tensorflow/passes/quantized_function_library_uniform_quantized.mlir

    // func.func func_name_${key1}_fn (...) {
    //   ...${key2}...
    // }
    // ```
    // The above template with generate two functions by substituting `key1` and
    // `key2` with given values.
    
    module {
    
      for main_op in ["Conv2D", "DepthwiseConv2D", "MatMul"] {
        parameters[
          {"quantized_ops": ["${main_op}", "BiasAdd"], "act_func": "internal_requantize_no_activation_fn", "output_type": "!tf_type.qint8"},
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Aug 29 01:13:58 UTC 2023
    - 19.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/tf_tfl_passes.cc

      // away ops that can't get constant folded after PrepareTF pass. For example,
      // tf.Conv2D is split into tf.Transpose and tfl.Conv2D.
      pass_manager->addNestedPass<mlir::func::FuncOp>(
          mlir::createCanonicalizerPass());
      pass_manager->addNestedPass<mlir::func::FuncOp>(mlir::createCSEPass());
      // This pass does dead code elimination based on symbol visibility.
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Jun 06 18:45:51 UTC 2024
    - 25.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/lite/experimental/tac/transforms/device_transform_patterns.cc

                          weight_new_shape, &rewriter);
    
      // Replace the fc with conv.
      // The output would be [1, 1, width, output].
      auto conv_output_type = RankedTensorType::get({1, 1, width, output_size},
                                                    output_type.getElementType());
      auto conv = rewriter.create<TFL::Conv2DOp>(
          fc_op.getLoc(), conv_output_type, reshaped_input, reshaped_weight,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 25.4K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/quantization/stablehlo/tests/passes/insert_calibration_statistics_saver.mlir

      %0 = "tf.Conv2D"(%output, %cst) <{data_format = "NHWC", dilations = [1, 1, 1, 1], explicit_paddings = [], padding = "SAME", strides = [1, 2, 2, 1], use_cudnn_on_gpu = true}> {attr_map = "0:strides,1:use_cudnn_on_gpu,2:padding,3:explicit_paddings,4:dilations", device...
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 01:09:50 UTC 2024
    - 24.3K bytes
    - Viewed (0)
Back to top