Search Options

Results per page
Sort
Preferred Languages
Advance

Results 21 - 30 of 58 for created (0.08 sec)

  1. tensorflow/compiler/mlir/quantization/stablehlo/cc/config.cc

        config.mutable_calibration_options()
            ->mutable_representative_datasets()
            ->Add(preset_datasets.begin(), preset_datasets.end());
      }
    
      // Create a new `QuantizationSpecs` to replace the existing one. The
      // expansion from `StaticRangePtqPreset` gets populated first and then
      // user-provided explicit `QuantizationSpec`s will be appended.
      QuantizationSpecs new_specs{};
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 17 03:36:50 UTC 2024
    - 8.3K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/quantization/common/ir/UniformSupport.cc

        return convert(attr);
      }
      return nullptr;
    }
    
    DenseElementsAttr UniformQuantizedPerAxisValueConverter::convert(
        DenseFPElementsAttr attr) {
      // Creates the converter for each chunk. Normally the size of the
      // quantization dim is 3, so we can cache all the converters.
      ShapedType type = attr.getType();
      std::size_t dim_size = type.getDimSize(quantization_dim_);
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed May 08 02:10:16 UTC 2024
    - 4.3K bytes
    - Viewed (0)
  3. tensorflow/compiler/mlir/lite/transforms/reduce_type_precision.cc

            dyn_cast<arith::ConstantOp>(op.getOperand(0).getDefiningOp());
        if (!input_op) {
          return failure();
        }
    
        Builder builder(op.getContext());
        auto new_gather_op = rewriter.create<TFL::GatherOp>(
            op.getLoc(),
            /*result=*/
            mlir::cast<TensorType>(op.getResult().getType())
                .clone(builder.getI4Type()),
            /*operand=*/op.getOperands(), op->getAttrs());
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.4K bytes
    - Viewed (0)
  4. tensorflow/compiler/mlir/tensorflow/transforms/set_tpu_infeed_layout.cc

        } else {
          /* If we're not running on a TPU node, we might not be able to
           * actually call the part of the TPU API that gives us layout.
           * This happens e.g. for unit tests. Below we just create a reasonable
           * layout.  We sort by dimension size, which makes the layout agree with
           * the "correct" TPU layout in surprisingly many cases.
           * Note that the corresponding InfeedEnqueue op will be generated
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 6.1K bytes
    - Viewed (0)
  5. tensorflow/c/eager/c_api_unified_experimental.cc

    using tensorflow::tracing::TracingTensorHandle;
    
    void TF_SetTracingImplementation(const char* name, TF_Status* s) {
      tsl::Set_TF_Status_from_Status(s, SetDefaultTracingEngine(name));
    }
    
    // Creates a new TensorFlow function, it is an execution context attached to a
    // given tracing context.
    TF_ExecutionContext* TF_CreateFunction(const char* fn_name, TF_Status* s) {
      return wrap(CreateTracingExecutionContext(fn_name, s));
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu May 09 10:15:17 UTC 2024
    - 9K bytes
    - Viewed (0)
  6. tensorflow/compiler/mlir/quantization/tensorflow/passes/convert_tpu_model_to_cpu.cc

        module_op.emitError() << "quant-convert-tpu-model-to-cpu pattern "
                                 "conversion did not converge.";
        signalPassFailure();
        return;
      }
    }
    
    }  // namespace
    
    // Creates an instance of `ConvertTpuModelToCpuPass`.
    std::unique_ptr<OperationPass<ModuleOp>> CreateConvertTpuModelToCpuPass() {
      return std::make_unique<ConvertTpuModelToCpuPass>();
    }
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 5.5K bytes
    - Viewed (0)
  7. tensorflow/compiler/mlir/quantization/stablehlo/passes/lift_quantizable_spots_as_functions.cc

        if (failed(ApplyQuantizationSpec(spec, module_op))) {
          signalPassFailure();
          return;
        }
      }
    }
    
    }  // namespace
    
    // Creates `LiftQuantizableSpotsAsFunctionsPass` with user-defined
    // `QuantizationSpecs`.
    std::unique_ptr<OperationPass<ModuleOp>>
    CreateLiftQuantizableSpotsAsFunctionsPass(
        const QuantizationSpecs& quantization_specs) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Fri May 10 04:07:09 UTC 2024
    - 9.3K bytes
    - Viewed (0)
  8. tensorflow/compiler/mlir/lite/transforms/modify_io_nodes.cc

        return;
      }
    
      auto new_func_type =
          builder.getFunctionType(new_input_types, new_output_types);
      func.setType(new_func_type);
    }
    }  // namespace
    
    // Creates an instance of the TensorFlow Lite modify io nodes pass.
    std::unique_ptr<OperationPass<func::FuncOp>> CreateModifyIONodesPass(
        Type input_type, Type output_type) {
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 8.9K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/transforms/xla_call_module_serialization.cc

        return WalkResult::advance();
      });
      if (result.wasInterrupted()) {
        return failure();
      }
    
      return builder.getArrayAttr(function_list);
    }
    
    // Creates a pruned module containing the XlaCallModule's entry function and
    // other functions transitively called by the entry function.
    FailureOr<OwningOpRef<ModuleOp>> PruneStablehloModule(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.5K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/lite/transforms/default_quant_params.cc

      } else {
        builder.setInsertionPointToStart(&block);
      }
      TypeAttr type_attr = TypeAttr::get(new_type);
      auto quantize = builder.create<TFL::QuantizeOp>(value.getLoc(), new_type,
                                                      value, type_attr);
      auto dequantize = builder.create<TFL::DequantizeOp>(
          value.getLoc(), expressed_type, quantize.getOutput());
      value.replaceAllUsesWith(dequantize);
    
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 9.4K bytes
    - Viewed (0)
Back to top