Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 7 of 7 for We (0.16 sec)

  1. tensorflow/c/experimental/grappler/grappler.h

    //       `TP_Optimizer`, `TP_OptimizerRegistrationParams`
    //     * Struct that should be filled by the proper: `TF_GrapplerItem`,
    //       `TF_GraphProperties`, `TF_FunctionLibraryDefinition`
    //   * We use `struct_size` for version checking. It should be set both by
    //     core and the plugin.
    //     * For example, `TF_InitGraph` function receives
    //       `TP_OptimizerRegistrationParams*` as input with `struct_size`
    C
    - Registered: Tue Feb 27 12:39:08 GMT 2024
    - Last Modified: Wed Aug 03 18:08:43 GMT 2022
    - 12.5K bytes
    - Viewed (0)
  2. tensorflow/c/eager/c_api_experimental.h

    // APIs for generically dealing with op attributes (e.g. when forwarding them
    // through custom device implementations).
    //
    // TODO(allenl): Currently these are black boxes, but we should have some way to
    // inspect values. This would let people e.g. copy over most attributes and then
    // modify some based on their values.
    
    // A reference to an op's name -> attribute mapping
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Wed Feb 21 22:37:46 GMT 2024
    - 39.5K bytes
    - Viewed (0)
  3. tensorflow/c/eager/parallel_device/parallel_device_lib.h

                  delete reinterpret_cast<DataType*>(data);
                },
                nullptr),
            TF_DeleteTensor);
        // TODO(allenl): Here and when executing regular operations, we could hold
        // on to one TFE_Op per device and just call TFE_ResetOp to avoid parsing
        // device names repeatedly.
        std::unique_ptr<TFE_Op, decltype(&TFE_DeleteOp)> const_op(
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Tue Apr 25 15:21:13 GMT 2023
    - 12.9K bytes
    - Viewed (0)
  4. tensorflow/c/experimental/filesystem/plugins/gcs/ram_file_block_cache.h

      // is always executed during Read.
      bool IsCacheEnabled() const { return block_size_ > 0 && max_bytes_ > 0; }
    
      // We can not pass a lambda with capture as a function pointer to
      // `TF_StartThread`, so we have to wrap `Prune` inside a static function.
      static void PruneThread(void* param) {
        auto ram_file_block_cache = static_cast<RamFileBlockCache*>(param);
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Mon Aug 31 04:46:34 GMT 2020
    - 10.6K bytes
    - Viewed (0)
  5. tensorflow/c/eager/c_api.h

    // error information in *status.
    TF_CAPI_EXPORT extern void TFE_ContextOptionsSetConfig(
        TFE_ContextOptions* options, const void* proto, size_t proto_len,
        TF_Status* status);
    
    // Controls how to act when we try to run an operation on a given device but
    // some input tensors are not on that device.
    // LINT.IfChange
    // Note: Keep in sync with internal copy of enum in eager/context.h.
    typedef enum TFE_ContextDevicePlacementPolicy {
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Apr 27 21:07:00 GMT 2023
    - 22.8K bytes
    - Viewed (1)
  6. tensorflow/c/eager/immediate_execution_context.h

      // Update the Eager Executor for current thread.
      virtual void SetExecutorForThread(EagerExecutor* executor) = 0;
    
      // Return a list of local tensorflow::Device*.
      // TODO(tfrt-devs): We shouldn't expose legacy device in this API.
      virtual std::vector<tensorflow::Device*> ListLocalTfDevices() = 0;
    
      // Return a list of all tensorflow::Device*.
      virtual std::vector<tensorflow::Device*> ListAllTfDevices() = 0;
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Thu Jul 06 08:34:00 GMT 2023
    - 12.3K bytes
    - Viewed (0)
  7. tensorflow/c/eager/tape.h

    //
    // Below here we do the gradient algorithm. It works as follows:
    //
    // First we filter the tape to just the subset of operations we want to
    // differentiate. In the process of doing so we count how many times each Tensor
    // is used as an input to an op (so we know when we're done computing gradients
    // for that Tensor). We also count, for each tape entry, how many of its output
    C
    - Registered: Tue Apr 23 12:39:09 GMT 2024
    - Last Modified: Tue Apr 02 12:40:29 GMT 2024
    - 47.2K bytes
    - Viewed (1)
Back to top