- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 14 for Howard (0.19 sec)
-
tensorflow/c/experimental/gradients/nn_grad.cc
namespace gradients { namespace { class ReluGradientFunction : public GradientFunction { public: explicit ReluGradientFunction(vector<AbstractTensorHandle*> f_outputs) : forward_outputs_(f_outputs) { for (auto output : forward_outputs_) { if (output) { output->Ref(); } } } Status Compute(AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> grad_outputs,
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5.7K bytes - Viewed (0) -
tensorflow/c/eager/gradients_test.cc
t.reset(x_raw); } AbstractOperationPtr check_numerics_op(ctx->CreateOperation()); ForwardOperation forward_op; Status s = Reset(check_numerics_op.get(), "CheckNumerics", /*raw_device_name=*/nullptr, &forward_op); ASSERT_EQ(errors::OK, s.code()) << s.message(); if (isa<TracingOperation>(check_numerics_op.get())) {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 7K bytes - Viewed (0) -
tensorflow/c/eager/gradient_checker.cc
Status RunAndMaybeSum(AbstractContext* ctx, Model forward, absl::Span<AbstractTensorHandle* const> inputs, absl::Span<AbstractTensorHandle*> outputs, bool use_function) { AbstractTensorHandle* model_outputs[1]; // Run the model. TF_RETURN_IF_ERROR( RunModel(forward, ctx, inputs, model_outputs, use_function));
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 7.3K bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.cc
device_threads_.emplace_back(new DeviceThread( devices[device_index].c_str(), is_async, in_flight_nodes_limit)); } } // Necessary for a unique_ptr to a forward-declared type. ParallelDevice::~ParallelDevice() = default; std::unique_ptr<ParallelTensor> ParallelDevice::CopyToParallelDevice( TFE_Context* context, TFE_TensorHandle* tensor, TF_Status* status) const {
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Feb 09 07:47:20 GMT 2024 - 25.4K bytes - Viewed (1) -
tensorflow/c/experimental/gradients/nn_grad_test.cc
ASSERT_EQ(errors::OK, status_.code()) << status_.message(); immediate_execution_ctx_.reset(ctx_raw); } // Computing numerical gradients with TensorFloat-32 is numerically // unstable. Some forward pass tests also fail with TensorFloat-32 due to // low tolerances enable_tensor_float_32_execution(false); } AbstractContextPtr immediate_execution_ctx_; GradientRegistry registry_; Status status_;
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 8.3K bytes - Viewed (0) -
tensorflow/c/c_api_test.cc
// | dz| | // | | | // | Const_3 | // | | // | ^ | // | z| | // MatMul Forward Graph // | | | // | MatMul | // | / \ | // | ^ ^ | // | | | | // |---x| y|----|
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Apr 15 03:35:10 GMT 2024 - 96.9K bytes - Viewed (3) -
tensorflow/c/eager/tape.h
forward_grads.resize(output_tensors.size()); TF_RETURN_IF_ERROR(ForwardpropFromTape( op_type, output_tensors, backward_function_getter, backward_function_deleter, in_grads, absl::MakeSpan(forward_grads))); } else { TF_RETURN_IF_ERROR( (*forward_function)(in_grads, &forward_grads, use_batch_)); } for (int i = 0; i < forward_grads.size(); ++i) { if (forward_grads[i] != nullptr) {
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1) -
tensorflow/c/experimental/gradients/grad_test_helper.cc
ASSERT_NEAR(manuals[j], danalytical[j], abs_error); } } TF_DeleteTensor(analytical_tensor); delete[] danalytical; } Model BuildGradModel(Model forward, GradientRegistry registry) { return [forward_model = std::move(forward), grad_registry = std::move(registry)]( AbstractContext* ctx, absl::Span<AbstractTensorHandle* const> inputs,
C++ - Registered: Tue Mar 26 12:39:09 GMT 2024 - Last Modified: Wed Feb 28 13:53:47 GMT 2024 - 5K bytes - Viewed (0) -
.bazelrc
# This is the same as the official TensorFlow builds. # See https://developer.nvidia.com/cuda-gpus#compute # `compute_XY` enables PTX embedding in addition to SASS. PTX # is forward compatible beyond the current compute capability major # release while SASS is only forward compatible inside the current # major release. Example: sm_80 kernels can run on sm_89 GPUs but # not on sm_90 GPUs. compute_80 kernels though can also run on sm_90 GPUs.
Plain Text - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Apr 24 20:50:35 GMT 2024 - 52.6K bytes - Viewed (2) -
tensorflow/c/eager/gradients.cc
const char* raw_device_name, ForwardOperation* forward_op_) { forward_op_->op_name = op; forward_op_->attrs.Reset(op); return op_->Reset(op, raw_device_name); } Status AddInput(AbstractOperation* op_, AbstractTensorHandle* input, ForwardOperation* forward_op_) { TF_RETURN_IF_ERROR(op_->AddInput(input)); forward_op_->inputs.push_back(input); return absl::OkStatus(); }
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 19.3K bytes - Viewed (0)