- Sort Score
- Result 10 results
- Languages All
Results 1 - 10 of 34 for Ehat (0.13 sec)
-
src/cmd/cgo/internal/test/callback_c_gccgo.c
// Copyright 2013 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. //go:build gccgo #include "_cgo_export.h" #include <stdint.h> #include <stdio.h> #include <stdlib.h> /* Test calling panic from C. This is what SWIG does. */ extern void _cgo_panic(const char *); extern void *_cgo_allocate(size_t); void callPanic(void) {
C - Registered: Tue Apr 23 11:13:09 GMT 2024 - Last Modified: Fri May 12 12:00:02 GMT 2023 - 452 bytes - Viewed (0) -
tensorflow/c/experimental/grappler/grappler.h
// // Conventions: // * Struct prefix indicates whether struct fields should be filled by the // plugin or core implementation: // * Struct that should be filled by the plugin: `TP_OptimizerConfigs`, // `TP_Optimizer`, `TP_OptimizerRegistrationParams` // * Struct that should be filled by the proper: `TF_GrapplerItem`, // `TF_GraphProperties`, `TF_FunctionLibraryDefinition`
C - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Wed Aug 03 18:08:43 GMT 2022 - 12.5K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental.h
TF_CAPI_EXPORT extern bool TFE_ExecutorIsAsync(TFE_Executor*); // Causes the calling thread to block till all ops dispatched in this executor // have been executed. Note that "execution" here refers to kernel execution / // scheduling of copies, etc. Similar to sync execution, it doesn't guarantee // that lower level device queues (like GPU streams) have been flushed. // // This call may not block for execution of ops enqueued concurrently with this // call.
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 39.5K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_operation.h
virtual ImmediateExecutionContext* GetContext() const = 0; // Following two methods are used to support custom device. // Return true if the inputs contain custom device tensor handle. It means // that the argument need to be handled by a custom device. virtual bool HasCustomDeviceInput() const = 0; virtual const tensorflow::OpDef* OpDef() const = 0;
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Mon Sep 26 22:40:32 GMT 2022 - 3.6K bytes - Viewed (0) -
tensorflow/c/eager/c_api_experimental_reader.h
// auto* streamz = tensorflow::monitoring::Counter<1>::New("name", // "description", "label"); // and then incremented that counter for various values of label: // streamz->GetCell("label-value")->IncrementBy(1); // // The test code can then read and test the value of that counter: // // auto* reader = TFE_MonitoringNewCounterReader("name"); // test();
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 20 03:14:47 GMT 2023 - 2.3K bytes - Viewed (0) -
tensorflow/c/c_api.h
// Get the number of operations that have `*oper` as a control input. // Note that this number can change when new operations are added to // the graph. TF_CAPI_EXPORT extern int TF_OperationNumControlOutputs(TF_Operation* oper); // Get the list of operations that have `*oper` as a control input. // `control_outputs` must point to an array of length at least
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Oct 26 21:08:15 GMT 2023 - 82.3K bytes - Viewed (3) -
src/cmd/cgo/internal/swig/testdata/callback/main.h
// Copyright 2011 The Go Authors. All rights reserved. // Use of this source code is governed by a BSD-style // license that can be found in the LICENSE file. class Callback { public: virtual ~Callback() { } virtual std::string run() { return "Callback::run"; } }; class Caller { private: Callback *callback_; public: Caller(): callback_(0) { } ~Caller() { delCallback(); } void delCallback() { delete callback_; callback_ = 0; }
C - Registered: Tue Apr 23 11:13:09 GMT 2024 - Last Modified: Fri May 12 12:00:07 GMT 2023 - 529 bytes - Viewed (0) -
tensorflow/c/eager/parallel_device/parallel_device_lib.h
TFE_TensorHandle* tensor(size_t index) const { return tensors_[index].get(); } // If the `shape` argument to `FromTensorHandles` is specified, returns that. // // Otherwise if all of the tensors have the same shape, returns that via the // `shape` output argument. This blocks waiting for async tensors, may return // a delayed bad status encountered during async execution, and will return a
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 25 15:21:13 GMT 2023 - 12.9K bytes - Viewed (0) -
tensorflow/c/c_api_experimental.h
TF_CAPI_EXPORT int TF_PickUnusedPortOrDie(void); // Fast path method that makes constructing a single scalar tensor require less // overhead and copies. TF_CAPI_EXPORT extern TFE_TensorHandle* TFE_NewTensorHandleFromScalar( TF_DataType data_type, void* data, size_t len, TF_Status* status); // Specify the server_def that enables collective ops. // This is different to the above function in that it doesn't create remote
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Apr 27 21:07:00 GMT 2023 - 15.1K bytes - Viewed (0) -
tensorflow/c/eager/tape.h
// the tape, so no gradient needs to be computed from it unless this // target is also a source. auto source_tensor = sources_that_are_targets.find(id); if (source_tensor != sources_that_are_targets.end()) { Gradient* ones_like = nullptr; TF_RETURN_IF_ERROR( vspace.BuildOnesLike(source_tensor->second, &ones_like));
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Apr 02 12:40:29 GMT 2024 - 47.2K bytes - Viewed (1)