- Sort Score
- Result 10 results
- Languages All
Results 31 - 40 of 110 for Chen (0.13 sec)
-
.github/ISSUE_TEMPLATE/tflite-converter-issue.md
- Provide links to your TensorFlow model and (optionally) TensorFlow Lite Model. ``` ### 3. Failure after conversion If the conversion is successful, but the generated model is wrong, then state what is wrong: - Model produces wrong results and/or has lesser accuracy. - Model produces correct results, but it is slower than expected. ### 4. (optional) RNN conversion support
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Wed Jun 15 03:35:58 GMT 2022 - 2.1K bytes - Viewed (0) -
ci/official/containers/linux_arm64/setup.python.sh
popd # Python 3.10 include headers fix: # sysconfig.get_path('include') incorrectly points to /usr/local/include/python # map /usr/include/python3.10 to /usr/local/include/python3.10 if [[ ! -f "/usr/local/include/$VERSION" ]]; then ln -sf /usr/include/$VERSION /usr/local/include/$VERSION fi # Install pip wget --retry-connrefused --waitretry=1 --read-timeout=20 --timeout=15 --tries=5 https://bootstrap.pypa.io/get-pip.py
Shell Script - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Sep 29 00:26:34 GMT 2023 - 2.1K bytes - Viewed (0) -
.github/workflows/release-branch-cherrypick.yml
# Usage: Go to # https://github.com/tensorflow/tensorflow/actions/workflows/release-branch-cherrypick.yml # and click "Run Workflow." Leave "Use Workflow From" set to "master", then # input the branch name and paste the cherry-pick commit and click Run. A PR # will be created. name: Release Branch Cherrypick on: workflow_dispatch: inputs:
Others - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Tue Sep 12 14:49:29 GMT 2023 - 3.1K bytes - Viewed (0) -
.github/bot_config.yml
* Try adding the CUDA, CUPTI, and cuDNN installation directories to the $LD_LIBRARY_PATH environment variable. * Refer [linux setup guide](https://www.tensorflow.org/install/gpu#linux_setup). * If error still persists then, apparently your CPU model does not support AVX instruction sets. * Refer [hardware requirements](https://www.tensorflow.org/install/pip#hardware-requirements).
Others - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Tue Oct 17 11:48:07 GMT 2023 - 4K bytes - Viewed (0) -
WORKSPACE
# # The cascade of load() statements and tf_workspace?() calls works around the # restriction that load() statements need to be at the top of .bzl files. # E.g. we can not retrieve a new repository with http_archive and then load() # a macro from that repository in the same file. load("@//tensorflow:workspace3.bzl", "tf_workspace3") tf_workspace3() load("@//tensorflow:workspace2.bzl", "tf_workspace2") tf_workspace2()
Plain Text - Registered: Tue May 07 12:40:20 GMT 2024 - Last Modified: Fri Apr 05 22:27:48 GMT 2024 - 3K bytes - Viewed (2) -
tensorflow/api_template_v1.__init__.py
# Delete modules that should be hidden from dir(). # Don't fail if these modules are not available. # For e.g. this file will be originally placed under tensorflow/_api/v1 which # does not have "python", "core" directories. Then, it will be copied # to tensorflow/ which does have these two directories. try: del python except NameError: pass try: del core except NameError: pass try: del compiler
Python - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Tue Jan 23 02:14:00 GMT 2024 - 7.4K bytes - Viewed (0) -
tensorflow/c/eager/dlpack.cc
#include "tensorflow/core/platform/logging.h" namespace tensorflow { namespace { // Managing context for the DLManagedTensor, will manage the lifetime of // DLManagedTensor. When calling DLManagedTensor::deleter, it will notify the // original framework of destruction, and this context will be deleted also. struct TfDlManagedTensorCtx { TensorReference reference; std::vector<int64_t> shape;
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Thu Feb 15 09:49:45 GMT 2024 - 12.8K bytes - Viewed (0) -
tensorflow/c/experimental/grappler/grappler.h
// nodes. This may cause incorrectness in graph analyses, but is useful for // simulation or scheduling. // If aggressive_shape_inference is true, nodes are executed on the host to // identify output values when possible and does other aggressive strategies. // This may cause incorrectness in graph analyses, but is useful for simulation // or scheduling. // If include_input_tensor_values is true, the values of constant
C - Registered: Tue Feb 27 12:39:08 GMT 2024 - Last Modified: Wed Aug 03 18:08:43 GMT 2022 - 12.5K bytes - Viewed (0) -
tensorflow/c/eager/immediate_execution_distributed_manager.h
class WorkerCacheInterface; class ImmediateExecutionDistributedManager { public: virtual ~ImmediateExecutionDistributedManager() {} // Set up distributed execution environment on local and remote tasks. // When `reset_context` is true, initialize new cluster context state based // on cluster configurations provided in `server_def`; otherwise, update // existing context state with the provided `server_def`. Contexts created
C - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Wed Feb 21 22:37:46 GMT 2024 - 2.9K bytes - Viewed (0) -
tensorflow/c/c_test_util.cc
&FloatDeallocator, nullptr); } // All the *Helper methods are used as a workaround for the restrictions that // one cannot call ASSERT_* methods in non-void-returning functions (when // exceptions are disabled during compilation) void PlaceholderHelper(TF_Graph* graph, TF_Status* s, const char* name, TF_DataType dtype, const std::vector<int64_t>& dims,
C++ - Registered: Tue Apr 30 12:39:09 GMT 2024 - Last Modified: Fri Oct 15 03:16:52 GMT 2021 - 17.8K bytes - Viewed (2)