- Sort Score
- Result 10 results
- Languages All
Results 11 - 20 of 30 for accelerate (0.25 sec)
-
src/crypto/aes/cipher_generic.go
//go:build (!amd64 && !s390x && !ppc64 && !ppc64le && !arm64) || purego package aes import ( "crypto/cipher" ) // newCipher calls the newCipherGeneric function // directly. Platforms with hardware accelerated // implementations of AES should implement their // own version of newCipher (which may then call // newCipherGeneric if needed). func newCipher(key []byte) (cipher.Block, error) { return newCipherGeneric(key) }
Registered: Wed Jun 12 16:32:35 UTC 2024 - Last Modified: Mon Mar 04 17:29:44 UTC 2024 - 772 bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/sparsecore_passes.h
#include "mlir/IR/BuiltinOps.h" // from @llvm-project #include "mlir/Pass/Pass.h" // from @llvm-project namespace mlir { namespace TFDevice { // For architectures that support accelerated embedding lookups, this pass will // rewrite the graph to use pipelining for better device utilization. std::unique_ptr<OperationPass<ModuleOp>> CreateEmbeddingSequencingPass();
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 28 23:42:09 UTC 2024 - 2.1K bytes - Viewed (0) -
tensorflow/compiler/jit/pjrt_device_context.h
#include "tensorflow/core/framework/device_base.h" #include "tensorflow/core/platform/status.h" namespace tensorflow { // Helper class for managing data transfers between host and accelerator // devices using PjRt. class PjRtDeviceContext : public DeviceContext { public: explicit PjRtDeviceContext( XlaShapeLayoutHelpers::ShapeDeterminationFns shape_determination_fns,
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Jul 19 19:27:39 UTC 2023 - 2.7K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/sparsecore/sparsecore_passes.td
let summary = "Rewrite graph for embedding pipelining"; let constructor = "TFDevice::CreateEmbeddingPipeliningPass()"; let description = [{ For architectures that support accelerated embedding lookups, this pass will rewrite the graph to use pipelining for better device utilization. }]; } def EmbeddingSequencingPass : Pass<"tf-embedding-sequencing", "mlir::ModuleOp"> {
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Thu Mar 28 23:42:09 UTC 2024 - 3.9K bytes - Viewed (0) -
tensorflow/compiler/jit/pjrt_base_device.h
#include "tensorflow/core/common_runtime/local_device.h" #include "tensorflow/core/framework/device_base.h" namespace tensorflow { // tensorflow::PjRtBaseDevice replaces the deprecated tensorflow::XlaDevice. // This accelerator agnostic device is mainly used to store metadata. class PjRtBaseDevice : public LocalDevice { public: // Stores metadata about the PjRtBaseDevice. class Metadata { public:
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Feb 21 12:19:41 UTC 2024 - 4K bytes - Viewed (0) -
.github/workflows/build.yml
uses: actions/setup-java@v4 with: distribution: 'zulu' java-version: 17 - name: Enable KVM group perms # https://github.blog/changelog/2023-02-23-hardware-accelerated-android-virtualization-on-actions-windows-and-linux-larger-hosted-runners/ run: | echo 'KERNEL=="kvm", GROUP="kvm", MODE="0666", OPTIONS+="static_node=kvm"' | sudo tee /etc/udev/rules.d/99-kvm4all.rules
Registered: Sun Jun 16 04:42:17 UTC 2024 - Last Modified: Mon Apr 15 01:51:50 UTC 2024 - 17.2K bytes - Viewed (0) -
cluster/gce/config-default.sh
# is a request for 2 SCSI formatted and mounted SSDs and 1 NVMe block device SSD. NODE_LOCAL_SSDS_EXT=${NODE_LOCAL_SSDS_EXT:-} # Accelerators to be attached to each node. Format "type=<accelerator-type>,count=<accelerator-count>" # More information on available GPUs here - https://cloud.google.com/compute/docs/gpus/ NODE_ACCELERATORS=${NODE_ACCELERATORS:-""} export REGISTER_MASTER_KUBELET=${REGISTER_MASTER:-true}
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Sat Mar 16 20:16:32 UTC 2024 - 26.9K bytes - Viewed (0) -
CHANGELOG/CHANGELOG-1.30.md
- Introduced a new alpha feature gate, `SELinuxMount`, which can now be enabled to accelerate SELinux relabeling. ([#123157](https://github.com/kubernetes/kubernetes/pull/123157), [@jsafrane](https://github.com/jsafrane))
Registered: Sat Jun 15 01:39:40 UTC 2024 - Last Modified: Wed Jun 12 04:05:28 UTC 2024 - 253.2K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/utils/cluster_util.cc
std::function<bool(Operation*)> is_ignored_op) { // Iteratively find clusters of different targets within the `block`. // Whenever we see an operation that is assigned to an accelerator target // (ie. get_target(op) != ""), we try to merge it into the last cluster // of same target. If that is infeasible (say because of violating // def-before-use), create a new cluster with that operation and move on.
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Fri Jul 28 00:32:55 UTC 2023 - 8.3K bytes - Viewed (0) -
tensorflow/compiler/mlir/tensorflow/transforms/tf_device_passes.td
let summary = "Lifting resource operations out of device computation"; let description = [{ This pass lifts resource variable operations outside of device computation. This is useful because a lot of accelerator devices can not interact with resource variables directly.. Here is a simple example in TensorFlow where a device doubles the value of a TensorFlow resource variable and returns new value: ```mlir
Registered: Sun Jun 16 05:45:23 UTC 2024 - Last Modified: Wed Apr 17 18:52:57 UTC 2024 - 12.5K bytes - Viewed (0)