Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 4 of 4 for sgdu (0.11 sec)

  1. src/cmd/asm/internal/asm/testdata/amd64enc.s

    	SETMI DL                                // 0f98c2
    	SETMI R11                               // 410f98c3
    	SFENCE                                  // 0faef8
    	SGDT (BX)                               // 0f0103
    	SGDT (R11)                              // 410f0103
    	SHLW $1, (BX)                           // 66d123
    	SHLW $1, (R11)                          // 6641d123
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Oct 08 21:38:44 UTC 2021
    - 581.9K bytes
    - Viewed (0)
  2. gradle/verification-keyring.keys

    8Cf6vooOcWYPPGYS0NR7aB5RKeV7NR0nHx5MMrlywgT0WUtdv0EmLigSR02iR0pV
    cWWCdv6/irriSC7KWBseKYkI27DBIhiNBTBe39cVgSd+RIlJ9ojEt7c7PWDQNFl2
    qYQQYGc0lMwTFp3+Y3aDFfEsRsTqzvMrCOzlzSXwlanYT/fbFL7iukkYYhA5lmfB
    sgdm+nSmwB35Bm7FBXMVSMvKD+SvAUGbkb7Kxutrbm76RQM7E4cNkNhQUcGqtJn5
    jRoulS6lXh/GN4V3pg20ds9NY4bcN8uuzyU1iqHfKVQo6l9KkCUItrNY6J/arE7+
    AuUSjKupp9fPjLgaM2PMTio1U8gdrhy3Yubfpxwo6Cswyys24QHfSMNYjjuEWhmI
    Registered: Wed Jun 12 18:38:38 UTC 2024
    - Last Modified: Mon Apr 01 11:46:17 UTC 2024
    - 525.2K bytes
    - Viewed (1)
  3. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      let results = (outs);
    }
    
    def TF_LoadTPUEmbeddingStochasticGradientDescentParametersOp : TF_Op<"LoadTPUEmbeddingStochasticGradientDescentParameters", [TF_MustExecute, TF_TPUEmbeddingReadEffect]> {
      let summary = "Load SGD embedding parameters.";
    
      let description = [{
    An op that loads optimization parameters into HBM for embedding. Must be
    preceded by a ConfigureTPUEmbeddingHost op that sets up the correct
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  4. RELEASE.md

        *   Added `tf.keras.optimizers.experimental.Optimizer`. The reworked
            optimizer gives more control over different phases of optimizer calls,
            and is easier to customize. We provide Adam, SGD, Adadelta, AdaGrad and
            RMSprop optimizers based on
            `tf.keras.optimizers.experimental.Optimizer`. Generally the new
            optimizers work in the same way as the old ones, but support new
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 730.3K bytes
    - Viewed (0)
Back to top