Search Options

Results per page
Sort
Preferred Languages
Advance

Results 11 - 20 of 44 for linear_1 (0.27 sec)

  1. src/compress/lzw/writer.go

    	// will make any future Write calls return errClosed
    	err error
    	// table is the hash table from 20-bit keys to 12-bit values. Each table
    	// entry contains key<<12|val and collisions resolve by linear probing.
    	// The keys consist of a 12-bit code prefix and an 8-bit byte suffix.
    	// The values are a 12-bit code.
    	table [tableSize]uint32
    }
    
    // writeLSB writes the code c for "Least Significant Bits first" data.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri Apr 26 13:32:40 UTC 2024
    - 7.9K bytes
    - Viewed (0)
  2. tensorflow/compiler/mlir/tensorflow/transforms/decompose_reduce_dataset.cc

          return WalkResult::advance();
        OpBuilder builder(reduce_dataset);
        Location loc = reduce_dataset.getLoc();
    
        // Get reduce function signature for dataset iteration types.
        // Note: lookupSymbol is a linear lookup which means the overall
        // complexity = # ReduceDataset ops x # of functions in module.
        func::FuncOp reduce_func =
            function->getParentOfType<ModuleOp>().lookupSymbol<FuncOp>(
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 14K bytes
    - Viewed (0)
  3. src/runtime/runtime.go

    	unlock(&ticks.lock)
    }
    
    // minTimeForTicksPerSecond is the minimum elapsed time we require to consider our ticksPerSecond
    // measurement to be of decent enough quality for profiling.
    //
    // There's a linear relationship here between minimum time and error from the true value.
    // The error from the true ticks-per-second in a linux/amd64 VM seems to be:
    // -   1 ms -> ~0.02% error
    // -   5 ms -> ~0.004% error
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Thu May 23 01:16:47 UTC 2024
    - 9.9K bytes
    - Viewed (0)
  4. src/internal/trace/gc.go

    	// We think of the mutator utilization over time as the
    	// box-filtered utilization function, which we call the
    	// "windowed mutator utilization function". The resulting
    	// function is continuous and piecewise linear (unless
    	// window==0, which we handle elsewhere), where the boundaries
    	// between segments occur when either edge of the window
    	// encounters a change in the instantaneous mutator
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Fri May 17 18:48:18 UTC 2024
    - 26K bytes
    - Viewed (0)
  5. tensorflow/compiler/mlir/tensorflow/ir/tf_generated_ops.td

      }];
    
      let description = [{
    That is for rows we have grad for, we update var, accum and linear as follows:
    accum_new = accum + grad * grad
    linear += grad - (accum_new^(-lr_power) - accum^(-lr_power)) / lr * var
    quadratic = 1.0 / (accum_new^(lr_power) * lr) + 2 * l2
    var = (sign(linear) * l1 - linear) / quadratic if |linear| > l1 else 0.0
    accum = accum_new
      }];
    
      let arguments = (ins
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Tue Jun 11 23:24:08 UTC 2024
    - 793K bytes
    - Viewed (0)
  6. samples/bookinfo/src/productpage/static/tailwind/tailwind.css

    e:i})=>i("colors"),backgroundImage:{none:"none","gradient-to-t":"linear-gradient(to top, var(--tw-gradient-stops))","gradient-to-tr":"linear-gradient(to top right, var(--tw-gradient-stops))","gradient-to-r":"linear-gradient(to right, var(--tw-gradient-stops))","gradient-to-br":"linear-gradient(to bottom right, var(--tw-gradient-stops))","gradient-to-b":"linear-gradient(to bottom, var(--tw-gradient-stops))","gradient-to-bl":"linear-gradient(to bottom left, var(--tw-gradient-stops))","gradient-to-...
    Registered: Fri Jun 14 15:00:06 UTC 2024
    - Last Modified: Tue May 28 14:48:01 UTC 2024
    - 357.1K bytes
    - Viewed (1)
  7. android/guava/src/com/google/common/collect/ImmutableSet.java

      private static final int CUTOFF = (int) (MAX_TABLE_SIZE * DESIRED_LOAD_FACTOR);
    
      /**
       * Returns an array size suitable for the backing array of a hash table that uses open addressing
       * with linear probing in its implementation. The returned size is the smallest power of two that
       * can hold setSize elements with the desired load factor. Always returns at least setSize + 2.
       */
      @VisibleForTesting
    Registered: Wed Jun 12 16:38:11 UTC 2024
    - Last Modified: Sun Jun 02 13:36:19 UTC 2024
    - 22.5K bytes
    - Viewed (0)
  8. src/cmd/internal/obj/util.go

    		arch:  arch,
    		cconv: cconv,
    	})
    }
    
    type regSet struct {
    	lo    int
    	hi    int
    	Rconv func(int) string
    }
    
    // Few enough architectures that a linear scan is fastest.
    // Not even worth sorting.
    var regSpace []regSet
    
    /*
    	Each architecture defines a register space as a unique
    	integer range.
    Registered: Wed Jun 12 16:32:35 UTC 2024
    - Last Modified: Wed May 15 15:44:14 UTC 2024
    - 17.5K bytes
    - Viewed (0)
  9. tensorflow/compiler/mlir/tensorflow/transforms/cluster_ops_by_policy.cc

    }
    
    LogicalResult ClusteringState::VerifyDominanceProperty(
        unsigned src_root, unsigned dst_root, Operation *insertion_point) {
      // TODO(ezhulenev): Optimize this linear scan with a map lookup.
      for (auto &member : members) {
        unsigned root = FindRoot(member.root);
        if (root != src_root) continue;
    
        // Block arguments do not really participate in clustering, they are only
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Thu Apr 25 16:01:03 UTC 2024
    - 27.9K bytes
    - Viewed (0)
  10. tensorflow/compiler/mlir/tensorflow/ir/tf_ops.td

    def TF_XlaSparseCoreFtrlOp : TF_Op<"XlaSparseCoreFtrl", []> {
      let summary = "aaa";
    
      let arguments = (ins
        TF_Float32Tensor:$embedding_table,
        TF_Float32Tensor:$accumulator,
        TF_Float32Tensor:$linear,
        TF_Float32Tensor:$learning_rate,
        TF_Int32Tensor:$indices,
        TF_Float32Tensor:$gradient,
        TF_Float32Tensor:$beta,
        TF_Float32Tensor:$learning_rate_power,
    Registered: Sun Jun 16 05:45:23 UTC 2024
    - Last Modified: Wed Apr 24 04:08:35 UTC 2024
    - 90.5K bytes
    - Viewed (0)
Back to top