Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 8 of 8 for Size (0.21 sec)

  1. okhttp-testing-support/src/main/kotlin/okhttp3/internal/concurrent/TaskFaker.kt

       */
      private inner class TaskFakerBlockingQueue<T>(
        val delegate: BlockingQueue<T>,
      ) : AbstractQueue<T>(), BlockingQueue<T> {
        override val size: Int = delegate.size
    
        private var editCount = 0
    
        override fun poll(): T = delegate.poll()
    
        override fun poll(
          timeout: Long,
          unit: TimeUnit,
        ): T? {
    Plain Text
    - Registered: Fri May 03 11:42:14 GMT 2024
    - Last Modified: Mon Apr 29 00:33:04 GMT 2024
    - 12.6K bytes
    - Viewed (0)
  2. okhttp/src/main/kotlin/okhttp3/internal/concurrent/TaskRunner.kt

        lock.withLock {
          return busyQueues + readyQueues
        }
      }
    
      fun cancelAll() {
        lock.assertHeld()
        for (i in busyQueues.size - 1 downTo 0) {
          busyQueues[i].cancelAllAndDecide()
        }
        for (i in readyQueues.size - 1 downTo 0) {
          val queue = readyQueues[i]
          queue.cancelAllAndDecide()
          if (queue.futureTasks.isEmpty()) {
            readyQueues.removeAt(i)
          }
    Plain Text
    - Registered: Fri May 03 11:42:14 GMT 2024
    - Last Modified: Mon Apr 29 00:33:04 GMT 2024
    - 10.6K bytes
    - Viewed (0)
  3. docs/metrics/v3.md

    | `resident_memory_bytes`       | `gauge`   | Resident memory size in bytes                                                                                  | `server` |
    | `virtual_memory_bytes`        | `gauge`   | Virtual memory size in bytes                                                                                   | `server` |
    Plain Text
    - Registered: Sun May 05 19:28:20 GMT 2024
    - Last Modified: Thu May 02 17:37:57 GMT 2024
    - 28.5K bytes
    - Viewed (0)
  4. .bazelrc

    build:nvcc_clang --@local_config_cuda//:cuda_compiler=nvcc
    
    
    # Debug config
    build:dbg -c dbg
    # Only include debug info for files under tensorflow/, excluding kernels, to
    # reduce the size of the debug info in the binary. This is because if the debug
    # sections in the ELF binary are too large, errors can occur. See
    # https://github.com/tensorflow/tensorflow/issues/48919.
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Thu May 02 19:34:20 GMT 2024
    - 52.8K bytes
    - Viewed (2)
  5. RELEASE.md

    *   `tensorflow.experimental.tensorrt`:
    
        *   `conversion_params` is now deprecated inside `TrtGraphConverterV2` in
            favor of direct arguments: `max_workspace_size_bytes`, `precision_mode`,
            `minimum_segment_size`, `maximum_cached_engines`, `use_calibration` and
            `allow_build_at_runtime`.
        *   Added a new parameter called `save_gpu_specific_engines` to the
    Plain Text
    - Registered: Tue May 07 12:40:20 GMT 2024
    - Last Modified: Mon Apr 29 19:17:57 GMT 2024
    - 727.7K bytes
    - Viewed (8)
  6. Makefile.core.mk

    LINUX_AGENT_BINARIES:=./cni/cmd/istio-cni \
      ./cni/cmd/install-cni \
      $(AGENT_BINARIES)
    
    BINARIES:=$(STANDARD_BINARIES) $(AGENT_BINARIES) $(LINUX_AGENT_BINARIES)
    
    # List of binaries that have their size tested
    RELEASE_SIZE_TEST_BINARIES:=pilot-discovery pilot-agent istioctl envoy ztunnel client server
    
    # agent: enables agent-specific files. Usually this is used to trim dependencies where they would be hard to trim through standard refactoring
    Plain Text
    - Registered: Wed May 08 22:53:08 GMT 2024
    - Last Modified: Wed May 08 20:25:15 GMT 2024
    - 22.5K bytes
    - Viewed (0)
  7. CONTRIBUTING.md

    - In the "File already exists" dialogue, choose "Yes" to overwrite
    - In the "Open Project" dialogue, choose "Delete Existing Project and Import"
    - Revert the Git changes to files in the `.idea` folder
    
    NOTE: Due to the project size, the very first import can take a while and IntelliJ might become unresponsive for several seconds during this period.
    
    Plain Text
    - Registered: Wed May 08 11:36:15 GMT 2024
    - Last Modified: Sat May 04 07:43:02 GMT 2024
    - 15.6K bytes
    - Viewed (0)
  8. docs/en/docs/deployment/concepts.md

    ### Server Memory
    
    For example, if your code loads a Machine Learning model with **1 GB in size**, when you run one process with your API, it will consume at least 1 GB of RAM. And if you start **4 processes** (4 workers), each will consume 1 GB of RAM. So in total, your API will consume **4 GB of RAM**.
    
    Plain Text
    - Registered: Sun May 05 07:19:11 GMT 2024
    - Last Modified: Thu May 02 22:37:31 GMT 2024
    - 18K bytes
    - Viewed (0)
Back to top