- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 199 for machine (0.04 seconds)
-
api/go1.11.txt
pkg debug/elf, const EM_C166 = 116 pkg debug/elf, const EM_C166 Machine pkg debug/elf, const EM_CDP = 215 pkg debug/elf, const EM_CDP Machine pkg debug/elf, const EM_CE = 119 pkg debug/elf, const EM_CE Machine pkg debug/elf, const EM_CLOUDSHIELD = 192 pkg debug/elf, const EM_CLOUDSHIELD Machine pkg debug/elf, const EM_COGE = 216 pkg debug/elf, const EM_COGE Machine pkg debug/elf, const EM_COOL = 217 pkg debug/elf, const EM_COOL Machine
Created: Tue Dec 30 11:13:12 GMT 2025 - Last Modified: Wed Aug 22 03:48:56 GMT 2018 - 25K bytes - Click Count (0) -
ci/devinfra/docker/windows2022/Dockerfile
Remove-Item $zulu_zip; \ $env:PATH = [Environment]::GetEnvironmentVariable(\"PATH\", \"Machine\") + \";${zulu_root}\\bin\"; \ [Environment]::SetEnvironmentVariable(\"PATH\", $env:PATH, \"Machine\"); \ $env:JAVA_HOME = $zulu_root; \ [Environment]::SetEnvironmentVariable(\"JAVA_HOME\", $env:JAVA_HOME, \"Machine\") # Point to the LLVM installation. # The Bazel Windows guide claims it can find LLVM automatically,Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Mon Jan 13 18:59:55 GMT 2025 - 10.3K bytes - Click Count (0) -
docs/en/docs/async.md
### Concurrency + Parallelism: Web + Machine Learning { #concurrency-parallelism-web-machine-learning } With **FastAPI** you can take advantage of concurrency that is very common for web development (the same main attraction of NodeJS). But you can also exploit the benefits of parallelism and multiprocessing (having multiple processes running in parallel) for **CPU bound** workloads like those in Machine Learning systems.Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Sun Aug 31 09:56:21 GMT 2025 - 24K bytes - Click Count (0) -
docs/en/docs/deployment/concepts.md
### Server Memory { #server-memory } For example, if your code loads a Machine Learning model with **1 GB in size**, when you run one process with your API, it will consume at least 1 GB of RAM. And if you start **4 processes** (4 workers), each will consume 1 GB of RAM. So in total, your API will consume **4 GB of RAM**. And if your remote server or virtual machine only has 3 GB of RAM, trying to load more than 4 GB of RAM will cause problems. 🚨Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Sun Aug 31 09:15:41 GMT 2025 - 18.6K bytes - Click Count (1) -
docs/es/docs/async.md
Eso, más el simple hecho de que Python es el lenguaje principal para **Data Science**, Machine Learning y especialmente Deep Learning, hacen de FastAPI una muy buena opción para APIs web de Data Science / Machine Learning y aplicaciones (entre muchas otras).
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Wed Dec 17 10:15:01 GMT 2025 - 25.4K bytes - Click Count (0) -
apache-maven/src/assembly/maven/conf/settings.xml
Created: Sun Dec 28 03:35:09 GMT 2025 - Last Modified: Wed Jan 22 07:44:50 GMT 2025 - 11.1K bytes - Click Count (0) -
README.md
researchers push the state-of-the-art in ML and developers easily build and deploy ML-powered applications. TensorFlow was originally developed by researchers and engineers working within the Machine Intelligence team at Google Brain to conduct research in machine learning and neural networks. However, the framework is versatile enough to be used in other areas as well. TensorFlow provides stable [Python](https://www.tensorflow.org/api_docs/python)
Created: Tue Dec 30 12:39:10 GMT 2025 - Last Modified: Fri Jul 18 14:09:03 GMT 2025 - 11.6K bytes - Click Count (0) -
docs/fr/docs/async.md
* L'apprentissage automatique (ou **Machine Learning**) : cela nécessite de nombreuses multiplications de matrices et vecteurs. Imaginez une énorme feuille de calcul remplie de nombres que vous multiplierez entre eux tous au même moment.
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Sun Aug 31 09:56:21 GMT 2025 - 25.4K bytes - Click Count (0) -
docs/en/docs/deployment/docker.md
Linux containers run using the same Linux kernel of the host (machine, virtual machine, cloud server, etc). This just means that they are very lightweight (compared to full virtual machines emulating an entire operating system). This way, containers consume **little resources**, an amount comparable to running the processes directly (a virtual machine would consume much more).
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Sat Sep 20 12:58:04 GMT 2025 - 29.5K bytes - Click Count (1) -
docs/en/docs/_llm-test.md
* <abbr title="Parallel Server Gateway Interface">PSGI</abbr> ### The abbr gives an explanation { #the-abbr-gives-an-explanation } * <abbr title="A group of machines that are configured to be connected and work together in some way.">cluster</abbr> * <abbr title="A method of machine learning that uses artificial neural networks with numerous hidden layers between input and output layers, thereby developing a comprehensive internal structure">Deep Learning</abbr>
Created: Sun Dec 28 07:19:09 GMT 2025 - Last Modified: Thu Dec 11 14:48:47 GMT 2025 - 11.4K bytes - Click Count (0)