- Sort Score
- Num 10 results
- Language All
Results 1 - 10 of 143 for streaming (0.73 seconds)
-
cmd/streaming-signature-v4.go
xhttp "github.com/minio/minio/internal/http" ) // Streaming AWS Signature Version '4' constants. const ( emptySHA256 = "e3b0c44298fc1c149afbf4c8996fb92427ae41e4649b934ca495991b7852b855" streamingContentSHA256 = "STREAMING-AWS4-HMAC-SHA256-PAYLOAD" streamingContentSHA256Trailer = "STREAMING-AWS4-HMAC-SHA256-PAYLOAD-TRAILER" signV4ChunkedAlgorithm = "AWS4-HMAC-SHA256-PAYLOAD"
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Sep 28 20:59:21 GMT 2025 - 18.2K bytes - Click Count (0) -
cmd/streaming-signature-v4_test.go
Klaus Post <******@****.***> 1756435188 +0200
Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Fri Aug 29 02:39:48 GMT 2025 - 5.7K bytes - Click Count (0) -
fastapi/.agents/skills/fastapi/references/streaming.md
# Streaming ## Stream JSON Lines To stream JSON Lines, declare the return type and use `yield` to return the data. ```python @app.get("/items/stream") async def stream_items() -> AsyncIterable[Item]: for item in items: yield item ``` ## Server-Sent Events (SSE) To stream Server-Sent Events, use `response_class=EventSourceResponse` and `yield` items from the endpoint.
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Sun Mar 01 10:05:57 GMT 2026 - 2.5K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/LlmClient.java
*/ LlmChatResponse chat(LlmChatRequest request); /** * Performs a streaming chat completion request. * The callback is invoked for each chunk of the response. * * @param request the chat request containing messages and parameters * @param callback the callback to receive streaming chunks * @throws LlmException if an error occurs during the request */Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Thu Mar 19 07:04:54 GMT 2026 - 7.3K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/LlmClientManager.java
throw new LlmException("LLM chat request failed", e); } } /** * Performs a streaming chat completion request using the configured LLM client. * * @param request the chat request * @param callback the callback to receive streaming chunks * @throws LlmException if LLM is not available or an error occurs */
Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Thu Mar 19 11:10:51 GMT 2026 - 17.4K bytes - Click Count (0) -
internal/event/target/nats.go
connOpts := []stan.Option{stan.NatsURL(addressURL)} if n.Streaming.MaxPubAcksInflight > 0 { connOpts = append(connOpts, stan.MaxPubAcksInflight(n.Streaming.MaxPubAcksInflight)) } if n.UserCredentials != "" { connOpts = append(connOpts, stan.NatsOptions(nats.UserCredentials(n.UserCredentials))) } return stan.Connect(n.Streaming.ClusterID, clientID, connOpts...) } // NATSTarget - NATS target.Created: Sun Apr 05 19:28:12 GMT 2026 - Last Modified: Sun Apr 27 04:30:57 GMT 2025 - 13.5K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/LlmStreamCallback.java
* governing permissions and limitations under the License. */ package org.codelibs.fess.llm; /** * Callback interface for receiving streaming chat responses from LLM. */ @FunctionalInterface public interface LlmStreamCallback { /** * Called for each chunk of the streaming response. * * @param chunk the text chunk from the LLM response * @param done true if this is the final chunk */Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Mon Jan 12 10:32:40 GMT 2026 - 1.2K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/LlmChatRequest.java
return this; } /** * Checks if streaming is enabled. * * @return true if streaming is enabled */ public boolean isStream() { return stream; } /** * Sets whether streaming is enabled. * * @param stream true to enable streaming * @return this request for method chaining */Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Thu Mar 05 03:38:31 GMT 2026 - 6.4K bytes - Click Count (0) -
tests/test_stream_cancellation.py
""" Test that async streaming endpoints can be cancelled without hanging. Ref: https://github.com/fastapi/fastapi/issues/14680 """ from collections.abc import AsyncIterable import anyio import pytest from fastapi import FastAPI from fastapi.responses import StreamingResponse pytestmark = [ pytest.mark.anyio, pytest.mark.filterwarnings("ignore::pytest.PytestUnraisableExceptionWarning"), ] app = FastAPI()Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Fri Feb 27 18:56:47 GMT 2026 - 2.7K bytes - Click Count (0) -
docs/en/docs/editor-support.md
For more details about the extension, refer to the README on the [GitHub repository](https://github.com/fastapi/fastapi-vscode). ## Setup and Installation { #setup-and-installation }
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 05 18:13:19 GMT 2026 - 2.5K bytes - Click Count (0)