- Sort Score
- Num 10 results
- Language All
Results 11 - 20 of 212 for prompt (0.15 seconds)
-
docs/tr/llm-prompt.md
Kader Miyanyedi <******@****.***> 1768941243 +0300
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Tue Jan 20 20:34:03 GMT 2026 - 2.7K bytes - Click Count (0) -
scripts/general-llm-prompt.md
Motov Yurii <******@****.***> 1773831336 +0100
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Wed Mar 18 10:55:36 GMT 2026 - 14.6K bytes - Click Count (0) -
compat/maven-compat/src/main/java/org/apache/maven/internal/compat/interactivity/LegacyPlexusInteractivity.java
public class LegacyPlexusInteractivity implements Prompter, InputHandler, OutputHandler { private final org.apache.maven.api.services.Prompter prompter; @Inject public LegacyPlexusInteractivity(org.apache.maven.api.services.Prompter prompter) { this.prompter = prompter; } @Override public String readLine() throws IOException { try { return prompter.prompt(null);Created: Sun Apr 05 03:35:12 GMT 2026 - Last Modified: Thu Apr 03 13:48:41 GMT 2025 - 5.2K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/AbstractLlmClient.java
} /** * Resolves the {{languageInstruction}} placeholder in a prompt. * * @param prompt the prompt template * @return the prompt with language instruction resolved */ protected String resolveLanguageInstruction(final String prompt) { if (prompt == null) { return null; }Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Sat Mar 21 06:04:58 GMT 2026 - 72K bytes - Click Count (0) -
api/maven-api-core/src/main/java/org/apache/maven/api/services/Prompter.java
*/ @Experimental public interface Prompter extends Service { /** * Prompts the user for a string. * * @param message the message to display to the user * @return the string entered by the user * @throws PrompterException if an exception occurs */ @Nonnull default String prompt(@Nullable String message) throws PrompterException { return prompt(message, null, null); } /**Created: Sun Apr 05 03:35:12 GMT 2026 - Last Modified: Fri Nov 17 15:52:15 GMT 2023 - 3.5K bytes - Click Count (0) -
docs_src/server_sent_events/tutorial005_py310.py
from fastapi.sse import EventSourceResponse, ServerSentEvent from pydantic import BaseModel app = FastAPI() class Prompt(BaseModel): text: str @app.post("/chat/stream", response_class=EventSourceResponse) async def stream_chat(prompt: Prompt) -> AsyncIterable[ServerSentEvent]: words = prompt.text.split() for word in words: yield ServerSentEvent(data=word, event="token")
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Sun Mar 01 09:21:52 GMT 2026 - 528 bytes - Click Count (0) -
docs/en/docs/contributing.md
Instead of adding suggestions to the translation PR, make the suggestions to the LLM prompt file for that language, in a new PR. For example, for Spanish, the LLM prompt file is at: [`docs/es/llm-prompt.md`](https://github.com/fastapi/fastapi/blob/master/docs/es/llm-prompt.md). /// tip
Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Mon Mar 23 13:59:26 GMT 2026 - 10.6K bytes - Click Count (0) -
docs/en/docs/css/termynal.css
display: inline-block; vertical-align: middle; } [data-ty="input"]:before, [data-ty-prompt]:before { margin-right: 0.75em; color: var(--color-text-subtle); } [data-ty="input"]:before { content: '$'; } [data-ty][data-ty-prompt]:before { content: attr(data-ty-prompt); } [data-ty-cursor]:after { content: attr(data-ty-cursor); font-family: monospace;Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Sun Nov 23 09:27:40 GMT 2025 - 2.2K bytes - Click Count (0) -
src/main/java/org/codelibs/fess/llm/LlmChatResponse.java
this.finishReason = finishReason; } /** * Gets the number of prompt tokens. * * @return the prompt tokens count */ public Integer getPromptTokens() { return promptTokens; } /** * Sets the number of prompt tokens. * * @param promptTokens the prompt tokens count */ public void setPromptTokens(final Integer promptTokens) {Created: Tue Mar 31 13:07:34 GMT 2026 - Last Modified: Mon Jan 12 10:32:40 GMT 2026 - 3.4K bytes - Click Count (0) -
docs/en/docs/_llm-test.md
Tests added here will be seen by all designers of language specific prompts. Use as follows: * Have a language specific prompt - `docs/{language code}/llm-prompt.md`.Created: Sun Apr 05 07:19:11 GMT 2026 - Last Modified: Thu Mar 05 18:13:19 GMT 2026 - 11K bytes - Click Count (0)