- Sort Score
- Num 10 results
- Language All
Results 41 - 50 of 84 for Roboto (0.05 seconds)
-
android-test/src/test/kotlin/okhttp/android/test/BaseOkHttpClientUnitTest.kt
.cache(Cache(FakeFileSystem(), "/cache".toPath(), 10_000_000)) .build() } @Test fun testRequestExternal() { assumeNetwork() val request = Request("https://www.google.com/robots.txt".toHttpUrl()) val networkRequest = request .newBuilder() .build() val call = client.newCall(networkRequest) call.execute().use { response ->
Created: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Tue Jul 22 20:03:31 GMT 2025 - 2.3K bytes - Click Count (0) -
fess-crawler/src/test/java/org/codelibs/fess/crawler/util/CrawlerWebServer.java
Created: Sat Dec 20 11:21:39 GMT 2025 - Last Modified: Sat Mar 15 06:52:00 GMT 2025 - 6.3K bytes - Click Count (0) -
CLAUDE.md
- **Testing**: JUnit 4, UTFlute, Mockito 5.7.0 - **Storage**: In-memory (default), OpenSearch (optional) ### Protocols - **HTTP/HTTPS**: Full crawling, cookies, auth, robots.txt - **File**: Local/network file systems - **FTP**: With authentication - **SMB/CIFS**: Windows shares (SMB1/SMB2+) - **Storage**: MinIO/S3-compatible ### Content Formats
Created: Sat Dec 20 11:21:39 GMT 2025 - Last Modified: Fri Nov 28 17:31:34 GMT 2025 - 10.7K bytes - Click Count (0) -
okhttp/src/jvmTest/kotlin/okhttp3/OpenJSSETest.kt
assertEquals(listOf("TLSv1.3", "TLSv1.2"), s.enabledProtocols.toList()) } @Test @Disabled fun testMozilla() { assumeNetwork() val request = Request.Builder().url("https://mozilla.org/robots.txt").build() client.newCall(request).execute().use { assertThat(it.protocol).isEqualTo(Protocol.HTTP_2) assertThat(it.handshake!!.tlsVersion).isEqualTo(TlsVersion.TLS_1_3) } } @TestCreated: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Wed Jun 18 12:28:21 GMT 2025 - 3.7K bytes - Click Count (0) -
android-test/src/androidTest/java/okhttp/android/test/letsencrypt/LetsEncryptClientTest.kt
handshakeCertificates.trustManager, ) } val client = clientBuilder.build() val request = Request .Builder() .url("https://valid-isrgrootx1.letsencrypt.org/robots.txt") .build() client.newCall(request).execute().use { response -> assertThat(response.code).isEqualTo(404) assertThat(response.protocol).isEqualTo(Protocol.HTTP_2) } }Created: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Wed Mar 19 19:25:20 GMT 2025 - 4.4K bytes - Click Count (0) -
src/main/webapp/css/admin/adminlte.min.css
c107;--danger:#dc3545;--light:#f8f9fa;--dark:#343a40;--breakpoint-xs:0;--breakpoint-sm:576px;--breakpoint-md:768px;--breakpoint-lg:992px;--breakpoint-xl:1200px;--font-family-sans-serif:"Source Sans Pro",-apple-system,BlinkMacSystemFont,"Segoe UI",Roboto,"Helvetica Neue",Arial,sans-serif,"Apple Color Emoji","Segoe UI Emoji","Segoe UI Symbol";--font-family-monospace:SFMono-Regular,Menlo,Monaco,Consolas,"Liberation Mono","Courier New",monospace}*,::after,::before{box-sizing:border-box}html{font-fam...
Created: Sat Dec 20 09:19:18 GMT 2025 - Last Modified: Sat Oct 26 01:49:09 GMT 2024 - 1.3M bytes - Click Count (0) -
fess-crawler/src/main/java/org/codelibs/fess/crawler/exception/RobotsTxtException.java
* governing permissions and limitations under the License. */ package org.codelibs.fess.crawler.exception; /** * RobotsTxtException is an exception class that represents an error related to robots.txt processing during web crawling. * It extends CrawlerSystemException and provides constructors to create instances with a message and/or a cause. * */ public class RobotsTxtException extends CrawlerSystemException {Created: Sat Dec 20 11:21:39 GMT 2025 - Last Modified: Sun Jul 06 02:13:03 GMT 2025 - 1.5K bytes - Click Count (0) -
src/main/resources/fess_indices/fess/cs/stopwords.txt
a s k o i u v z dnes cz tímto budeš budem byli jseš můj svým ta tomto tohle tuto tyto jej zda proč máte tato kam tohoto kdo kteří mi nám tom tomuto mít nic proto kterou byla toho protože asi ho naši napište re což tím takže svých její svými jste aj tu
Created: Sat Dec 20 09:19:18 GMT 2025 - Last Modified: Thu Jul 19 06:31:02 GMT 2018 - 992 bytes - Click Count (0) -
okhttp/src/jvmTest/kotlin/okhttp3/BouncyCastleTest.kt
OkHttpDebugLogging.enable("org.bouncycastle.jsse") platform.assumeBouncyCastle() } @Test fun testMozilla() { assumeNetwork() val request = Request.Builder().url("https://mozilla.org/robots.txt").build() client.newCall(request).execute().use { assertThat(it.protocol).isEqualTo(Protocol.HTTP_2) assertThat(it.handshake!!.tlsVersion).isEqualTo(TlsVersion.TLS_1_3) } }
Created: Fri Dec 26 11:42:13 GMT 2025 - Last Modified: Wed Jun 18 12:28:21 GMT 2025 - 1.6K bytes - Click Count (0) -
fess-crawler/src/test/resources/org/codelibs/fess/crawler/helper/robots_malformed.txt
# Test file for malformed robots.txt parsing # This file contains various malformed directives that should be handled gracefully # Case 1: Directives before any User-agent (should be ignored) Disallow: /orphaned1/ Allow: /orphaned2/ # Case 2: Valid user-agent with various malformed directives User-agent: GoodBot Disallow: /admin/ InvalidDirective: some-value unknown-field: test Disallow /missing-colon
Created: Sat Dec 20 11:21:39 GMT 2025 - Last Modified: Fri Nov 14 12:52:01 GMT 2025 - 2.6K bytes - Click Count (0)