Search Options

Results per page
Sort
Preferred Languages
Advance

Results 51 - 60 of 268 for Extract (0.03 sec)

  1. build-logic-commons/gradle-plugin/src/main/kotlin/gradlebuild/testcleanup/TestFilesCleanupService.kt

        override fun close() {
            val projectPathToLeftoverFiles = mutableMapOf<String, LeftoverFiles>()
            // First run: delete any temporary directories used to extract resources from jars
            parameters.projectStates.get().values.forEach { projectExtension ->
                cleanUp(projectExtension.tmpExtractedResourcesDirs())
            }
    
    Registered: Wed Dec 31 11:36:14 UTC 2025
    - Last Modified: Wed Nov 05 11:43:49 UTC 2025
    - 12.5K bytes
    - Viewed (1)
  2. src/main/java/org/codelibs/fess/crawler/transformer/FessTransformer.java

         *
         * @return the logger instance
         */
        Logger getLogger();
    
        /**
         * Extracts the host name from a URL string.
         * Removes protocol and path components to return just the hostname.
         *
         * @param u the URL string to extract host from
         * @return the host name, or empty string if URL is blank, or unknown hostname if parsing fails
         */
    Registered: Sat Dec 20 09:19:18 UTC 2025
    - Last Modified: Thu Dec 11 09:47:03 UTC 2025
    - 14.1K bytes
    - Viewed (0)
  3. fess-crawler/src/main/java/org/codelibs/fess/crawler/extractor/impl/MsPowerPointExtractor.java

    import org.codelibs.fess.crawler.exception.ExtractException;
    
    /**
     * Extracts text content from Microsoft PowerPoint documents.
     */
    public class MsPowerPointExtractor extends AbstractExtractor {
    
        /**
         * Creates a new MsPowerPointExtractor instance.
         */
        public MsPowerPointExtractor() {
            super();
        }
    
        /**
         * Extracts text from the PowerPoint input stream.
         * @param in The input stream.
    Registered: Sat Dec 20 11:21:39 UTC 2025
    - Last Modified: Mon Nov 24 03:59:47 UTC 2025
    - 2K bytes
    - Viewed (0)
  4. fess-crawler/src/test/java/org/codelibs/fess/crawler/extractor/ExtractorFactoryTest.java

        public void test_builder() {
            assertEquals("test", extractorFactory.builder(new ByteArrayInputStream("test".getBytes()), null).extract().getContent());
            assertEquals("test",
                    extractorFactory.builder(new ByteArrayInputStream("test".getBytes()), null).filename("test.txt").extract().getContent());
        }
    
        public void test_addExtractor_weight() {
            final String key = "application/test";
    Registered: Sat Dec 20 11:21:39 UTC 2025
    - Last Modified: Sat Mar 15 06:52:00 UTC 2025
    - 6.9K bytes
    - Viewed (0)
  5. fess-crawler/src/main/java/org/codelibs/fess/crawler/extractor/impl/TextExtractor.java

                final String content = new String(InputStreamUtil.getBytes(in), getEncoding());
                return new ExtractData(content);
            } catch (final Exception e) {
                throw new ExtractException("Failed to extract text content using encoding: " + getEncoding(), e);
            }
        }
    
        /**
         * Returns the encoding used for text extraction.
         * @return the encoding
         */
        public String getEncoding() {
    Registered: Sat Dec 20 11:21:39 UTC 2025
    - Last Modified: Thu Dec 11 08:38:29 UTC 2025
    - 2K bytes
    - Viewed (0)
  6. fess-crawler/src/main/java/org/codelibs/fess/crawler/extractor/impl/MsPublisherExtractor.java

            if (in == null) {
                throw new CrawlerSystemException("Microsoft Publisher input stream is null. Cannot extract text from null input.");
            }
            try {
                @SuppressWarnings("resource")
                final PublisherTextExtractor publisherTextExtractor = new PublisherTextExtractor(in);
                return new ExtractData(publisherTextExtractor.getText());
            } catch (final IOException e) {
    Registered: Sat Dec 20 11:21:39 UTC 2025
    - Last Modified: Sun Nov 23 12:19:14 UTC 2025
    - 2K bytes
    - Viewed (0)
  7. docs/extensions/s3zip/examples/boto3/main.py

            aws_secret_access_key='YOUR-SECRETACCESSKEY',
            config=Config(signature_version='s3v4'),
            region_name='us-east-1')
    
    
    def _add_header(request, **kwargs):
        request.headers.add_header('x-minio-extract', 'true')
    event_system = s3.meta.events
    event_system.register_first('before-sign.s3.*', _add_header)
    
    # List zip contents
    response = s3.list_objects_v2(Bucket="your-bucket", Prefix="path/to/file.zip/")
    print(response)
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Wed Aug 04 21:15:45 UTC 2021
    - 771 bytes
    - Viewed (0)
  8. build-logic/documentation/src/main/groovy/gradlebuild/docs/dsl/source/ExtractDslMetaDataTask.groovy

        /**
         * {@inheritDoc}
         */
        @Override
        @PathSensitive(PathSensitivity.NAME_ONLY)
        FileTree getSource() {
            return super.getSource();
        }
    
        @TaskAction
        def extract() {
            Date start = new Date()
    
            //parsing all input files into metadata
            //and placing them in the repository object
    Registered: Wed Dec 31 11:36:14 UTC 2025
    - Last Modified: Mon Jan 08 12:45:57 UTC 2024
    - 4.4K bytes
    - Viewed (0)
  9. docs/en/docs/tutorial/query-param-models.md

    Declare the **query parameters** that you need in a **Pydantic model**, and then declare the parameter as `Query`:
    
    {* ../../docs_src/query_param_models/tutorial001_an_py310.py hl[9:13,17] *}
    
    **FastAPI** will **extract** the data for **each field** from the **query parameters** in the request and give you the Pydantic model you defined.
    
    ## Check the Docs { #check-the-docs }
    
    You can see the query parameters in the docs UI at `/docs`:
    
    Registered: Sun Dec 28 07:19:09 UTC 2025
    - Last Modified: Sun Aug 31 09:15:41 UTC 2025
    - 2.2K bytes
    - Viewed (0)
  10. docs/resiliency/resiliency-verify-healing-script.sh

    #!/usr/bin/env bash
    
    echo "script failed" >resiliency-verify-healing.log # assume initial state
    
    # Extract arguments from json object ...
    FILE=$(echo $1 | jq -r '.args.file')
    DIR=$(echo $1 | jq -r '.args.dir')
    DEEP=$(echo $1 | jq -r '.args.deep')
    WANT=$(echo $1 | jq 'del(.args)') # ... and remove args from wanted result
    
    ALIAS_NAME=myminio
    BUCKET="test-bucket"
    Registered: Sun Dec 28 19:28:13 UTC 2025
    - Last Modified: Sat Dec 21 04:24:45 UTC 2024
    - 996 bytes
    - Viewed (0)
Back to top