Search Options

Results per page
Sort
Preferred Languages
Advance

Results 1 - 10 of 68 for Shleifer (0.63 sec)

  1. examples/research_projects/seq2seq-distillation/README.md

    + For the XSUM dataset, training on pseudo-labels worked best for Pegasus (`sshleifer/distill-pegasus-16-4`), while training with KD worked best for `distilbart-xsum-12-6`
    + For `sshleifer/dbart-xsum-12-3`
    Plain Text
    - Registered: 2021-04-25 10:36
    - Last Modified: 2020-12-11 15:07
    - 19.2K bytes
    - Viewed (2)
  2. setup.py

    setup(
        name="transformers",
        version="4.6.0.dev0",  # expected format is one of x.y.z.dev0, or x.y.z.rc1 or x.y.z (no to dashes, yes to dots)
        author="Thomas Wolf, Lysandre Debut, Victor Sanh, Julien Chaumond, Sam Shleifer, Patrick von Platen, Sylvain Gugger, Google AI Language Team Authors, Open AI team Authors, Facebook AI Authors, Carnegie Mellon University Authors",
        author_email="******@****.***",
    Python
    - Registered: 2021-04-25 10:36
    - Last Modified: 2021-04-21 23:17
    - 11.8K bytes
    - Viewed (0)
  3. README.md

        author = "Thomas Wolf and Lysandre Debut and Victor Sanh and Julien Chaumond and Clement Delangue and Anthony Moi and Pierric Cistac and Tim Rault and Rémi Louf and Morgan Funtowicz and Joe Davison and Sam Shleifer and Patrick von Platen and Clara Ma and Yacine Jernite and Julien Plu and Canwen Xu and Teven Le Scao and Sylvain Gugger and Mariama Drame and Quentin Lhoest and Alexander M. Rush",
    Plain Text
    - Registered: 2021-04-25 10:36
    - Last Modified: 2021-04-12 22:07
    - 33.8K bytes
    - Viewed (0)
  4. doc/whats_new/v0.18.rst

    practicalswift, Preston Parry, Qimu Zheng, Rachit Kansal, Raghav RV,
    Ralf Gommers, Ramana.S, Rammig, Randy Olson, Rob Alexander, Robert Lutz,
    Robin Schucker, Rohan Jain, Ruifeng Zheng, Ryan Yu, Rémy Léone, saihttam,
    Saiwing Yeung, Sam Shleifer, Samuel St-Jean, Sartaj Singh, Sasank Chilamkurthy,
    saurabh.bansod, Scott Andrews, Scott Lowe, seales, Sebastian Raschka, Sebastian
    Saeger, Sebastián Vanrell, Sergei Lebedev, shagun Sodhani, shanmuga cv,
    Plain Text
    - Registered: 2021-05-07 09:24
    - Last Modified: 2018-06-11 07:06
    - 35.9K bytes
    - Viewed (0)
  5. doc/whats_new/v0.19.rst

    Stephen Hoover, AishwaryaRK, Steven C. Howell, Gary Foreman, Neeraj Gangwar,
    Tahar, Jon Crall, dokato, Kathy Chen, ferria, Thomas Moreau, Charlie Brummitt,
    Nicolas Goix, Adam Kleczewski, Sam Shleifer, Nikita Singh, Basil Beirouti,
    Giorgio Patrini, Manoj Kumar, Rafael Possas, James Bourbeau, James A. Bednar,
    Janine Harper, Jaye, Jean Helie, Jeremy Steward, Artsiom, John Wei, Jonathan
    Plain Text
    - Registered: 2021-05-07 09:24
    - Last Modified: 2020-08-31 09:43
    - 47.2K bytes
    - Viewed (0)
  6. tests/test_benchmark.py

                    result = model_result["result"][batch_size][sequence_length]
                    self.assertIsNotNone(result)
    
        def test_inference_no_configs(self):
            MODEL_ID = "sshleifer/tiny-gpt2"
            benchmark_args = PyTorchBenchmarkArguments(
                models=[MODEL_ID],
                training=False,
                inference=True,
                sequence_lengths=[8],
                batch_sizes=[1],
    Python
    - Registered: 2021-04-25 10:36
    - Last Modified: 2020-12-07 23:36
    - 10.4K bytes
    - Viewed (1)
  7. tests/test_benchmark_tf.py

                    result = model_result["result"][batch_size][sequence_length]
                    self.assertIsNotNone(result)
    
        def test_inference_no_configs_eager(self):
            MODEL_ID = "sshleifer/tiny-gpt2"
            benchmark_args = TensorFlowBenchmarkArguments(
                models=[MODEL_ID],
                training=False,
                inference=True,
                sequence_lengths=[8],
                batch_sizes=[1],
    Python
    - Registered: 2021-04-25 10:36
    - Last Modified: 2020-12-07 23:36
    - 8.8K bytes
    - Viewed (0)
  8. examples/research_projects/seq2seq-distillation/_test_seq2seq_examples.py

    SUMMARIES = ["A very interesting story about what I ate for lunch.", "Avocado, celery, turkey, coffee"]
    T5_TINY = "patrickvonplaten/t5-tiny-random"
    T5_TINIER = "sshleifer/t5-tinier-random"
    BART_TINY = "sshleifer/bart-tiny-random"
    MBART_TINY = "sshleifer/tiny-mbart"
    MARIAN_TINY = "sshleifer/tiny-marian-en-de"
    FSMT_TINY = "stas/tiny-wmt19-en-de"
    
    
    stream_handler = logging.StreamHandler(sys.stdout)
    logger.addHandler(stream_handler)
    Python
    - Registered: 2021-04-25 10:36
    - Last Modified: 2020-12-11 15:07
    - 16.2K bytes
    - Viewed (0)
  9. docs/source/model_doc/pegasus.rst

      extractive summary.
    - Pegasus achieves SOTA summarization performance on all 12 downstream tasks, as measured by ROUGE and human eval.
    
    This model was contributed by `sshleifer <https://huggingface.co/sshleifer>`__. The Authors' code can be found `here
    <https://github.com/google-research/pegasus>`__.
    
    
    Checkpoints
    Plain Text
    - Registered: 2021-04-25 10:36
    - Last Modified: 2021-04-21 15:11
    - 7K bytes
    - Viewed (0)
  10. examples/research_projects/seq2seq-distillation/train_distilbart_cnn.sh

        --freeze_encoder --freeze_embeds --data_dir cnn_dm \
        --max_target_length 142 --val_max_target_length=142 \
        --train_batch_size=$BS --eval_batch_size=$BS --gradient_accumulation_steps=$GAS \
        --model_name_or_path sshleifer/student_cnn_12_6 \
        --tokenizer_name facebook/bart-large \
        --warmup_steps 500 \
        --output_dir distilbart-cnn-12-6 \
    Shell Script
    - Registered: 2021-04-25 10:36
    - Last Modified: 2020-12-11 15:07
    - 649 bytes
    - Viewed (0)
Back to top