repo_name
stringclasses
6 values
pr_number
int64
99
20.3k
pr_title
stringlengths
8
158
pr_description
stringlengths
0
6.54k
author
stringlengths
4
18
date_created
timestamp[ns, tz=UTC]
date_merged
timestamp[ns, tz=UTC]
previous_commit
stringlengths
40
40
pr_commit
stringlengths
40
40
query
stringlengths
37
6.57k
filepath
stringlengths
8
153
before_content
stringlengths
0
876M
after_content
stringlengths
0
876M
label
int64
-1
1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./js/lykke.js
'use strict'; // --------------------------------------------------------------------------- const Exchange = require ('./base/Exchange'); // --------------------------------------------------------------------------- module.exports = class lykke extends Exchange { describe () { return this.deepExtend...
'use strict'; // --------------------------------------------------------------------------- const Exchange = require ('./base/Exchange'); // --------------------------------------------------------------------------- module.exports = class lykke extends Exchange { describe () { return this.deepExtend...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./js/static_dependencies/node-rsa/formats/pkcs1.js
var ber = require('../asn1/index').Ber; var _ = require('../utils')._; var utils = require('../utils'); const PRIVATE_OPENING_BOUNDARY = '-----BEGIN RSA PRIVATE KEY-----'; const PRIVATE_CLOSING_BOUNDARY = '-----END RSA PRIVATE KEY-----'; const PUBLIC_OPENING_BOUNDARY = '-----BEGIN RSA PUBLIC KEY-----'; const PUBLIC_C...
var ber = require('../asn1/index').Ber; var _ = require('../utils')._; var utils = require('../utils'); const PRIVATE_OPENING_BOUNDARY = '-----BEGIN RSA PRIVATE KEY-----'; const PRIVATE_CLOSING_BOUNDARY = '-----END RSA PRIVATE KEY-----'; const PUBLIC_OPENING_BOUNDARY = '-----BEGIN RSA PUBLIC KEY-----'; const PUBLIC_C...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./php/bitcoincom.php
<?php namespace ccxt; // PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: // https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code use Exception; // a common import class bitcoincom extends hitbtc { public function describe() { return $this->deep_extend(p...
<?php namespace ccxt; // PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: // https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code use Exception; // a common import class bitcoincom extends hitbtc { public function describe() { return $this->deep_extend(p...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./python/ccxt/vbtc.py
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.foxbit import foxbit class vbtc(foxbit): def describe(self): return self.deep_extend(super(vbtc, self).describe(),...
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.foxbit import foxbit class vbtc(foxbit): def describe(self): return self.deep_extend(super(vbtc, self).describe(),...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./examples/js/binance-fetch-all-deposits.js
"use strict"; const ccxt = require ('../../ccxt.js') ;(async function main () { const exchange = new ccxt.binance ({ 'apiKey': 'YOUR_API_KEY', 'secret': 'YOUR_SECRET', 'enableRateLimit': true, }) await exchange.loadMarkets () // exchange.verbose = true // uncomment for debug...
"use strict"; const ccxt = require ('../../ccxt.js') ;(async function main () { const exchange = new ccxt.binance ({ 'apiKey': 'YOUR_API_KEY', 'secret': 'YOUR_SECRET', 'enableRateLimit': true, }) await exchange.loadMarkets () // exchange.verbose = true // uncomment for debug...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./doc/.gitignore
_build
_build
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./python/ccxt/async_support/bcex.py
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.async_support.base.exchange import Exchange from ccxt.base.errors import ExchangeError from ccxt.base.errors import Authenticatio...
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.async_support.base.exchange import Exchange from ccxt.base.errors import ExchangeError from ccxt.base.errors import Authenticatio...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./examples/py/async-okex-fetch-margin-balance-with-options.py
# -*- coding: utf-8 -*- import asyncio import os import sys from pprint import pprint root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root + '/python') import ccxt.async_support as ccxt # noqa: E402 async def loop(exchange): while True: try: ...
# -*- coding: utf-8 -*- import asyncio import os import sys from pprint import pprint root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__)))) sys.path.append(root + '/python') import ccxt.async_support as ccxt # noqa: E402 async def loop(exchange): while True: try: ...
-1
ccxt/ccxt
8,318
SBTC mapping
https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
ndubel
2021-01-21T10:35:23Z
2021-01-21T20:04:45Z
2614db0ebd43f3cf9e1222bde6cefbabb955f681
05f5feeaaac3d213bc1314ba47b814db9ac30852
SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/ https://coinmarketcap.com/currencies/super-bitcoin/ https://www.coingecko.com/en/coins/siambitcoin
./python/ccxt/yobit.py
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.base.exchange import Exchange # ----------------------------------------------------------------------------- try: basestri...
# -*- coding: utf-8 -*- # PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN: # https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code from ccxt.base.exchange import Exchange # ----------------------------------------------------------------------------- try: basestri...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/bert/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/wav2vec2_conformer/configuration_wav2vec2_conformer.py
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/roformer/tokenization_utils.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/pytorch/audio-classification/run_audio_classification.py
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
#!/usr/bin/env python # coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LI...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/research_projects/information-gain-filtration/run_clm_igf.py
# Copyright 2022 - Intel Corp. All rights reserved. # Authors: Mayank Kumar Raunak, Javier Turek, Nicole Beckage """ Implementation of a new method for fine-tuning transformer models that we call Information Gain Filtration 'IGF' on WikiText data set and compared the results with the standard fine-tuning method Steps...
# Copyright 2022 - Intel Corp. All rights reserved. # Authors: Mayank Kumar Raunak, Javier Turek, Nicole Beckage """ Implementation of a new method for fine-tuning transformer models that we call Information Gain Filtration 'IGF' on WikiText data set and compared the results with the standard fine-tuning method Steps...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/mctct/modeling_mctct.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/research_projects/seq2seq-distillation/make_student.py
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging logger = logging.get_logger(__name__) def copy_layers(src_layers: nn.ModuleList, des...
import warnings from pathlib import Path from typing import List, Tuple, Union import fire from torch import nn from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, PreTrainedModel from transformers.utils import logging logger = logging.get_logger(__name__) def copy_layers(src_layers: nn.ModuleList, des...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/auto/modeling_tf_auto.py
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/distilbert/tokenization_distilbert.py
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./utils/test_module/custom_tokenization_fast.py
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class CustomTokenizerFast(BertTokenizerFast): slow_tokenizer_class = CustomTokenizer pass
from transformers import BertTokenizerFast from .custom_tokenization import CustomTokenizer class CustomTokenizerFast(BertTokenizerFast): slow_tokenizer_class = CustomTokenizer pass
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/tapex/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/data2vec/configuration_data2vec_audio.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/gpt2/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/conditional_detr/feature_extraction_conditional_detr.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/t5/configuration_t5.py
# coding=utf-8 # Copyright 2020, The T5 Authors and HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by ...
# coding=utf-8 # Copyright 2020, The T5 Authors and HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by ...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/speech_to_text_2/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/deformable_detr/convert_deformable_detr_to_pytorch.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/segformer/feature_extraction_segformer.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/modeling_outputs.py
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/markuplm/configuration_markuplm.py
# coding=utf-8 # Copyright 2021, The Microsoft Research Asia MarkupLM Team authors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
# coding=utf-8 # Copyright 2021, The Microsoft Research Asia MarkupLM Team authors # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/marian/tokenization_marian.py
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/trocr/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./utils/get_modified_files.py
# coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2020 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/splinter/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/mvp/configuration_mvp.py
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/barthez/tokenization_barthez.py
# coding=utf-8 # Copyright 2020 Ecole Polytechnique and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
# coding=utf-8 # Copyright 2020 Ecole Polytechnique and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unles...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/test_feature_extraction_common.py
# coding=utf-8 # Copyright 2021 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
# coding=utf-8 # Copyright 2021 HuggingFace Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or ag...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/convnext/test_modeling_tf_convnext.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/data2vec/test_modeling_data2vec_vision.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/funnel/test_modeling_funnel.py
# coding=utf-8 # Copyright 2020 HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
# coding=utf-8 # Copyright 2020 HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/opt/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/speech_to_text/test_modeling_speech_to_text.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/sagemaker/test_multi_node_data_parallel.py
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface im...
import json import os import subprocess import unittest from ast import literal_eval import pytest from parameterized import parameterized, parameterized_class from . import is_sagemaker_available if is_sagemaker_available(): from sagemaker import Session, TrainingJobAnalytics from sagemaker.huggingface im...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/xlm_roberta/modeling_tf_xlm_roberta.py
# coding=utf-8 # Copyright 2019 Facebook AI Research and the HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the Licens...
# coding=utf-8 # Copyright 2019 Facebook AI Research and the HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the Licens...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/perceiver/tokenization_perceiver.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/flax/question-answering/utils_qa.py
# coding=utf-8 # Copyright 2020 The HuggingFace Team All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
# coding=utf-8 # Copyright 2020 The HuggingFace Team All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless require...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/deberta/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/en/model_doc/audio-spectrogram-transformer.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/pytorch/text-classification/run_glue_no_trainer.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/research_projects/distillation/train.py
# coding=utf-8 # Copyright 2019-present, the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
# coding=utf-8 # Copyright 2019-present, the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/squeezebert/configuration_squeezebert.py
# coding=utf-8 # Copyright 2020 The SqueezeBert authors and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# coding=utf-8 # Copyright 2020 The SqueezeBert authors and The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/research_projects/lxmert/demo.ipynb
{ "cells": [ { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [], "source": [ "# %pip install-r requirements.txt" ] }, { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", ...
{ "cells": [ { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [], "source": [ "# %pip install-r requirements.txt" ] }, { "cell_type": "code", "execution_count": 1, "metadata": {}, "outputs": [ { "name": "stderr", "output_type": "stream", ...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py
# coding=utf-8 # Copyright 2021 Google Research The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
# coding=utf-8 # Copyright 2021 Google Research The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/legacy/run_openai_gpt.py
#!/usr/bin/env python # coding=utf-8 # Copyright 2018 Google AI, Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in co...
#!/usr/bin/env python # coding=utf-8 # Copyright 2018 Google AI, Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in co...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/mobilenet_v2/convert_original_tf_checkpoint_to_pytorch.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/mbart/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/clipseg/processing_clipseg.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/xlnet/test_modeling_xlnet.py
# coding=utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/xlm/test_modeling_xlm.py
# coding=utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/en/tasks/token_classification.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/speech_encoder_decoder/modeling_flax_speech_encoder_decoder.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/it/pipeline_tutorial.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/gpt_neox_japanese/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/en/model_doc/layoutxlm.mdx
<!--Copyright 2021 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2021 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/roformer/test_modeling_tf_roformer.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/utils/test_add_new_model_like.py
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
# Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicabl...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/utils/dummy_sentencepiece_and_speech_objects.py
# This file is autogenerated by the command `make fix-copies`, do not edit. # flake8: noqa from ..utils import DummyObject, requires_backends class Speech2TextProcessor(metaclass=DummyObject): _backends = ["sentencepiece", "speech"] def __init__(self, *args, **kwargs): requires_backends(self, ["sente...
# This file is autogenerated by the command `make fix-copies`, do not edit. # flake8: noqa from ..utils import DummyObject, requires_backends class Speech2TextProcessor(metaclass=DummyObject): _backends = ["sentencepiece", "speech"] def __init__(self, *args, **kwargs): requires_backends(self, ["sente...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/vision_text_dual_encoder/configuration_vision_text_dual_encoder.py
# coding=utf-8 # Copyright The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/commands/transformers_cli.py
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/legacy/seq2seq/minify_dataset.py
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
#!/usr/bin/env python # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/opt/modeling_opt.py
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
# coding=utf-8 # Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./templates/adding_a_missing_tokenization_test/cookiecutter-template-{{cookiecutter.modelname}}/test_tokenization_{{cookiecutter.lowercase_modelname}}.py
# coding=utf-8 # Copyright 2022 {{cookiecutter.authors}}. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
# coding=utf-8 # Copyright 2022 {{cookiecutter.authors}}. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless re...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/bert_japanese/__init__.py
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/layoutlmv2/processing_layoutlmv2.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/it/preprocessing.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/unispeech/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2021 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./.github/workflows/self-nightly-scheduled.yml
name: Self-hosted runner (nightly) # Note that each job's dependencies go into a corresponding docker file. # # For example for `run_all_tests_torch_cuda_extensions_gpu` the docker image is # `huggingface/transformers-pytorch-deepspeed-latest-gpu`, which can be found at # `docker/transformers-pytorch-deepspeed-latest-...
name: Self-hosted runner (nightly) # Note that each job's dependencies go into a corresponding docker file. # # For example for `run_all_tests_torch_cuda_extensions_gpu` the docker image is # `huggingface/transformers-pytorch-deepspeed-latest-gpu`, which can be found at # `docker/transformers-pytorch-deepspeed-latest-...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./.github/ISSUE_TEMPLATE/config.yml
blank_issues_enabled: true version: 2.1 contact_links: - name: Model checkpoints on the Hugging Face Hub url: https://huggingface.co/models about: Open a Pull request / Discussion related to a specific model checkpoint directly on the Hugging Face Hub - name: Website Related url: https://github.com/hugg...
blank_issues_enabled: true version: 2.1 contact_links: - name: Model checkpoints on the Hugging Face Hub url: https://huggingface.co/models about: Open a Pull request / Discussion related to a specific model checkpoint directly on the Hugging Face Hub - name: Website Related url: https://github.com/hugg...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/poolformer/feature_extraction_poolformer.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/convert_slow_tokenizers_checkpoints_to_fast.py
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./examples/research_projects/codeparrot/scripts/arguments.py
from dataclasses import dataclass, field from typing import Optional @dataclass class TrainingArguments: """ Configuration for training model. """ model_ckpt: Optional[str] = field( default="codeparrot/codeparrot", metadata={"help": "Model name or path of model to be trained."} ) save...
from dataclasses import dataclass, field from typing import Optional @dataclass class TrainingArguments: """ Configuration for training model. """ model_ckpt: Optional[str] = field( default="codeparrot/codeparrot", metadata={"help": "Model name or path of model to be trained."} ) save...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/de/pipeline_tutorial.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./templates/adding_a_missing_tokenization_test/cookiecutter.json
{ "modelname": "BrandNewBERT", "uppercase_modelname": "BRAND_NEW_BERT", "lowercase_modelname": "brand_new_bert", "camelcase_modelname": "BrandNewBert", "has_slow_class": ["True", "False"], "has_fast_class": ["True", "False"], "slow_tokenizer_use_sentencepiece": ["True", "False"], "authors": "The Hugging...
{ "modelname": "BrandNewBERT", "uppercase_modelname": "BRAND_NEW_BERT", "lowercase_modelname": "brand_new_bert", "camelcase_modelname": "BrandNewBert", "has_slow_class": ["True", "False"], "has_fast_class": ["True", "False"], "slow_tokenizer_use_sentencepiece": ["True", "False"], "authors": "The Hugging...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/bert_japanese/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2020 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/big_bird/tokenization_big_bird_fast.py
# coding=utf-8 # Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
# coding=utf-8 # Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # U...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/generation/tf_utils.py
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
# coding=utf-8 # Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a cop...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/big_bird/tokenization_big_bird.py
# coding=utf-8 # Copyright 2021 Google Research and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
# coding=utf-8 # Copyright 2021 Google Research and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICE...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/mpnet/configuration_mpnet.py
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License ...
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation. # Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License ...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./src/transformers/models/wavlm/configuration_wavlm.py
# coding=utf-8 # Copyright 2021 The Fairseq Authors, Microsoft Research, and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www...
# coding=utf-8 # Copyright 2021 The Fairseq Authors, Microsoft Research, and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./docs/source/en/perf_train_tpu.mdx
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
<!--Copyright 2022 The HuggingFace Team. All rights reserved. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed...
-1
huggingface/transformers
20,340
[FLAX] Add dtype to embedding for bert/bart/opt/t5
## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @LysandreJik, @sanchit-gandhi
merrymercy
2022-11-21T00:53:45Z
2022-11-28T15:21:43Z
667ccea72235504ab7876024e4f8c113ca62190f
ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e
[FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do? This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies. This dtype is necessary for mixed precision training. ## Who can review? @patrickvonplaten, @...
./tests/models/layoutlmv2/test_modeling_layoutlmv2.py
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2021 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
-1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/dinat/configuration_dinat.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/dinat/modeling_dinat.py
# coding=utf-8 # Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0...
# coding=utf-8 # Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0...
1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/nat/__init__.py
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
# flake8: noqa # There's no way to ignore "F401 '...' imported but unused" warnings in this # module, but to preserve other warnings. So, don't check this module at all. # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use thi...
1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/nat/configuration_nat.py
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
# coding=utf-8 # Copyright 2022 The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless r...
1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/nat/modeling_nat.py
# coding=utf-8 # Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0...
# coding=utf-8 # Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0...
1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./examples/legacy/pytorch-lightning/lightning_base.py
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, ...
import argparse import logging import os from pathlib import Path from typing import Any, Dict import pytorch_lightning as pl from pytorch_lightning.utilities import rank_zero_info from transformers import ( AdamW, AutoConfig, AutoModel, AutoModelForPreTraining, AutoModelForQuestionAnswering, ...
-1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./src/transformers/models/roberta/convert_roberta_original_pytorch_checkpoint_to_pytorch.py
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
# coding=utf-8 # Copyright 2018 The HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable...
-1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./examples/research_projects/distillation/scripts/extract_distilbert.py
# coding=utf-8 # Copyright 2019-present, the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
# coding=utf-8 # Copyright 2019-present, the HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by a...
-1
huggingface/transformers
20,325
Add LayerScale to NAT/DiNAT
# What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] This PR fixes a typo or improv...
alihassanijr
2022-11-18T22:01:53Z
2022-11-21T14:08:35Z
d28448c5cd8fa8dfb64190c7f55275d80e256a9e
11f3ec7224c83c9e5c379a774b9d3984e68d26fa
Add LayerScale to NAT/DiNAT. # What does this PR do? This follows PR #20219 . I completely dropped the ball on LayerScale in the original PR. This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability. ## Before submitting - [ ] T...
./tests/models/data2vec/test_modeling_data2vec_audio.py
# coding=utf-8 # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
# coding=utf-8 # Copyright 2022 The HuggingFace Team. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless requir...
-1