repo_name stringclasses 6
values | pr_number int64 99 20.3k | pr_title stringlengths 8 158 | pr_description stringlengths 0 6.54k | author stringlengths 4 18 | date_created timestamp[ns, tz=UTC] | date_merged timestamp[ns, tz=UTC] | previous_commit stringlengths 40 40 | pr_commit stringlengths 40 40 | query stringlengths 37 6.57k | filepath stringlengths 8 153 | before_content stringlengths 0 876M | after_content stringlengths 0 876M | label int64 -1 1 |
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./js/lykke.js | 'use strict';
// ---------------------------------------------------------------------------
const Exchange = require ('./base/Exchange');
// ---------------------------------------------------------------------------
module.exports = class lykke extends Exchange {
describe () {
return this.deepExtend... | 'use strict';
// ---------------------------------------------------------------------------
const Exchange = require ('./base/Exchange');
// ---------------------------------------------------------------------------
module.exports = class lykke extends Exchange {
describe () {
return this.deepExtend... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./js/static_dependencies/node-rsa/formats/pkcs1.js | var ber = require('../asn1/index').Ber;
var _ = require('../utils')._;
var utils = require('../utils');
const PRIVATE_OPENING_BOUNDARY = '-----BEGIN RSA PRIVATE KEY-----';
const PRIVATE_CLOSING_BOUNDARY = '-----END RSA PRIVATE KEY-----';
const PUBLIC_OPENING_BOUNDARY = '-----BEGIN RSA PUBLIC KEY-----';
const PUBLIC_C... | var ber = require('../asn1/index').Ber;
var _ = require('../utils')._;
var utils = require('../utils');
const PRIVATE_OPENING_BOUNDARY = '-----BEGIN RSA PRIVATE KEY-----';
const PRIVATE_CLOSING_BOUNDARY = '-----END RSA PRIVATE KEY-----';
const PUBLIC_OPENING_BOUNDARY = '-----BEGIN RSA PUBLIC KEY-----';
const PUBLIC_C... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./php/bitcoincom.php | <?php
namespace ccxt;
// PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
// https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
use Exception; // a common import
class bitcoincom extends hitbtc {
public function describe() {
return $this->deep_extend(p... | <?php
namespace ccxt;
// PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
// https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
use Exception; // a common import
class bitcoincom extends hitbtc {
public function describe() {
return $this->deep_extend(p... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./python/ccxt/vbtc.py | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.foxbit import foxbit
class vbtc(foxbit):
def describe(self):
return self.deep_extend(super(vbtc, self).describe(),... | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.foxbit import foxbit
class vbtc(foxbit):
def describe(self):
return self.deep_extend(super(vbtc, self).describe(),... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./examples/js/binance-fetch-all-deposits.js | "use strict";
const ccxt = require ('../../ccxt.js')
;(async function main () {
const exchange = new ccxt.binance ({
'apiKey': 'YOUR_API_KEY',
'secret': 'YOUR_SECRET',
'enableRateLimit': true,
})
await exchange.loadMarkets ()
// exchange.verbose = true // uncomment for debug... | "use strict";
const ccxt = require ('../../ccxt.js')
;(async function main () {
const exchange = new ccxt.binance ({
'apiKey': 'YOUR_API_KEY',
'secret': 'YOUR_SECRET',
'enableRateLimit': true,
})
await exchange.loadMarkets ()
// exchange.verbose = true // uncomment for debug... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./doc/.gitignore | _build
| _build
| -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./python/ccxt/async_support/bcex.py | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.async_support.base.exchange import Exchange
from ccxt.base.errors import ExchangeError
from ccxt.base.errors import Authenticatio... | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.async_support.base.exchange import Exchange
from ccxt.base.errors import ExchangeError
from ccxt.base.errors import Authenticatio... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./examples/py/async-okex-fetch-margin-balance-with-options.py | # -*- coding: utf-8 -*-
import asyncio
import os
import sys
from pprint import pprint
root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
sys.path.append(root + '/python')
import ccxt.async_support as ccxt # noqa: E402
async def loop(exchange):
while True:
try:
... | # -*- coding: utf-8 -*-
import asyncio
import os
import sys
from pprint import pprint
root = os.path.dirname(os.path.dirname(os.path.dirname(os.path.abspath(__file__))))
sys.path.append(root + '/python')
import ccxt.async_support as ccxt # noqa: E402
async def loop(exchange):
while True:
try:
... | -1 |
ccxt/ccxt | 8,318 | SBTC mapping | https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ndubel | 2021-01-21T10:35:23Z | 2021-01-21T20:04:45Z | 2614db0ebd43f3cf9e1222bde6cefbabb955f681 | 05f5feeaaac3d213bc1314ba47b814db9ac30852 | SBTC mapping. https://coinmarketcap.com/currencies/sbtc/markets/
https://coinmarketcap.com/currencies/super-bitcoin/
https://www.coingecko.com/en/coins/siambitcoin | ./python/ccxt/yobit.py | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.base.exchange import Exchange
# -----------------------------------------------------------------------------
try:
basestri... | # -*- coding: utf-8 -*-
# PLEASE DO NOT EDIT THIS FILE, IT IS GENERATED AND WILL BE OVERWRITTEN:
# https://github.com/ccxt/ccxt/blob/master/CONTRIBUTING.md#how-to-contribute-code
from ccxt.base.exchange import Exchange
# -----------------------------------------------------------------------------
try:
basestri... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/bert/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/wav2vec2_conformer/configuration_wav2vec2_conformer.py | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/roformer/tokenization_utils.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/pytorch/audio-classification/run_audio_classification.py | #!/usr/bin/env python
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LI... | #!/usr/bin/env python
# coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LI... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/research_projects/information-gain-filtration/run_clm_igf.py | # Copyright 2022 - Intel Corp. All rights reserved.
# Authors: Mayank Kumar Raunak, Javier Turek, Nicole Beckage
"""
Implementation of a new method for fine-tuning transformer models that we call
Information Gain Filtration 'IGF' on WikiText data set and compared the results
with the standard fine-tuning method
Steps... | # Copyright 2022 - Intel Corp. All rights reserved.
# Authors: Mayank Kumar Raunak, Javier Turek, Nicole Beckage
"""
Implementation of a new method for fine-tuning transformer models that we call
Information Gain Filtration 'IGF' on WikiText data set and compared the results
with the standard fine-tuning method
Steps... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/mctct/modeling_mctct.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/research_projects/seq2seq-distillation/make_student.py | import warnings
from pathlib import Path
from typing import List, Tuple, Union
import fire
from torch import nn
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, PreTrainedModel
from transformers.utils import logging
logger = logging.get_logger(__name__)
def copy_layers(src_layers: nn.ModuleList, des... | import warnings
from pathlib import Path
from typing import List, Tuple, Union
import fire
from torch import nn
from transformers import AutoModelForSeq2SeqLM, AutoTokenizer, PreTrainedModel
from transformers.utils import logging
logger = logging.get_logger(__name__)
def copy_layers(src_layers: nn.ModuleList, des... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/auto/modeling_tf_auto.py | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/distilbert/tokenization_distilbert.py | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./utils/test_module/custom_tokenization_fast.py | from transformers import BertTokenizerFast
from .custom_tokenization import CustomTokenizer
class CustomTokenizerFast(BertTokenizerFast):
slow_tokenizer_class = CustomTokenizer
pass
| from transformers import BertTokenizerFast
from .custom_tokenization import CustomTokenizer
class CustomTokenizerFast(BertTokenizerFast):
slow_tokenizer_class = CustomTokenizer
pass
| -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/tapex/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/data2vec/configuration_data2vec_audio.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/gpt2/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/conditional_detr/feature_extraction_conditional_detr.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/t5/configuration_t5.py | # coding=utf-8
# Copyright 2020, The T5 Authors and HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by ... | # coding=utf-8
# Copyright 2020, The T5 Authors and HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by ... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/speech_to_text_2/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/deformable_detr/convert_deformable_detr_to_pytorch.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/segformer/feature_extraction_segformer.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/modeling_outputs.py | # Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | # Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/markuplm/configuration_markuplm.py | # coding=utf-8
# Copyright 2021, The Microsoft Research Asia MarkupLM Team authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | # coding=utf-8
# Copyright 2021, The Microsoft Research Asia MarkupLM Team authors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/marian/tokenization_marian.py | # Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | # Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/trocr/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./utils/get_modified_files.py | # coding=utf-8
# Copyright 2020 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2020 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/splinter/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/mvp/configuration_mvp.py | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/barthez/tokenization_barthez.py | # coding=utf-8
# Copyright 2020 Ecole Polytechnique and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | # coding=utf-8
# Copyright 2020 Ecole Polytechnique and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unles... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/test_feature_extraction_common.py | # coding=utf-8
# Copyright 2021 HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or ag... | # coding=utf-8
# Copyright 2021 HuggingFace Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or ag... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/convnext/test_modeling_tf_convnext.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/data2vec/test_modeling_data2vec_vision.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/funnel/test_modeling_funnel.py | # coding=utf-8
# Copyright 2020 HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law... | # coding=utf-8
# Copyright 2020 HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/opt/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/speech_to_text/test_modeling_speech_to_text.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/sagemaker/test_multi_node_data_parallel.py | import json
import os
import subprocess
import unittest
from ast import literal_eval
import pytest
from parameterized import parameterized, parameterized_class
from . import is_sagemaker_available
if is_sagemaker_available():
from sagemaker import Session, TrainingJobAnalytics
from sagemaker.huggingface im... | import json
import os
import subprocess
import unittest
from ast import literal_eval
import pytest
from parameterized import parameterized, parameterized_class
from . import is_sagemaker_available
if is_sagemaker_available():
from sagemaker import Session, TrainingJobAnalytics
from sagemaker.huggingface im... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/xlm_roberta/modeling_tf_xlm_roberta.py | # coding=utf-8
# Copyright 2019 Facebook AI Research and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the Licens... | # coding=utf-8
# Copyright 2019 Facebook AI Research and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the Licens... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/perceiver/tokenization_perceiver.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/flax/question-answering/utils_qa.py | # coding=utf-8
# Copyright 2020 The HuggingFace Team All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless require... | # coding=utf-8
# Copyright 2020 The HuggingFace Team All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless require... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/deberta/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/en/model_doc/audio-spectrogram-transformer.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/pytorch/text-classification/run_glue_no_trainer.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/research_projects/distillation/train.py | # coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | # coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/squeezebert/configuration_squeezebert.py | # coding=utf-8
# Copyright 2020 The SqueezeBert authors and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | # coding=utf-8
# Copyright 2020 The SqueezeBert authors and The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/research_projects/lxmert/demo.ipynb | {
"cells": [
{
"cell_type": "code",
"execution_count": 1,
"metadata": {},
"outputs": [],
"source": [
"# %pip install-r requirements.txt"
]
},
{
"cell_type": "code",
"execution_count": 1,
"metadata": {},
"outputs": [
{
"name": "stderr",
"output_type": "stream",
... | {
"cells": [
{
"cell_type": "code",
"execution_count": 1,
"metadata": {},
"outputs": [],
"source": [
"# %pip install-r requirements.txt"
]
},
{
"cell_type": "code",
"execution_count": 1,
"metadata": {},
"outputs": [
{
"name": "stderr",
"output_type": "stream",
... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/bigbird_pegasus/modeling_bigbird_pegasus.py | # coding=utf-8
# Copyright 2021 Google Research The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-... | # coding=utf-8
# Copyright 2021 Google Research The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/legacy/run_openai_gpt.py | #!/usr/bin/env python
# coding=utf-8
# Copyright 2018 Google AI, Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in co... | #!/usr/bin/env python
# coding=utf-8
# Copyright 2018 Google AI, Google Brain and Carnegie Mellon University Authors and the HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in co... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/mobilenet_v2/convert_original_tf_checkpoint_to_pytorch.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/mbart/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/clipseg/processing_clipseg.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/xlnet/test_modeling_xlnet.py | # coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | # coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/xlm/test_modeling_xlm.py | # coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | # coding=utf-8
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/en/tasks/token_classification.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/speech_encoder_decoder/modeling_flax_speech_encoder_decoder.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/it/pipeline_tutorial.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/gpt_neox_japanese/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/en/model_doc/layoutxlm.mdx | <!--Copyright 2021 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2021 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/roformer/test_modeling_tf_roformer.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/utils/test_add_new_model_like.py | # Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | # Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicabl... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/utils/dummy_sentencepiece_and_speech_objects.py | # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends
class Speech2TextProcessor(metaclass=DummyObject):
_backends = ["sentencepiece", "speech"]
def __init__(self, *args, **kwargs):
requires_backends(self, ["sente... | # This file is autogenerated by the command `make fix-copies`, do not edit.
# flake8: noqa
from ..utils import DummyObject, requires_backends
class Speech2TextProcessor(metaclass=DummyObject):
_backends = ["sentencepiece", "speech"]
def __init__(self, *args, **kwargs):
requires_backends(self, ["sente... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/vision_text_dual_encoder/configuration_vision_text_dual_encoder.py | # coding=utf-8
# Copyright The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | # coding=utf-8
# Copyright The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/commands/transformers_cli.py | #!/usr/bin/env python
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | #!/usr/bin/env python
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/legacy/seq2seq/minify_dataset.py | #!/usr/bin/env python
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | #!/usr/bin/env python
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/opt/modeling_opt.py | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | # coding=utf-8
# Copyright 2022 The Fairseq Authors and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./templates/adding_a_missing_tokenization_test/cookiecutter-template-{{cookiecutter.modelname}}/test_tokenization_{{cookiecutter.lowercase_modelname}}.py | # coding=utf-8
# Copyright 2022 {{cookiecutter.authors}}. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | # coding=utf-8
# Copyright 2022 {{cookiecutter.authors}}. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless re... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/bert_japanese/__init__.py | -1 | ||
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/layoutlmv2/processing_layoutlmv2.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/it/preprocessing.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/unispeech/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2021 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./.github/workflows/self-nightly-scheduled.yml | name: Self-hosted runner (nightly)
# Note that each job's dependencies go into a corresponding docker file.
#
# For example for `run_all_tests_torch_cuda_extensions_gpu` the docker image is
# `huggingface/transformers-pytorch-deepspeed-latest-gpu`, which can be found at
# `docker/transformers-pytorch-deepspeed-latest-... | name: Self-hosted runner (nightly)
# Note that each job's dependencies go into a corresponding docker file.
#
# For example for `run_all_tests_torch_cuda_extensions_gpu` the docker image is
# `huggingface/transformers-pytorch-deepspeed-latest-gpu`, which can be found at
# `docker/transformers-pytorch-deepspeed-latest-... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./.github/ISSUE_TEMPLATE/config.yml | blank_issues_enabled: true
version: 2.1
contact_links:
- name: Model checkpoints on the Hugging Face Hub
url: https://huggingface.co/models
about: Open a Pull request / Discussion related to a specific model checkpoint directly on the Hugging Face Hub
- name: Website Related
url: https://github.com/hugg... | blank_issues_enabled: true
version: 2.1
contact_links:
- name: Model checkpoints on the Hugging Face Hub
url: https://huggingface.co/models
about: Open a Pull request / Discussion related to a specific model checkpoint directly on the Hugging Face Hub
- name: Website Related
url: https://github.com/hugg... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/poolformer/feature_extraction_poolformer.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/convert_slow_tokenizers_checkpoints_to_fast.py | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./examples/research_projects/codeparrot/scripts/arguments.py | from dataclasses import dataclass, field
from typing import Optional
@dataclass
class TrainingArguments:
"""
Configuration for training model.
"""
model_ckpt: Optional[str] = field(
default="codeparrot/codeparrot", metadata={"help": "Model name or path of model to be trained."}
)
save... | from dataclasses import dataclass, field
from typing import Optional
@dataclass
class TrainingArguments:
"""
Configuration for training model.
"""
model_ckpt: Optional[str] = field(
default="codeparrot/codeparrot", metadata={"help": "Model name or path of model to be trained."}
)
save... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/de/pipeline_tutorial.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./templates/adding_a_missing_tokenization_test/cookiecutter.json | {
"modelname": "BrandNewBERT",
"uppercase_modelname": "BRAND_NEW_BERT",
"lowercase_modelname": "brand_new_bert",
"camelcase_modelname": "BrandNewBert",
"has_slow_class": ["True", "False"],
"has_fast_class": ["True", "False"],
"slow_tokenizer_use_sentencepiece": ["True", "False"],
"authors": "The Hugging... | {
"modelname": "BrandNewBERT",
"uppercase_modelname": "BRAND_NEW_BERT",
"lowercase_modelname": "brand_new_bert",
"camelcase_modelname": "BrandNewBert",
"has_slow_class": ["True", "False"],
"has_fast_class": ["True", "False"],
"slow_tokenizer_use_sentencepiece": ["True", "False"],
"authors": "The Hugging... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/bert_japanese/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2020 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/big_bird/tokenization_big_bird_fast.py | # coding=utf-8
# Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | # coding=utf-8
# Copyright 2018 Google AI, Google Brain and the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# U... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/generation/tf_utils.py | # coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | # coding=utf-8
# Copyright 2018 The Google AI Language Team Authors and The HuggingFace Inc. team.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a cop... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/big_bird/tokenization_big_bird.py | # coding=utf-8
# Copyright 2021 Google Research and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICE... | # coding=utf-8
# Copyright 2021 Google Research and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICE... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/mpnet/configuration_mpnet.py | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License ... | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team, Microsoft Corporation.
# Copyright (c) 2018, NVIDIA CORPORATION. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License ... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./src/transformers/models/wavlm/configuration_wavlm.py | # coding=utf-8
# Copyright 2021 The Fairseq Authors, Microsoft Research, and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www... | # coding=utf-8
# Copyright 2021 The Fairseq Authors, Microsoft Research, and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./docs/source/en/perf_train_tpu.mdx | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | <!--Copyright 2022 The HuggingFace Team. All rights reserved.
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with
the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed... | -1 |
huggingface/transformers | 20,340 | [FLAX] Add dtype to embedding for bert/bart/opt/t5 | ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @LysandreJik, @sanchit-gandhi | merrymercy | 2022-11-21T00:53:45Z | 2022-11-28T15:21:43Z | 667ccea72235504ab7876024e4f8c113ca62190f | ac2f6674a33e8eaffdf868e1fa6cbc8e722f469e | [FLAX] Add dtype to embedding for bert/bart/opt/t5. ## What does this PR do?
This PR is the follow-up of #18462. It adds dtype to `nn.Embed` for more common Flax models, including bert, bart, opt, t5, and their copies.
This dtype is necessary for mixed precision training.
## Who can review?
@patrickvonplaten, @... | ./tests/models/layoutlmv2/test_modeling_layoutlmv2.py | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2021 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | -1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/dinat/configuration_dinat.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | 1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/dinat/modeling_dinat.py | # coding=utf-8
# Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | # coding=utf-8
# Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | 1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/nat/__init__.py | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | # flake8: noqa
# There's no way to ignore "F401 '...' imported but unused" warnings in this
# module, but to preserve other warnings. So, don't check this module at all.
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use thi... | 1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/nat/configuration_nat.py | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | # coding=utf-8
# Copyright 2022 The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless r... | 1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/nat/modeling_nat.py | # coding=utf-8
# Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | # coding=utf-8
# Copyright 2022 SHI Labs and The HuggingFace Inc. team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0... | 1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./examples/legacy/pytorch-lightning/lightning_base.py | import argparse
import logging
import os
from pathlib import Path
from typing import Any, Dict
import pytorch_lightning as pl
from pytorch_lightning.utilities import rank_zero_info
from transformers import (
AdamW,
AutoConfig,
AutoModel,
AutoModelForPreTraining,
AutoModelForQuestionAnswering,
... | import argparse
import logging
import os
from pathlib import Path
from typing import Any, Dict
import pytorch_lightning as pl
from pytorch_lightning.utilities import rank_zero_info
from transformers import (
AdamW,
AutoConfig,
AutoModel,
AutoModelForPreTraining,
AutoModelForQuestionAnswering,
... | -1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./src/transformers/models/roberta/convert_roberta_original_pytorch_checkpoint_to_pytorch.py | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | # coding=utf-8
# Copyright 2018 The HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable... | -1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./examples/research_projects/distillation/scripts/extract_distilbert.py | # coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | # coding=utf-8
# Copyright 2019-present, the HuggingFace Inc. team.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by a... | -1 |
huggingface/transformers | 20,325 | Add LayerScale to NAT/DiNAT |
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] This PR fixes a typo or improv... | alihassanijr | 2022-11-18T22:01:53Z | 2022-11-21T14:08:35Z | d28448c5cd8fa8dfb64190c7f55275d80e256a9e | 11f3ec7224c83c9e5c379a774b9d3984e68d26fa | Add LayerScale to NAT/DiNAT.
# What does this PR do?
This follows PR #20219 .
I completely dropped the ball on LayerScale in the original PR.
This is just an optional argument in both models, and is only activated for larger variants in order to provide training stability.
## Before submitting
- [ ] T... | ./tests/models/data2vec/test_modeling_data2vec_audio.py | # coding=utf-8
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | # coding=utf-8
# Copyright 2022 The HuggingFace Team. All rights reserved.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless requir... | -1 |