python3Packages.transformers: 3.3.1 -> 3.4.0
Changelog: https://github.com/huggingface/transformers/releases/tag/v3.4.0
This commit is contained in:
parent
d2e918cc12
commit
05f6de94cf
@ -7,6 +7,7 @@
|
|||||||
, requests
|
, requests
|
||||||
, numpy
|
, numpy
|
||||||
, parameterized
|
, parameterized
|
||||||
|
, protobuf
|
||||||
, sacremoses
|
, sacremoses
|
||||||
, sentencepiece
|
, sentencepiece
|
||||||
, timeout-decorator
|
, timeout-decorator
|
||||||
@ -17,19 +18,19 @@
|
|||||||
|
|
||||||
buildPythonPackage rec {
|
buildPythonPackage rec {
|
||||||
pname = "transformers";
|
pname = "transformers";
|
||||||
version = "3.3.1";
|
version = "3.4.0";
|
||||||
|
|
||||||
src = fetchFromGitHub {
|
src = fetchFromGitHub {
|
||||||
owner = "huggingface";
|
owner = "huggingface";
|
||||||
repo = pname;
|
repo = pname;
|
||||||
rev = "v${version}";
|
rev = "v${version}";
|
||||||
sha256 = "1j9nzhl0zw5z9rnvzfih7v6bax353rxp05b3f0cvkii3b5dbkc2j";
|
sha256 = "1v09gryxsg57d2cjwagna1535m8mbxlazdbhsww210lxa818m5qj";
|
||||||
};
|
};
|
||||||
|
|
||||||
propagatedBuildInputs = [
|
propagatedBuildInputs = [
|
||||||
boto3
|
|
||||||
filelock
|
filelock
|
||||||
numpy
|
numpy
|
||||||
|
protobuf
|
||||||
regex
|
regex
|
||||||
requests
|
requests
|
||||||
sacremoses
|
sacremoses
|
||||||
@ -44,11 +45,6 @@ buildPythonPackage rec {
|
|||||||
timeout-decorator
|
timeout-decorator
|
||||||
];
|
];
|
||||||
|
|
||||||
postPatch = ''
|
|
||||||
substituteInPlace setup.py \
|
|
||||||
--replace "tokenizers == 0.8.1.rc2" "tokenizers>=0.8"
|
|
||||||
'';
|
|
||||||
|
|
||||||
preCheck = ''
|
preCheck = ''
|
||||||
export HOME="$TMPDIR"
|
export HOME="$TMPDIR"
|
||||||
|
|
||||||
@ -67,8 +63,10 @@ buildPythonPackage rec {
|
|||||||
|
|
||||||
# Disable tests that require network access.
|
# Disable tests that require network access.
|
||||||
disabledTests = [
|
disabledTests = [
|
||||||
"PegasusTokenizationTest"
|
"BlenderbotSmallTokenizerTest"
|
||||||
"T5TokenizationTest"
|
"Blenderbot3BTokenizerTests"
|
||||||
|
"TokenizationTest"
|
||||||
|
"TestTokenizationBart"
|
||||||
"test_all_tokenizers"
|
"test_all_tokenizers"
|
||||||
"test_batch_encoding_is_fast"
|
"test_batch_encoding_is_fast"
|
||||||
"test_batch_encoding_pickle"
|
"test_batch_encoding_pickle"
|
||||||
|
Loading…
Reference in New Issue
Block a user