## Environment info
<!-- You can run the command `transformers-cli env` and copy-and-paste its output below.
Don't forget to fill out the missing fields in that output! -->
- `transformers` version: 4.6.0.dev0
- Platform: Linux-4.15.0-140-generic-x86_64-with-debian-buster-sid
- Python version: 3.7.9
- PyTorch version (GPU?): 1.8.1 (True)
- Tensorflow version (GPU?): not installed (NA)
- Using GPU in script?: Yes
- Using distributed or parallel set-up in script?: Yes
### Who can help
<!-- Your issue will be replied to more quickly if you can figure out the right person to tag with @
If you know how to use git blame, that is the easiest way, otherwise, here is a rough guide of **who to tag**.
Please tag fewer than 3 people.
Models:
- albert, bert, xlm: @LysandreJik
- blenderbot, bart, marian, pegasus, encoderdecoder, t5: @patrickvonplaten, @patil-suraj
- longformer, reformer, transfoxl, xlnet: @patrickvonplaten
- fsmt: @stas00
- funnel: @sgugger
- gpt2: @patrickvonplaten, @LysandreJik
- rag: @patrickvonplaten, @lhoestq
- tensorflow: @Rocketknight1
Library:
- benchmarks: @patrickvonplaten
- deepspeed: @stas00
- ray/raytune: @richardliaw, @amogkam
- text generation: @patrickvonplaten
- tokenizers: @LysandreJik
- trainer: @sgugger
- pipelines: @LysandreJik
Documentation: @sgugger
Model hub:
- for issues with a model report at https://discuss.huggingface.co/ and tag the model's creator.
HF projects:
- datasets: [different repo](https://github.com/huggingface/datasets)
- rust tokenizers: [different repo](https://github.com/huggingface/tokenizers)
Examples:
- maintained examples (not research project or legacy): @sgugger, @patil-suraj
- research_projects/bert-loses-patience: @JetRunner
- research_projects/distillation: @VictorSanh
-->
## Information
Model I am using (Bert, XLNet ...):
The problem arises when using:
* [ ] the official example scripts: (give details below)
* [ ] my own modified scripts: (give details below)
The tasks I am working on is:
* [ ] an official GLUE/SQUaD task: (give the name)
* [ ] my own task or dataset: (give details below)
## To reproduce
Steps to reproduce the behavior:
1.
2.
3.
<!-- If you have code snippets, error messages, stack traces please provide them here as well.
Important! Use code tags to correctly format your code. See https://help.github.com/en/github/writing-on-github/creating-and-highlighting-code-blocks#syntax-highlighting
Do not use screenshots, as they are hard to read and (more importantly) don't allow others to copy-and-paste your code.-->
Error message:
```shell=
Using /home/tomihira/.cache/torch_extensions as PyTorch extensions root...
Traceback (most recent call last):
File "finetuning.py", line 329, in <module>
main(args)
File "finetuning.py", line 298, in main
train_result = trainer.train(resume_from_checkpoint=checkpoint)
File "/home/tomihira/workspace/transformers/src/transformers/trainer.py", line 1067, in train
self, num_training_steps=max_steps, resume_from_checkpoint=resume_from_checkpoint
File "/home/tomihira/workspace/transformers/src/transformers/integrations.py", line 519, in deepspeed_init
lr_scheduler=lr_scheduler,
File "/home/tomihira/workspace/DeepSpeed/deepspeed/__init__.py", line 130, in initialize
config_params=config_params)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/runtime/engine.py", line 198, in __init__
util_ops = UtilsBuilder().load()
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 215, in load
return self.jit_load(verbose)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 252, in jit_load
verbose=verbose)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1091, in load
keep_intermediates=keep_intermediates)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1302, in _jit_compile
is_standalone=is_standalone)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1373, in _write_ninja_file_and_build_library
verify_ninja_availability()
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1429, in verify_ninja_availability
raise RuntimeError("Ninja is required to load C++ extensions")
RuntimeError: Ninja is required to load C++ extensions
Using /home/tomihira/.cache/torch_extensions as PyTorch extensions root...
Traceback (most recent call last):
File "finetuning.py", line 329, in <module>
main(args)
File "finetuning.py", line 298, in main
train_result = trainer.train(resume_from_checkpoint=checkpoint)
File "/home/tomihira/workspace/transformers/src/transformers/trainer.py", line 1067, in train
self, num_training_steps=max_steps, resume_from_checkpoint=resume_from_checkpoint
File "/home/tomihira/workspace/transformers/src/transformers/integrations.py", line 519, in deepspeed_init
lr_scheduler=lr_scheduler,
File "/home/tomihira/workspace/DeepSpeed/deepspeed/__init__.py", line 130, in initialize
config_params=config_params)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/runtime/engine.py", line 198, in __init__
util_ops = UtilsBuilder().load()
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 215, in load
return self.jit_load(verbose)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 252, in jit_load
verbose=verbose)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1091, in load
keep_intermediates=keep_intermediates)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1302, in _jit_compile
is_standalone=is_standalone)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1373, in _write_ninja_file_and_build_library
verify_ninja_availability()
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1429, in verify_ninja_availability
raise RuntimeError("Ninja is required to load C++ extensions")
RuntimeError: Ninja is required to load C++ extensions
Loading extension module utils...
Time to load utils op: 0.13512039184570312 seconds
Using /home/tomihira/.cache/torch_extensions as PyTorch extensions root...
Traceback (most recent call last):
File "finetuning.py", line 329, in <module>
main(args)
File "finetuning.py", line 298, in main
train_result = trainer.train(resume_from_checkpoint=checkpoint)
File "/home/tomihira/workspace/transformers/src/transformers/trainer.py", line 1067, in train
self, num_training_steps=max_steps, resume_from_checkpoint=resume_from_checkpoint
File "/home/tomihira/workspace/transformers/src/transformers/integrations.py", line 519, in deepspeed_init
lr_scheduler=lr_scheduler,
File "/home/tomihira/workspace/DeepSpeed/deepspeed/__init__.py", line 130, in initialize
config_params=config_params)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/runtime/engine.py", line 198, in __init__
util_ops = UtilsBuilder().load()
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 215, in load
return self.jit_load(verbose)
File "/home/tomihira/workspace/DeepSpeed/deepspeed/ops/op_builder/builder.py", line 252, in jit_load
verbose=verbose)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1091, in load
keep_intermediates=keep_intermediates)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1302, in _jit_compile
is_standalone=is_standalone)
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1373, in _write_ninja_file_and_build_library
verify_ninja_availability()
File "/home/tomihira/.conda/envs/hf/lib/python3.7/site-packages/torch/utils/cpp_extension.py", line 1429, in verify_ninja_availability
raise RuntimeError("Ninja is required to load C++ extensions")
RuntimeError: Ninja is required to load C++ extensions
```
## Expected behavior
<!-- A clear and concise description of what you would expect to happen. -->