@@ -17,47 +17,6 @@ gpu: &gpu
17
17
# -------------------------------------------------------------------------------------
18
18
cache_key : &cache_key cache-key-{{ .Environment.CIRCLE_JOB }}-{{ checksum ".circleci/config.yml" }}-{{ checksum "setup.py"}}
19
19
20
- install_dep_common : &install_dep_common
21
- - run :
22
- name : Install Common Dependencies
23
- command : |
24
- source activate fairseq
25
- pip install --upgrade setuptools
26
- pip install bitarray boto3 deepspeed editdistance fastBPE iopath ipdb ipython pyarrow pytest sacremoses sentencepiece subword-nmt hydra-core==1.0.7 omegaconf==2.0.6
27
- pip install --progress-bar off pytest
28
- pip install --progress-bar off fairscale
29
- pip install -i https://test.pypi.org/simple/ bitsandbytes-cuda111 -U
30
- python -c 'import torch; print("Torch version:", torch.__version__)'
31
- python -m torch.utils.collect_env
32
-
33
- install_dep_fused_ops : &install_dep_fused_ops
34
- # this version of Apex is from Feb 2021 and doesn't work with torch>=1.12
35
- - run :
36
- name : Install Megatron/Apex Dependencies
37
- working_directory : ~/
38
- command : |
39
- source activate fairseq
40
- git clone https://github.com/NVIDIA/apex
41
- cd apex
42
- git checkout e2083df5eb96643c61613b9df48dd4eea6b07690
43
- sed -i '101,107 s/^/#/' setup.py
44
- pip install -v --no-cache-dir --global-option="--cpp_ext" --global-option="--cuda_ext" --global-option="--deprecated_fused_adam" --global-option="--xentropy" --global-option="--fast_multihead_attn" ./
45
- cd ~/
46
- git clone --depth=1 --branch v2.4 https://github.com/NVIDIA/Megatron-LM.git
47
- cd Megatron-LM
48
- pip install -e .
49
-
50
- install_dep_xformers : &install_dep_xformers
51
- - run :
52
- name : Install xFormers Dependencies
53
- working_directory : ~/
54
- command : |
55
- source activate fairseq
56
- git clone https://github.com/facebookresearch/xformers.git
57
- cd xformers
58
- pip install -r requirements.txt
59
- pip install -e .
60
-
61
20
install_dep_pt1_10 : &install_dep_pt1_10
62
21
- run :
63
22
name : Install Pytorch Dependencies
@@ -81,8 +40,9 @@ install_repo: &install_repo
81
40
name : Install Repository
82
41
command : |
83
42
source activate fairseq
84
- pip install .
85
- python setup.py build_ext --inplace
43
+ python -m pip install fairscale
44
+ python -m pip install -e '.[dev,docs]'
45
+ python -c 'import torch; print("Torch version:", torch.__version__)'
86
46
87
47
run_unittests : &run_unittests
88
48
- run :
134
94
- restore_cache :
135
95
key : *cache_key
136
96
- << : *install_dep_pt1_10
137
- - << : *install_dep_common
138
- - << : *install_dep_fused_ops
139
97
- save_cache :
140
98
paths :
141
99
- ~/miniconda/
@@ -155,8 +113,6 @@ jobs:
155
113
- restore_cache :
156
114
key : *cache_key
157
115
- << : *install_dep_pt1_12
158
- - << : *install_dep_common
159
- - << : *install_dep_fused_ops
160
116
- save_cache :
161
117
paths :
162
118
- ~/miniconda/
@@ -168,6 +124,5 @@ workflows:
168
124
version : 2
169
125
build :
170
126
jobs :
171
- # TODO: Figure out how to run APEX on torch 1.12
172
- # - gpu_tests_pt1_12
127
+ - gpu_tests_pt1_12
173
128
- gpu_tests_pt1_10
0 commit comments