@@ -182,48 +182,48 @@ jobs:
182
182
cd src/ci/citool
183
183
cargo build
184
184
185
- - name : run the build
186
- # Redirect stderr to stdout to avoid reordering the two streams in the GHA logs.
187
- run : src/ci/scripts/run-build-from-ci.sh 2>&1
188
- env :
189
- AWS_ACCESS_KEY_ID : ${{ env.CACHES_AWS_ACCESS_KEY_ID }}
190
- AWS_SECRET_ACCESS_KEY : ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.CACHES_AWS_ACCESS_KEY_ID)] }}
191
-
192
- - name : create github artifacts
193
- run : src/ci/scripts/create-doc-artifacts.sh
194
-
195
- - name : print disk usage
196
- run : |
197
- echo "disk usage:"
198
- df -h
199
-
200
- - name : upload artifacts to github
201
- uses : actions/upload-artifact@v4
202
- with :
203
- # name is set in previous step
204
- name : ${{ env.DOC_ARTIFACT_NAME }}
205
- path : obj/artifacts/doc
206
- if-no-files-found : ignore
207
- retention-days : 5
208
-
209
- - name : upload artifacts to S3
210
- run : src/ci/scripts/upload-artifacts.sh
211
- env :
212
- AWS_ACCESS_KEY_ID : ${{ env.ARTIFACTS_AWS_ACCESS_KEY_ID }}
213
- AWS_SECRET_ACCESS_KEY : ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.ARTIFACTS_AWS_ACCESS_KEY_ID)] }}
214
- # Adding a condition on DEPLOY=1 or DEPLOY_ALT=1 is not needed as all deploy
215
- # builders *should* have the AWS credentials available. Still, explicitly
216
- # adding the condition is helpful as this way CI will not silently skip
217
- # deploying artifacts from a dist builder if the variables are misconfigured,
218
- # erroring about invalid credentials instead.
219
- if : github.event_name == 'push' || env.DEPLOY == '1' || env.DEPLOY_ALT == '1'
185
+ # - name: run the build
186
+ # # Redirect stderr to stdout to avoid reordering the two streams in the GHA logs.
187
+ # run: src/ci/scripts/run-build-from-ci.sh 2>&1
188
+ # env:
189
+ # AWS_ACCESS_KEY_ID: ${{ env.CACHES_AWS_ACCESS_KEY_ID }}
190
+ # AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.CACHES_AWS_ACCESS_KEY_ID)] }}
191
+ #
192
+ # - name: create github artifacts
193
+ # run: src/ci/scripts/create-doc-artifacts.sh
194
+ #
195
+ # - name: print disk usage
196
+ # run: |
197
+ # echo "disk usage:"
198
+ # df -h
199
+ #
200
+ # - name: upload artifacts to github
201
+ # uses: actions/upload-artifact@v4
202
+ # with:
203
+ # # name is set in previous step
204
+ # name: ${{ env.DOC_ARTIFACT_NAME }}
205
+ # path: obj/artifacts/doc
206
+ # if-no-files-found: ignore
207
+ # retention-days: 5
208
+ #
209
+ # - name: upload artifacts to S3
210
+ # run: src/ci/scripts/upload-artifacts.sh
211
+ # env:
212
+ # AWS_ACCESS_KEY_ID: ${{ env.ARTIFACTS_AWS_ACCESS_KEY_ID }}
213
+ # AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.ARTIFACTS_AWS_ACCESS_KEY_ID)] }}
214
+ # # Adding a condition on DEPLOY=1 or DEPLOY_ALT=1 is not needed as all deploy
215
+ # # builders *should* have the AWS credentials available. Still, explicitly
216
+ # # adding the condition is helpful as this way CI will not silently skip
217
+ # # deploying artifacts from a dist builder if the variables are misconfigured,
218
+ # # erroring about invalid credentials instead.
219
+ # if: github.event_name == 'push' || env.DEPLOY == '1' || env.DEPLOY_ALT == '1'
220
220
221
221
- name : upload job metrics to DataDog
222
222
if : needs.calculate_matrix.outputs.run_type != 'pr'
223
223
env :
224
224
DATADOG_API_KEY : ${{ secrets.DATADOG_API_KEY }}
225
225
DD_GITHUB_JOB_NAME : ${{ matrix.full_name }}
226
- run : ./src/ci/citool/target/debug/citool upload-build-metrics ../../ build/cpu-usage.csv
226
+ run : ./src/ci/citool/target/debug/citool upload-build-metrics build/cpu-usage.csv
227
227
228
228
# This job isused to tell bors the final status of the build, as there is no practical way to detect
229
229
# when a workflow is successful listening to webhooks only in our current bors implementation (homu).
0 commit comments