Skip to content

Commit 38ad38d

Browse files
committed
WIP
1 parent 9043ce8 commit 38ad38d

File tree

1 file changed

+36
-36
lines changed

1 file changed

+36
-36
lines changed

.github/workflows/ci.yml

+36-36
Original file line numberDiff line numberDiff line change
@@ -182,48 +182,48 @@ jobs:
182182
cd src/ci/citool
183183
cargo build
184184
185-
- name: run the build
186-
# Redirect stderr to stdout to avoid reordering the two streams in the GHA logs.
187-
run: src/ci/scripts/run-build-from-ci.sh 2>&1
188-
env:
189-
AWS_ACCESS_KEY_ID: ${{ env.CACHES_AWS_ACCESS_KEY_ID }}
190-
AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.CACHES_AWS_ACCESS_KEY_ID)] }}
191-
192-
- name: create github artifacts
193-
run: src/ci/scripts/create-doc-artifacts.sh
194-
195-
- name: print disk usage
196-
run: |
197-
echo "disk usage:"
198-
df -h
199-
200-
- name: upload artifacts to github
201-
uses: actions/upload-artifact@v4
202-
with:
203-
# name is set in previous step
204-
name: ${{ env.DOC_ARTIFACT_NAME }}
205-
path: obj/artifacts/doc
206-
if-no-files-found: ignore
207-
retention-days: 5
208-
209-
- name: upload artifacts to S3
210-
run: src/ci/scripts/upload-artifacts.sh
211-
env:
212-
AWS_ACCESS_KEY_ID: ${{ env.ARTIFACTS_AWS_ACCESS_KEY_ID }}
213-
AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.ARTIFACTS_AWS_ACCESS_KEY_ID)] }}
214-
# Adding a condition on DEPLOY=1 or DEPLOY_ALT=1 is not needed as all deploy
215-
# builders *should* have the AWS credentials available. Still, explicitly
216-
# adding the condition is helpful as this way CI will not silently skip
217-
# deploying artifacts from a dist builder if the variables are misconfigured,
218-
# erroring about invalid credentials instead.
219-
if: github.event_name == 'push' || env.DEPLOY == '1' || env.DEPLOY_ALT == '1'
185+
# - name: run the build
186+
# # Redirect stderr to stdout to avoid reordering the two streams in the GHA logs.
187+
# run: src/ci/scripts/run-build-from-ci.sh 2>&1
188+
# env:
189+
# AWS_ACCESS_KEY_ID: ${{ env.CACHES_AWS_ACCESS_KEY_ID }}
190+
# AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.CACHES_AWS_ACCESS_KEY_ID)] }}
191+
#
192+
# - name: create github artifacts
193+
# run: src/ci/scripts/create-doc-artifacts.sh
194+
#
195+
# - name: print disk usage
196+
# run: |
197+
# echo "disk usage:"
198+
# df -h
199+
#
200+
# - name: upload artifacts to github
201+
# uses: actions/upload-artifact@v4
202+
# with:
203+
# # name is set in previous step
204+
# name: ${{ env.DOC_ARTIFACT_NAME }}
205+
# path: obj/artifacts/doc
206+
# if-no-files-found: ignore
207+
# retention-days: 5
208+
#
209+
# - name: upload artifacts to S3
210+
# run: src/ci/scripts/upload-artifacts.sh
211+
# env:
212+
# AWS_ACCESS_KEY_ID: ${{ env.ARTIFACTS_AWS_ACCESS_KEY_ID }}
213+
# AWS_SECRET_ACCESS_KEY: ${{ secrets[format('AWS_SECRET_ACCESS_KEY_{0}', env.ARTIFACTS_AWS_ACCESS_KEY_ID)] }}
214+
# # Adding a condition on DEPLOY=1 or DEPLOY_ALT=1 is not needed as all deploy
215+
# # builders *should* have the AWS credentials available. Still, explicitly
216+
# # adding the condition is helpful as this way CI will not silently skip
217+
# # deploying artifacts from a dist builder if the variables are misconfigured,
218+
# # erroring about invalid credentials instead.
219+
# if: github.event_name == 'push' || env.DEPLOY == '1' || env.DEPLOY_ALT == '1'
220220

221221
- name: upload job metrics to DataDog
222222
if: needs.calculate_matrix.outputs.run_type != 'pr'
223223
env:
224224
DATADOG_API_KEY: ${{ secrets.DATADOG_API_KEY }}
225225
DD_GITHUB_JOB_NAME: ${{ matrix.full_name }}
226-
run: ./src/ci/citool/target/debug/citool upload-build-metrics ../../build/cpu-usage.csv
226+
run: ./src/ci/citool/target/debug/citool upload-build-metrics build/cpu-usage.csv
227227

228228
# This job isused to tell bors the final status of the build, as there is no practical way to detect
229229
# when a workflow is successful listening to webhooks only in our current bors implementation (homu).

0 commit comments

Comments
 (0)