SDK Samples Run generated from 4402058 #1452
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
# This code is autogenerated. | |
# Code is generated by running custom script: python3 readme.py | |
# Any manual changes to this file may cause incorrect behavior. | |
# Any manual changes will be overwritten if the code is regenerated. | |
name: sdk-jobs-pipelines-1i_pipeline_with_spark_nodes-pipeline_with_spark_nodes | |
# This file is created by sdk/python/readme.py. | |
# Please do not edit directly. | |
on: | |
workflow_dispatch: | |
schedule: | |
- cron: "0 10/12 * * *" | |
pull_request: | |
branches: | |
- main | |
paths: | |
- sdk/python/jobs/pipelines/1i_pipeline_with_spark_nodes/** | |
- .github/workflows/sdk-jobs-pipelines-1i_pipeline_with_spark_nodes-pipeline_with_spark_nodes.yml | |
- sdk/python/dev-requirements.txt | |
- infra/bootstrapping/** | |
- sdk/python/setup.sh | |
permissions: | |
id-token: write | |
concurrency: | |
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }} | |
cancel-in-progress: true | |
jobs: | |
build: | |
runs-on: ubuntu-latest | |
steps: | |
- name: check out repo | |
uses: actions/checkout@v2 | |
- name: setup python | |
uses: actions/setup-python@v2 | |
with: | |
python-version: "3.10" | |
- name: pip install notebook reqs | |
run: pip install -r sdk/python/dev-requirements.txt | |
- name: azure login | |
uses: azure/login@v1 | |
with: | |
client-id: ${{ secrets.OIDC_AZURE_CLIENT_ID }} | |
tenant-id: ${{ secrets.OIDC_AZURE_TENANT_ID }} | |
subscription-id: ${{ secrets.OIDC_AZURE_SUBSCRIPTION_ID }} | |
- name: bootstrap resources | |
run: | | |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}'; | |
bash bootstrap.sh | |
working-directory: infra/bootstrapping | |
continue-on-error: false | |
- name: setup SDK | |
run: | | |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh"; | |
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh"; | |
bash setup.sh | |
working-directory: sdk/python | |
continue-on-error: true | |
- name: validate readme | |
run: | | |
python check-readme.py "${{ github.workspace }}/sdk/python/jobs/pipelines/1i_pipeline_with_spark_nodes" | |
working-directory: infra/bootstrapping | |
continue-on-error: false | |
- name: setup-cli | |
run: | | |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh"; | |
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh"; | |
bash setup.sh | |
working-directory: cli | |
continue-on-error: true | |
- name: Eagerly cache access tokens for required scopes | |
run: | | |
# Workaround for azure-cli's lack of support for ID token refresh | |
# Taken from: https://github.com/Azure/login/issues/372#issuecomment-2056289617 | |
# Management | |
az account get-access-token --scope https://management.azure.com/.default --output none | |
# ML | |
az account get-access-token --scope https://ml.azure.com/.default --output none | |
- name: setup spark resources | |
run: | | |
bash -x jobs/spark/setup_spark.sh jobs/spark/ jobs/pipelines/1i_pipeline_with_spark_nodes/pipeline_with_spark_nodes.ipynb | |
working-directory: sdk/python | |
continue-on-error: true | |
- name: run jobs/pipelines/1i_pipeline_with_spark_nodes/pipeline_with_spark_nodes.ipynb | |
run: | | |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh"; | |
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh"; | |
bash "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh" generate_workspace_config "../../.azureml/config.json"; | |
bash "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh" replace_template_values "pipeline_with_spark_nodes.ipynb"; | |
[ -f "../../.azureml/config" ] && cat "../../.azureml/config"; | |
papermill -k python pipeline_with_spark_nodes.ipynb pipeline_with_spark_nodes.output.ipynb | |
working-directory: sdk/python/jobs/pipelines/1i_pipeline_with_spark_nodes | |
- name: upload notebook's working folder as an artifact | |
if: ${{ always() }} | |
uses: ./.github/actions/upload-artifact | |
with: | |
name: pipeline_with_spark_nodes | |
path: sdk/python/jobs/pipelines/1i_pipeline_with_spark_nodes |