Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Automation test for spark CLI samples #2377

Merged
merged 42 commits into from
Jul 6, 2023
Merged
Show file tree
Hide file tree
Changes from 36 commits
Commits
Show all changes
42 commits
Select commit Hold shift + click to select a range
e4838c9
Enable test for submit_spark_standalone_jobs
fredms2 Jun 13, 2023
3d92c07
Generate workflow yaml
fredms2 Jun 13, 2023
f41b50c
update spark job files for automation test
fredms2 Jun 14, 2023
3099208
Add workflow for serverless spark with user identity job
fredms2 Jun 15, 2023
e20a24b
Add scripts to upload input data
fredms2 Jun 15, 2023
04d78f3
Update workflow to refer the script
fredms2 Jun 15, 2023
d452e9d
Update source file path
fredms2 Jun 15, 2023
0956274
Update workflow with correct file path
fredms2 Jun 15, 2023
f56be56
Update working directory
fredms2 Jun 15, 2023
f5f3975
Update workflow
fredms2 Jun 15, 2023
70f6deb
Update the path
fredms2 Jun 16, 2023
ec30c16
Update the script to upload data
fredms2 Jun 16, 2023
ec9da6e
Update the overwrite mode
fredms2 Jun 16, 2023
f08200b
Update destination blob name
fredms2 Jun 16, 2023
76fb70d
Use blob upload batch
fredms2 Jun 16, 2023
4caf851
Add spark pipeline tests
fredms2 Jun 16, 2023
b5682d7
Update spark component extension
fredms2 Jun 16, 2023
70e86c1
Add script to attache uai
fredms2 Jun 19, 2023
f3df676
Update property name in workflow
fredms2 Jun 19, 2023
ea7b996
Update script parameters
fredms2 Jun 19, 2023
7425097
Update assign uai script
fredms2 Jun 19, 2023
75e217f
Format the script
fredms2 Jun 19, 2023
7e9cd2a
Update setup identities script
fredms2 Jun 19, 2023
b9c11ad
Update path to infra bootstraping
fredms2 Jun 19, 2023
1b3f848
Enable automation test for attached spark job
fredms2 Jun 20, 2023
319423e
Update resource path
fredms2 Jun 20, 2023
0e0b4ec
Update setup attached resource script
fredms2 Jun 20, 2023
ddad2f6
Update script of setup resources
fredms2 Jun 20, 2023
7ee07f7
Update setup attached resource script2
fredms2 Jun 21, 2023
5bce143
Add logic to assign identity role
fredms2 Jun 21, 2023
0fb9cbb
Format the empty check
fredms2 Jun 21, 2023
5b06d82
Check if identity is empty
fredms2 Jun 21, 2023
1b57095
Update to get compute properties
fredms2 Jun 21, 2023
ac35f70
Merge branch 'main' into feli1/spark-automation
fredms2 Jun 21, 2023
a2c5e6e
update readme
fredms2 Jun 21, 2023
9464d9b
Reformat the script
fredms2 Jun 21, 2023
805e744
Update schema location and revert sdk notebook changes
fredms2 Jun 22, 2023
9f50ca4
Attach pool first
fredms2 Jun 22, 2023
8b125a8
Merge branch 'main' into feli1/spark-automation
fredms2 Jul 6, 2023
bd06bf7
Rename resources and merge main
fredms2 Jul 6, 2023
279a9ee
Update format in yml
fredms2 Jul 6, 2023
01e1d38
Add role assigment to uid
fredms2 Jul 6, 2023
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-pipeline-default-identity
on:
workflow_dispatch:
schedule:
- cron: "30 9/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-pipeline-default-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-pipeline-default-identity.yml
working-directory: cli/jobs/spark
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-pipeline-managed-identity
on:
workflow_dispatch:
schedule:
- cron: "43 7/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-pipeline-managed-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup identities
run: |
bash -x setup-identities.sh
working-directory: cli/jobs/spark
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark-system-identity.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-pipeline-managed-identity.yml
working-directory: cli/jobs/spark
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-pipeline-user-identity
on:
workflow_dispatch:
schedule:
- cron: "15 4/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-pipeline-user-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark-user-identity.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-pipeline-user-identity.yml
working-directory: cli/jobs/spark
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-standalone-default-identity
on:
workflow_dispatch:
schedule:
- cron: "15 0/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-standalone-default-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-standalone-default-identity.yml
working-directory: cli/jobs/spark
Original file line number Diff line number Diff line change
@@ -0,0 +1,66 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-standalone-managed-identity
on:
workflow_dispatch:
schedule:
- cron: "16 1/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-standalone-managed-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup identities
run: |
bash -x setup-identities.sh
working-directory: cli/jobs/spark
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark-system-identity.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-standalone-managed-identity.yml
working-directory: cli/jobs/spark
Original file line number Diff line number Diff line change
@@ -0,0 +1,61 @@
# This code is autogenerated.
# Code is generated by running custom script: python3 readme.py
# Any manual changes to this file may cause incorrect behavior.
# Any manual changes will be overwritten if the code is regenerated.

name: cli-jobs-spark-attached-spark-standalone-user-identity
on:
workflow_dispatch:
schedule:
- cron: "7 1/12 * * *"
pull_request:
branches:
- main
paths:
- cli/jobs/spark/**
- infra/bootstrapping/**
- .github/workflows/cli-jobs-spark-attached-spark-standalone-user-identity.yml
- cli/jobs/spark/data/titanic.csv
- cli/setup.sh
concurrency:
group: ${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}
cancel-in-progress: true
jobs:
build:
runs-on: ubuntu-latest
steps:
- name: check out repo
uses: actions/checkout@v2
- name: azure login
uses: azure/login@v1
with:
creds: ${{secrets.AZUREML_CREDENTIALS}}
- name: bootstrap resources
run: |
echo '${{ github.workflow }}-${{ github.event.pull_request.number || github.ref }}';
bash bootstrap.sh
working-directory: infra/bootstrapping
continue-on-error: false
- name: setup-cli
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash setup.sh
working-directory: cli
continue-on-error: true
- name: upload data
run: |
bash -x upload-data-to-blob.sh jobs/spark/
working-directory: cli
continue-on-error: true
- name: setup attached spark
working-directory: cli
continue-on-error: true
run: |
bash -x jobs/spark/setup-attached-resources.sh resources/compute/attached-spark-user-identity.yml
- name: run job
run: |
source "${{ github.workspace }}/infra/bootstrapping/sdk_helpers.sh";
source "${{ github.workspace }}/infra/bootstrapping/init_environment.sh";
bash -x ../../run-job.sh attached-spark-standalone-user-identity.yml
working-directory: cli/jobs/spark
Loading