You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The overview in the AML Studio shows the correct input data and I can navigate to it.
Now the issue: Inside my job which runs without problems the input data argument is expanded to something like:
/mnt/var/hadoop/tmp/nm-local-dir/usercache/trusted-service-user/appcache/application_1724844698277_0001/container_1724844698277_0001_01_000001/azureml:/subscriptions/xxxxxx/resourcegroups/weg-aml-v2/workspaces/weg-aml-v2/datastores/workspaceblobstore/paths/data/titanic.csv
However this path does not exist.
There is no indication in the logs what failed.
Expected behavior
The input data should be passed to the spark application, such that it can be accessed.
Actual behavior
The issue is that the input data is not passed to the spark application.
Inside the spark application which runs without problems the input data argument is expanded to something like:
/mnt/var/hadoop/tmp/nm-local-dir/usercache/trusted-service-user/appcache/application_1724844698277_0001/container_1724844698277_0001_01_000001/azureml:/subscriptions/xxxxxx/resourcegroups/weg-aml-v2/workspaces/weg-aml-v2/datastores/workspaceblobstore/paths/data/titanic.csv
However this path does not exist.
There is no indication in the logs what failed.
Addition information
No response
The text was updated successfully, but these errors were encountered:
Operating System
Linux
Version Information
az --version
ml 2.28.0
Steps to reproduce
I am trying to submit a spark job as shown in https://github.com/Azure/azureml-examples/tree/main/cli/jobs/spark
The data has been uploaded as described here: https://github.com/Azure/azureml-examples/blob/main/cli/jobs/spark/data/README.md
This is the yml definition (partially shown)
The overview in the AML Studio shows the correct input data and I can navigate to it.
Now the issue: Inside my job which runs without problems the input data argument is expanded to something like:
/mnt/var/hadoop/tmp/nm-local-dir/usercache/trusted-service-user/appcache/application_1724844698277_0001/container_1724844698277_0001_01_000001/azureml:/subscriptions/xxxxxx/resourcegroups/weg-aml-v2/workspaces/weg-aml-v2/datastores/workspaceblobstore/paths/data/titanic.csv
However this path does not exist.
There is no indication in the logs what failed.
Expected behavior
The input data should be passed to the spark application, such that it can be accessed.
Actual behavior
The issue is that the input data is not passed to the spark application.
Inside the spark application which runs without problems the input data argument is expanded to something like:
/mnt/var/hadoop/tmp/nm-local-dir/usercache/trusted-service-user/appcache/application_1724844698277_0001/container_1724844698277_0001_01_000001/azureml:/subscriptions/xxxxxx/resourcegroups/weg-aml-v2/workspaces/weg-aml-v2/datastores/workspaceblobstore/paths/data/titanic.csv
However this path does not exist.
There is no indication in the logs what failed.
Addition information
No response
The text was updated successfully, but these errors were encountered: