You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I'm repeatedly launching jobs on the same machine that use the same S3 mount point. On subsequent invocations, I run into an error when re-mounting:
echo "Mounting $SOURCE_BUCKET to $MOUNT_PATH with $MOUNT_BINARY..."
goofys -o allow_other --stat-cache-ttl 5s --type-cache-ttl 5s cloud-av-workloads /data
echo "Mounting done."
) && chmod +x ~/.sky/mount_785106.sh && bash ~/.sky/mount_785106.sh && rm ~/.sky/mount_785106.sh failed with return code 1.
Failed to run command before rsync cloud-av-workloads -> /data.
And here is the output of storage_mounts.log:
bash: warning: here-document at line 37 delimited by end-of-file (wanted `EOF')
Path already mounted - unmounting...
fusermount: failed to unmount /data: Device or resource busy
Successfully unmounted /data.
goofys already installed. Proceeding...
Mount path /data is not empty. Please make sure its empty.
I think this could be resolved by detecting whether the mount points are identical and keeping the file mount alive between jobs rather than unmounting and re-mounting.
The text was updated successfully, but these errors were encountered:
First sky launch is successful, but the second one fails because the unmounting fails (and the mounting script incorrectly continues). Will fix this.
Also (you might know this already) you can use sky exec <cluster_name> <task.yaml> to submit jobs to the cluster without re-executing file mounts and setup :)
Hi,
I'm repeatedly launching jobs on the same machine that use the same S3 mount point. On subsequent invocations, I run into an error when re-mounting:
And here is the output of
storage_mounts.log
:I think this could be resolved by detecting whether the mount points are identical and keeping the file mount alive between jobs rather than unmounting and re-mounting.
The text was updated successfully, but these errors were encountered: