Syncing Code and Artifacts#
SkyPilot simplifies transferring code, data, and artifacts to and from cloud clusters:
To upload code and project files - use
workdir
To upload files outside of workdir (e.g., dotfiles) - use
file_mounts
To upload/reuse large files (e.g., datasets) - use SkyPilot bucket mounting
To download files and artifacts from a cluster - use SkyPilot bucket mounting or
rsync
Here, “upload” means uploading files from your local machine (or a cloud object storage) to a SkyPilot cluster, while “download” means the reverse direction. The same mechanisms work for both files and directories.
Uploading code and project files#
SkyPilot automatically syncs a local working directory to a cluster on every
sky launch
and sky exec
. The workdir contains a project’s
code and other files, and is typically a Git folder.
The working directory can be configured either
by the
workdir
field in a task YAML file, orby the command line option
--workdir
:
$ # Assuming task.yaml has a 'workdir: <path>' field, these commands
$ # sync the workdir to the cluster:
$ sky launch -c dev task.yaml
$ sky exec dev task.yaml
$ # Add a --workdir flag if the yaml doesn't contain the field, or
$ # to override it.
These commands sync the working directory to ~/sky_workdir
on the remote
VMs. The task is invoked under that working directory (so that it can call
scripts, access checkpoints, etc.).
Note
For large, multi-gigabyte workdirs, uploading may be slow because they
are synced to the remote VM(s) with rsync
. To exclude large files in
your workdir from being uploaded, add them to the .gitignore
file
(or a .git/info/exclude
file) under the workdir.
Note
You can keep and edit code in one central place—the local machine where
sky
is used—and have them transparently synced to multiple remote
clusters for execution:
$ sky exec cluster0 task.yaml
$ # Make local edits to the workdir...
$ # cluster1 will get the updated code.
$ sky exec cluster1 task.yaml
Uploading files outside of workdir#
Use the file_mounts
field in a task YAML to upload to a cluster
local files outside of the working directory (e.g., dotfiles)
cloud object storage URIs (currently, SkyPilot supports AWS S3, GCP GCS, Cloudflare R2 and IBM COS)
Every sky launch
invocation reruns the sync up of these files.
Example file mounts:
file_mounts:
# Format: <cluster path>: <local path/cloud object URI>
# Upload from local machine to the cluster via rsync.
/remote/datasets: ~/local/datasets
~/.vimrc: ~/.vimrc
~/.ssh/id_rsa.pub: ~/.ssh/id_rsa.pub
# Download from S3 to the cluster.
/s3-data-test: s3://fah-public-data-covid19-cryptic-pockets/human/il6/PROJ14534/RUN999/CLONE0/results0
For more details, see this example and YAML Configuration.
If you have edited the file_mounts
section and would like to have it reflected on an existing cluster without rerunning the setup
commands,
pass the --no-setup
flag to sky launch
. For example, sky launch --no-setup -c <cluster_name> <task.yaml>
.
Note
Items listed in a .gitignore
file (or a .git/info/exclude
file)
under a local file_mount source are also ignored (the same behavior as
handling workdir
).
Note
If relative paths are used in file_mounts
or workdir
, they
are evaluated relative to the location from which the sky
command
is run.
Uploading or reusing large files#
For large files (e.g., 10s or 100s of GBs), putting them into the workdir or a
file_mount may be slow because they are processed by rsync
. Use
SkyPilot bucket mounting to efficiently handle
large files.
Downloading files and artifacts#
Task artifacts, such as logs and checkpoints, can either be transparently uploaded to a cloud object storage, or directly copied from the remote cluster.
Writing artifacts to cloud object storage#
In addition to handling datasets and other task inputs,
SkyPilot bucket mounting can also be used to directly upload artifacts
generated by tasks. This is achieved by creating a MOUNT
mode Storage
mount like so in your task YAML:
file_mounts:
/outputs:
name: my-sky-outputs # Can be existing S3 bucket or a new bucket
store: s3
mode: MOUNT
This file_mount
will mount the bucket s3://my-sky-outputs/
(creating it if it doesn’t exist) at /outputs
. Since this is specified
with mode: MOUNT
, any files written to /outputs
will also be
automatically written to the s3://my-sky-outputs/
bucket.
Thus, if you point your code to produce files at /outputs/
, they
will be available on the S3 bucket when they are written to /outputs/
.
You can then fetch those files either using the S3 web console or aws-cli
(e.g., aws s3 ls my-sky-outputs
).
Transferring directly with rsync#
Files and artifacts can also be directly transferred from remote clusters to the local machine.
To transfer files from cluster nodes, use rsync
(or scp
):
$ # Rsync from head
$ rsync -Pavz dev:/path/to/checkpoints local/
$ # Rsync from worker nodes (1-based indexing)
$ rsync -Pavz dev-worker1:/path/to/checkpoints local/