-
Notifications
You must be signed in to change notification settings - Fork 56
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Usage with AWS S3 and Ray #59
Comments
This was referenced Sep 16, 2023
This comment was marked as off-topic.
This comment was marked as off-topic.
Hey why did you close it ? I think it's a good improvement and people will review the PRs soon |
Hello! I closed the issue because it wasn't quite actionable, but rather a “note to my future self” that could eventually become documentation. 🙈 I'll reopen it if you wish, though. |
Alternative version, without containers. cluster_name: datacomp-downloader
min_workers: 0
max_workers: 10
upscaling_speed: 1.0
provider:
type: aws
region: us-east-1
cache_stopped_nodes: false
available_node_types:
ray.head.default:
resources: {}
node_config:
InstanceType: m5.12xlarge
ImageId: ami-068d304eca3399469
BlockDeviceMappings:
- DeviceName: /dev/sda1
Ebs:
DeleteOnTermination: true
VolumeSize: 200
VolumeType: gp2
ray.worker.default:
resources: {}
node_config:
InstanceType: m5.12xlarge
ImageId: ami-068d304eca3399469
BlockDeviceMappings:
- DeviceName: /dev/sda1
Ebs:
DeleteOnTermination: true
VolumeSize: 200
VolumeType: gp2
initialization_commands:
# Knot Resolver
- wget https://secure.nic.cz/files/knot-resolver/knot-resolver-release.deb
- sudo dpkg --install knot-resolver-release.deb
- rm knot-resolver-release.deb
- sudo apt-get update
- sudo apt-get install --yes knot-resolver
- echo $(hostname --all-ip-addresses) $(hostname) | sudo tee --append /etc/hosts
- sudo systemctl start kresd@{1..48}.service
- echo nameserver 127.0.0.1 | sudo tee /etc/resolv.conf
- sudo systemctl stop systemd-resolved
# Anaconda
- sudo mkdir /opt/miniconda3 && sudo chown $USER /opt/miniconda3
- wget https://repo.anaconda.com/miniconda/Miniconda3-py39_22.11.1-1-Linux-x86_64.sh
- bash Miniconda3-py39_22.11.1-1-Linux-x86_64.sh -f -b -p /opt/miniconda3
- rm Miniconda3-py39_22.11.1-1-Linux-x86_64.sh
- /opt/miniconda3/bin/conda init bash
# Ray
- conda create --yes --name=ray python=3.10.8
- echo conda activate ray >> ~/.bashrc
- pip install ray[all]==2.7.0
setup_commands:
- sudo apt-get update
- sudo apt-get install --yes build-essential ffmpeg |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Usage
Cluster creation
Job submission
Note
Image shards would be saved to the
datacomp-small
AWS S3 bucket, specified with the--data_dir
option.Cluster deletion
Configuration
Sample
cluster.yml
Obscure details
When
--data_dir
points to a cloud storage like S3, we also have to specify a local--metadata_dir
because the downloader script doesn't support saving metadata to cloud storage.The last
pip install
on thesetup_commands
section is needed for compatibility with AWS S3, because the required libraries aren't included in theconda
environment file.There is no need to provide additional AWS credentials if the destination bucket is on the same account as the cluster, because it already has S3 full access through an instance profile.The Python version in
environment.yml
must match the Python version of the Ray cluster; make sure thatdocker.image
oncluster.yaml
has exactly the same version as theenvironment.yml
from this project.The text was updated successfully, but these errors were encountered: