# Don't include large datasets in Docker image# Instead, access from cloud storageimport pandas as pddataset = pd.read_csv("/orwd_data/datasets/large_dataset.csv")
2. Configuration Files:
Copy
import yaml# Store environment config in cloud storageconfig_path = Path("/orwd_data/config.yaml")if config_path.exists(): with open(config_path) as f: config = yaml.safe_load(f)
Configure storage access via bucket_config in SandboxSettings:
Copy
from openreward import OpenReward, SandboxSettings, SandboxBucketConfigclient = OpenReward(api_key="your-api-key")settings = SandboxSettings( environment="username/env-name", image="python:3.11-slim", machine_size="1:2", bucket_config=SandboxBucketConfig( mount_path="/workspace", # Where to mount in container read_only=True, # Buckets are always read-only only_dir="datasets/subset", # Optional: mount only subdirectory implicit_dirs=False # Optional: show all subdirectories ))async with client.sandbox(settings) as sandbox: # Storage is mounted at /workspace output, _ = await sandbox.run("ls -la /workspace") print(output)
# Without implicit_dirs (default, faster):bucket_config=SandboxBucketConfig( mount_path="/workspace", implicit_dirs=False # Default)# Shows only directories that explicitly exist# Better performance# With implicit_dirs (slower, more complete):bucket_config=SandboxBucketConfig( mount_path="/workspace", implicit_dirs=True)# Shows all directories implied by file paths# More complete directory tree