You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
It would be great to have standalone R function utilities to manage batch jobs. These would run in the user's interactive session outside the targets pipeline / crew controller. I am thinking of covering the same functionality as qsub, qstat, and qdel in SGE (sbatch, squeue, and scancel in SLURM), plus log files. Proposal:
crew_aws_batch_submit(): submit a job that runs some code (R or shell). This could help e.g. submit a targets pipeline as a Batch job which submits other Batch jobs.
crew_aws_batch_status(): get the status of jobs in a given job queue / job definition.
crew_aws_batch_terminate(): terminate one or more jobs with specific job names/IDs/ARNs.
crew_aws_batch_logs(): log files for one or more jobs, or for an entire job definition. This would really help detect tricky worker-level errors such as running out of memory or hitting a price spike that terminates spot instances.
The text was updated successfully, but these errors were encountered:
Hi @wlandau for crew_aws_batch_logs you could hack smdocker logging method. In short when it is building a docker using AWS CodeBuild it returns the AWS CloudWatch logs back the to the console for R users to monitor and check.
It would be great to have standalone R function utilities to manage batch jobs. These would run in the user's interactive session outside the
targets
pipeline /crew
controller. I am thinking of covering the same functionality asqsub
,qstat
, andqdel
in SGE (sbatch
,squeue
, andscancel
in SLURM), plus log files. Proposal:crew_aws_batch_submit()
: submit a job that runs some code (R or shell). This could help e.g. submit atargets
pipeline as a Batch job which submits other Batch jobs.crew_aws_batch_status()
: get the status of jobs in a given job queue / job definition.crew_aws_batch_terminate()
: terminate one or more jobs with specific job names/IDs/ARNs.crew_aws_batch_logs()
: log files for one or more jobs, or for an entire job definition. This would really help detect tricky worker-level errors such as running out of memory or hitting a price spike that terminates spot instances.The text was updated successfully, but these errors were encountered: