{"id":96668,"date":"2025-08-26T08:06:34","date_gmt":"2025-08-26T08:06:34","guid":{"rendered":"https:\/\/www.newsbeep.com\/au\/96668\/"},"modified":"2025-08-26T08:06:34","modified_gmt":"2025-08-26T08:06:34","slug":"announcing-aws-parallel-computing-service-to-run-hpc-workloads-at-virtually-any-scale","status":"publish","type":"post","link":"https:\/\/www.newsbeep.com\/au\/96668\/","title":{"rendered":"Announcing AWS Parallel Computing Service to run HPC workloads at virtually any scale"},"content":{"rendered":"\n<p>            <a href=\"https:\/\/aws.amazon.com\/polly\/\" target=\"_blank\" rel=\"noopener noreferrer nofollow\"><img loading=\"lazy\" decoding=\"async\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/Voiced_by_Amazon_Polly_EN.png\" alt=\"Voiced by Polly\" width=\"554\" height=\"56\"\/><\/a> <\/p>\n<p>Today we are announcing <a href=\"https:\/\/aws.amazon.com\/pcs\" rel=\"nofollow noopener\" target=\"_blank\">AWS Parallel Computing Service (AWS PCS)<\/a>, a new managed service that helps customers set up and manage <a href=\"https:\/\/aws.amazon.com\/hpc\/\" rel=\"nofollow noopener\" target=\"_blank\">high performance computing (HPC)<\/a> clusters so they seamlessly run their simulations at virtually any scale on AWS. Using the <a href=\"https:\/\/slurm.schedmd.com\/man_index.html\" rel=\"nofollow noopener\" target=\"_blank\">Slurm<\/a> scheduler, they can work in a familiar HPC environment, accelerating their time to results instead of worrying about infrastructure.<\/p>\n<p>In November 2018, we introduced <a href=\"https:\/\/aws.amazon.com\/about-aws\/whats-new\/2018\/11\/AWSParallelCluster\/\" rel=\"nofollow noopener\" target=\"_blank\">AWS ParallelCluster<\/a>, an AWS supported open-source cluster management tool that helps you to deploy and manage HPC clusters in the AWS Cloud. With AWS ParallelCluster, customers can also quickly build and deploy proof of concept and production HPC compute environments. They can use <a href=\"https:\/\/docs.aws.amazon.com\/parallelcluster\/latest\/ug\/commands-v3.html\" rel=\"nofollow noopener\" target=\"_blank\">AWS ParallelCluster Command-Line interface<\/a>, <a href=\"https:\/\/docs.aws.amazon.com\/parallelcluster\/latest\/ug\/api-ref-v3.html\" rel=\"nofollow noopener\" target=\"_blank\">API<\/a>, <a href=\"https:\/\/docs.aws.amazon.com\/parallelcluster\/latest\/ug\/pc-py-library-v3.html\" rel=\"nofollow noopener\" target=\"_blank\">Python library<\/a>, and the user interface installed from open source packages. They are responsible for updates, which can include tearing down and redeploying clusters. Many customers, though, have asked us for a fully managed AWS service to eliminate operational jobs in building and operating HPC environments.<\/p>\n<p>AWS PCS simplifies HPC environments managed by AWS and is accessible through the <a href=\"https:\/\/console.aws.amazon.com\/pcs\" rel=\"nofollow noopener\" target=\"_blank\">AWS Management Console<\/a>, AWS SDK, and <a href=\"https:\/\/aws.amazon.com\/cli\" rel=\"nofollow noopener\" target=\"_blank\">AWS Command-Line Interface (AWS CLI)<\/a>. Your system administrators can create managed Slurm clusters that use their compute and storage configurations, identity, and job allocation preferences. AWS PCS uses Slurm, a highly scalable, fault-tolerant job scheduler used across a wide range of HPC customers, for scheduling and orchestrating simulations. End users such as scientists, researchers, and engineers can log in to AWS PCS clusters to run and manage HPC jobs, use interactive software on virtual desktops, and access data. You can bring their workloads to AWS PCS quickly, without significant effort to port code.<\/p>\n<p>You can use fully managed <a href=\"https:\/\/aws.amazon.com\/hpc\/dcv\/\" rel=\"nofollow noopener\" target=\"_blank\">NICE DCV<\/a> remote desktops for remote visualization, and access job telemetry or application logs to enable specialists to manage your HPC workflows in one place.<\/p>\n<p><img loading=\"lazy\" decoding=\"async\" class=\"aligncenter size-full wp-image-87272\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-1-diagram.png\" alt=\"\" width=\"1984\" height=\"1114\"\/><\/p>\n<p>AWS PCS is designed for a wide range of traditional and emerging, compute or data-intensive, engineering and scientific workloads across areas such as computational fluid dynamics, weather modeling, finite element analysis, electronic design automation, and reservoir simulations using familiar ways of preparing, executing, and analyzing simulations and computations.<\/p>\n<p>Getting started with AWS Parallel Computing Service<br \/>To try out AWS PCS, you can use our <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started.html\" rel=\"nofollow noopener\" target=\"_blank\">tutorial for creating a simple cluster<\/a> in the AWS documentation. First, you create a virtual private cloud (VPC) with an AWS CloudFormation template and shared storage in <a href=\"https:\/\/aws.amazon.com\/efs\" rel=\"nofollow noopener\" target=\"_blank\">Amazon Elastic File System (Amazon EFS)<\/a> within your account for the AWS Region where you will try AWS PCS. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-vpc.html\" rel=\"nofollow noopener\" target=\"_blank\">Create a VPC<\/a> and <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-efs.html\" rel=\"nofollow noopener\" target=\"_blank\">Create shared storage<\/a> in the AWS documentation.<\/p>\n<p>1. Create a cluster<br \/>In the <a href=\"https:\/\/console.aws.amazon.com\/pcs\/home\" rel=\"nofollow noopener\" target=\"_blank\">AWS PCS console<\/a>, choose Create cluster, a persistent resource for managing resources and running workloads.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87277\" style=\"width: 100%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-2-cluster-list.png\" alt=\"\" width=\"2258\" height=\"626\"\/><\/p>\n<p>Next, enter your cluster name and choose the controller size of your Slurm scheduler. You can choose Small (up to 32 nodes and 256 jobs), Medium (up to 512 nodes and 8,192 jobs), or Large (up to 2,048 nodes and 16,384 jobs) for the limits of cluster workloads. In the Networking section, choose your created VPC, subnet to launch the cluster, and security group applied to your cluster.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87276\" style=\"width: 90%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-3-create-a-cluster.png\" alt=\"\" width=\"1693\" height=\"2540\"\/><\/p>\n<p>Optionally, you can set the Slurm configuration such as an idle time before compute nodes will scale down, a Prolog and Epilog scripts directory on launched compute nodes, and a resource selection algorithm parameter used by Slurm.<\/p>\n<p>Choose Create cluster. It takes some time for the cluster to be provisioned.<\/p>\n<p>2. Create compute node groups<br \/>After creating your cluster, you can create compute node groups, a virtual collection of <a href=\"https:\/\/aws.amazon.com\/ec2\" rel=\"nofollow noopener\" target=\"_blank\">Amazon Elastic Compute Cloud (Amazon EC2)<\/a> instances that AWS PCS uses to provide interactive access to a cluster or run jobs in a cluster. When you define a compute node group, you specify common traits such as EC2 instance types, minimum and maximum instance count, target VPC subnets, <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/working-with_ami.html\" rel=\"nofollow noopener\" target=\"_blank\">Amazon Machine Image (AMI)<\/a>, purchase option, and custom launch configuration. Compute node groups require an instance profile to pass an <a href=\"https:\/\/aws.amazon.com\/iam\" rel=\"nofollow noopener\" target=\"_blank\">AWS Identity and Access Management (IAM)<\/a> role to an EC2 instance and an EC2 launch template that AWS PCS uses to configure EC2 instances it launches. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-cng_launch-templates.html\" rel=\"nofollow noopener\" target=\"_blank\">Create a launch template<\/a> And <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-cng_instance-profile.html\" rel=\"nofollow noopener\" target=\"_blank\">Create an instance profile<\/a> in the AWS documentation.<\/p>\n<p>To create a compute node group in the console, go to your cluster and choose the Compute node groups tab and the Create compute node group button.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87273\" style=\"width: 90%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-4-compute-node-groups.png\" alt=\"\" width=\"1855\" height=\"1445\"\/><\/p>\n<p>You can create two compute node groups: a login node group to be accessed by end users and a job node group to run HPC jobs.<\/p>\n<p>To create a compute node group running HPC jobs, enter a compute node name and select a previously-created EC2 launch template, IAM instance profile, and subnets to launch compute nodes in your cluster VPC.<\/p>\n<p>Next, choose your preferred EC2 instance types to use when launching compute nodes and the minimum and maximum instance count for scaling. I chose the hpc6a.48xlarge instance type and scale limit up to eight instances. For a login node, you can choose a smaller instance, such as one c6i.xlarge instance. You can also choose either the On-demand or Spot EC2 purchase option if the instance type supports. Optionally, you can choose a specific AMI.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87275\" style=\"width: 90%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-5-create-compute-note-group.png\" alt=\"\" width=\"1678\" height=\"3112\"\/><\/p>\n<p>Choose Create. It takes some time for the compute node group to be provisioned. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-cng_workers.html\" rel=\"nofollow noopener\" target=\"_blank\">Create a compute node group to run jobs<\/a> and <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-cng_login-nodes.html\" rel=\"nofollow noopener\" target=\"_blank\">Create a compute node group for login nodes<\/a> in the AWS documentation.<\/p>\n<p>3. Create and run your HPC jobs<br \/>After creating your compute node groups, you submit a job to a queue to run it. The job remains in the queue until AWS PCS schedules it to run on a compute node group, based on available provisioned capacity. Each queue is associated with one or more compute node groups, which provide the necessary EC2 instances to do the processing.<\/p>\n<p>To create a queue in the console, go to your cluster and choose the Queues tab and the Create queue button.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87281\" style=\"width: 90%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-create-a-queue.png\" alt=\"\" width=\"1838\" height=\"586\"\/><\/p>\n<p>Enter your queue name and choose your compute node groups assigned to your queue.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87282\" style=\"width: 90%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-create-a-queue-detail.png\" alt=\"\" width=\"1680\" height=\"1324\"\/><\/p>\n<p>Choose Create and wait while the queue is being created.<\/p>\n<p>When the login compute node group is active, you can use <a href=\"https:\/\/aws.amazon.com\/systems-manager\/\" rel=\"nofollow noopener\" target=\"_blank\">AWS Systems Manager<\/a> to connect to the EC2 instance it created. Go to the <a href=\"https:\/\/console.aws.amazon.com\/ec2\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon EC2 console<\/a> and choose your EC2 instance of the login compute node group. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_create-queue.html\" rel=\"nofollow noopener\" target=\"_blank\">Create a queue to submit and manage jobs<\/a> and <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_connect.html\" rel=\"nofollow noopener\" target=\"_blank\">Connect to your cluster<\/a> in the AWS documentation.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87308\" style=\"width: 100%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-login-instance-1.png\" alt=\"\" width=\"2170\" height=\"674\"\/><\/p>\n<p>To run a job using Slurm, you prepare a submission script that specifies the job requirements and submit it to a queue with the sbatch command. Typically, this is done from a shared directory so the login and compute nodes have a common space for accessing files.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87278\" style=\"width: 100%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-run-jobs.png\" alt=\"\" width=\"1656\" height=\"1119\"\/><\/p>\n<p>You can also run a message passing interface (MPI) job in AWS PCS using Slurm. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_run-job.html\" rel=\"nofollow noopener\" target=\"_blank\">Run a single node job with Slurm<\/a> or <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_run-mpi-job.html\" rel=\"nofollow noopener\" target=\"_blank\">Run a multi-node MPI job with Slurm<\/a> in the AWS documentation.<\/p>\n<p>You can connect a fully-managed NICE DCV remote desktop for visualization. To get started, use the CloudFormation template from <a href=\"https:\/\/github.com\/aws-samples\/aws-hpc-recipes\/tree\/main\/recipes\/pcs\/cfd_cluster\" rel=\"nofollow noopener\" target=\"_blank\">HPC Recipes for AWS GitHub repository<\/a>.<\/p>\n<p>In this example, I used the <a href=\"https:\/\/www.openfoam.com\/\" rel=\"nofollow noopener\" target=\"_blank\">OpenFOAM<\/a> <a href=\"https:\/\/develop.openfoam.com\/Development\/openfoam\/-\/tree\/master\/tutorials\/incompressible\/simpleFoam\/motorBike\" rel=\"nofollow noopener\" target=\"_blank\">motorBike simulation<\/a> to calculate the steady flow around a motorcycle and rider. This simulation was run with 288 cores of three hpc6a instances. The output can be visualized in the <a href=\"https:\/\/www.paraview.org\/\" rel=\"nofollow noopener\" target=\"_blank\">ParaView<\/a> session after logging in to the web interface of DCV instance.<\/p>\n<p><img decoding=\"async\" loading=\"lazy\" class=\"aligncenter size-full wp-image-87280\" style=\"width: 100%; border: solid 1px #ccc;\" src=\"https:\/\/www.newsbeep.com\/au\/wp-content\/uploads\/2025\/08\/2024-aws-pcs-nice-dcv-paraview.png\" alt=\"\" width=\"2466\" height=\"1486\"\/><\/p>\n<p>Finally, after you are done HPC jobs with the cluster and node groups that you created, you should delete the resources that you created to avoid unnecessary charges. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/getting-started_delete.html\" rel=\"nofollow noopener\" target=\"_blank\">Delete your AWS resources<\/a> in the AWS documentation.<\/p>\n<p>Things to know<br \/>Here are a couple of things that you should know about this feature:<\/p>\n<p>        Slurm versions \u2013 AWS PCS initially supports Slurm 23.11 and o\ufb00ers mechanisms designed to enable customers to upgrade their Slurm major versions once new versions are added. Additionally, AWS PCS is designed to automatically update the Slurm controller with patch versions. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/slurm-versions.html\" rel=\"nofollow noopener\" target=\"_blank\">Slurm versions<\/a> in the AWS documentation.<br \/>\n        Capacity Reservations \u2013 You can reserve EC2 capacity in a specific Availability Zone and for a specific duration using On-Demand Capacity Reservations to make sure that you have the necessary compute capacity available when you need it. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/working-with_capacity-reservations.html\" rel=\"nofollow noopener\" target=\"_blank\">Capacity Reservations<\/a> in the AWS documentation.<br \/>\n        Network file systems \u2013 You can attach network storage volumes where data and files can be written and accessed, including <a href=\"https:\/\/aws.amazon.com\/fsx\/netapp-ontap\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon FSx for NetApp ONTAP<\/a>, <a href=\"https:\/\/aws.amazon.com\/fsx\/openzfs\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon FSx for OpenZFS<\/a>, and <a href=\"https:\/\/aws.amazon.com\/filecache\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon File Cache<\/a> as well as <a href=\"https:\/\/aws.amazon.com\/efs\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon EFS<\/a> and <a href=\"https:\/\/aws.amazon.com\/fsx\/lustre\/\" rel=\"nofollow noopener\" target=\"_blank\">Amazon FSx for Lustre<\/a>. You can also use self-managed volumes, such as NFS servers. To learn more, visit <a href=\"https:\/\/docs.aws.amazon.com\/pcs\/latest\/userguide\/working-with_file-systems.html\" rel=\"nofollow noopener\" target=\"_blank\">Network file systems<\/a> in the AWS documentation. <\/p>\n<p>Now available<br \/><a href=\"https:\/\/aws.amazon.com\/pcs\" rel=\"nofollow noopener\" target=\"_blank\">AWS Parallel Computing Service<\/a> is now available in the US East (N. Virginia), AWS US East (Ohio), US West (Oregon), Asia Pacific (Singapore), Asia Pacific (Sydney), Asia Pacific (Tokyo), Europe (Frankfurt), Europe (Ireland), Europe (Stockholm) Regions.<\/p>\n<p>AWS PCS launches all resources in your AWS account. You will be billed appropriately for those resources. For more information, see the <a href=\"https:\/\/aws.amazon.com\/pcs\/pricing\/\" rel=\"nofollow noopener\" target=\"_blank\">AWS PCS Pricing page<\/a>.<\/p>\n<p>Give it a try and send feedback to <a href=\"https:\/\/repost.aws\/\" rel=\"nofollow noopener\" target=\"_blank\">AWS re:Post<\/a> or through your usual AWS Support contacts.<\/p>\n<p>\u2014 <a href=\"https:\/\/twitter.com\/channyun\" rel=\"nofollow noopener\" target=\"_blank\">Channy<\/a><\/p>\n<p>P.S. Special thanks to <a href=\"https:\/\/www.linkedin.com\/in\/mattdotvaughn\/\" rel=\"nofollow noopener\" target=\"_blank\">Matthew Vaughn<\/a>, a principal developer advocate at AWS for his contribution in creating a HPC testing environment.<\/p>\n<p>      <script async src=\"https:\/\/platform.twitter.com\/widgets.js\" charset=\"utf-8\"><\/script><\/p>\n","protected":false},"excerpt":{"rendered":"Today we are announcing AWS Parallel Computing Service (AWS PCS), a new managed service that helps customers set&hellip;\n","protected":false},"author":2,"featured_media":96669,"comment_status":"","ping_status":"","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[21],"tags":[64,63,257,105],"class_list":{"0":"post-96668","1":"post","2":"type-post","3":"status-publish","4":"format-standard","5":"has-post-thumbnail","7":"category-computing","8":"tag-au","9":"tag-australia","10":"tag-computing","11":"tag-technology"},"_links":{"self":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/96668","targetHints":{"allow":["GET"]}}],"collection":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts"}],"about":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/comments?post=96668"}],"version-history":[{"count":0,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/posts\/96668\/revisions"}],"wp:featuredmedia":[{"embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media\/96669"}],"wp:attachment":[{"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/media?parent=96668"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/categories?post=96668"},{"taxonomy":"post_tag","embeddable":true,"href":"https:\/\/www.newsbeep.com\/au\/wp-json\/wp\/v2\/tags?post=96668"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}