Skip to content

This example allows you to quickly evaluate some typical solution areas of a Game Production in the Cloud pipeline (GPIC). Remote workstation, version control and acceleration of computationally expensive process.

License

Notifications You must be signed in to change notification settings

aws-solutions-library-samples/guidance-for-game-production-environment-on-aws

Guidance for a Game Production Environment on AWS

Overview

Quickly deploy a Cloud Game Development environment for you and your team, with this AWS Sample. Once deployed, you will have high performance virtual workstation, central version control system, and acceleration of compute heavy tasks by distributing the work to other machines on demand.

The AWS Cloud Development Kit(CDK), or optionally AWS CloudFormation templates, creates a new AWS Virtual Private Cloud (VPC), a customizable virtual workstation, a Perforce Helix Core server, and an Unreal Swarm cluster. The latter can be used to accelerate Unreal Engine 5 lighting builds using the vast compute resources available in AWS.

Please note that this guidance was written using CDK v1 and the customer is responsible for updating to CDK v2 if they wish.

Optionally you can add on other solutions from partners such as Epic Games, HP Anyware, Parsec, and Incredibuild. These partner solutions further accelerate your development with:

Contents

This project repository contains:

  • Infrastructure deployment automation that deploys required infrastructure resources using AWS CDK or AWS CloudFormation.
  • PowerShell Scripts that are used to:
    • Collect dependencies from your Unreal Engine 5 installation
    • Setup the Unreal Engine 5 Swarm coordinator and agents.
    • Setup the Virtual Workstation
  • Bash Scripts to setup a secure Perforce Helix Core Server

You can deploy this multiple ways, with the easiest being the AWS CloudFormation web interface to deploy the following CloudFormation stacks from the cloudformation folder in this repository:

  • cloudformation/gpic-pipeline-foundation.yaml
  • cloudformation/gpic-pipeline-perforce-helix-core.yaml
  • cloudformation/gpic-pipeline-virtual-workstation.yaml
  • cloudformation/gpic-pipeline-swarm-cluster.yaml

Download these four files locally and follow the instructions below, only using the AWS CDK or AWS CLI as described below if you are familiar with that process.

Architecture Diagram

Architecture

Cost

You are responsible for the cost of the AWS services used while running this Guidance. As of 7-November-2023, the cost for running this Guidance with the default settings in the US East (N. Virginia) region is approximately $544 per month for 40 hours per week of use (instances stopped when not in use).

Note that this estimate does not include costs of third party software installed with this guidance, including Unreal Engine and associated tools, Perforce Helix Core, Incredibuild, Parsec, HP Anyware, and any others.

Prerequisites

Operating System

This guidance can be deployed from any Windows, macOS, or Linux environment with an internet connection and the ability to run Python 3 and the AWS Cloud Development Kit (AWS CDK).

Deployment steps

On a high level the deployment is split into four (4) parts.

The first step is to deploy the gpic-pipeline-foundation stack. This will setup a VPC, private- and public -subnets, the network routing and a S3 bucket which is needed in subsequent steps.

The second step is to deploy the gpic-pipeline-perforce-helix-core stack. This will deploy a secure Perforce Helix Core Server into one of the private subnets and configure it according to best-practices. Once this machine is up and running you can connect to it as user perforce using P4V/P4/P4Admin. The password for the user perforce can be retrieved via AWS Secrets Manager.

The third step is to deploy the gpic-pipeline-virtual-workstation stack. Once this machine is up and running you can connect to it as user Administrator via RDP using the public IP address (or the public DNS name). The password for the user Administrator can be retrieved via the AWS Secrets Manager.

The last step is the deployment of the gpic-pipeline-swarm-cluster stack, which will deploy a Unreal Engine 5 Swarm Coordinator into one of the private subnets as well as an Amazon EC2 AutoScaling consisting of Unreal Engine 5 Swarm agents.

For the best possible experience, we recommend to deploy this example in the AWS region closest to you.

Initiate CDK (Optional)

This step is optional if you decide to use the CloudFormation templates in the cloudformation directory, which is recommended for most users.

This project uses a standard Python CDK project to deploy this example. If you don't have CDK installed, please follow the instructions at AWS Cloud Development Kit documentation.

When you have CDK and Python installed you can use following steps to initiate this project:

Create a python virtualenv

MacOS and Linux:

 python3 -m venv .venv

Windows

python -m venv .venv

After the virtualenv is created, you can use the following step to activate your virtualenv.

MacOS and Linux:

source .venv/bin/activate

Windows:

source.bat

Once the virtualenv is activated, you can install the required dependencies.

pip install -r requirements.txt

If you run on a MacOS or Linux, you potentialy need to modify the file cdk.json to use the correct python version. This CDK application expects that python is a symlink to python3 or the binary for python verions 3.

Deployment of the Foundations Stack

The Foundation stacks deploys an AWS Virtual Private Cloud (VPC) with two public subnets, two private subnets and a NAT Gateway. It also deploys a S3 bucket, which is utilized by the other stacks of this example.

You can use the cloudformation/gpic-pipeline-foundation.yaml template to create a CloudFormation stack called gpic-pipeline-foundation using the AWS Console and AWS CLI. If you want to customize the VPC CIDR you will need to edit the CloudFormation template or use the CDK to synthesize new one with the CDK context variable (see the Tips section).

Alternatively can deploy this stack with CDK or by AWS CLI. By default we will create a VPC with a CIDR prefix of 10.0.0.0/16 and allow traffic to the Virtual Workstation, the Perforce Helix Core Server and the Unreal Engine 5 Swarm coordinator and agents from CIDR Prefix 10.0.0.0/8. In addition we will allow traffic from the CIDR prefix 0.0.0.0/0. to the Virtual Workstation for ease of use. These values can be changed by editing the cdk.json-file and giving new values for foundation_vpc_cidr, virtual_workstation_trusted_internal_cidr, virtual_workstation_trusted_remote_cidr, unreal_engine_swarm_cluster_trusted_internal_cidr and perforce_trusted_internal_cidr.

To deploy the foundation stack please run the following command:

cdk deploy gpic-pipeline-foundation

CLI deployment command:

aws cloudformation create-stack --stack-name gpic-pipeline-foundation --template-body file://cloudformation/gpic-pipeline-foundation.yaml

Once the creation of resources is ready you can move to next step.

Deployment of the Perforce Helix Core Stack

The Perforce Helix Core stack deploys an EC2 instance based on the latest Amazon Linux 2 AMI into one of the private subnets and setup the Perforce Helix Core daemon using Perforce Helix Server Deployment Package (SDP). In the default configuration access to this machine is only allowed by the CIDR prefix 10.0.0.0/8 on port 1666. These CIDR prefix can be changed by editing the cdk.json and modifying the value of perforce_trusted_internal_cidr.

To deploy the Perforce Helix Core stack you can use the cloudformation/gpic-pipeline-perforce-helix-core.yaml template to create a CloudFormation stack called gpic-pipeline-perforce-helix-core using the AWS Console.

Alternatively you can deploy the Perforce Helix Core stack using AWS CDK with the following command:

cdk deploy gpic-pipeline-perforce-helix-core

You can also use the AWS CLI. If you want to customize the VPC CIDR you will need to edit the CloudFormation template or use the CDK to synthesize new one with the CDK context variable (see the Tips section).

AWS CLI command:

aws cloudformation create-stack --stack-name gpic-pipeline-perforce-helix-core --template-body file://cloudformation/gpic-pipeline-perforce-helix-core.yaml --capabilities CAPABILITY_NAMED_IAM

To access the Perforce Helix Core server with the P4/P4V/P4Admin please use the user perforce. The password for this user can be obtained via the AWS Secrets Manager. To access the server via Secure Shell (SSH), please use the AWS Systems Manager Session Manager.

Once the creation of resources is ready you can move to next step.

Deployment and Setup of the Virtual Workstation Stack

This deployment will not only launch a virtual workstation, it will also deploy the necessary security groups and additional AWS infrastructure to use when deploying other workstations from the AWS Marketplace as in the Description.

Deployment of the Virtual Workstation

The Virtual Workstation stack deploys an EC2 Instance based on the latest Microsoft Windows Server 2019 Base AMI in one of the public subnets. In this process we also:

  • Set the password for the user "Administrator"
  • Install the latest NVIDIA Grid driver
  • Setup Windows 2019 Server firewall and allow ingress traffic for the following applications/protocols:
    • ICMP
    • PCoIP
    • Parsec
    • NICE-DCV
    • HP Anyware
    • Unreal Engine 5 Swarm

You can use the cloudformation/gpic-pipeline-virtual-workstation.yaml template to create a CloudFormation stack called gpic-pipeline-virtual-workstation using the AWS Console. To customize this workstation further, you will need to use the AWS CDK or AWS CLI. It is recommended to launch this stack with defaults then deploy Marketplace AMIs into the same VPC created in the first stack.

Access to this virtual workstation is controlled by the variables virtual_workstation_trusted_internal_cidr and virtual_workstation_trusted_remote_cidr in cdk.json.

To deploy the Virtual Workstation stack please via AWS CDK, run the following command:

cdk deploy gpic-pipeline-virtual-workstation

If you want to customize the trusted CIDR's you will need to edit the CloudFormation template or use the CDK to synthesize new one with the CDK context variable (see the Tips section).

AWS CLI command:

aws cloudformation create-stack --stack-name gpic-pipeline-virtual-workstation --template-body file://cloudformation/gpic-pipeline-virtual-workstation.yaml --capabilities CAPABILITY_IAM

Once the creation of resources is finished you can move to next step.

Connecting to initial Virtual Workstation

If you are using this initial virtual workstation, you can configure the instance by first connecting via Remote Desktop Protocol (RDP) by doing the following:

Open a client capable of connection to a Windows computer using RDP.

  • Windows has the Remote Desktop Connection software built in. Search for Remote Desktop Connection in the Windows taskbar
  • For macOS, you'll need to download the Remote Desktop client
  • For Linux, you'll need to obtain an RDP capable like Remmina or TeamViewer

To access the Virtual Workstation use the public IP/DNS, which can be obtained via the Output values from CDK/CloudFormation stack or alternatively via the AWS Management Console.

Once you're connected, you'll need to log in with credentials that you retrieve from AWS Secrets Manager. To find the credentials:

  • Open the AWS console and navigate to the AWS CloudFormation console
  • In the CloudFormation console, select the gpic-pipeline-virtual-workstation stack and expand the Virtual Workstation Password
  • Select the Physical ID and you will be navigated to the AWS Secrets Manager console with the correct secret open
  • Select Retrieve secret value to see the password
  • Copy the password, use administrator for the user name and paste the password

Once you are logged in, you have several options for installing high performance connection protocols:

  • NICE DCV
    Follow the installation steps in the NICE DCV Administrator Guide.

  • Parsec

    • Open a PowerShell and run the script, which is provided by Parsec in their GitHub repository
    • Question: Do you want this computer to log on to Windows automatically? Yes
    • Once the installation is finished, the script will open up an additional Powershell session to update the GPU driver. Please close this Powershell session and don't provide a ACCESS_KEY & SECRET_ACCESS_KEY. We have already-installed the latest NVIDIA Grid driver on the virtual workstation.
    • Close all remaining Powershell sessions
    • Sign-Up for a Parsec account or use an existing Parsec account to login
  • HP Anyware
    Visit the HP Anyware free trial page for more information on installation.

Setup of the Virtual Workstation

There are some Windows Firewall rules that need to be added when using some software on an AWS virtual workstation. You can run the below PowerShell commands to add these rules:

Parsec

New-NetFirewallRule -DisplayName 'Allow Parsec' -Direction Inbound -Action Allow -Protocol TCP -LocalPort 1666

HP Anyware

New-NetFirewallRule -DisplayName 'Allow PCoIP' -Direction Inbound -Action Allow -Protocol TCP -LocalPort 4172

New-NetFirewallRule -DisplayName 'Allow PCoIP' -Direction Inbound -Action Allow -Protocol UDP -LocalPort 4172

New-NetFirewallRule -DisplayName 'Allow PCoIP' -Direction Inbound -Action Allow -Protocol TCP -LocalPort 443

NICE DCV

New-NetFirewallRule -DisplayName 'Allow PCoIP' -Direction Inbound -Action Allow -Protocol TCP -LocalPort 8443

Unreal Engine Swarm Communication

New-NetFirewallRule -DisplayName 'Allow UE5 Swarm TCP' -Direction Inbound -Action Allow -Protocol TCP -LocalPort 8008-8009

New-NetFirewallRule -DisplayName 'Allow UE5 Swarm UDP' -Direction Inbound -Action Allow -Protocol UDP -LocalPort 8008-8009

New-NetFirewallRule -DisplayName 'Allow ICMP' -Direction Inbound -Action Allow -Protocol ICMPv4

Below you will find instructions on adding additional software to your virtual workstation:

  • Perforce Helix
    • Steps:
      • Download and Install the Helix Visual Client (P4V) from here
      • Steps in the Installation Wizard:
        • Keep all application selected and hit Next
        • For the Server please enter the private IP/DNS name of the Perforce Helix Core server we deployed earlier. You can find the private IP/DNS name in the CDK/CloudFormation Outputs or via the AWS Management Console. Example: ip-10-0-XXX-XXX.eu-central-1.compute.internal:1666
        • The User Name is perforce
        • Hit Next
        • Hit Next
        • Hit Install
        • Uncheck "Start P4V" and hit Exit
  • Unreal Engine 5 You need to Sign-Up for an EPIC account or use your existing one. Alternatively you can clone the Unreal Engine 5 source from Github if you have connected your EPIC Account with your Github account.
  • Incredibuild Cloud (optional)
    To install Incredibuild Cloud, you can either install from the AWS Marketplace or follow the Incredibuild installation instructions.

Now, start the PV4 application and connect to the Peforce Helix Core server. The password of the user perforce can be found in the AWS Secrets Manager.

Once you are done please continue to the next step

Deployment of the Unreal Engine 5 Swarm Cluster

Collecting dependencies

Each Windows instance that will act as a Swarm Coordinator or as a Swarm Agent will need a set of prerequisites installed. We can collect these prerequisites from the Unreal Engine 5 version you installed on the Virtual Workstation with the provided assets/unreal-engine-swarm-create-dependencies-archive.ps1 script in this repository.

  • If you are no longer logged into the Virtual Workstation, please login again.
  • Please download the script unreal-engine-swarm-create-dependencies-archive.ps1 from the assets folder to the virtual workstation
  • This PowerShell script will copy all the components that are needed to customize a fresh Window installation
  • The script assumes that your Unreal Engine is installed to C:\Program Files\Epic Games\UE_5.2 directory but you can customize the script to match your location
  • Script will create a compressed archive called ue5-swarm-archive.zip under your My Documents directory
  • You can find more details about these prerequisites at:

After you have created the ue5-swarm-archive.zip -archive you need to upload it into the root directory of the newly created S3 bucket. It will be downloaded from that location and used during the EC2 Image Builder process. The name of the bucket is available as an output called BucketName from the gpic-pipeline-foundation stack.

With AWS Tools for PowerShell you can use following command to first list buckets and fetch the name of the bucket created by the foundation stack:

Get-S3Bucket

Then write the local file to the specified bucket and key:

Write-S3Object -BucketName <NAME OF THE S3 BUCKET> -Key ue5-swarm-archive.zip -File ue5-swarm-archive.zip

Deployment

To deploy the Swarm Cluster you can use the cloudformation/gpic-pipeline-swarm-cluster.yaml template to create a CloudFormation stack called gpic-pipeline-swarm-cluster using the AWS Web Console.

You can deploy with AWS CDK using the following command:

cdk deploy gpic-pipeline-swarm-cluster

Alternatively you can use the AWS CLI with this command:

CLI example:

aws cloudformation create-stack --stack-name gpic-pipeline-swarm-cluster --template-body file://cloudformation/gpic-pipeline-swarm-cluster.yaml --capabilities CAPABILITY_NAMED_IAM

This step will take 30 minutes on average as it's baking the Window AMI for Swarm. The steps to install all dependencies does take some time to complete. While the deployment is running you can read below for details on what's happening during stack creation.

Once this stack is deployed, please proceed to the final step

A look behind the curtain

Baking custom Windows AMI for UE5 Swarm

The gpic-pipeline-swarm-cluster -stack will first configure EC2 Image Builder to use latest "Microsoft Windows Server 2019 Base" image as the base image. It also creates a EC2 Image Builder component defining the build steps. These steps will download the Zip archive from S3, install .Net runtime, run the UEPrereqSetup_x64.exe installer and then open Windows Firewall for the Swarm ports. You can view the assets/unreal-engine-swarm-cluster-component.yaml file for details.

Once the EC2 Image Builder completes it will create a private AMI under your account. This AMI contains all the required Unreal Engine 5 Swarm build dependencies and can be used to quickly launch the Swarm Coordinator and Swarm Agents.

Deploying UE5 Swarm Coordinator

The Swarm Coordinator will be launched as a single EC2 Instance. The launch will use User Data to configure the Windows to start SwarmCoordinator.exe on boot. You can view the contents of the User Data in assets/setup-unreal-engine-swarm-coordinator.ps1 - Powershell script.

Deploying UE5 Swarm Agent Auto Scaling Group

The Swarm Agents are going to be launched as Auto Scaling Group. Enabling us to quickly scale the number of nodes up and down. As the Swarm Agents need to be already online and registered when you submit a UE5 build, we can't use any metrics to scale the cluster on demand. Instead you can use for example Schedule or some script to scale the cluster before submit a job. With a schedule you could for example configure the cluster to scale up to certain number of nodes in the morning and then after office hours scale the cluster back to zero.

The Swarm Agent will also use User Data to configure the Windows to start SwarmAgent.exe on boot and injects a Swarm configuration file into the Instance. This configuration file will set number of threads to equal amount of CPU Core and also will set the Coordinator IP address. You can view the contents of the User Data in assets/setup-unreal-engine-swarm-agent.ps1 - Powershell script.

Deployment Validation

Now that the gpic-pipeline-swarm-cluster.yaml stack has completed deployment you should see two additional EC2 Instances running in your new VPC. Also the CDK/CloudFormation stack should have output the private IP address of the Unreal Engine 5 Swarm Coordinator.

On your Virtual Workstation you have to configure the local Swarm Agent. You can launch it from C:\Program Files\Epic Games\UE_5.27\Engine\Binaries\DotNET directory. The application can be accessed by double clicking the Swarm Agent icon in the Taskbar (System Tray). After this you will need to configure the following values accessible in the Settings tab:

  • AgentGroupName: ue5-swarm-aws
  • AllowedRemoteAgentGroup: ue5-swarm-aws
  • AllowedRemoteAgentNames: *
  • CoordinatorRemotingHost: <Add coordinator private IP from the Stack output>

Once this is done start experimenting with your environment. Here are some example tasks that you might want to test:

  • Create a streaming depot
  • Create a mainline stream
  • Create additional Perforce user and grant them access to the newly created streaming depot
  • Create a new workspace
  • Create a .p4ignore file for your Unreal project, mark it for add and submit it to the mainline stream
  • Set the correct typemap for an Unreal project
  • Start Unreal Engine 5
  • Create a new Unreal project
  • Close Unreal Engine 5
  • Move your Unreal project into your Perforce Workspace
  • Mark your Unreal project folder for add and submit it to the mainline stream
  • Open the Unreal project and setup Perforce integration
  • Reconfigure the EC2 Autoscaling Group to use more instances
  • Submit a lightning build and see your Unreal Engine 5 Swarm agents in action

If you have any questions regarding the steps outlined above feel free to reach out to us.

Please, don't forget stop the resources if you are not working with them, otherwise you will incur unnecessary cost. You can simply stop the Virtual Workstation, the Perforce Helix Core server and the Unreal Engine 5 Swarm Coordinator if you are taking a break and restart these instances when you want to continue to work with them. The same applies to the Unreal Engine 5 Swarm agents, with the only caveat that the start and termination of these instances is managed by an EC2 Auto Scaling group. To change the amount of running EC2 instances you need to modify the EC2 Auto Scaling group, see here.

Cleanup

To clean up this example you need to delete the CloudFormation stacks. Start by deleting the gpic-pipeline-virtual-workstation, followed by the gpic-pipeline-perforce-helix-core and gpic-pipeline-swarm-cluster stacks. Once all of these stacks are completely removed you can delete the gpic-pipeline-foundation stack. Additionally terminate any additional workstations launched from the AWS Marketplace.

With CDK you can delete the stacks with: Example commands:

 $  cdk destroy gpic-pipeline-virtual-workstation

After you have removed all stacks, two resources need to be deleted manually. First the S3 bucket will need to be deleted manually and second the AMI that was created in context of the Unreal Engine 5 Swarm stack needs to be deleted.

Example commands:

 $  aws s3 rb s3://<bucket-name> --force
 $
 $  aws ec2 deregister-image --image-id <AMI ID>

Extra Tips

How to access the EC2 instances in the private subnets

The Perforce Helix Core Server can be accessed via the AWS Systems Manager - Session Manager.

The Unreal Engine 5 Coordinator and agents can be accessed via RDP utilizing the Virtual Workstation. Alternatively you can use the AWS Systems Manager - Session Manager to open a Powershell session to these instances.

How to acess the Unreal Engine 5 Swarm Agent logs?

The Swarm Agent writes logs to: C:\ue5-swarm\SwarmCache\Logs. See the section above on how to connect to these EC2 instances.

Updating CloudFormation templates after code changes

If you do changes to the CDK code and want to generate new CloudFormation templates you will need to use following commands to keep the stack references in sync:

cdk synth gpic-pipeline-foundation -e > cloudformation/gpic-pipeline-foundation.yaml
cdk synth gpic-pipeline-perforce-helix-core -e > cloudformation/gpic-pipeline-perforce-helix-core.yaml
cdk synth gpic-pipeline-swarm-cluster -e > cloudformation/gpic-pipeline-swarm-cluster.yaml
cdk synth gpic-pipeline-virtual-workstation -e > cloudformation/gpic-pipeline-virtual-workstation.yaml

Security

See CONTRIBUTING for more information.

License

This library is licensed under the MIT-0 License. See the LICENSE file.

The AWS CloudFormation template downloads and installs Perforce Helix Core on EC2 instances during the process. Helix Core is a proprietary software and is subject to the terms and conditions of Perforce. Please refer to EULA in the following page for details. Perforce Terms of Use : https://www.perforce.com/terms-use

About

This example allows you to quickly evaluate some typical solution areas of a Game Production in the Cloud pipeline (GPIC). Remote workstation, version control and acceleration of computationally expensive process.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published