A foundational decision an organization will make when adopting Azure is settling on a networking architecture. The Microsoft Azure Cloud Adoption Framework and Azure Architecture Center can help you to align your organizational requirements for security and operations with the appropriate architecture. While these resources do a great job explaining the benefits and considerations of each architecture, they often lack details as to how a packet gets from point A to point B.
The traffic flows documented in this repository seek to fill this gap to provide the details of the packet flow and the options available to influence these flows to achieve security and operational goals. Additionally, they can act as a tool for learning the platform and troubleshooting issues with Azure networking.
The term NVA (network virtual appliance) is used in this repository to reference a security appliance, most commonly a next-gen firewall. NVAs will typically have a separate management interface in a dedicated management subnet which is not shown in these images. Examples of common NVAs used in Azure include Azure Firewall, Palo Alto, Fortinet, and Imperva.
For the purposes of this repository:
- Northbound traffic is traffic destined to the Internet
- Southbound traffic is traffic destined to Azure coming a source on the Internet
- Eastbound and westbound traffic is traffic between on-premises and Azure or between Azure virtual networks.
There are many viable patterns in Azure and this repository does not contain them all. Its goal is to document the most common packet flows to help aid in the understanding of Azure networking.
- On-premises to Azure
- Azure to Azure
- Azure to Internet using Public IP
- Azure to Internet using NAT Gateway
- Internet to Azure Non HTTP/HTTPS Traffic
- Cross Region Azure to Azure
- Dual NIC On-Premises to Azure
- Internet to Azure with HTTP/HTTPS Traffic Option 1 - Virtual Machines
- Internet to Azure with HTTP/HTTPS Traffic Option 2 - Virtual Machines
- Internet to Azure with HTTP/HTTPS Traffic with IDS IPS Option 1 - Virtual Machines
- Internet to Azure with HTTP/HTTPS Traffic with IDS IPS Option 2 - Virtual Machines
- Internet to Azure with HTTP/HTTPS Traffic Option 1 - Private Endpoints
- Internet to Azure with HTTP/HTTPS Traffic Option 2 - Private Endpoints
- Internet to Azure with HTTP/HTTPS Traffic with IDS IPS Option 1 - Private Endpoints
- Internet to Azure with HTTP/HTTPS Traffic with IDS IPS Option 2 - Private Endpoints
The patterns in this section assume the organization is deploying a single NVA stack that will handle north/south and east/west traffic. Each NVA is configured with three network interfaces. The first interface handles internal traffic (to and from on-premises to Azure or between Azure virtual networks). The second interface handles external traffic (to and from the Internet). The third interface, which is not pictured in these images, is used for management of the NVA.
Scenario: Machine on-premises initiates a connection to an application running in Azure.
Step | Path | Description |
---|---|---|
1 | A -> B | Machine traffic passes over ExpressRoute circuit to Virtual Network Gateway |
2 | B -> F | User defined route in route table assigned to GatewaySubnet directs traffic to internal load balancer for NVA |
3 | F -> E | Internal load balancer passes traffic to private interface for NVA |
4 | E -> G | NVA evaluates its rules, allows traffic, and passes it to internal load balancer for frontend application |
5 | G -> H | Internal load balancer for frontend application passes traffic to frontend application virtual machine |
6 | H -> F | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for NVA |
7 | F -> E | Internal load balancer passes traffic to originating NVA's private interface to preserve traffic symmetry |
8 | E -> B | NVA passes traffic to Virtual Network Gateway |
9 | B -> A | Virtual Network Gateway passes traffic over ExpressRoute circuit back to machine on-premises |
Scenario: Virtual machine in one spoke initiates connection to virtual machine in another spoke.
Step | Path | Description |
---|---|---|
1 | H -> F | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for NVA |
2 | F -> E | Internal load balancer passes traffic to NVA's private interface |
3 | E -> K | NVA evaluates its rules, allows traffic, and passes it to Active Directory domain controller virtual machine |
4 | K -> F | User defined route in route table assigned to domain controller subnet directs traffic to internal load balancer for NVA |
5 | F -> E | Internal load balancer passes traffic to originating NVA's private interface to preserve traffic symmetry |
6 | E -> H | NVA passes traffic back to frontend virtual machine |
Scenario: Virtual machine in Azure initiates a connection to a third-party website on the Internet and the NVA is configured with public IPs.
Step | Path | Description |
---|---|---|
1 | H -> F | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for NVA |
2 | F -> E | Internal load balancer passes traffic to NVA's private interface |
3 | E -> D | NVA evaluates its rules, allows traffic, routes to its public interface |
4 | D -> @ | Public interface NATs to its public IP and traffic is routed over the Internet to the third-party website |
5 | @ -> D | Third-party website passes traffic back to public IP of NVA's public interface |
6 | D -> E | NVA passes traffic rom its public interface to its private interface |
7 | E -> H | NVA passes traffic back to frontend virtual machine |
Scenario: Virtual machine in Azure initiates a connection to a third-party website on the Internet and the NVAs are configured to use NAT Gateway.
Step | Path | Description |
---|---|---|
1 | H -> F | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for NVA |
2 | F -> E | Internal load balancer passes traffic to NVA's private interface |
3 | E -> D | NVA evaluates its rules, allows traffic, routes to its public interface |
4 | D -> M | Public interface routes traffic to NAT Gateway via System Route |
5 | M -> @ | NAT Gateway NATs to its public IP and traffic is routed over the Internet to the third-party website |
6 | @ -> M | Third-party website passes traffic back to public IP of NAT Gateway |
7 | M -> D | NAT Gateway passes traffic back to NVA's public interface |
8 | D -> E | NVA routes traffic from its public interface to its private interface |
9 | E -> H | NVA passes traffic back to frontend virtual machine |
Scenario: User on the Internet initiates a connection to an application running in Azure. The application has been secured behind an Application Gateway for intra-region security and application-layer load balancing. The Application Gateway is located in the transit virtual network and is provided as a centralized service to all workloads. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration.
Benefits of this pattern include:
- Centralized administration of the Application Gateway which may fit the operational model of organizations new to Azure.
- Traffic sourced from the Internet is centrally ingressed through the transit virtual network which can be tightly controlled by central IT.
Considerations of this pattern include:
- Each instance of Application Gateway is limited to 100 backend address pools. This means that each instance of Application Gateway can provide its services for a maximum of 100 different applications where each application is an address pool.
- Misconfiguration or failure of the Application Gateway in a centralized model can result in a large blast radius.
- There is not an easy way to democratize control of the Application Gateway to development teams. This means Central IT will need to manage and modify the Application Gateway.
- Chargebacks to individual business units will be challenging.
Step | Path | Description |
---|---|---|
1 | @ -> O | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | O -> N | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | N -> M | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | M -> G | Application Gateway passes traffic from its internal interface to the internal load balancer for the web frontend |
5 | G -> H | Internal load balancer passes traffic to the web frontend virtual machine |
6 | H -> M | Web frontend virtual machine passes traffic to the Application Gateway private IP |
7 | M -> N | Application Gateway NATs to it's public IP |
8 | N -> O | Application Gateway passes traffic to Azure Front Door |
9 | O -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure. The application has been secured behind an Application Gateway for intra-region security and load application-level balancing. The Application Gateway is located in the workload virtual network and is dedicated to the workload. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration.
Benefits of this pattern include:
- The blast radius for misconfigurations or failures of the Application Gateway instance are limited to the individual workload.
- The reduction in risk can allow for further democratization of Azure resources providing more agility to workload owners.
- Enabling DDoS Network Protection on workload virtual network provides cost benefits of billing Application Gateway with WAF at the non-WAF rate.
- Chargebacks to individual business units is straightforward.
Considerations of this pattern include:
- Additional costs of an Application Gateway per workload
- Additional costs for DDoS Network Protection. DDoS Standard is licensed per 100 Public IPs and these IPs can be across multiple Virtual Networks in different subscriptions in the same Entra Tenant. Each Azure Application Gateway will consume at least one Public IP.
- Additional Azure Policy may also need to be introduced to ensure appropriate guardrails are put in place around secure configuration of Application Gateway.
Step | Path | Description |
---|---|---|
1 | @ -> O | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | O -> N | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | N -> M | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | M -> G | Application Gateway passes traffic from its internal interface to the internal load balancer for the web frontend |
5 | G -> H | Internal load balancer passes traffic to the web frontend virtual machine |
6 | H -> M | Web frontend virtual machine passes traffic to the Application Gateway private IP |
7 | M -> N | Application Gateway NATs to it's public IP |
8 | N -> O | Application Gateway passes traffic to Azure Front Door |
9 | O -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure. The application has been secured behind an Application Gateway for intra-region security and load balancing. The Application Gateway is located in the transit virtual network and is provided as a centralized service to all workloads. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration. An NVA is placed between the Application Gateway and the application to provide IDS/IPS functionality.
Reference the public documentation for additional ways to achieve this pattern.
Benefits of this pattern include:
- Traffic can be inspected and processed through an NVA's IDS/IPS engine.
- Centralized administration of the Application Gateway which may fit the operational model of organizations new to Azure.
- Traffic sourced from the Internet is centrally ingressed through the transit virtual network which can be tightly controlled by central IT.
Considerations of this pattern include:
- Additional hop through the NVA can introduce additional latency or act as a bottleneck.
- Each instance of Application Gateway is limited to 100 backend address pools. This means that each instance of Application Gateway can provide its services for a maximum of 100 different applications where each application is an address pool.
- Misconfiguration or failure of the Application Gateway in a centralized model can result in a large blast radius.
- There is not an easy way to democratize control of the Application Gateway to development teams. This means Central IT will need to manage and modify the Application Gateway.
- Chargebacks to individual business units will be challenging.
Step | Path | Description |
---|---|---|
1 | @ -> O | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | O -> N | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | N -> M | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | M -> F | User defined route in route table assigned to Application Gateway subnet directs traffic from Application Gateway internal interface to internal load balancer for NVA |
5 | F -> E | Internal load balancer passes traffic to private interface for NVA |
6 | E -> G | NVA evaluates its rules, allows traffic, and passes it to internal load balancer for frontend application |
7 | G -> H | Internal load balancer passes traffic to the web frontend virtual machine |
8 | H -> F | User defined route in route table assigned to web frontend subnet directs traffic to the internal load balancer for the NVA |
9 | F -> E | Internal load balancer passes traffic to the NVA |
10 | E -> M | NVA passes traffic to the Application Gateway internal interface |
11 | M -> N | Application Gateway NATs to it's public IP |
12 | N -> O | Application Gateway passes traffic to Azure Front Door |
13 | O -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure. The application has been secured behind an Application Gateway for intra-region security and load balancing. The Application Gateway is located in the workload virtual network and is dedicated to the workload. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration. An NVA is placed between the Application Gateway and the application to provide IDS/IPS functionality.
Reference the public documentation for additional ways to achieve this pattern.
Benefits of this pattern include:
- Traffic can be inspected and processed through an NVA's IDS/IPS engine.
- The blast radius for misconfigurations or failures of the Application Gateway instance are limited to the individual workload.
- The reduction in risk can allow for further democratization of Azure resources providing more agility to workload owners.
- Enabling DDoS Network Protection on workload virtual network causes Application Gateway with WAF to be billed at non-WAF rate.
- Chargebacks to individual business unit is straightforward.
Considerations of this pattern include:
- Additional hop through the NVA can introduce additional latency or act as a bottleneck.
- Additional costs of traffic traversing the peering to the transit virtual network to be processed by the NVA.
- Additional costs of an Application Gateway per workload
- Additional costs for DDoS Network Protection. DDoS Standard is licensed per 100 Public IPs and these IPs can be across multiple Virtual Networks in different subscriptions in the same Entra Tenant. Each Azure Application Gateway will consume at least one Public IP.
- Additional Azure Policy may also need to be introduced to ensure appropriate guardrails are put in place around secure configuration of Application Gateway.
Step | Path | Description |
---|---|---|
1 | @ -> O | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | O -> N | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | N -> M | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | M -> F | User defined route in route table assigned to Application Gateway subnet directs traffic from Application Gateway internal interface to internal load balancer for NVA |
5 | F -> E | Internal load balancer passes traffic to private interface for NVA |
6 | E -> G | NVA evaluates its rules, allows traffic, and passes it to internal load balancer for frontend application |
7 | G -> H | Internal load balancer passes traffic to the web frontend virtual machine |
8 | H -> F | User defined route in route table assigned to web frontend subnet directs traffic to the internal load balancer for the NVA |
9 | F -> E | Internal load balancer passes traffic to the NVA |
10 | E -> M | NVA passes traffic to the Application Gateway internal interface |
11 | M -> N | Application Gateway NATs to it's public IP |
12 | N -> O | Application Gateway passes traffic to Azure Front Door |
13 | O -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure in Azure App Services which has been secured behind a Private Endpoint. The application has been secured behind an Application Gateway for intra-region security and application-layer load balancing. The Application Gateway is located in the transit virtual network and is provided as a centralized service to all workloads. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration.
Benefits of this pattern include:
- Centralized administration of the Application Gateway which may fit the operational model of organizations new to Azure.
- Traffic sourced from the Internet is centrally ingressed through the transit virtual network which can be tightly controlled by central IT.
Considerations of this pattern include:
- Each instance of Application Gateway is limited to 100 backend address pools. This means that each instance of Application Gateway can provide its services for a maximum of 100 different applications where each application is an address pool.
- Misconfiguration or failure of the Application Gateway in a centralized model can result in a large blast radius.
- There is not an easy way to democratize control of the Application Gateway to development teams. This means Central IT will need to manage and modify the Application Gateway.
- Chargebacks to individual business units will be challenging.
Step | Path | Description |
---|---|---|
1 | @ -> N | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | N -> M | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | M -> L | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | L -> H | Application Gateway passes traffic from its internal interface to the private endpoint of the App Service |
5 | H -> L | App Service returns traffic to the Application Gateway private IP |
6 | L -> M | Application Gateway NATs to it's public IP |
7 | M -> N | Application Gateway returns traffic to Azure Front Door |
8 | N -> @ | Azure Front Door returns traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure in Azure App Services which has been secured behind a Private Endpoint. The application has been secured behind an Application Gateway for intra-region security and load application-level balancing. The Application Gateway is located in the workload virtual network and is dedicated to the workload. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration.
Benefits of this pattern include:
- The blast radius for misconfigurations or failures of the Application Gateway instance are limited to the individual workload.
- The reduction in risk can allow for further democratization of Azure resources providing more agility to workload owners.
- Enabling DDoS Network Protection on workload virtual network provides cost benefits of billing Application Gateway with WAF at the non-WAF rate.
- Chargebacks to individual business units is straightforward.
Considerations of this pattern include:
- Additional costs of an Application Gateway per workload
- Additional costs for DDoS Network Protection. DDoS Standard is licensed per 100 Public IPs and these IPs can be across multiple Virtual Networks in different subscriptions in the same Entra Tenant. Each Azure Application Gateway will consume at least one Public IP.
- Additional Azure Policy may also need to be introduced to ensure appropriate guardrails are put in place around secure configuration of Application Gateway.
Step | Path | Description |
---|---|---|
1 | @ -> N | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | N -> M | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | M -> L | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | L -> H | Application Gateway passes traffic from its internal interface to the private endpoint of the App Service |
5 | H -> L | App Service returns traffic to the Application Gateway private IP |
6 | L -> M | Application Gateway NATs to it's public IP |
7 | M -> N | Application Gateway returns traffic to Azure Front Door |
8 | N -> @ | Azure Front Door returns traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure in Azure App Services which has been secured behind a Private Endpoint. The application has been secured behind an Application Gateway for intra-region security and load balancing. The Application Gateway is located in the transit virtual network and is provided as a centralized service to all workloads. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration. An NVA is placed between the Application Gateway and the application to provide IDS/IPS functionality. Private Endpoint network policies have been enabled on the frontend subnet containing the App Services Private Endpoint enabling improved routing and network security group support.
Reference the public documentation for additional ways to achieve this pattern.
Benefits of this pattern include:
- Traffic can be inspected and processed through an NVA's IDS/IPS engine.
- Centralized administration of the Application Gateway which may fit the operational model of organizations new to Azure.
- Traffic sourced from the Internet is centrally ingressed through the transit virtual network which can be tightly controlled by central IT.
Considerations of this pattern include:
- Additional hop through the NVA can introduce additional latency or act as a bottleneck.
- Approriate settings for Private Endpoint Network Policies must be configured on Private Endpoint subnet
- Each instance of Application Gateway is limited to 100 backend address pools. This means that each instance of Application Gateway can provide its services for a maximum of 100 different applications where each application is an address pool.
- Misconfiguration or failure of the Application Gateway in a centralized model can result in a large blast radius.
- There is not an easy way to democratize control of the Application Gateway to development teams. This means Central IT will need to manage and modify the Application Gateway.
- Chargebacks to individual business units will be challenging.
Step | Path | Description |
---|---|---|
1 | @ -> N | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | N -> M | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | M -> L | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | L -> F | User defined route in route table assigned to Application Gateway subnet directs traffic from Application Gateway internal interface to internal load balancer for NVA |
5 | F -> E | Internal load balancer passes traffic to private interface for NVA |
6 | E -> H | NVA evaluates its rules, allows traffic, SNATs to its IP, and passes it to the private endpoint for the App Service |
7 | H -> E | App Service returns the traffic to the NVAs private IP |
8 | E -> L | NVA passes traffic to the Application Gateway internal interface |
9 | L -> M | Application Gateway NATs to it's public IP |
10 | M -> N | Application Gateway passes traffic to Azure Front Door |
11 | N -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure in Azure App Services which has been secured behind a Private Endpoint. The application has been secured behind an Application Gateway for intra-region security and load balancing. The Application Gateway is located in the workload virtual network and is dedicated to the workload. Azure Front Door is placed in front of the Application Gateway to provide inter-region security, load balancing, and site acceleration. An NVA is placed between the Application Gateway and the application to provide IDS/IPS functionality. Private Endpoint network policies have been enabled on the frontend subnet containing the App Services Private Endpoint enabling improved routing and network security group support.
Reference the public documentation for additional ways to achieve this pattern.
Benefits of this pattern include:
- The blast radius for misconfigurations or failures of the Application Gateway instance are limited to the individual workload.
- The reduction in risk can allow for further democratization of Azure resources providing more agility to workload owners.
- Enabling DDoS Network Protection on workload virtual network provides cost benefits of billing Application Gateway with WAF at the non-WAF rate.
- Chargebacks to individual business units is straightforward.
Considerations of this pattern include:
- Additional costs of an Application Gateway per workload
- Additional costs for DDoS Network Protection. DDoS Standard is licensed per 100 Public IPs and these IPs can be across multiple Virtual Networks in different subscriptions in the same Entra Tenant. Each Azure Application Gateway will consume at least one Public IP.
- Additional Azure Policy may also need to be introduced to ensure appropriate guardrails are put in place around secure configuration of Application Gateway.h Azure Application Gateway will consume at least one Public IP.
- Traffic must be source NATed at the firewall to ensure traffic symmetry.
Step | Path | Description |
---|---|---|
1 | @ -> N | User's machine sends traffic to Azure Front Door which terminates the TCP connection |
2 | N -> M | Azure Front Door establishes a new TCP connection with the Application Gateway's public IP and adds the user's public IP to the X-Forwarded-For header |
3 | M -> L | Application Gateway NATs to its private IP, appends the X-Forwarded-Header with the Azure Front Door public IP, performs its security and load balancing function |
4 | L -> F | User defined route in route table assigned to Application Gateway subnet directs traffic from Application Gateway internal interface to internal load balancer for NVA |
5 | F -> E | Internal load balancer passes traffic to private interface for NVA |
6 | E -> H | NVA evaluates its rules, allows traffic, SNATs to its IP, and passes it to the private endpoint for the App Service |
7 | H -> E | App Service returns the traffic to the NVAs private IP |
8 | E -> L | NVA passes traffic to the Application Gateway internal interface |
9 | L -> M | Application Gateway NATs to it's public IP |
10 | M -> N | Application Gateway passes traffic to Azure Front Door |
11 | N -> @ | Azure Front Door passes traffic to user's machine |
Scenario: User on the Internet initiates a connection to an application running in Azure. The application is served up using a protocol that IS NOT HTTP/HTTPS. An NVA is placed between the Internet and the application.
Step | Path | Description |
---|---|---|
1 | @ -> C | User's machine sends traffic to the public IP address of an external Azure Load Balancer |
2 | C -> D | External load balancer passes the traffic through the Azure software defined network which preserves the user's machine's public IP as the source and passes it on to the NVA's public interface |
3 | D -> E | NVA evaluates its rules, allows traffic, routes the traffic to its private interface and NATs to its private IP |
4 | E -> G | NVA passes traffic to the internal load balancer for the web frontend |
5 | G -> H | Web frontend internal load balancer passes traffic to web frontend virtual machine |
6 | H -> E | Web frontend virtual machine passes traffic to NVA's private interface |
7 | E -> D | NVA routes traffic from its private interface to its public interface |
8 | D -> @ | NVA passes traffic back through the Azure software defined network where its source IP address is rewritten to the external load balancer's public IP |
Scenario: Workload running in spoke 1 region 1 needs to communicate with workload in spoke 2 region 2.
Step | Path | Description |
---|---|---|
1 | H -> F | User defined route in route table assigned to web frontend subnet directs traffic to the internal load balancer for NVA in region 1 |
2 | F -> E | Internal load balancer passes traffic to private interface of NVA in region 1 |
3 | E -> FF | NVA in region 1 evaluates its rules, allows traffic, and user defined route in route table assigned to private interface subnet directs traffic to internal load balancer for NVA in region 2 |
4 | FF -> EE | Internal load balancer passes traffic to private interface of NVA in region 2 |
5 | EE -> GG | NVA in region 2 evaluates its rules, allows traffic, and passes it to internal load balancer for frontend application in region 2 |
6 | GG -> HH | Web frontend internal load balancer passes traffic to web frontend virtual machine |
7 | HH -> FF | User defined route in route table assigned to web frontend subnet directs traffic to the internal load balancer for NVA in region 2 |
8 | FF -> EE | Internal load balancer passes traffic to private interface of NVA in region 2 |
9 | EE -> F | User defined route in route table assigned to NVA in region 2 private interface subnet directs traffic to internal load balancer of NVA in region 1 |
10 | F -> E | Internal load balancer passes traffic to private interface of NVA in region 1 |
11 | E -> G | NVA in region 1 passes traffic to the internal load balancer for the web frontend in region 1 |
12 | G -> H | Web frontend internal load balancer passes traffic to web frontend virtual machine |
This pattern adds an additional network interface to the NVA to handle east/west traffic. In some NVAs this allows the user to establish separate security zones for each network interface allowing for a more traditional administration experience. It is often adopted by organizations who have not yet adopted traditional processes to cloud.
Organizations should avoid this pattern and instead use a single network interface for all east and west traffic as illustrated in the other patterns in this repository. Using this pattern will introduce significant additional administration overhead because traffic will be passes through separate load balancers attached to the same NVA forcing SNAT to be used. This is due to mismatches in the five-tuple hash algorithm Azure Load Balancers use. This video by John Savill goes into more detail.
Microsoft does not recommend this pattern and organizations should consult with their NVA vendor for best practices on implementing NVAs in Azure.
Note that this pattern does not show the public interface or management interface in this diagram.
Scenario: Machine on-premises initiates a connection to an application running in Azure.
Step | Path | Description |
---|---|---|
1 | A -> B | Machine traffic passes over ExpressRoute circuit to Virtual Network Gateway |
2 | B -> C | User defined route in route table assigned to GatewaySubnet directs traffic to internal load balancer for NVA's untrusted interface |
3 | C -> D | Internal load balancer passes traffic to untrusted interface of NVA |
4 | D -> E | NVA evaluates its rules, allows traffic, routes to its trusted interface |
5 | E -> G | NVA SNATs to its trusted interface IP and passes traffic to internal load balancer for frontend application |
6 | G -> H | Internal load balancer for frontend application passes traffic to frontend application virtual machine |
7 | H -> E | Frontend application virtual machine passes traffic directly back to NVA's trusted interface |
8 | E -> D | NVA routes traffic from its trusted interface to its untrusted interface |
9 | D -> B | NVA passes traffic to Virtual Network Gateway |
10 | B -> A | Virtual Network Gateway passes traffic over ExpressRoute circuit back to machine on-premises |
The patterns in this section assume the organization is deploying a separate NVA stack for north/south and east/west traffic. This is done to mitigate the risk of a bottle neck or failure to north/south traffic affecting east/west traffic and vice versa.
The north/south NVAs are configured wwith three network interfaces. The first interface handles private traffic (to and from on-premises to Azure or within Azure). The second interface handles public traffic (to and from the Internet). The third interface, which is not pictured in these images, is used for management of the NVA.
The east/west NVAs are configured with two interfaces. The first interface handles payload traffic (to and from on-premises to Azure or within Azure). The second interface, which is not pictured in these images, is used for management of the NVA.
Scenario: Virtual machine in one spoke initiates connection to virtual machine in another spoke.
Step | Path | Description |
---|---|---|
1 | A -> B | Machine traffic passes over ExpressRoute circuit to Virtual Network Gateway |
2 | B -> C | User defined route in route table assigned to GatewaySubnet directs traffic to internal load balancer for east/west NVA |
3 | C -> D | Internal load balancer passes traffic to payload interface for east/west NVA |
4 | D -> I | NVA evaluates its rules, allows traffic, and passes it to internal load balancer for frontend application |
5 | I -> J | Internal load balancer for frontend application passes traffic to frontend application virtual machine |
6 | J -> C | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for east/west NVA |
7 | C -> D | Internal load balancer passes traffic to originating east/west NVA's payload interface to preserve traffic symmetry |
8 | D -> B | NVA passes traffic to Virtual Network Gateway |
9 | B -> A | Virtual Network Gateway passes traffic over ExpressRoute circuit back to machine on-premises |
Scenario: Virtual machine in Azure initiates a connection to a third-party website on the Internet and the NVA is configured with public IPs.
Step | Path | Description |
---|---|---|
1 | J -> H | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for north/south NVA |
2 | H -> G | Internal load balancer passes traffic to north/south NVA's private interface |
3 | G -> F | North/south NVA evaluates its rules, allows traffic, routes to its public interface |
4 | F -> @ | Public interface NATs to its public IP and traffic is routed over the Internet to the third-party website |
5 | @ -> F | Third-party website passes traffic back to public IP of north/south NVA's public interface |
6 | F -> G | North/south NVA routes traffic from its public interface to its private interface |
7 | G -> J | North/south NVA passes traffic back to frontend virtual machine |
Scenario: Virtual machine in Azure initiates a connection to a third-party website on the Internet and the NVAs are configured to use NAT Gateway.
Step | Path | Description |
---|---|---|
1 | J -> H | User defined route in route table assigned to frontend subnet directs traffic to internal load balancer for north/south NVA |
2 | H -> G | Internal load balancer passes traffic to north/south NVA's private interface |
3 | G -> F | North/south NVA evaluates its rules, allows traffic, routes to its public interface |
4 | F -> M | Public interface routes traffic to NAT Gateway via System Route |
5 | M -> @ | NAT Gateway NATs to its public IP and traffic is routed over the Internet to the third-party website |
6 | @ -> M | Third-party website passes traffic back to public IP of NAT Gateway |
7 | M -> F | NAT Gateway passes traffic back to north/south NVA's public interface |
8 | F -> G | North/south NVA routes traffic from its public interface to its private interface |
9 | G -> J | North/south NVA passes traffic back to frontend virtual machine |