Presentation is loading. Please wait.

Presentation is loading. Please wait.

Windows Server 2012 Networking Performance and Management Ahmed Talat Senior Program Manager Lead Microsoft Corporation Prabu Rambadran Product Marketing.

Similar presentations


Presentation on theme: "Windows Server 2012 Networking Performance and Management Ahmed Talat Senior Program Manager Lead Microsoft Corporation Prabu Rambadran Product Marketing."— Presentation transcript:

1 Windows Server 2012 Networking Performance and Management Ahmed Talat Senior Program Manager Lead Microsoft Corporation Prabu Rambadran Product Marketing Manager Microsoft Corporation WSV304

2

3

4 Beyond Virtualization Windows Server 2012 offers a dynamic, multi-tenant infrastructure that goes beyond virtualization to provide maximum flexibility for delivering and connecting to cloud services. Modern Workstyle, Enabled Windows Server 2012empowers IT to provide users with flexible access to data and applications from virtually anywhere on any device with a rich user experience, while simplifying management and helping maintain security, control and compliance. The Power of Many Servers, the Simplicity of One Windows Server 2012 offers excellent economics by integrating a highly available and easy to manage multi-server platform with breakthrough efficiency and ubiquitous automation. Every App, Any Cloud Windows Server 2012 is a broad, scalable and elastic server platform that gives you the flexibility to build and deploy applications and websites on-premises, in the cloud and in a hybrid environment, using a consistent set of tools and frameworks.

5

6

7 title Resource Metering

8 A two-tenant environment built with Hyper-V in Windows Server 2012

9 Basic model of Resource Metering PowerShell Cmdlets Enable resource Metering: Get-VM -ComputerName | Enable-VMResourceMetering Get VM metering data: Get-VM -ComputerName -Name | Measure-VM

10 title IP Address Management(IPAM)

11

12

13 Performance Features and Improvements

14 Processor. One physical processor, which can consist of one or more nodes. A physical processor is the same as a package, a socket, or a CPU. Non-uniform memory architecture (NUMA) node. A set of logical processors and cache that are close to one another. Core. One processing unit, which can consist of one or more logical processors. Logical processor (LP). One logical computing engine from the perspective of the operating system, application, or driver. In effect, a logical processor is a thread. Kernel Group. A set of up to 64 logical processors. Affinity. A preference indicated by a thread, process, or interrupt for operation on a particular processor, node, or group. Processor or Socket NUMA Node Core LP Kernel Group (KGroup) Processor or Socket NUMA Node Core LP Processor or Socket NUMA Node Core LP Processor or Socket NUMA Node Core LP Processor or Socket NUMA Node Core LP

15 Latency. The time required for an operation to complete. Lower is better. Scalability. The ability to adapt to increasing demand on system resources. Higher is better. Throughput. The amount of data transferred or processed in a given time period. Higher is better. Path Length. The number of CPU cycles divided by the throughput. Lower is better. Variability / Jitter. Fluctuation in throughput and/or latency. Lower is better. L STPJ

16 IT Pro has received a heavy workload and CPU utilization is almost 100%. Unable to increase receive throughput Examples: Backup Server or FTP Server RSC can help by coalescing multiple inbound packets into a larger buffer and effectively reduce per packet CPU cost Preliminary data shows up to a 20% reduction in CPU cycles depending on the workload All inbox 10G network adapters support the feature and have it on by default RSC: Multiple Inbound Packets Coalesced to reduce CPU load TP

17 demo Receive Segment Coalescing (RSC)

18 Developer is writing an Application that needs to send and receive data at microsecond granularity. Examples: 3D image processing or stock trading application Registered I/O (RIO) delivers low latency by pinning the application memory and reduces CPU cost RIO improves predictability with message latency (jitter) Example APIs: RIORegisterBuffer and RIOReceiveRIORegisterBufferRIOReceive TPLSJ

19 IT Pro wants to fine tune their system for low latency application and minimize jitter Best Practices include: Disable C-states from BIOS Set Power Policy to High Performance Use powercfg utility Disable NIC Interrupt Moderation setting Disable Flow Control Disable Receive Side Scaling Disable Receive Segment Coalescing (RSC) Use thread affinity in Task Manager to partition work by nodes Check out us/library/hh aspxhttp://technet.microsoft.com/en- us/library/hh aspx

20 Winsock Application (E.g. Using localhost as TCP destination) Winsock Application (E.g. Using localhost as TCP destination) AFD TCP/IP NDIS kernel user Win08 R2 Loopback Path Windows 2008 R2 Windows Server 2012 Loopback Path New: Shorter Path Developer looking for an efficient and fast method to share data between processes on their Server. Examples: Database Servers and Business Applications like SAP Customer should use new IOCTL SIO_LOOPBACK_FAST_PATH socket option to take new shorter path Both sides of connection must set new option but cant have SO_OOBINLINE, TCP_EXPEDITED_1122, or TCP_STDURG socket options TPL

21 demo TCP Loopback fast path

22 IT Pro has received a heavy workload that is CPU bound and wants to scale their workload across CPUs Examples: Web Server and File Server Use RSS to distribute the receive network traffic either across multiple processors or multiple NUMA nodes IT Pros should use new Power Shell cmdlets to control RSS behavior based on their workload characteristics Dynamic Static NUMA aware Non NUMA Profile 1 Profile 2Profile 3 (default) Profile 0 (like WS08 R2) TS

23 NIC with 8 RSS queues Node 0Node 1 Node 2 Node 3 queues Incoming Packets New RSS profiles improve scalability on multi-node servers by distributing TCP/UDP receive traffic across nodes

24 Throughput Single Process Hardware: 40LPs, 4 NUMA nodes, 10GigE, 128GB RAM Workload: mixed Web scenario Note: Measurements reflect release defaults 50% 201% Multiple Processes (2 and 4) Throughput 43% 115% Overall Web Server Scalability Improvements New K-Group awareness in the IIS thread pool Partitioning network I/O processing by K-Groups when using multiple NICs RSS NUMA profiles distribute network I/O processing across NUMA nodes

25 An IT Pro wants to run more VMs on a physical machine but cant because the incoming packet processing is saturating a limited set of CPUs on the physical host Examples: Private Cloud Deployments Use VMQ to dynamically distribute the receive network traffic across multiple VMs based on destination and network load Most inbox 10Gig Ethernet adapters support the feature which is enabled by default TS

26 Network IO path without VMQ Network IO path with VMQ Windows Server 2008 R2: Offloading routing and filtering of network packets to the network card (enabled by hardware-based receive queues) to reduce host overhead New in Windows Server 2012: Dynamically distributing incoming network traffic processing to host processors (based on processor utilization and network load)

27 1.8x scaling from 40VMs on 80 LPs to 80VMs on 160 LPs with dynamic VMQ Hardware: 160LPs, 8 NUMA nodes, 10Gigabit NIC, 512GB RAM Workload: 2VPs/VM mixed Web scenario Scenarios: Web Hosters 80LPs160LPs Throughput Larger is better

28 title Performance Analysis and Tools

29 demo Identifying performance bottlenecks using SPA 3.0

30 Original Setting is derived from system tuning, empirical evidence EXAMPLE RULE

31

32 EVERYTHING is configurable through PowerShell

33

34 Performance Metric Loopback Fast Path Registered I/O (RIO) Large Send Offload (LSO) Receive Segmentation Offload (RSC) Receive Side Scaling (RSS) Virtual Machine Queues (VMQ) Remote DMA (RDMA) Single Root I/O Virtual (SR-IOV) Lower End- to-End Latency XXXX Higher Scalability XXX Higher Throughput XXXXXXXX Lower Path Length XXXXXX Lower Variability X

35

36 VIR303 – An Overview of Hyper-V Networking in Windows Server 2012 VIR305 – Hyper-V Network Virtualization for Scalable Multi-Tenancy in Windows MGT315 - Network Management in Microsoft System Center 2012 SP1 WSV314 – Windows Server 2012 NIC Teaming and Multichannel Solutions Visit networking booth at the partner pavillion to learn more

37 WSV321: Windows PowerShell Crash Course WSV307: Windows Server 2012 IP Address Management WSV314: Windows Server 2012 NIC Teaming and Multichannel Solutions VIR303: An Overview of Hyper-V Networking in Windows Server 2012 VIR305: Hyper-V Network Virtualization for Scalable Multi-Tenancy in Windows VIR307: Get Hands-on with the New Hyper-V Extensible Switch in Windows Server 2012 WSV14 – HOL: Managing Your Network Infrastructure with IP Address Management WSV24 – HOL: Managing Network Infrastructure with Windows Server 2012

38 Talk to our Experts at the TLC #TE(sessioncode) DOWNLOAD Windows Server 2012 Release Candidate microsoft.com/windowsserver Hands-On Labs DOWNLOAD Windows Azure Windowsazure.com/ teched

39 Connect. Share. Discuss. Learning Microsoft Certification & Training Resources TechNet Resources for IT Professionals Resources for Developers

40 Required Slide Complete an evaluation on CommNet and enter to win!

41 Scan the Tag to evaluate this session now on myTechEd Mobile

42

43


Download ppt "Windows Server 2012 Networking Performance and Management Ahmed Talat Senior Program Manager Lead Microsoft Corporation Prabu Rambadran Product Marketing."

Similar presentations


Ads by Google