Data Infrastructure

Azure, Data Infrastructure

Azure SQL VM Best Practices: VM Sizing

  Overview When we run SQL Server on Azure Virtual Machines, we can use the same database performance tuning options applicable to SQL Server in an on-premises environment.   For Azure SQL server VM, best practice starts with collecting the baseline requirements for the processing unit and the storage; for this, we have to know the IOPS, throughput, and latency of the source. Also noteworthy is that the CPU and Memory requirement also gather data during peak hours, such as workloads during your business hours, and other high load processes such as end-of-day processing or weekend ETL workloads. Then, use the performance analysis to select the VM Size that can scale to your workload performance requirements. However, the performance of a relational(SQL) database in the cloud depends on many factors, such as the size of a virtual machine, the configuration of the data disks, SQL Server features, Azure Features, High availability, and disaster recovery (HADR) features and also the Security optimization.   Consider scaling up your resources for typically heavy workloads such as end-of-quarter processing, and then scale done once the workload completes. There is typically a trade-off between reducing costs and optimizing for performance. This performance best practices series focuses on getting the best performance for SQL Server on Azure Virtual Machines. If your workload is less demanding, you might not require every recommended optimization. Consider your performance needs, costs, and workload patterns as you evaluate these recommendations. It is about setting up or opting for the best infrastructure for a SQL Server on the Azure cloud. However, if the peak workload at your end is less demanding, you might not require all recommended optimizations. Inspect your performance needs, work pattern, and most importantly, the cost involved as you evaluate these recommendations.   The following is a check-list for optimal performance of SQL Server on Azure Virtual Machines: – VM Size – Storage and Scalability – Baseline Requirements – Security Considerations – SQL Server Features   Since this article is Azure SQL VM and no one knows Azure and SQL better than Microsoft, we will post the link to relevant Microsoft documentation for better understanding and transparency wherever needed. VM Size Here is a quick check-list of VM size best practices for running your SQL Server on Azure VM: · Use VM sizes with four or more vCPU. · Use memory-optimized virtual machine sizes for the best performance of SQL Server workloads. · Consider a higher memory-to-vCore ratio for mission-critical and data warehouse workloads. · Collect the target workload performance characteristics (IOPS) and use them to determine the appropriate VM size for your business.   A vCPU is the abbreviation for virtual centralized processing unit and represents a portion or share of the underlying, physical CPU assigned to a particular virtual machine (VM). A Physical CPU can have multiple vCPU, and this vCPU can be managed by controller software like Hyper-V. In the past, there was a standard procedure that there were eight vCPUs per core. Today, vCPU count is primarily determined by the manufacturer. It is determined by taking the number of processing threads that a chipset offers per core and multiplying the occupied sockets. This is how it looks: (Threads x Cores) x Physical CPU = Number vCPU. If you have small VMs that barely use CPU time, you could quickly get 20-30 VMs from an 8-core server. But, if you have larger workloads such as a database server, you will have far fewer VMs from that same eight-core server. Thus, it is all about resource utilization and allocation.   A few questions to consider before we get started: · Do our apps run at 100% CPU utilization all the time? · Do they have periods where utilization bursts? · Do they have maintenance windows? · When you know the requirements, you can make an informed decision on the underlying hardware. There is a large array of machines in the Azure cloud that we can select from; still, we will emphasize only on D, E, and M family machines since we refer only to SQL workloads. SQL Server database engine performance is more memory-driven as compared to a number of cores available .eg: Dsv2 11-15, Edsv4, Mv2, Msv2, M, Mdsv2, DCsv2, Dasv4, Dsv3, Ddsv4, Lsv2. These can be broadly categorized into General purpose, Memory, and Storage Optimized machines.   Some of the naming convention for machines are as follows: A – AMD based processor D – Disk(local temp, the temporary disk is present) M – The most amount of Memory in a particular size S – Premium Storage capable. I – Isolated storage V – Version   D series: Has a consistent memory to core ratio of 7:1, medium to large caches, and in-memory analytics. Supports Premium Storage and are entry-level SQL Server virtual machines. E Series: Has a Memory to core ratio of 8:1; these are ideal for Memory intensive enterprise applications, also have sizable local storage SSD capacity and throughput. This acts as a solid all-purpose SQL Server virtual machine. M Series: Has a core to memory ratio of 30:1 without constrained cores and 122:1 with constrained cores. These offer vCore counts and Memory for the largest SQL Server workloads. These are mission-critical and Data Warehouse Virtual Machines. Also, it is the point to note that these costs are enormous, so we need to manage them accordingly.   Also, it is essential to understand the concept of constrained cores. Some database workloads like SQL Server or Oracle require high Memory, storage, and I/O bandwidth, but not a high core count. Conversely, many database workloads are not CPU-intensive. Azure offers specific VM sizes to constrain the VM vCPU count to reduce software licensing costs while conserving the same Memory, storage, and I/O bandwidth. For example, the vCPU count can be constrained to one-half or one-quarter of the original VM size. For Example – Standard_M16ms comes with 16 cores and 437.5 GB of Memory(RAM).   Screenshot taken from M-series – Azure Virtual Machines| Microsoft Docs  

AI/ML, Azure, Data Infrastructure, Data Management, Data Science, Data/Performance Analytics, Decision Making, Microsoft Power BI

Data Democratization: Stage 6: Lock them up – Secure your data

Data needs to be protected just as networks are protected from malicious actors (internal and external). Data security should not be an afterthought but a design decision from the start. Enterprises have spent a considerable amount of resources securing their data centers, monitoring access to sensitive data, and yet there are significant data breaches every year. It has been reported that between January and September 2019, there were over 7.9 billion data records exposed — a 33% increase from the same time in 2018! The year 2020 turned out to be worse than 2019, and 2021 is on pace to be the worst year so far.   Modern data architectures in the cloud make it easy for enterprises of any size to adopt these standards, but it needs to be augmented with resources from the internal IT team. In addition, the modern cloud infrastructures are monitored, deploying the best processes, top-notch technology, and expert security analysts. Hence, everyone reaps the benefit of best practices deployed universally in the cloud and continued fixes for any security flaws in the technology stack.   The innovation of modern cloud computing, such as Azure, has lowered the barrier to setting up world-class secure decision support systems.

AI/ML, Data Infrastructure, Data Management, Data Science, Data/Performance Analytics, Decision Making, Oracle

Data Democratization: Stage 5, Executive Actions

It is said that there are no million-dollar ideas, only million-dollar actions! Hence, all thoughts, plans, and decisions need follow-up actions. So for these actions to be most effective, organizations need to drive activities driven by data.   The executive actions in successful enterprises are taken by the executives and undertaken every day by all employees. Enterprise change management is a complex process, and it needs a change of culture. Data needs to tell a story that resonates with employees at every level, motivating them to continue to take incremental steps (actions) to move the ball towards the goal post.   While such actions are set in motion at every department, team level, there must be continual monitoring of whether efforts achieve desired results. As mentioned in earlier sections, the change is constant. Defining measurable key performance indices and continuous tracking of teams provide a path to make mid-flight adjustments. Such feedback loops improve success chances while mitigating the risk posed by changing conditions, both internal and external.   Accelerant has created data-centric playbooks for various departments such as Finance, Supply Chain, Operations, Sales, and Marketing. Our solutions are hosted on Microsoft Azure and reap the benefits of scalable cloud from Microsoft. We can assist you in defining measurable KPIs and putting together a framework to create closed-loop monitoring for continuous improvement.

AI/ML, Azure, Data Infrastructure, Data Management, Data Science, Data/Performance Analytics, Decision Making

Data Democratization: Stage 4, Best is yet to come (Artificial Intelligence)

Top-performing organizations extract incremental values from various business processes. A combination of small incremental gains results in a significant change that benefits multiple departments across the organization. For example, incremental improvements in the supply chain result in better costs while also improving customer satisfaction with faster delivery. This results in more customers coming back for repeat business, thus delivering better sales. However, it must be emphasized that data initiatives need continuous effort from the various stakeholders responsible for the business performance. It is a company culture where every employee gets access to data relevant to his/her job. Every employee understands that decisions or opinions need to be backed by data and the “gut-feel.”   When data is harmonized and analyzed descriptively, organizations should start investing in the ultimate prize of achieving higher automation levels via AI and machine learning. AI and machine learning open doors to opportunities that offer unprecedented efficiency gains, e.g., autonomous driving. Organizations must do a self-assessment of data maturity, without which all such initiatives are bound to fail. AI/Machine Learning techniques provide exceptional value provided the training data sets are pristine. Such models need periodic adjustments to account for various data drag.   Data is the rocket fuel that leapfrogs businesses ahead of the competition. Investments in data infrastructure, analysis capabilities, and data-driven decision-making culture are vital to keeping a competitive edge.

Activate Your Free 15-Day Infominer Trial

X
Scroll to Top