Many People know FastTrack as a reference architecture for relational databases. The goal of this guideline is to provide a reference architecture for scalable and fast Analysis Services solutions.
The document discusses troubleshooting performance issues for SQL Server. It begins with an introduction and case study on the MS Society of Canada's website. It then discusses optimizing the environment, using Performance Monitor (PerfMon) to monitor performance, and concludes with recommendations to address issues like high CPU usage, slow disk speeds, and insufficient memory.
The document discusses using Automatic Workload Repository (AWR) to analyze IO subsystem performance. It provides examples of AWR reports including foreground and background wait events, operating system statistics, wait histograms. The document recommends using this data to identify IO bottlenecks and guide tuning efforts like optimizing indexes to reduce full table scans.
Performance Tuning With Oracle ASH and AWR. Part 1 How And Whatudaymoogala
The document discusses various techniques for identifying and analyzing SQL performance issues in an Oracle database, including gathering diagnostic data from AWR reports, ASH reports, SQL execution plans, and real-time SQL monitoring reports. It provides an overview of how to use these tools to understand what is causing performance problems by identifying what is slow, quantifying the impact, determining the component involved, and analyzing the root cause.
This document provides an overview of SQL Server performance tuning. It discusses monitoring tools and dynamic management views that can be used to identify performance issues. Several common performance problems are described such as those related to CPU, memory, I/O, and blocking. The document also covers query tuning, indexing, and optimizing joins. Overall it serves as a guide to optimizing SQL Server performance through monitoring, troubleshooting, and addressing issues at the server, database, and query levels.
This document discusses how to optimize performance in SQL Server. It covers:
1) Why performance tuning is necessary to allow systems to scale, improve performance, and save costs.
2) How to optimize SQL Server performance by addressing CPU, memory, I/O, and other factors like compression and partitioning.
3) How to optimize the database for performance through techniques like schema design, indexing, locking, and query optimization.
Sql server performance tuning and optimizationManish Rawat
Sql server performance tuning and optimization
SQL Server Concepts/Structure
Performance Measuring & Troubleshooting Tools
Locking
Performance Problem : CPU
Performance Problem : Memory
Performance Problem : I/O
Performance Problem : Blocking
Query Tuning
Indexing
SQL Server Query Tuning Tips - Get it Right the First TimeDean Richards
Whether you are a developer or DBA, this presentation will outline a method for determining the best approach for tuning a query every time by utilizing response time analysis and SQL Diagramming techniques. Regardless of the complexity of the statement or database platform being utilized (this method works on all), this quick and systematic approach will lead you down the correct
tuning path with no guessing. If you are a beginner or expert, this approach will save you countless hours tuning a query.
This document summarizes the main parts of an Oracle AWR report, including the snapshot details, load profile, top timed foreground events, time model statistics, and SQL section. The time model statistics indicate that 86.45% of database time was spent executing SQL statements. The top foreground event was waiting for database file sequential reads, taking up 62% of database time.
Jugal Shah has over 14 years of experience in IT working in roles such as manager, solution architect, DBA, developer and software engineer. He has worked extensively with database technologies including SQL Server, MySQL, PostgreSQL and others. He has received the MVP award from Microsoft for SQL Server in multiple years. Common causes of SQL Server performance problems include configuration issues, design problems, bottlenecks and poorly written queries or code. Various tools can be used to diagnose issues including dynamic management views, Performance Monitor, SQL Server Profiler and DBCC commands.
Progress OpenEdge database administration guide and referenceVinh Nguyen
The document is a guide and reference for Progress Database Administration. It discusses Progress database architecture, storage design, configuration variables, limits, and procedures for creating and deleting databases. The guide contains information to help plan, administer, and optimize Progress databases.
This document provides an overview of performance tuning for SQL Server databases. It discusses eliminating obvious issues, understanding common bottlenecks like CPU, memory, disk and locking, and using tools like wait statistics, performance counters and dynamic management views to analyze workload and identify tuning opportunities. The document outlines a 12 step process for capturing a production workload, replaying it on a test server, making tuning changes and reanalyzing. It also demonstrates various tools and techniques like SQL Profiler, SQL Diagnostic Manager and RML Utilities.
Hardware planning & sizing for sql serverDavide Mauri
This document provides an overview of hardware planning and sizing considerations for SQL Server. It discusses that performance is the typical requirement for relational database management systems. While high performance is expected, typical server hardware configurations often result in unbalanced systems that are not optimized. The document advocates for balanced systems with no single bottleneck. It provides guidance on evaluating CPU, memory, I/O capabilities and storage to ensure a system can handle peak resource consumption. Baseline testing is recommended to compare hardware performance.
This document provides an overview of performance monitoring and optimization for SQL Server databases. It discusses monitoring database activity using tools like SQL Profiler and Activity Monitor, identifying bottlenecks, using the Database Engine Tuning Advisor to generate optimization recommendations, and addressing issues related to processes, locking, and deadlocks. Best practices emphasized establishing a performance baseline, making incremental changes while measuring impact, and focusing on specific issues to optimize real-world workloads.
This document discusses PostgreSQL parameter tuning, specifically related to memory and optimizer parameters. It provides guidance on setting parameters like shared_buffer, work_mem, temp_buffer, maintenance_work_mem, random_page_cost, sequential_page_cost, and effective_cache_size to optimize performance based on hardware characteristics like available RAM and disk speed. It also covers force_plan parameters that can include or exclude certain query optimization techniques.
Whitepaper: Exadata Consolidation Success StoryKristofferson A
1. The document discusses database and server consolidation using Oracle Exadata and describes the challenges of managing highly consolidated environments to ensure quality of service.
2. It outlines a 4-step process for accurate provisioning and capacity planning using a tool called the Provisioning Worksheet: collecting database details, defining the target Exadata hardware capacity, creating a provisioning plan, and reviewing resource utilization.
3. The process relies on basic capacity planning to ensure workload requirements fit available capacity. Database CPU and storage requirements are gathered, a target Exadata configuration is set, databases are mapped to nodes in the plan, and final utilization is summarized to identify any capacity shortfalls.
The document discusses SQL Server performance monitoring and tuning. It recommends taking a holistic view of the entire system landscape, including hardware, software, systems and networking components. It outlines various tools for performance monitoring, and provides guidance on identifying and addressing common performance issues like high CPU utilization, disk I/O issues and poorly performing queries.
SQL Server Tuning to Improve Database PerformanceMark Ginnebaugh
SQL Server tuning is a process to eliminate performance bottlenecks and improve application service. This presentation from Confio Software discusses SQL diagramming, wait type data, column selectivity, and other solutions that will help make tuning projects a success, including:
•SQL Tuning Methodology
•Response Time Tuning Practices
•How to use SQL Diagramming techniques to tune SQL statements
•How to read executions plans
The document discusses Oracle Database performance tuning. It begins by defining performance as the accepted throughput for a given workload. Performance tuning is defined as optimizing resource use to increase throughput and minimize contention. A performance problem occurs when database tasks do not complete in a timely manner, such as SQL running longer than usual or users facing slowness. Performance problems can be caused by contention for resources, overutilization of the system, or poorly written SQL. The document discusses various performance diagnostics tools and concepts like wait events, enqueues, I/O performance, and provides examples of how to analyze issues related to these areas.
Find and fix SQL Server performance problems fasterSolarWinds
Great DBAs must be able to quickly identify problems with SQL Server instances. In this presentation, you will learn how to quickly identify where your problems are using tools such as:
*Dynamic Management Views
*Query Execution Plans
*Windows Performance Monitor
*Extended Events
*Third-party tools (including SolarWinds Database Performance Analyzer)
This document provides an overview and interpretation of the Automatic Workload Repository (AWR) report in Oracle database. Some key points:
- AWR collects snapshots of database metrics and performance data every 60 minutes by default and retains them for 7 days. This data is used by tools like ADDM for self-management and diagnosing issues.
- The top timed waits in the AWR report usually indicate where to focus tuning efforts. Common waits include I/O waits, buffer busy waits, and enqueue waits.
- Other useful AWR metrics include parse/execute ratios, wait event distributions, and top activities to identify bottlenecks like parsing overhead, locking issues, or inefficient SQL.
Parallel processing involves executing multiple tasks simultaneously using multiple cores or processors. It can provide performance benefits over serial processing by reducing execution time. When developing parallel applications, developers must identify independent tasks that can be executed concurrently and avoid issues like race conditions and deadlocks. Effective parallelization requires analyzing serial code to find optimization opportunities, designing and implementing concurrent tasks, and testing and tuning to maximize performance gains.
This document summarizes the results of a proof of concept test of PolyServe's solution for scalable shared databases with SQL Server 2005. The test involved running a set of reporting queries against a database with 100 million rows on 1, 2, and 4 SQL Server instances attached to the same database via PolyServe's software. Performance improved linearly as more servers were added, with completion time reducing by up to 16 times on 4 servers compared to running the queries on 1 server. PolyServe's approach simplified database administration by allowing servers to directly access shared storage and avoiding the need to manually create and manage read-only database copies or volumes.
Hotsos 2011: Mining the AWR repository for Capacity Planning, Visualization, ...Kristofferson A
The document discusses mining the Automatic Workload Repository (AWR) in Oracle databases for capacity planning, visualization, and other real-world uses. It introduces Karl Arao as a speaker and discusses topics he will cover including AWR, diagnosing performance issues using AWR data, visualization of AWR data, capacity planning, and tools for working with AWR data like scripts and linear regression. References and resources on working with AWR are also provided.
This document discusses CPU monitoring and capacity planning in a massively consolidated environment. It begins with comparing CPU speeds between hardware platforms using benchmarks like TPC-C and SPECint_rate2006. It then discusses measuring the performance of cores vs threads using tools like cputoolkit. Finally, it outlines the different types of CPU events that can be monitored in Oracle. The key takeaways are how to use benchmarks to compare CPU performance for capacity planning, understanding the impact of features like hyper-threading, and being aware of the various CPU metrics available.
A Gentle Introduction to Microsoft SSASJohn Paredes
This document provides an overview of Analysis Services and how to create an OLAP cube. It discusses why data is stored in cubes rather than tables, including better query performance, efficient storage and calculations. It then outlines the steps to create an Analysis Services project, define data sources and dimensions, create and process the cube, and optimize query performance through partitioning and pre-aggregating data.
Karan Gulati has over 5 years of experience as a SQL Server Analysis Services expert at Microsoft. In this presentation, he provides an overview of key data warehousing and OLAP concepts, including: defining a data warehouse and why OLAP is used; the components of a cube like measures, dimensions, and schemas; and slowly changing dimension types like Type 1, 2, and 3. He explains these concepts at a high level to help attendees understand the terminology in the SQL and data warehousing fields.
This document summarizes the main parts of an Oracle AWR report, including the snapshot details, load profile, top timed foreground events, time model statistics, and SQL section. The time model statistics indicate that 86.45% of database time was spent executing SQL statements. The top foreground event was waiting for database file sequential reads, taking up 62% of database time.
Jugal Shah has over 14 years of experience in IT working in roles such as manager, solution architect, DBA, developer and software engineer. He has worked extensively with database technologies including SQL Server, MySQL, PostgreSQL and others. He has received the MVP award from Microsoft for SQL Server in multiple years. Common causes of SQL Server performance problems include configuration issues, design problems, bottlenecks and poorly written queries or code. Various tools can be used to diagnose issues including dynamic management views, Performance Monitor, SQL Server Profiler and DBCC commands.
Progress OpenEdge database administration guide and referenceVinh Nguyen
The document is a guide and reference for Progress Database Administration. It discusses Progress database architecture, storage design, configuration variables, limits, and procedures for creating and deleting databases. The guide contains information to help plan, administer, and optimize Progress databases.
This document provides an overview of performance tuning for SQL Server databases. It discusses eliminating obvious issues, understanding common bottlenecks like CPU, memory, disk and locking, and using tools like wait statistics, performance counters and dynamic management views to analyze workload and identify tuning opportunities. The document outlines a 12 step process for capturing a production workload, replaying it on a test server, making tuning changes and reanalyzing. It also demonstrates various tools and techniques like SQL Profiler, SQL Diagnostic Manager and RML Utilities.
Hardware planning & sizing for sql serverDavide Mauri
This document provides an overview of hardware planning and sizing considerations for SQL Server. It discusses that performance is the typical requirement for relational database management systems. While high performance is expected, typical server hardware configurations often result in unbalanced systems that are not optimized. The document advocates for balanced systems with no single bottleneck. It provides guidance on evaluating CPU, memory, I/O capabilities and storage to ensure a system can handle peak resource consumption. Baseline testing is recommended to compare hardware performance.
This document provides an overview of performance monitoring and optimization for SQL Server databases. It discusses monitoring database activity using tools like SQL Profiler and Activity Monitor, identifying bottlenecks, using the Database Engine Tuning Advisor to generate optimization recommendations, and addressing issues related to processes, locking, and deadlocks. Best practices emphasized establishing a performance baseline, making incremental changes while measuring impact, and focusing on specific issues to optimize real-world workloads.
This document discusses PostgreSQL parameter tuning, specifically related to memory and optimizer parameters. It provides guidance on setting parameters like shared_buffer, work_mem, temp_buffer, maintenance_work_mem, random_page_cost, sequential_page_cost, and effective_cache_size to optimize performance based on hardware characteristics like available RAM and disk speed. It also covers force_plan parameters that can include or exclude certain query optimization techniques.
Whitepaper: Exadata Consolidation Success StoryKristofferson A
1. The document discusses database and server consolidation using Oracle Exadata and describes the challenges of managing highly consolidated environments to ensure quality of service.
2. It outlines a 4-step process for accurate provisioning and capacity planning using a tool called the Provisioning Worksheet: collecting database details, defining the target Exadata hardware capacity, creating a provisioning plan, and reviewing resource utilization.
3. The process relies on basic capacity planning to ensure workload requirements fit available capacity. Database CPU and storage requirements are gathered, a target Exadata configuration is set, databases are mapped to nodes in the plan, and final utilization is summarized to identify any capacity shortfalls.
The document discusses SQL Server performance monitoring and tuning. It recommends taking a holistic view of the entire system landscape, including hardware, software, systems and networking components. It outlines various tools for performance monitoring, and provides guidance on identifying and addressing common performance issues like high CPU utilization, disk I/O issues and poorly performing queries.
SQL Server Tuning to Improve Database PerformanceMark Ginnebaugh
SQL Server tuning is a process to eliminate performance bottlenecks and improve application service. This presentation from Confio Software discusses SQL diagramming, wait type data, column selectivity, and other solutions that will help make tuning projects a success, including:
•SQL Tuning Methodology
•Response Time Tuning Practices
•How to use SQL Diagramming techniques to tune SQL statements
•How to read executions plans
The document discusses Oracle Database performance tuning. It begins by defining performance as the accepted throughput for a given workload. Performance tuning is defined as optimizing resource use to increase throughput and minimize contention. A performance problem occurs when database tasks do not complete in a timely manner, such as SQL running longer than usual or users facing slowness. Performance problems can be caused by contention for resources, overutilization of the system, or poorly written SQL. The document discusses various performance diagnostics tools and concepts like wait events, enqueues, I/O performance, and provides examples of how to analyze issues related to these areas.
Find and fix SQL Server performance problems fasterSolarWinds
Great DBAs must be able to quickly identify problems with SQL Server instances. In this presentation, you will learn how to quickly identify where your problems are using tools such as:
*Dynamic Management Views
*Query Execution Plans
*Windows Performance Monitor
*Extended Events
*Third-party tools (including SolarWinds Database Performance Analyzer)
This document provides an overview and interpretation of the Automatic Workload Repository (AWR) report in Oracle database. Some key points:
- AWR collects snapshots of database metrics and performance data every 60 minutes by default and retains them for 7 days. This data is used by tools like ADDM for self-management and diagnosing issues.
- The top timed waits in the AWR report usually indicate where to focus tuning efforts. Common waits include I/O waits, buffer busy waits, and enqueue waits.
- Other useful AWR metrics include parse/execute ratios, wait event distributions, and top activities to identify bottlenecks like parsing overhead, locking issues, or inefficient SQL.
Parallel processing involves executing multiple tasks simultaneously using multiple cores or processors. It can provide performance benefits over serial processing by reducing execution time. When developing parallel applications, developers must identify independent tasks that can be executed concurrently and avoid issues like race conditions and deadlocks. Effective parallelization requires analyzing serial code to find optimization opportunities, designing and implementing concurrent tasks, and testing and tuning to maximize performance gains.
This document summarizes the results of a proof of concept test of PolyServe's solution for scalable shared databases with SQL Server 2005. The test involved running a set of reporting queries against a database with 100 million rows on 1, 2, and 4 SQL Server instances attached to the same database via PolyServe's software. Performance improved linearly as more servers were added, with completion time reducing by up to 16 times on 4 servers compared to running the queries on 1 server. PolyServe's approach simplified database administration by allowing servers to directly access shared storage and avoiding the need to manually create and manage read-only database copies or volumes.
Hotsos 2011: Mining the AWR repository for Capacity Planning, Visualization, ...Kristofferson A
The document discusses mining the Automatic Workload Repository (AWR) in Oracle databases for capacity planning, visualization, and other real-world uses. It introduces Karl Arao as a speaker and discusses topics he will cover including AWR, diagnosing performance issues using AWR data, visualization of AWR data, capacity planning, and tools for working with AWR data like scripts and linear regression. References and resources on working with AWR are also provided.
This document discusses CPU monitoring and capacity planning in a massively consolidated environment. It begins with comparing CPU speeds between hardware platforms using benchmarks like TPC-C and SPECint_rate2006. It then discusses measuring the performance of cores vs threads using tools like cputoolkit. Finally, it outlines the different types of CPU events that can be monitored in Oracle. The key takeaways are how to use benchmarks to compare CPU performance for capacity planning, understanding the impact of features like hyper-threading, and being aware of the various CPU metrics available.
A Gentle Introduction to Microsoft SSASJohn Paredes
This document provides an overview of Analysis Services and how to create an OLAP cube. It discusses why data is stored in cubes rather than tables, including better query performance, efficient storage and calculations. It then outlines the steps to create an Analysis Services project, define data sources and dimensions, create and process the cube, and optimize query performance through partitioning and pre-aggregating data.
Karan Gulati has over 5 years of experience as a SQL Server Analysis Services expert at Microsoft. In this presentation, he provides an overview of key data warehousing and OLAP concepts, including: defining a data warehouse and why OLAP is used; the components of a cube like measures, dimensions, and schemas; and slowly changing dimension types like Type 1, 2, and 3. He explains these concepts at a high level to help attendees understand the terminology in the SQL and data warehousing fields.
Microsoft SSAS: Should I Use Tabular or Multidimensional?Senturus
Learn the right version Microsoft SQL Server Analysis services to use to easily migrate the work to the other version. View the webinar video recording and download this deck: http://www.senturus.com/resources/microsoft-ssas/.
During this webinar, Senturus discussed how to choose between the tabular and multi-dimensional versions of SSAS for your analytic needs and the various features and benefits that each version provides.
Senturus, a business analytics consulting firm, has a resource library with hundreds of free recorded webinars, trainings, demos and unbiased product reviews. Take a look and share them with your colleagues and friends: http://www.senturus.com/resources/.
Microsoft SQL Server Analysis Services (SSAS) - A Practical Introduction Mark Ginnebaugh
Patrick Sheehan of Microsoft covers platform architecture, data warehousing methodology, and multi-dimensional cube development.
You will learn:
* How to develop and deploy data cubes using SQL Server Analysis Services (SSAS)
* Optimal data warehouse methodology for use with SSAS
* Tips/tricks for designing & building cubes over no warehouse/suboptimal source system (it happens)
* Cube processing types - How/why to use each
* Cube design practices + How to build and deploy cubes!
The document discusses SQL Parallel Data Warehouse (PDW), which is a massively parallel processing appliance for large data warehousing workloads. It describes the different types of nodes in PDW, including control nodes that manage query execution, compute nodes that store and process data, and administrative nodes. The document also explains how PDW uses a hub and spoke architecture with the PDW appliance acting as a central data hub and individual data marts acting as spokes optimized for different user groups.
Building an Effective Data Warehouse ArchitectureJames Serra
Why use a data warehouse? What is the best methodology to use when creating a data warehouse? Should I use a normalized or dimensional approach? What is the difference between the Kimball and Inmon methodologies? Does the new Tabular model in SQL Server 2012 change things? What is the difference between a data warehouse and a data mart? Is there hardware that is optimized for a data warehouse? What if I have a ton of data? During this session James will help you to answer these questions.
Mapping Data Flows Perf Tuning April 2021Mark Kromer
This document discusses optimizing performance for data flows in Azure Data Factory. It provides sample timing results for various scenarios and recommends settings to improve performance. Some best practices include using memory optimized Azure integration runtimes, maintaining current partitioning, scaling virtual cores, and optimizing transformations and sources/sinks. The document also covers monitoring flows to identify bottlenecks and global settings that affect performance.
Leveraging Functional Tools and AWS for Performance TestingThoughtworks
This document discusses leveraging functional test tools and AWS for performance testing. It describes challenges with functional testing like needing quick reusable tools for continuous integration. It also covers using AWS to help with performance testing by allowing different customer environments to be easily setup and configured. Key aspects of performance testing discussed include measuring response times, concurrency, and failover testing using tools like SOAP UI, custom code, and analyzing performance counters.
"Эффективность и оптимизация кода в Java 8" Сергей МоренецFwdays
Если мы захотим понять, что такое совершенный(идеальный) код, то одной из его характеристик будет эффективность. Это включает в себя и быстродействие кода, и объем потребляемых ресурсов(память, дисковых, I/O).
Зачастую эффективность отодвигается на второй план, поскольку ее не так просто рассчитать заранее, а также точно определить на ревью кода. В то же время это единственная характеристика, которая затрагивает конечного пользователя наших проектов.
В моем докладе я рассмотрю, что такое эффективность, как ее правильно измерять, мы коснемся мифов об эффективности, которые очень популярны сейчас, рассмотрим примеры эффективного и неэффективного кода, нужной и бессмысленной оптимизации кода.
Главный упор будет сделан на функциональности, которая была добавлена в Java 8.
Initially presented at Software Architecture Conference in Boston, MA on 3/18/15.
Distributed systems are complex beasts. Breaking your application into multiple services introduces new types of errors, cascading failures, and CAP theorem limitations. Unfortunately, your uptime and sanity both suffer. This session will focus on various tactics and learnings from Lucid Software's migration to a service oriented architecture.
Help, I need to migrate my On Premise Database to Azure, which Database Tier ...Erwin de Kreuk
Azure SQL Database provides several deployment options including single databases and elastic pools. The single database option provides resource guarantees at the database level while elastic pools allow for sharing of resources across multiple databases for better cost efficiency. Azure SQL Database offers different service tiers including Basic, Standard, and Premium that provide different performance levels and features. Customers can choose between DTU-based and vCore-based purchasing models, with vCores offering more flexibility and control over compute and storage. The Data Migration Assistant and Data Migration Service can help customers assess, plan, and execute migrations of databases to Azure SQL Database.
Windows Azure Storage: Overview, Internals, and Best PracticesAnton Vidishchev
This document provides an overview of Windows Azure Storage, including its internals and best practices. It describes the key abstractions in Azure Storage like blobs, disks, tables, and queues. It then discusses the internal architecture including how data is distributed across data centers and storage stamps for high availability and scalability. The document concludes with best practices for Azure Storage like optimizing for concurrency versus parallelism and using common design patterns for blobs, tables, and queues.
Sql saturday azure storage by Anton VidishchevAlex Tumanoff
This document provides an overview of Windows Azure Storage, including its internals and best practices. It describes the key abstractions in Azure Storage like blobs, disks, tables, and queues. It then discusses the internal architecture and design goals around high availability, durability, and scalability. Specific topics covered include storage stamps, the partition layer, dynamic load balancing, and availability with consistency for writing and reading. The document concludes with best practices for .NET, blobs, tables, queues, and general usage of Azure Storage.
Cassandra CLuster Management by Japan Cassandra CommunityHiromitsu Komatsu
This document discusses best practices for managing Cassandra clusters based on Instaclustr's experience managing over 500 nodes and 3 million node-hours. It covers choosing the right Cassandra version, hardware configuration, cost estimation, load testing, data modeling practices, common issues like modeling errors and overload, and important monitoring techniques like logs, metrics, cfstats and histograms. Maintaining a well-designed cluster and proactively monitoring performance are keys to avoiding issues with Cassandra.
Datadog: a Real-Time Metrics Database for One Quadrillion Points/DayC4Media
Video and slides synchronized, mp3 and slide download available at URL http://bit.ly/2mAKgJi.
Ian Nowland and Joel Barciauskas talk about the challenges Datadog faces as the company has grown its real-time metrics systems that collect, process, and visualize data to the point they now handle trillions of points per day. They also talk about how the architecture has evolved, and what they are looking to in the future as they architect for a quadrillion points per day. Filmed at qconnewyork.com.
Ian Nowland is the VP Engineering Metrics and Alerting at Datadog. Joel Barciauskas currently leads Datadog's distribution metrics team, providing accurate, low latency percentile measures for customers across their infrastructure.
Azure Data Factory Data Flow Performance Tuning 101Mark Kromer
The document provides performance timing results and recommendations for optimizing Azure Data Factory data flows. Sample 1 processed a 421MB file with 887k rows in 4 minutes using default partitioning on an 80-core Azure IR. Sample 2 processed a table with the same size and transforms in 3 minutes using source and derived column partitioning. Sample 3 processed the same size file in 2 minutes with default partitioning. The document recommends partitioning strategies, using memory optimized clusters, and scaling cores to improve performance.
Speeding Up Atlas Deep Learning Platform with Alluxio + FluidAlluxio, Inc.
Data Orchestration Summit 2020 organized by Alluxio
https://www.alluxio.io/data-orchestration-summit-2020/
Speeding Up Atlas Deep Learning Platform with Alluxio + Fluid
Yuandong Xie, Platform Researcher (Unisound)
About Alluxio: alluxio.io
Engage with the open source community on slack: alluxio.io/slack
Caracteristicas de estudio de SQL Server 2008 y MOSS
Ing. Eduardo Castro Martinez, PhD
Microsoft SQL Server MVP
http://ecastrom.blogspot.com
http://comunidadwindows.org
The document discusses database capacity planning and analysis. It covers collecting and analyzing resource utilization data, developing mathematical models to predict performance, using queueing theory and response time analysis. Steps outlined include determining goals, gathering workload data, characterizing and modeling data, validating forecasts, and conducting case studies on deletion performance and evaluating MySQL capacity. The overall aim is to accurately measure current capacity, predict future growth, maintain balanced performance, and identify risks.
Michigan Information Retrieval Enthusiasts Group Meetup - August 19, 2010ivan provalov
Two presentation from the Michigan Information Retrieval Enthusiasts Group Meetup on August 19 by Cengage Learning search platform development team.
Scaling Performance Tuning With Lucene by John Nader discusses primary performance hot spots related to scaling to a multi-million document collection. This includes the team's experiences with memory consumption, GC tuning, query expansion, and filter performance. Discusses both the tools used to identify issues and the techniques used to address them.
Relevance Tuning Using TREC Dataset by Rohit Laungani and Ivan Provalov describes the TREC dataset used by the team to improve the relevance of the Lucene-based search platform. Goes over IBM paper and describe the approaches tried: Lexical Affinities, Stemming, Pivot Length Normalization, Sweet Spot Similarity, Term Frequency Average Normalization. Talks about Pseudo Relevance Feedback.
Presentation delivered by Matt Done, Head Of Platform Development at expanz Pty. Ltd. during DDD Sydney event on 2 July 2011.
Matt demonstrates what it takes to setup a highly sophisticated load test, using the Azure environment and how to use the results to optimise a fully blown application development platform and application server running on Azure.
Recording of this presentation can be found at www.youtube.com/expanzTV
This document discusses database performance core principles and how moving business logic out of the database and into application servers can negatively impact performance. The key points are:
1. Oracle databases are process-based and processes need to get CPU time quickly, stay on the CPU, and experience few involuntary context switches.
2. Moving business logic out of the database and into many small calls results in high call counts, more voluntary sleeps as processes switch contexts, and potential oversubscription of processes leading to involuntary sleeps.
3. This violates database performance principles and leads to inconsistent response times, lower throughput, and inefficient use of computing resources across database and application servers due to increased overhead of switching processes.
Keeping
This document outlines a performance test plan for Sakai 2.5.0. It describes the objectives, approach, test types, metrics, goals, tools, and data preparation. The objectives are to validate Sakai meets minimum performance standards and test any new or changed tools. Tests include capacity, consistent load, and single function stress tests. Metrics like response time, CPU utilization, and errors will be measured. Goals include average response time under 2.5s and max under 30s, CPU under 75%, and 500 concurrent users supported. Silk Performer will be used to run tests against a Sakai/Tomcat/Oracle environment. Over 92,000 students and 1,557 instructors of data will be preloaded
The document discusses Oracle Database result caching. It provides an overview of database caches including the result cache. It then describes a hand-made result cache implementation for a retailer case study and how it improved performance from 20 minutes to 4 minutes for a report. It also discusses using the Oracle Database result cache explicitly with hints and annotations, how to monitor and manage it using views and packages, limitations, and best practices.
The document discusses DeepDB, a storage engine plugin for MySQL that aims to address MySQL's performance and scaling limitations for large datasets and heavy indexing. It does this through techniques like a Cache Ahead Summary Index Tree, Segmented Column Store, Streaming I/O, Extreme Concurrency, and Intelligent Caching. The document provides examples showing DeepDB significantly outperforming MySQL's InnoDB storage engine for tasks like data loading, transactions, queries, backups and more. It positions DeepDB as a drop-in replacement for InnoDB that can scale MySQL to support billions of rows and queries 2x faster while reducing data footprint by 50%.
"Rebranding for Growth", Anna VelykoivanenkoFwdays
Since there is no single formula for rebranding, this presentation will explore best practices for aligning business strategy and communication to achieve business goals.
Special Meetup Edition - TDX Bengaluru Meetup #52.pptxshyamraj55
We’re bringing the TDX energy to our community with 2 power-packed sessions:
🛠️ Workshop: MuleSoft for Agentforce
Explore the new version of our hands-on workshop featuring the latest Topic Center and API Catalog updates.
📄 Talk: Power Up Document Processing
Dive into smart automation with MuleSoft IDP, NLP, and Einstein AI for intelligent document workflows.
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxJustin Reock
Building 10x Organizations with Modern Productivity Metrics
10x developers may be a myth, but 10x organizations are very real, as proven by the influential study performed in the 1980s, ‘The Coding War Games.’
Right now, here in early 2025, we seem to be experiencing YAPP (Yet Another Productivity Philosophy), and that philosophy is converging on developer experience. It seems that with every new method we invent for the delivery of products, whether physical or virtual, we reinvent productivity philosophies to go alongside them.
But which of these approaches actually work? DORA? SPACE? DevEx? What should we invest in and create urgency behind today, so that we don’t find ourselves having the same discussion again in a decade?
Spark is a powerhouse for large datasets, but when it comes to smaller data workloads, its overhead can sometimes slow things down. What if you could achieve high performance and efficiency without the need for Spark?
At S&P Global Commodity Insights, having a complete view of global energy and commodities markets enables customers to make data-driven decisions with confidence and create long-term, sustainable value. 🌍
Explore delta-rs + CDC and how these open-source innovations power lightweight, high-performance data applications beyond Spark! 🚀
Role of Data Annotation Services in AI-Powered ManufacturingAndrew Leo
From predictive maintenance to robotic automation, AI is driving the future of manufacturing. But without high-quality annotated data, even the smartest models fall short.
Discover how data annotation services are powering accuracy, safety, and efficiency in AI-driven manufacturing systems.
Precision in data labeling = Precision on the production floor.
Hands On: Create a Lightning Aura Component with force:RecordDataLynda Kane
Slide Deck from the 3/26/2020 virtual meeting of the Cleveland Developer Group presentation on creating a Lightning Aura Component using force:RecordData.
The Evolution of Meme Coins A New Era for Digital Currency ppt.pdfAbi john
Analyze the growth of meme coins from mere online jokes to potential assets in the digital economy. Explore the community, culture, and utility as they elevate themselves to a new era in cryptocurrency.
Big Data Analytics Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
Leading AI Innovation As A Product Manager - Michael JidaelMichael Jidael
Unlike traditional product management, AI product leadership requires new mental models, collaborative approaches, and new measurement frameworks. This presentation breaks down how Product Managers can successfully lead AI Innovation in today's rapidly evolving technology landscape. Drawing from practical experience and industry best practices, I shared frameworks, approaches, and mindset shifts essential for product leaders navigating the unique challenges of AI product development.
In this deck, you'll discover:
- What AI leadership means for product managers
- The fundamental paradigm shift required for AI product development.
- A framework for identifying high-value AI opportunities for your products.
- How to transition from user stories to AI learning loops and hypothesis-driven development.
- The essential AI product management framework for defining, developing, and deploying intelligence.
- Technical and business metrics that matter in AI product development.
- Strategies for effective collaboration with data science and engineering teams.
- Framework for handling AI's probabilistic nature and setting stakeholder expectations.
- A real-world case study demonstrating these principles in action.
- Practical next steps to begin your AI product leadership journey.
This presentation is essential for Product Managers, aspiring PMs, product leaders, innovators, and anyone interested in understanding how to successfully build and manage AI-powered products from idea to impact. The key takeaway is that leading AI products is about creating capabilities (intelligence) that continuously improve and deliver increasing value over time.
Automation Dreamin' 2022: Sharing Some Gratitude with Your UsersLynda Kane
Slide Deck from Automation Dreamin'2022 presentation Sharing Some Gratitude with Your Users on creating a Flow to present a random statement of Gratitude to a User in Salesforce.
AI and Data Privacy in 2025: Global TrendsInData Labs
In this infographic, we explore how businesses can implement effective governance frameworks to address AI data privacy. Understanding it is crucial for developing effective strategies that ensure compliance, safeguard customer trust, and leverage AI responsibly. Equip yourself with insights that can drive informed decision-making and position your organization for success in the future of data privacy.
This infographic contains:
-AI and data privacy: Key findings
-Statistics on AI data privacy in the today’s world
-Tips on how to overcome data privacy challenges
-Benefits of AI data security investments.
Keep up-to-date on how AI is reshaping privacy standards and what this entails for both individuals and organizations.
3. TEAM
Thomas
Kejser
CTO EMEA at Fusion-
io
Alexei
Khalyako
SQL CAT Program
Manager at Microsoft
Marcel Franke
Practice Lead at pmOne
Gerhard Brückl
Practice Lead at pmOne
SSAS Maestro
Hannes Mayer
Practice Lead at pmOne
SSAS Maestro
4. Build scalable solutions on user demand
Choose of an OLAP optimized hardware configuration
Align SSAS configuration to hardware settings
CUSTOMER CHALLENGES
7. What makes a good hardware for SQL Server Analysis Services
CPU
More cores more parallelism for Storage Engine bound queries
High clock rate faster Formula Engine bound queries
I/O
SSAS uses random reads
No spinning disks - use Fusion-IO drives instead
Use local disks to avoid unnecessary fiber channel/iSCSI cost
Memory
More memory more data can be loaded into memory
In ideal case: entire cube could be loaded into memory
GENERAL HARDWARE RECOMMENDATIONS
8. Server: HP DL380 Gen8
CPU
2.90 GHz. 8 Cores. 16 Logical Processors
Good mix between speed and number of cores
Hyper threading to 32 logical cores to the OS
I/O
Local Fixed Disk (4,38 TB). 4 x Fusion IO Drive 2 Stripe Set (4x1,2TB)
Very high IOPs (random reads: 140.000/s)
Very high throughput (random reads: 4,3GB/s)
Memory
256 GB physical memory
OS: Windows Server 2012 Datacenter
REFERENCE HARDWARE CONFIGURATION
11. 2 SQL Server Analysis Services Databases
100 GB – fits into memory
1.000 GB – does not fit into memory
3 query-patterns
7 Standard queries for common business questions
Ratio-to-Total, Rolling-12-month, Year-over-Year growth, …
1 DistinctCount query
1 Many-to-Many query
Increasing number of users
Incrementally grow number of connected users – 1 every 10 sec
Up to 200 concurrent users
TEST OVERVIEW
13. “How many queries can be answered
within a given time period“
DEFINING CONCURRENCY
14. EXPECTED RESULTS
(1) Constant test duration
(2) Until saturation point is reached
(3) Linear increase of test duration together with concurrent users
(1) (3)(2)
Avg. query
response time
15. More than 100+ concurrent users (connections) supported
For standard queries
With Average response time <3 seconds
Regardless of database size
Complex queries impact average response time
DistinctCount and Many-to-Many queries may not finish before the
next user connects – especially for the 1,000 GB cube
Usually:
- not every query has the pattern of a complex query
- executed rarely compare to the standard queries
- can be avoided by appropriate cube design
Scalability depends on CPU resources. I/O is not a limiting factor
anymore
FINAL RESULTS – OVERVIEW
16. Constant query execution time till CPU limit reached with growing
numbers of concurrent users
FINAL RESULTS – OVERVIEW
* Query 20
17. An average of 285 concurrent users with a response time below 3
seconds for standard queries
Only 2 concurrent users for more complex queries
FINAL RESULTS – DETAILS 100 GB
Query-Pattern Query
AvgTests/s
(@SP)
Max
Avg Tests/s
Median
Avg Tests/s
Queries
/ Test
Supported Concurrent Users/Queries
3s Response Time 10s Response Time
Standard Query2 360.0 513.2 429.8 1 1.080 3,600
Standard Query3 4.5 9.4 4.0 1 14 45
Standard Query20 22.5 41.0 22.0 1 68 225
Standard Query21 14.0 17.2 13.0 1 42 140
Standard Query22 1.5 4.2 1.0 1 5 15
Standard Query52 250.0 505.6 323.4 1 750 2,500
Standard Query77 13.0 88.0 14.8 1 39 130
DistinctCount Query100 0.6 2.8 0.6 1 2 6
Many-to-Many Query101 6.8 112.0 76.2 1 20 68
AllQueries AllQueries 0.8 260.4 0.2 9 2 8
(@SP) = @saturation point
18. An average of 210 concurrent users with a response time below 3
seconds for standard queries
more complex queries run longer than our required response time
which means we cannot even satisfy 1 concurrent user
FINAL RESULTS – DETAILS 1000 GB
Query-Pattern Query
AvgTests/s
(@SP)
Max
Avg Tests/s
Median
Avg Tests/s
Queries
/ Test
Supported Concurrent Users/Queries
3s Response Time 10s Response Time
Standard Query2 250,0 409,8 373,2 1 750 2.500
Standard Query3 0,4 3,4 - 1 1 4
Standard Query20 10,0 16,6 10,5 1 30 100
Standard Query21 6,6 11,4 6,6 1 20 66
Standard Query22 1,0 4,4 0,8 1 3 10
Standard Query52 220,0 388,6 304,2 1 660 2.200
Standard Query77 0,1 5,4 0,2 1 0 1
DistinctCount Query100 - 0,2 - 1 - -
Many-to-Many Query101 - 0,6 - 1 - -
AllQueries AllQueries - 2,6 - 9 - -
(@SP) = @saturation point
19. Shown great linear, CPU bound scale for fast queries
Concurrency is related to the number of queries not to the number of
users
No user would run more than one query per second
Caching helps to avoid expensive IO operations
NUMBER OF USERS VS. QUERY RESPONSE TIME
20. Bigger cube more IO
slower query response times
Bigger cube not everything can be loaded into memory
Constant IO going on slower query response times
2 possible solutions to handle more data
Better IO system
More memory
(More CPU to handle additional threads)
QUERY RESPONSE TIME VS. DATA VOLUME
22. This sample configuration can be used as reference architecture for
building high scalable OLAP solutions.
With given configuration we satisfied the SLAs for the cube up to 1TB
in size and 200+ concurrent queries.
Capacity calculation
The average query per core ratio is 6,5.
If more than 200 concurrent queries are expected then this ratio can be
used to calculated necessary number of cores.
Sample: 400 concurrent queries = 400/6,5 = 64 cores
CONCLUSION
#6: Accelerate Data Warehouse projects with scalable reference architecture for SQL Server Analysis ServicesReduce costs, save time and reduce risk with reliable reference architectures and best practices Suitable for typical workloads (OLAP-DB size, query pattern, number of users)Reference Architecture consists of:Best Practices for SQL Server Analysis Services settingsHardware recommendations for typical workloadsAligned configurations of software, hardware and OSShow Scalability of SSAS and reference architectureProvide test cases & results