This document provides an overview and introduction to NoSQL databases. It discusses key-value stores like Dynamo and BigTable, which are distributed, scalable databases that sacrifice complex queries for availability and performance. It also explains column-oriented databases like Cassandra that scale to massive workloads. The document compares the CAP theorem and consistency models of these databases and provides examples of their architectures, data models, and operations.
The Information Technology have led us into an era where the production, sharing and use of information are now part of everyday life and of which we are often unaware actors almost: it is now almost inevitable not leave a digital trail of many of the actions we do every day; for example, by digital content such as photos, videos, blog posts and everything that revolves around the social networks (Facebook and Twitter in particular). Added to this is that with the "internet of things", we see an increase in devices such as watches, bracelets, thermostats and many other items that are able to connect to the network and therefore generate large data streams. This explosion of data justifies the birth, in the world of the term Big Data: it indicates the data produced in large quantities, with remarkable speed and in different formats, which requires processing technologies and resources that go far beyond the conventional systems management and storage of data. It is immediately clear that, 1) models of data storage based on the relational model, and 2) processing systems based on stored procedures and computations on grids are not applicable in these contexts. As regards the point 1, the RDBMS, widely used for a great variety of applications, have some problems when the amount of data grows beyond certain limits. The scalability and cost of implementation are only a part of the disadvantages: very often, in fact, when there is opposite to the management of big data, also the variability, or the lack of a fixed structure, represents a significant problem. This has given a boost to the development of the NoSQL database. The website NoSQL Databases defines NoSQL databases such as "Next Generation Databases mostly addressing some of the points: being non-relational, distributed, open source and horizontally scalable." These databases are: distributed, open source, scalable horizontally, without a predetermined pattern (key-value, column-oriented, document-based and graph-based), easily replicable, devoid of the ACID and can handle large amounts of data. These databases are integrated or integrated with processing tools based on the MapReduce paradigm proposed by Google in 2009. MapReduce with the open source Hadoop framework represent the new model for distributed processing of large amounts of data that goes to supplant techniques based on stored procedures and computational grids (step 2). The relational model taught courses in basic database design, has many limitations compared to the demands posed by new applications based on Big Data and NoSQL databases that use to store data and MapReduce to process large amounts of data.
Course Website http://pbdmng.datatoknowledge.it/
Contact me for other informations and to download
In these slides we introduce Column-Oriented Stores. We deeply analyze Google BigTable. We discuss about features, data model, architecture, components and its implementation. In the second part we discuss all the major open source implementation for column-oriented databases.
This document provides an overview of NoSQL databases and summarizes key information about several NoSQL databases, including HBase, Redis, Cassandra, MongoDB, and Memcached. It discusses concepts like horizontal scalability, the CAP theorem, eventual consistency, and data models used by different NoSQL databases like key-value, document, columnar, and graph structures.
NoSQL databases take a different approach to data storage than traditional RDBMS systems. There are several categories of NoSQL databases including key-value stores, wide column stores, document stores, and graph databases. Each has different strengths such as flexibility, performance, or suitability for certain types of data. Choosing the right data model depends on factors like the relationships between data elements, scalability needs, and query requirements.
An overview of various database technologies and their underlying mechanisms over time.
Presentation delivered at Alliander internally to inspire the use of and forster the interest in new (NOSQL) technologies. 18 September 2012
Oracle has evolved from its first release in 1979 to become a leading database with various editions that can be used by individuals, workgroups or enterprises, and it provides developer tools and supports different database structures, security mechanisms, SQL for data access and transactions. Key components of an Oracle database include control files, data files, redo log files, tablespaces that logically organize storage, and various memory and file structures.
This document provides an overview of different database types including relational, NoSQL, document, key-value, graph, and column family databases. It discusses the history and drivers behind the development of NoSQL databases, as well as concepts like horizontal scaling, the CAP theorem, and eventual consistency. Specific databases are also summarized, including MongoDB, Redis, Neo4j, and HBase.
The presentation provides an overview of NoSQL databases, including a brief history of databases, the characteristics of NoSQL databases, different data models like key-value, document, column family and graph databases. It discusses why NoSQL databases were developed as relational databases do not scale well for distributed applications. The CAP theorem is also explained, which states that only two out of consistency, availability and partition tolerance can be achieved in a distributed system.
Cloud Deployments with Apache Hadoop and Apache HBaseDATAVERSITY
The document discusses deploying Apache Hadoop and Apache HBase cloud deployments. It begins with introducing the speaker and their background with Cloudera and various Apache projects. It then provides an overview of Cloudera and what they do. The majority of the document discusses Apache Hadoop and Apache HBase, what they are, how they are open source and horizontally scalable. It also discusses deploying a Hadoop and HBase cluster on Amazon EC2 using Apache Whirr to provision the machines. Real-world examples of using these technologies include building a web index for a search engine.
In this lecture we analyze document oriented databases. In particular we consider why there are the first approach to nosql and what are the main features. Then, we analyze as example MongoDB. We consider the data model, CRUD operations, write concerns, scaling (replication and sharding).
Finally we presents other document oriented database and when to use or not document oriented databases.
The document discusses object-relational impedance mismatch and various data source patterns for mapping objects to relational databases in a way that minimizes this mismatch. It describes the table data gateway, row data gateway, active record, and data mapper patterns. The table data gateway acts as a gateway to a database table, while the row data gateway acts as a gateway to a single record. Active record wraps a database row and adds domain logic, and data mapper provides object-relational mapping to keep the object model independent from the database schema. Spring JDBC is also introduced as a framework that can help implement these patterns.
The initiation of The Hadoop Apache Hive began in 2007 by Facebook due to its data growth.
This ETL system began to fail over few years as more people joined Facebook.
In August 2008, Facebook decided to move to scalable a more scalable open-source Hadoop environment; Hive
Facebook, Netflix and Amazons support the Apache Hive SQL now known as the HiveQL
This document provides an overview of the course content for an online SAS training course. The course covers topics such as SAS basics, statistical analysis, data management, SQL, macro programming, and debugging SAS programs. It explores how to use SAS for clinical research studies and banking analysis. The course aims to teach students how to manage, analyze, and report on data with SAS.
This document provides an overview of NoSQL databases, including:
- Key-value stores store data as maps or hashmaps and are efficient for data access but limited in query capabilities.
- Column-oriented stores group attributes into column families and store data efficiently but are operationally challenging.
- Document databases store loosely structured data like JSON and allow retrieving documents by keys or contents.
- Graph databases are suited for interaction networks and path finding but are less suited for tabular data.
The document discusses the MariaDB CONNECT storage engine, which allows querying external file formats from within MariaDB. It was created by database expert Olivier Bertrand and brings business intelligence capabilities to MariaDB by enabling access to data sources like CSV, XML, Excel and other formats without needing ETL processes. The storage engine uses the MySQL plugin architecture and implements features like indexing, condition pushdown, and support for ODBC, MySQL tables, and various file types.
This document compares SQL and NoSQL databases. It defines databases, describes different types including relational and NoSQL, and explains key differences between SQL and NoSQL in areas like scaling, modeling, and query syntax. SQL databases are better suited for projects with logical related discrete data requirements and data integrity needs, while NoSQL is more ideal for projects with unrelated, evolving data where speed and scalability are important. MongoDB is provided as an example of a NoSQL database, and the CAP theorem is introduced to explain tradeoffs in distributed systems.
Apache Storm is a distributed, real-time computational framework used to process unbounded streams of data from sources like messaging systems or databases. It allows building topologies with spouts that act as data sources and bolts that perform computations. Data flows between nodes as tuples through streams. Apache Kafka is a distributed publish-subscribe messaging system that stores feeds of messages in topics, allowing producers to write data and consumers to read it.
This document discusses different types of distributed databases. It covers data models like relational, aggregate-oriented, key-value, and document models. It also discusses different distribution models like sharding and replication. Consistency models for distributed databases are explained including eventual consistency and the CAP theorem. Key-value stores are described in more detail as a simple but widely used data model with features like consistency, scaling, and suitable use cases. Specific key-value databases like Redis, Riak, and DynamoDB are mentioned.
This presentation explains why NoSQL databases came over SQL databases although SQL databases has been successfully technology for more than twenty years. Moreover, This presentation discuses the characteristics and classifications of NoSQL databases. Finally, These slides cover four NoSQL databases briefly.
MySQL: Know more about open Source DatabaseMahesh Salaria
- As a developer, it is important to understand MySQL's storage engines, data types, indexing, and normalization to build high-performing applications.
- MySQL has several storage engines that handle different table types differently in terms of transactions, locking, storage, and memory usage. Choosing the right engine depends on data usage.
- Properly normalizing data, using optimal data types, and adding indexes improves performance by reducing storage needs, memory usage, and speeding up queries.
The document discusses factors to consider when selecting a NoSQL database management system (DBMS). It provides an overview of different NoSQL database types, including document databases, key-value databases, column databases, and graph databases. For each type, popular open-source options are described, such as MongoDB for document databases, Redis for key-value, Cassandra for columnar, and Neo4j for graph databases. The document emphasizes choosing a NoSQL solution based on application needs and recommends commercial support for production systems.
The document discusses moving from traditional ETL processes to "analytics with no ETL" using Hadoop. It describes how Hadoop currently supports some ETL functions by storing raw and transformed data together. However, this still requires periodic loading of new data. The vision is to support complex schemas, perform background format conversion incrementally, and enable schema inference and evolution to allow analyzing data as it arrives without explicit ETL steps. This would provide an up-to-date, performant single view of all data.
This presentation discusses using WorldView-2 satellite imagery to classify land cover in Atlanta, Georgia. It combined multi-spectral data with multi-angle observations from 13 images. Four experiments classified imagery using a nadir multi-spectral image only, full multi-angle data, and dimensionality reduction techniques. The multi-angle data improved classification accuracy by 14% over using a single nadir image alone. Specific classes like cars and highways benefited more from the multi-angle information.
Este documento presenta una introducción al municipio de Carral en España. Incluye secciones sobre su ubicación geográfica, historia desde la prehistoria hasta la época contemporánea, monumentos destacados, gastronomía local y rutas turísticas recomendadas para visitar la región.
El grupo de estudiantes visita Madrid y realiza varias actividades como ver animales en el zoo, comer en una terraza cerca de la Puerta de Alcalá y sacarse una foto con un elefante. A lo largo del día expresan opiniones sobre la comida, el clima frío y lo cansados que están.
The presentation provides an overview of NoSQL databases, including a brief history of databases, the characteristics of NoSQL databases, different data models like key-value, document, column family and graph databases. It discusses why NoSQL databases were developed as relational databases do not scale well for distributed applications. The CAP theorem is also explained, which states that only two out of consistency, availability and partition tolerance can be achieved in a distributed system.
Cloud Deployments with Apache Hadoop and Apache HBaseDATAVERSITY
The document discusses deploying Apache Hadoop and Apache HBase cloud deployments. It begins with introducing the speaker and their background with Cloudera and various Apache projects. It then provides an overview of Cloudera and what they do. The majority of the document discusses Apache Hadoop and Apache HBase, what they are, how they are open source and horizontally scalable. It also discusses deploying a Hadoop and HBase cluster on Amazon EC2 using Apache Whirr to provision the machines. Real-world examples of using these technologies include building a web index for a search engine.
In this lecture we analyze document oriented databases. In particular we consider why there are the first approach to nosql and what are the main features. Then, we analyze as example MongoDB. We consider the data model, CRUD operations, write concerns, scaling (replication and sharding).
Finally we presents other document oriented database and when to use or not document oriented databases.
The document discusses object-relational impedance mismatch and various data source patterns for mapping objects to relational databases in a way that minimizes this mismatch. It describes the table data gateway, row data gateway, active record, and data mapper patterns. The table data gateway acts as a gateway to a database table, while the row data gateway acts as a gateway to a single record. Active record wraps a database row and adds domain logic, and data mapper provides object-relational mapping to keep the object model independent from the database schema. Spring JDBC is also introduced as a framework that can help implement these patterns.
The initiation of The Hadoop Apache Hive began in 2007 by Facebook due to its data growth.
This ETL system began to fail over few years as more people joined Facebook.
In August 2008, Facebook decided to move to scalable a more scalable open-source Hadoop environment; Hive
Facebook, Netflix and Amazons support the Apache Hive SQL now known as the HiveQL
This document provides an overview of the course content for an online SAS training course. The course covers topics such as SAS basics, statistical analysis, data management, SQL, macro programming, and debugging SAS programs. It explores how to use SAS for clinical research studies and banking analysis. The course aims to teach students how to manage, analyze, and report on data with SAS.
This document provides an overview of NoSQL databases, including:
- Key-value stores store data as maps or hashmaps and are efficient for data access but limited in query capabilities.
- Column-oriented stores group attributes into column families and store data efficiently but are operationally challenging.
- Document databases store loosely structured data like JSON and allow retrieving documents by keys or contents.
- Graph databases are suited for interaction networks and path finding but are less suited for tabular data.
The document discusses the MariaDB CONNECT storage engine, which allows querying external file formats from within MariaDB. It was created by database expert Olivier Bertrand and brings business intelligence capabilities to MariaDB by enabling access to data sources like CSV, XML, Excel and other formats without needing ETL processes. The storage engine uses the MySQL plugin architecture and implements features like indexing, condition pushdown, and support for ODBC, MySQL tables, and various file types.
This document compares SQL and NoSQL databases. It defines databases, describes different types including relational and NoSQL, and explains key differences between SQL and NoSQL in areas like scaling, modeling, and query syntax. SQL databases are better suited for projects with logical related discrete data requirements and data integrity needs, while NoSQL is more ideal for projects with unrelated, evolving data where speed and scalability are important. MongoDB is provided as an example of a NoSQL database, and the CAP theorem is introduced to explain tradeoffs in distributed systems.
Apache Storm is a distributed, real-time computational framework used to process unbounded streams of data from sources like messaging systems or databases. It allows building topologies with spouts that act as data sources and bolts that perform computations. Data flows between nodes as tuples through streams. Apache Kafka is a distributed publish-subscribe messaging system that stores feeds of messages in topics, allowing producers to write data and consumers to read it.
This document discusses different types of distributed databases. It covers data models like relational, aggregate-oriented, key-value, and document models. It also discusses different distribution models like sharding and replication. Consistency models for distributed databases are explained including eventual consistency and the CAP theorem. Key-value stores are described in more detail as a simple but widely used data model with features like consistency, scaling, and suitable use cases. Specific key-value databases like Redis, Riak, and DynamoDB are mentioned.
This presentation explains why NoSQL databases came over SQL databases although SQL databases has been successfully technology for more than twenty years. Moreover, This presentation discuses the characteristics and classifications of NoSQL databases. Finally, These slides cover four NoSQL databases briefly.
MySQL: Know more about open Source DatabaseMahesh Salaria
- As a developer, it is important to understand MySQL's storage engines, data types, indexing, and normalization to build high-performing applications.
- MySQL has several storage engines that handle different table types differently in terms of transactions, locking, storage, and memory usage. Choosing the right engine depends on data usage.
- Properly normalizing data, using optimal data types, and adding indexes improves performance by reducing storage needs, memory usage, and speeding up queries.
The document discusses factors to consider when selecting a NoSQL database management system (DBMS). It provides an overview of different NoSQL database types, including document databases, key-value databases, column databases, and graph databases. For each type, popular open-source options are described, such as MongoDB for document databases, Redis for key-value, Cassandra for columnar, and Neo4j for graph databases. The document emphasizes choosing a NoSQL solution based on application needs and recommends commercial support for production systems.
The document discusses moving from traditional ETL processes to "analytics with no ETL" using Hadoop. It describes how Hadoop currently supports some ETL functions by storing raw and transformed data together. However, this still requires periodic loading of new data. The vision is to support complex schemas, perform background format conversion incrementally, and enable schema inference and evolution to allow analyzing data as it arrives without explicit ETL steps. This would provide an up-to-date, performant single view of all data.
This presentation discusses using WorldView-2 satellite imagery to classify land cover in Atlanta, Georgia. It combined multi-spectral data with multi-angle observations from 13 images. Four experiments classified imagery using a nadir multi-spectral image only, full multi-angle data, and dimensionality reduction techniques. The multi-angle data improved classification accuracy by 14% over using a single nadir image alone. Specific classes like cars and highways benefited more from the multi-angle information.
Este documento presenta una introducción al municipio de Carral en España. Incluye secciones sobre su ubicación geográfica, historia desde la prehistoria hasta la época contemporánea, monumentos destacados, gastronomía local y rutas turísticas recomendadas para visitar la región.
El grupo de estudiantes visita Madrid y realiza varias actividades como ver animales en el zoo, comer en una terraza cerca de la Puerta de Alcalá y sacarse una foto con un elefante. A lo largo del día expresan opiniones sobre la comida, el clima frío y lo cansados que están.
Windows Server 2008 and 2008 R2 provide greater control, protection, and flexibility for IT. Key features include Hyper-V virtualization, improved management with PowerShell and Server Core, enhanced security with BitLocker and OS hardening, and technologies like BranchCache and DirectAccess that improve the mobile workforce experience. The products also integrate well with Windows 7 to provide a consistent experience both inside and outside the corporate network.
The document introduces Microsoft's Windows Azure cloud platform. It summarizes that Windows Azure provides an operating system for the cloud that abstracts away hardware and provides services for automated management, scalable computing and storage. It allows developers to build applications and services that can easily scale across large, connected data centers. The talk demonstrates how Windows Azure allows building complex service architectures from simple components like web and worker roles that interact through a durable storage system. It emphasizes that the platform aims to provide a familiar development experience while handling all the complexities of highly scalable cloud services.
This document discusses Windows Small Business Server and Windows Essential Business Server solutions. It provides an overview of the different editions available and who they are designed for based on business size and growth. The Small Business Server is designed for companies with up to 75 PCs, while Essential Business Server is designed for midsize businesses with up to 300 PCs and the ability to grow. Both solutions provide integrated server technologies at a lower price point than comparable standalone products and are aimed at reducing IT complexity for small and midsize businesses.
from the Commonwealth Transportation Board's June meeting
presented to the Hampton Roads Partnership Annual Meeting, June 19, 2009
by Pierce Homer, the Commonwealth's Secretary of Transportation
Transportation Sub-Committee Meeting of 10 Dec 08 reports on alternatives and their ramifications to improving road congestion in Hampton Roads and the need to be ready to claim infrastructure stimulus money from the Obama administration. http://www.vmasc.odu.edu
This document discusses xRM (extended relationship management) applications built using Microsoft Dynamics CRM. It provides examples of how CRM can be extended beyond traditional customer relationship management to manage other types of relationships. These include applications for healthcare patient relationship management, education student information systems, government constituent relationship management, and more. The document also discusses the benefits of building xRM applications on the Dynamics CRM platform in terms of leveraging existing investments, rapid development and deployment cycles compared to custom solutions.
Windows Server 2008 and 2008 R2 provide greater control, protection, and flexibility for IT. Key features include Hyper-V virtualization, improved management with PowerShell and Server Core, enhanced security with BitLocker and OS hardening, and technologies like BranchCache and DirectAccess that improve the mobile workforce experience. The products also integrate well with Windows 7 to provide a consistent experience both inside and outside the corporate network.
1) Plastic roads use shredded plastic waste that is mixed with hot bitumen and laid like conventional tar roads.
2) Laboratory studies have shown plastic roads have improved properties like increased stability and strength compared to ordinary roads.
3) Using plastic waste in road construction provides an effective solution for plastic disposal while enhancing road quality in a more environmentally friendly manner.
The document discusses different NoSQL data models including key-value, document, column family, and graph models. It provides examples of popular NoSQL databases that implement each model such as Redis, MongoDB, Cassandra, and Neo4j. The document argues that these NoSQL databases address limitations of relational databases in supporting modern web applications with requirements for scalability, flexibility, and high performance.
Enterprise geodatabase sql access and administrationbrentpierce
The document provides an overview of accessing and administering an enterprise geodatabase through SQL and Python. It discusses how the geodatabase is based on relational database principles with user data stored in tables and system metadata stored in system tables. It describes how spatial types store geometry data and the benefits of using SQL to access and edit geodatabase content. The document also outlines how Python can be used for geodatabase administration tasks like schema creation, maintenance, and publishing tools.
Business intelligence and data warehousesDhani Ahmad
This chapter discusses business intelligence and data warehouses. It covers how operational data differs from decision support data, the components of a data warehouse including facts, dimensions and star schemas, and how online analytical processing (OLAP) and SQL extensions support analysis of multidimensional decision support data. The chapter also discusses data mining, requirements for decision support databases, and considerations for implementing a successful data warehouse project.
This document provides a summary of Oracle OpenWorld 2014 discussions on database cloud, in-memory database, native JSON support, big data, and Internet of Things (IoT) technologies. Key points include:
- Database Cloud on Oracle offers pay-as-you-go pricing and self-service provisioning similar to on-premise databases.
- Oracle Database 12c includes an in-memory option that can provide up to 100x faster analytics queries and 2-4x faster transaction processing.
- Native JSON support in 12c allows storing and querying JSON documents within the database.
- Big data technologies like Oracle Big Data SQL and Oracle Big Data Discovery help analyze large and diverse data sets from sources like
Researching an alternative to the MS SQL database - first of all in order to gain additional technological benefits, secondly moving towards an open source way of development.
The idea behind this presentation was to introduce PostgreSQL (ver. 9.4+) in a different manner than a conventional "Pros Vs. Cons" style, it is more likely to be a "Buzz Word" thesaurus (of course based on a deep research).
P.S. Since it's a presentation, there was no intention going over and covering all of the PostgreSQL features - most of the interesting parts.
An AMIS Overview of Oracle database 12c (12.1)Marco Gralike
Presentation used by Lucas Jellema and Marco Gralike during the AMIS Oracle Database 12c Launch event on Monday the 15th of July 2013 (much thanks to Tom Kyte, Oracle, for being allowed to use some of his material)
M.
AMIS organiseerde op maandagavond 15 juli het seminar ‘Oracle database 12c revealed’. Deze avond bood AMIS Oracle professionals de eerste mogelijkheid om de vernieuwingen in Oracle database 12c in actie te zien! De AMIS specialisten die meer dan een jaar bèta testen hebben uitgevoerd lieten zien wat er nieuw is en hoe we dat de komende jaren gaan inzetten!
Deze presentatie is deze avond gegeven als een plenaire sessie!
This document provides an overview and summary of key concepts related to advanced databases. It discusses relational databases including MySQL, SQL, transactions, and ODBC. It also covers database topics like triggers, indexes, and NoSQL databases. Alternative database systems like graph databases, triplestores, and linked data are introduced. Web services, XML, and data journalism are also briefly summarized. The document provides definitions and examples of these technical database terms and concepts.
The document provides an overview of database systems, including their purpose, components, and history. It discusses how database systems address issues with using file systems to store data, such as data redundancy, difficulty of accessing data, integrity problems, and concurrent access. The key components of a database system are the database management system (DBMS), data models, data definition and manipulation languages, database design, storage and querying, transaction management, architecture, users, and administrators. The relational model and SQL are introduced as widely used standards. A brief history outlines the evolution from early data processing using tapes and cards to modern database systems.
Cheetah is a custom data warehouse system built on top of Hadoop that provides high performance for storing and querying large datasets. It uses a virtual view abstraction over star and snowflake schemas to provide a simple yet powerful SQL-like query language. The system architecture utilizes MapReduce to parallelize query execution across many nodes. Cheetah employs columnar data storage and compression, multi-query optimization, and materialized views to improve query performance. Based on evaluations, Cheetah can efficiently handle both small and large queries and outperforms single-query execution when processing batches of queries together.
SKILLWISE-SSIS DESIGN PATTERN FOR DATA WAREHOUSINGSkillwise Group
This document provides an overview of the SSIS design pattern for data warehousing and change data capture. It discusses what design patterns are and how they are commonly used for SSIS and data warehousing projects. It then covers 13 specific patterns including truncate and load, slowly changing dimensions, hashbytes, change data capture, merge, and master/child workflows. The document explains when each pattern is best used and provides pros and cons. It also provides guidance on configuring and using SQL Server change data capture functionality.
The document provides an overview of database systems and their components. It discusses the purpose of database systems, database languages, data models, database internals including storage management, query processing and transaction management. It also describes different types of database users and the role of the database administrator.
SQL, NoSQL, Distributed SQL: Choose your DataStore carefullyMd Kamaruzzaman
In modern Software Development and Software Architecture, selecting the right DataStore is one of the most challenging and important task. In this presentation, I have summarized the major DataStores and the decision criteria to select the right DataStore according to the use case.
NoSQL databases were developed to address the need for databases that can handle big data and scale horizontally to support massive amounts of data and high user loads. NoSQL databases are non-relational and support high availability through horizontal scaling and replication across commodity servers to allow for continuous availability. Popular types of NoSQL databases include key-value stores, document stores, column-oriented databases, and graph databases, each suited for different use cases depending on an application's data model and query requirements.
FOSSASIA 2015 - 10 Features your developers are missing when stuck with Propr...Ashnikbiz
Ashnik Database Solution Architect, Sameer Kumar, an Open Source evangelist presented at FOSSASIA 2015 about the features of open source database like PostgreSQL which are missed by developers stuck on proprietary databases.
10 Features you would love as an Open Source developer!
- New JSON Datatype
- Vast set of datatypes supported
- Rich support for foreign Data Wrap
- User Defined Operators
- User Defined Extensions
- Filter Based Indexes or Partial Indexes
- Granular control of parameters at User, Database, Connection or Transaction Level
- Use of indexes to get statistics
- JDBC API for COPY -Command
- Full Text Search
NoSQL databases provide an alternative to traditional relational databases that is well-suited for large datasets, high scalability needs, and flexible, changing schemas. NoSQL databases sacrifice strict consistency for greater scalability and availability. The document model is well-suited for semi-structured data and allows for embedding related data within documents. Key-value stores provide simple lookup of data by key but do not support complex queries. Graph databases effectively represent network-like connections between data elements.
This document discusses data-intensive computing and provides examples of technologies used for processing large datasets. It defines data-intensive computing as concerned with manipulating and analyzing large datasets ranging from hundreds of megabytes to petabytes. It then characterizes challenges including scalable algorithms, metadata management, and high-performance computing platforms and file systems. Specific technologies discussed include distributed file systems like Lustre, MapReduce frameworks like Hadoop, and NoSQL databases like MongoDB.
The document discusses data-intensive computing and provides details about related technologies. It defines data-intensive computing as concerned with large-scale data in the hundreds of megabytes to petabytes range. Key challenges include scalable algorithms, metadata management, high-performance computing platforms, and distributed file systems. Technologies discussed include MapReduce frameworks like Hadoop, Pig, and Hive; NoSQL databases like MongoDB, Cassandra, and HBase; and distributed file systems like Lustre, GPFS, and HDFS. The document also covers programming models, scheduling, and an example application to parse Aneka logs using MapReduce.
Hyper-V is Microsoft's server virtualization technology that is included with Windows Server 2008. It allows multiple virtual machines to run on a single physical machine. Key capabilities of Hyper-V include support for large memory virtual machines up to 64GB, live migration of virtual machines between physical servers, and integration with the Windows hypervisor for security and isolation of virtual machines. System Center Virtual Machine Manager 2008 provides centralized management of virtualized and physical infrastructure across Hyper-V, Virtual Server and VMware environments.
Silverlight 2 allows developers to create rich internet applications using XAML and .NET code with the ability to access local storage, network resources, graphics, and multimedia; it provides a lightweight runtime for deploying smart client applications within web pages similar to Adobe Flash but with the added benefits of the .NET framework and local file access capabilities. Key controls and APIs include data binding, animations, graphics, audio/video playback and networking functionality for building rich interactive applications.
The document discusses Microsoft Dynamics NAV 5.0 SP1, including the overall strategy and new features. It outlines enhancements to application functionality, productivity tools, and integration capabilities. Key points include 300+ general improvements, planning engine refactoring, new document archiving and commenting features, and SQL Server performance optimizations. Upgrade considerations are also addressed.
Mesh Services are one of the underlying core services of the Live Framework. They manage users, devices, applications, and synchronization across a user's digital experiences. Key functions include identity management, directory services, storage, communications, and search capabilities to enable sharing of resources like contacts, files, and data across devices and applications.
Mogens Larsen will give a presentation on Dynamics AX 2009 Supply Chain Management from 14:15-15:15. The presentation will cover the user interface of Dynamics AX 2009, inventory management, order flow, and warehouse management. It will also include a short introduction to Dynamics AX 2009 through a PowerPoint presentation and discuss the product roadmap.
This document discusses how Visual Studio Team System can maximize ROI and drive IT governance through an integrated Application Lifecycle Management (ALM) solution. It provides concise summaries of key points, including how VSTS improves collaboration, ensures quality, integrates work frequently, and enables real-time decision making. IT governance is also discussed at a high level, focusing on compliance, metrics/reporting, and aligning IT with business needs. Examples are given of organizations seeing improvements in areas like productivity, quality, and cost reductions through an ALM approach.
This document discusses cloud computing options and dispels common myths about the cloud. It presents a cloud maturity model and suggests that CIOs should focus on reducing costs, attracting customers, and stimulating innovation. The document advocates evaluating different computing options based on needs and observing the evolving maturity of cloud computing. It warns of risks like dependency on vendors and issues with migrating systems. Overall, the document provides an overview of cloud computing and advice for CIOs on developing strategies regarding the cloud.
SOA involves exposing business functions as reusable services. This allows for greater agility, flexibility and reuse of services across different applications. SOA breaks down monolithic applications into discrete services that can be accessed over the network in a standardized way. This trend is driving the development of loosely coupled, interoperable services that can be discovered and orchestrated to meet business needs.
AI Changes Everything – Talk at Cardiff Metropolitan University, 29th April 2...Alan Dix
Talk at the final event of Data Fusion Dynamics: A Collaborative UK-Saudi Initiative in Cybersecurity and Artificial Intelligence funded by the British Council UK-Saudi Challenge Fund 2024, Cardiff Metropolitan University, 29th April 2025
https://alandix.com/academic/talks/CMet2025-AI-Changes-Everything/
Is AI just another technology, or does it fundamentally change the way we live and think?
Every technology has a direct impact with micro-ethical consequences, some good, some bad. However more profound are the ways in which some technologies reshape the very fabric of society with macro-ethical impacts. The invention of the stirrup revolutionised mounted combat, but as a side effect gave rise to the feudal system, which still shapes politics today. The internal combustion engine offers personal freedom and creates pollution, but has also transformed the nature of urban planning and international trade. When we look at AI the micro-ethical issues, such as bias, are most obvious, but the macro-ethical challenges may be greater.
At a micro-ethical level AI has the potential to deepen social, ethnic and gender bias, issues I have warned about since the early 1990s! It is also being used increasingly on the battlefield. However, it also offers amazing opportunities in health and educations, as the recent Nobel prizes for the developers of AlphaFold illustrate. More radically, the need to encode ethics acts as a mirror to surface essential ethical problems and conflicts.
At the macro-ethical level, by the early 2000s digital technology had already begun to undermine sovereignty (e.g. gambling), market economics (through network effects and emergent monopolies), and the very meaning of money. Modern AI is the child of big data, big computation and ultimately big business, intensifying the inherent tendency of digital technology to concentrate power. AI is already unravelling the fundamentals of the social, political and economic world around us, but this is a world that needs radical reimagining to overcome the global environmental and human challenges that confront us. Our challenge is whether to let the threads fall as they may, or to use them to weave a better future.
Noah Loul Shares 5 Steps to Implement AI Agents for Maximum Business Efficien...Noah Loul
Artificial intelligence is changing how businesses operate. Companies are using AI agents to automate tasks, reduce time spent on repetitive work, and focus more on high-value activities. Noah Loul, an AI strategist and entrepreneur, has helped dozens of companies streamline their operations using smart automation. He believes AI agents aren't just tools—they're workers that take on repeatable tasks so your human team can focus on what matters. If you want to reduce time waste and increase output, AI agents are the next move.
#StandardsGoals for 2025: Standards & certification roundup - Tech Forum 2025BookNet Canada
Book industry standards are evolving rapidly. In the first part of this session, we’ll share an overview of key developments from 2024 and the early months of 2025. Then, BookNet’s resident standards expert, Tom Richardson, and CEO, Lauren Stewart, have a forward-looking conversation about what’s next.
Link to recording, transcript, and accompanying resource: https://bnctechforum.ca/sessions/standardsgoals-for-2025-standards-certification-roundup/
Presented by BookNet Canada on May 6, 2025 with support from the Department of Canadian Heritage.
Big Data Analytics Quick Research Guide by Arthur MorganArthur Morgan
This is a Quick Research Guide (QRG).
QRGs include the following:
- A brief, high-level overview of the QRG topic.
- A milestone timeline for the QRG topic.
- Links to various free online resource materials to provide a deeper dive into the QRG topic.
- Conclusion and a recommendation for at least two books available in the SJPL system on the QRG topic.
QRGs planned for the series:
- Artificial Intelligence QRG
- Quantum Computing QRG
- Big Data Analytics QRG
- Spacecraft Guidance, Navigation & Control QRG (coming 2026)
- UK Home Computing & The Birth of ARM QRG (coming 2027)
Any questions or comments?
- Please contact Arthur Morgan at [email protected].
100% human made.
Procurement Insights Cost To Value Guide.pptxJon Hansen
Procurement Insights integrated Historic Procurement Industry Archives, serves as a powerful complement — not a competitor — to other procurement industry firms. It fills critical gaps in depth, agility, and contextual insight that most traditional analyst and association models overlook.
Learn more about this value- driven proprietary service offering here.
Andrew Marnell: Transforming Business Strategy Through Data-Driven InsightsAndrew Marnell
With expertise in data architecture, performance tracking, and revenue forecasting, Andrew Marnell plays a vital role in aligning business strategies with data insights. Andrew Marnell’s ability to lead cross-functional teams ensures businesses achieve sustainable growth and operational excellence.
Special Meetup Edition - TDX Bengaluru Meetup #52.pptxshyamraj55
We’re bringing the TDX energy to our community with 2 power-packed sessions:
🛠️ Workshop: MuleSoft for Agentforce
Explore the new version of our hands-on workshop featuring the latest Topic Center and API Catalog updates.
📄 Talk: Power Up Document Processing
Dive into smart automation with MuleSoft IDP, NLP, and Einstein AI for intelligent document workflows.
Mobile App Development Company in Saudi ArabiaSteve Jonas
EmizenTech is a globally recognized software development company, proudly serving businesses since 2013. With over 11+ years of industry experience and a team of 200+ skilled professionals, we have successfully delivered 1200+ projects across various sectors. As a leading Mobile App Development Company In Saudi Arabia we offer end-to-end solutions for iOS, Android, and cross-platform applications. Our apps are known for their user-friendly interfaces, scalability, high performance, and strong security features. We tailor each mobile application to meet the unique needs of different industries, ensuring a seamless user experience. EmizenTech is committed to turning your vision into a powerful digital product that drives growth, innovation, and long-term success in the competitive mobile landscape of Saudi Arabia.
TrustArc Webinar: Consumer Expectations vs Corporate Realities on Data Broker...TrustArc
Most consumers believe they’re making informed decisions about their personal data—adjusting privacy settings, blocking trackers, and opting out where they can. However, our new research reveals that while awareness is high, taking meaningful action is still lacking. On the corporate side, many organizations report strong policies for managing third-party data and consumer consent yet fall short when it comes to consistency, accountability and transparency.
This session will explore the research findings from TrustArc’s Privacy Pulse Survey, examining consumer attitudes toward personal data collection and practical suggestions for corporate practices around purchasing third-party data.
Attendees will learn:
- Consumer awareness around data brokers and what consumers are doing to limit data collection
- How businesses assess third-party vendors and their consent management operations
- Where business preparedness needs improvement
- What these trends mean for the future of privacy governance and public trust
This discussion is essential for privacy, risk, and compliance professionals who want to ground their strategies in current data and prepare for what’s next in the privacy landscape.
HCL Nomad Web – Best Practices und Verwaltung von Multiuser-Umgebungenpanagenda
Webinar Recording: https://www.panagenda.com/webinars/hcl-nomad-web-best-practices-und-verwaltung-von-multiuser-umgebungen/
HCL Nomad Web wird als die nächste Generation des HCL Notes-Clients gefeiert und bietet zahlreiche Vorteile, wie die Beseitigung des Bedarfs an Paketierung, Verteilung und Installation. Nomad Web-Client-Updates werden “automatisch” im Hintergrund installiert, was den administrativen Aufwand im Vergleich zu traditionellen HCL Notes-Clients erheblich reduziert. Allerdings stellt die Fehlerbehebung in Nomad Web im Vergleich zum Notes-Client einzigartige Herausforderungen dar.
Begleiten Sie Christoph und Marc, während sie demonstrieren, wie der Fehlerbehebungsprozess in HCL Nomad Web vereinfacht werden kann, um eine reibungslose und effiziente Benutzererfahrung zu gewährleisten.
In diesem Webinar werden wir effektive Strategien zur Diagnose und Lösung häufiger Probleme in HCL Nomad Web untersuchen, einschließlich
- Zugriff auf die Konsole
- Auffinden und Interpretieren von Protokolldateien
- Zugriff auf den Datenordner im Cache des Browsers (unter Verwendung von OPFS)
- Verständnis der Unterschiede zwischen Einzel- und Mehrbenutzerszenarien
- Nutzung der Client Clocking-Funktion
Enhancing ICU Intelligence: How Our Functional Testing Enabled a Healthcare I...Impelsys Inc.
Impelsys provided a robust testing solution, leveraging a risk-based and requirement-mapped approach to validate ICU Connect and CritiXpert. A well-defined test suite was developed to assess data communication, clinical data collection, transformation, and visualization across integrated devices.
Dev Dives: Automate and orchestrate your processes with UiPath MaestroUiPathCommunity
This session is designed to equip developers with the skills needed to build mission-critical, end-to-end processes that seamlessly orchestrate agents, people, and robots.
📕 Here's what you can expect:
- Modeling: Build end-to-end processes using BPMN.
- Implementing: Integrate agentic tasks, RPA, APIs, and advanced decisioning into processes.
- Operating: Control process instances with rewind, replay, pause, and stop functions.
- Monitoring: Use dashboards and embedded analytics for real-time insights into process instances.
This webinar is a must-attend for developers looking to enhance their agentic automation skills and orchestrate robust, mission-critical processes.
👨🏫 Speaker:
Andrei Vintila, Principal Product Manager @UiPath
This session streamed live on April 29, 2025, 16:00 CET.
Check out all our upcoming Dev Dives sessions at https://community.uipath.com/dev-dives-automation-developer-2025/.
UiPath Community Berlin: Orchestrator API, Swagger, and Test Manager APIUiPathCommunity
Join this UiPath Community Berlin meetup to explore the Orchestrator API, Swagger interface, and the Test Manager API. Learn how to leverage these tools to streamline automation, enhance testing, and integrate more efficiently with UiPath. Perfect for developers, testers, and automation enthusiasts!
📕 Agenda
Welcome & Introductions
Orchestrator API Overview
Exploring the Swagger Interface
Test Manager API Highlights
Streamlining Automation & Testing with APIs (Demo)
Q&A and Open Discussion
Perfect for developers, testers, and automation enthusiasts!
👉 Join our UiPath Community Berlin chapter: https://community.uipath.com/berlin/
This session streamed live on April 29, 2025, 18:00 CET.
Check out all our upcoming UiPath Community sessions at https://community.uipath.com/events/.
DevOpsDays Atlanta 2025 - Building 10x Development Organizations.pptxJustin Reock
Building 10x Organizations with Modern Productivity Metrics
10x developers may be a myth, but 10x organizations are very real, as proven by the influential study performed in the 1980s, ‘The Coding War Games.’
Right now, here in early 2025, we seem to be experiencing YAPP (Yet Another Productivity Philosophy), and that philosophy is converging on developer experience. It seems that with every new method we invent for the delivery of products, whether physical or virtual, we reinvent productivity philosophies to go alongside them.
But which of these approaches actually work? DORA? SPACE? DevEx? What should we invest in and create urgency behind today, so that we don’t find ourselves having the same discussion again in a decade?
Designing Low-Latency Systems with Rust and ScyllaDB: An Architectural Deep DiveScyllaDB
Want to learn practical tips for designing systems that can scale efficiently without compromising speed?
Join us for a workshop where we’ll address these challenges head-on and explore how to architect low-latency systems using Rust. During this free interactive workshop oriented for developers, engineers, and architects, we’ll cover how Rust’s unique language features and the Tokio async runtime enable high-performance application development.
As you explore key principles of designing low-latency systems with Rust, you will learn how to:
- Create and compile a real-world app with Rust
- Connect the application to ScyllaDB (NoSQL data store)
- Negotiate tradeoffs related to data modeling and querying
- Manage and monitor the database for consistently low latencies
2. SQL Server 2008 for relational and
multi-dimentional solution
developers
Silvano Coriani
[email protected]
Developer Evangelist
Microsoft
2
3. Agenda
• SQL Server 2008 support for next generation
application development
– Geospatial data type
– Filestream
– Date & Time
– Large UDT
• Simplify existing application scenario
– Table Valued Parameters
– Change Tracking
– Hierarchy ID
• Going multi-dimensional
– Developer’s roadbook to SSIS, SSAS and SSRS
3
4. Relational and Non-Relational Data
• Relational data uses simple data types
– Each type has a single value
– Generic operations work well with the types
• Relational storage/query may not be optimal for
– Hierarchical data
– Spatial data
– Sparse, variable, property bags
• Some types
– benefit by using a custom library
– use extended type system (complex types, inheritance)
– use custom storage and non-SQL APIs
– use non-relational queries and indexing
4
5. Spatial Data
• Spatial data provides answers to location-
based queries
– Which roads intersect the Microsoft campus?
– Does my land claim overlap yours?
– List all of the Italian restaurants within 5
kilometers
• Spatial data is part of almost every database
– If your database includes an address
5
6. SQL Server 2008 and Spatial Data
• SQL Server supports two spatial data types
– GEOMETRY - flat earth model
– GEOGRAPHY - round earth model
• Both types support all of the instanciable OGC
types
– InstanceOf method can distinguish between them
• Supports two dimension data
– X and Y or Lat and Long members
– Z member - elevation (user-defined semantics)
– M member - measure (user-defined semantics)
6
7. Sample Query
Which roads intersect Microsoft’s main
SELECT *
FROM roads
campus? roads.geom.STIntersects(@ms)=1
WHERE
7
8. Filestream storage
• Storing large binary objects in databases is
suboptimal
– Large objects take buffers in database memory
– Updating large objects cause database fragmentation
• In file system however, "update" is delete and insert
• "Before image" in an update is not deleted immediately
• Storing all related data in a database adds
– Transactional consistency
– Integrated, point-in-time backup and restore
– Single storage and query vehicle
8
9. SQL Server 2008 Filestream
Implementation
• A filegroup for filestream storage is declared using DDL
– Filestream storage is tied to a database
• The filegroup is mapped to a directory
– Must be NTFS file system
– Caution: Files deleteable from file system if you have
appropriate permissions
• VARBINARY(MAX) columns can be defined with
FILESTREAM attribute
– Table must also have UNIQUEIDENTIFIER column
– Filestream storage not available for other large types
• Data is stored in the file system
9
10. Programming with Filestreams
• Filestream columns are available with SQL methods
– If SQL is used, indistinguishable from varbinary(max)
• Filestream can be accessed and modified using file IO
– PathName function retrieves a symbolic path name
– Acquire context with
• GET_FILESTREAM_TRANSACTION_CONTEXT
– Use OpenSqlFilestream to get a file handle based on
• File Name
• Required Access
• Access Options
• FilestreamTransaction context
10
11. Filestream behaviors
• File IO methods are available using
– Win32 APIs (usually with SQL Native Client)
– .NET Wrapper
• Handle can use
– ReadFile, WriteFile, TransmitFile, FlushFileBuffers...
– Handle must be closed before transaction commits
– FileIO supported with ReadCommitted isolation level
• File is required for handle, so to insert
– Insert a zero-length value
– Retrieve path and transaction context
– Write using streamed IO
11
14. Table-valued Parameters (TVP)
• Input parameters of Table type on
SPs/Functions
• Optimized to scale and perform better for
large data CREATE TYPE myTableType AS TABLE
(id INT, name NVARCHAR(100),qty
• Behaves like BCP in server INT);
CREATE PROCEDURE myProc (@tvp
• Simple programming model myTableType READONLY) AS
UPDATE Inventory SET
• Strongly typed
qty += s.qty
FROM Inventory AS i INNER JOIN
@tvp AS tvp
• Reduce client/server round trips GO
ON i.id = tvp.id
• Do not cause a statement to recompile
14
15. Table-valued Parameters (TVP)
TVP Client Stack Support
• Fully supported in ADO.NET 3
• New Parameter type:
SqlDbType.Structured
• Parameters can be passed in multiple ways
– DataTable
– IEnumerable<SqlDataRecord> (fully streamed)
– DbDataReader
15
16. Hierarchical Data
• Hierarchical data consists of nodes and edges
– In employee-boss relationship, employee and boss are
each nodes, the relationship between them is an edge
• Hierarchical data can be modeled in relational as
– Adjacency model - separate column for edge
• Most common, column can either be in same or separate
table
– Path Enumeration model - column w/hierarchical path
– Nested Set model - adds "left" and "right" columns to
represent edges, which must be maintained
separately
16
17. SQL Server 2008 and Hierarchical Data
• New Built-In Data Type - HierarchyID
• SQLCLR based system UDT
– Useable on .NET clients directly as SqlHierarchyId
• An implementation of path enumeration
model
– Uses ORDPATH internally for speed
17
18. HierarchyID
• Depth-first indexing
• "Level" property - allows breadth-first indexing
• Methods for common hierarchical operations
– GetRoot
– GetLevel
– IsDescendant
– GetDescendant, GetAncestor
– Reparent
• Does not enforce tree structure
– Can enforce tree using constraints
18
20. Sparse Properties
• Many designs require sparse properties
– Hardware store has different attributes for each
product
– Lab tests have different readings for each test
– Directory systems have different attributes for
each item
• These are name-value pairs (property bags)
• Because they don't appear on each tuple
(row) they are difficult to model
20
21. Modeling Sparse Properties
• Sparse Properties often modeled as separate table
– Base table has one row per item - common properties
– Property table has N rows per item - one per property
– Known as Entity-Attribute-Value
• Can be modeled as sparse tables
– 256 table limit in SQL Server JOIN
• Can be modeled as sparse columns
– 1024 column limit in SQL Server tables
• Can be modeled as XML
– Common properties are elements, sparse are attributes
21
22. SQL Server 2008 and Sparse Columns
• Sparse Column extends column limit
• Still 1024 column limit for "non-sparse"
columns
• Over 1024 (10000) for sparse columns
• Column marked as SPARSE in table definition
• Additional column represents all sparse
column name value pairs as attributes in a
single XML element
22
23. Change Tracking
• 3 different “flavor” of tracking data changes in SQL
Server 2008
– Change Tracking, CDC (used in DW), Auditing (security-
oriented)
• Keeps track of data modifications in a table
– Lightweight (No trigger, No schema changes)
• Overhead similar to a traditional index
– Synchronous at commit time
– Gives you access to “net changes” from T0
• Doesn’t keep track of “historical” changes
23
24. Why go multi-dimensional?
• Organizations have large volumes of related data stored in a
variety of data systems, often in different formats
• Data systems may not…
– Be optimized for analytical queries
– Contain all the data required by design or by time
– Manage historical context
– Be available or accessible
• Non-technical employees and managers may not have
sufficient skills, tools, or permissions to query data systems
• Systems may not have universal definitions of an entity
• Analytical queries & reporting can impact operational system
performance
24
25. A realistic scenario
• Data source
independence
– Can survive OLTP
system changes
– Heterogeneous data
source
• Single version of the truth
– Data Warehouse data
centralization
– Data Mart as specific
model for analysis
– Data Mart is user
oriented, not Data
Warehouse
• Some tools can be used
also by OLTP solutions
– Reporting Services
– OLTP queries
25 25
26. The Microsoft BI Platform
SQL Server 2008
Integrate Store
Analyze Report
26
27. New with Microsoft SQL Server 2008
Integration & Data Warehousing
• Scale and Manage large number of users and
data
– Improved Query performance on large tables Enhanced Partitioning
– Queries Optimized for data warehousing
scenarios DW Query Optimizations
– Increase I/O performance with efficient and
cost effective data storage Data Compression
– Manage concurrent workloads of ad-hoc
queries, reporting and analysis Resource Governor
• Integrate growing volumes of data
Persistent Lookups
– Optimize your ETL performance by identifying
data in your largest tables
Change Data Capture
– Reduce the data load volumes by capturing
operational changes in data
MERGE SQL Statement
– Simplify your insert and update data processing
– Profile your information to identify dirty data Data Profiling
27
28. Enterprise-class Data Integration with
SQL Server Integration Services
• Scalable Integrations
– Connect to data
– Multi-threaded architecture
– Comprehensive transformations
– Profile your data
– Cleanse your data
• Data Quality
– Cleanse data
– Text Mining
– Identify dirty data
28
29. Rich Connectivity
• Extensive Connectivity
– Standards based support
Unstructured data
– XML, Flat Files and Excel
– Binary Files
Legacy data: Binary files – BizTalk, MS Message Queues
– Oracle, DB2 and SQL Server
Application database
– Partner Ecosystem
OLTP • Change Data Capture
– Transparently capture changes
Change
Tables – Real time integration
DW
29
30. Rich Connectivity
Data Providers
ODBC
SQL Server SAP
NetWeaver BI SQL Server
Report Server Models
SQL Server
Integration Services Teradata
XML
OLE DB
DB2
MySAP SQL Server
Data Mining Models
Oracle
SQL Server
Analysis Services
Hyperion Essbase
30
32. New with Microsoft SQL Server 2008
Analysis Services
Innovative Cube Designer
Best Practice Design Alerts
Enhanced Dimension Design
Enhanced Aggregation Design
New Subspace Computations
MOLAP Enabled Write-Back
Enhanced Back-Up Scalability
New Resource Monitor
Execution Plan
32
34. New with Microsoft SQL Server 2008
Reporting Services
New Report Designer
Enhanced Data Visualization
New Flexible Report Layout
Scalable Report Engine
Single Service Architecture
New Word Rendering
Improved Excel Rendering
New End User Design Experience
SharePoint Integration
34
35. The complete flow
OLTP Client Portal
Analytical Applications
(MBS, third-party)
Office/SharePoint/PPS
Query and
CRM DW, Reporting
ERP ODS
Integration Analytical Devices
Services Data Analysis Components
LOB (ETL) (OLAP, DM)
Analytic Platform
.NET Framework (IIS, ASP, Net, CLR) and SQL Server
(Relational, Multidimensional, XML)
BI Development and Management Tools SQL Server Management Tools
35
36. Languages, APIs, And SDKs
• MDX + DMX
• ADO MD.NET
– AdoMdClient and AdoMdServer
• XML/A
• AMO
• RDL
• Report Server Web Service, RS URL Access,
and RS Extensions
36
37. Develop Custom Client Applications
• Using ADO MD.NET, AMO, and XMLA in your
own applications
• Front-ending RS and ProClarity
• Integrating with AdoMdServer and
server-side assemblies
• Using Data Mining Model Viewer controls
• Visualization with WPF and Silverlight
37
38. Summary
• Microsoft SQL Server and his services are the basement
for a complete solution, from data access to analysis, from
data consolidation to performance management
• Together with other Microsoft technologies can be used
by Developers and IT Professionals to build powerful and
flexible reporting and analysis solutions for the end users
• Several class libraries and protocols helps solution
developers to integrate these components in line of
business applications in a easy and natural way
– .NET Framework languages and technologies are the glue that
connect these building blocks together
38
39. Don’t forget the evalutations!!
• Fill the evaluations and you’ll get
– Windows Home Server (1st day)
– Windows 7 Beta (2nd day)
39