Your cart is empty
A practical guide to making good decisions in a world of missing data In the era of big data, it is easy to imagine that we have all the information we need to make good decisions. But in fact the data we have are never complete, and may be only the tip of the iceberg. Just as much of the universe is composed of dark matter, invisible to us but nonetheless present, the universe of information is full of dark data that we overlook at our peril. In Dark Data, data expert David Hand takes us on a fascinating and enlightening journey into the world of the data we don't see. Dark Data explores the many ways in which we can be blind to missing data and how that can lead us to conclusions and actions that are mistaken, dangerous, or even disastrous. Examining a wealth of real-life examples, from the Challenger shuttle explosion to complex financial frauds, Hand gives us a practical taxonomy of the types of dark data that exist and the situations in which they can arise, so that we can learn to recognize and control for them. In doing so, he teaches us not only to be alert to the problems presented by the things we don't know, but also shows how dark data can be used to our advantage, leading to greater understanding and better decisions. Today, we all make decisions using data. Dark Data shows us all how to reduce the risk of making bad ones.
An introduction to the field of applied ontology with examples derived particularly from biomedicine, covering theoretical components, design practices, and practical applications. In the era of "big data," science is increasingly information driven, and the potential for computers to store, manage, and integrate massive amounts of data has given rise to such new disciplinary fields as biomedical informatics. Applied ontology offers a strategy for the organization of scientific information in computer-tractable form, drawing on concepts not only from computer and information science but also from linguistics, logic, and philosophy. This book provides an introduction to the field of applied ontology that is of particular relevance to biomedicine, covering theoretical components of ontologies, best practices for ontology design, and examples of biomedical ontologies in use. After defining an ontology as a representation of the types of entities in a given domain, the book distinguishes between different kinds of ontologies and taxonomies, and shows how applied ontology draws on more traditional ideas from metaphysics. It presents the core features of the Basic Formal Ontology (BFO), now used by over one hundred ontology projects around the world, and offers examples of domain ontologies that utilize BFO. The book also describes Web Ontology Language (OWL), a common framework for Semantic Web technologies. Throughout, the book provides concrete recommendations for the design and construction of domain ontologies.
Three books by the bestselling authors on Data Warehousing The most authoritative guides from the inventor of the technique all for a value price.
The Data Warehouse Toolkit, 3rd Edition (9781118530801) Ralph Kimball invented a data warehousing technique called "dimensional modeling" and popularized it in his first Wiley book, The Data Warehouse Toolkit. Since this book was first published in 1996, dimensional modeling has become the most widely accepted technique for data warehouse design. Over the past 10 years, Kimball has improved on his earlier techniques and created many new ones. In this 3rd edition, he will provide a comprehensive collection of all of these techniques, from basic to advanced.
The Data Warehouse Lifecycle Toolkit, 2nd Edition (9780470149775) Complete coverage of best practices from data warehouse project inception through on-going program management. Updates industry best practices to be in sync with current recommendations of Kimball Group. Streamlines the lifecycle methodology to be more efficient and user-friendly
The Data Warehouse ETL Toolkit (9780764567575) shows data warehouse developers how to effectively manage the ETL (Extract, Transform, Load) phase of the data warehouse development lifecycle. The authors show developers the best methods for extracting data from scattered sources throughout the enterprise, removing obsolete, redundant, and innaccurate data, transforming the remaining data into correctly formatted data structures, and then physically loading them into the data warehouse. This book provides complete coverage of proven, time-saving ETL techniques. It begins with a quick overview of ETL fundamentals and the role of the ETL development team. It then quickly moves into an overview of the ETL data structures, both relational and dimensional. The authors show how to build useful dimensional stuctures, providing practical examples of beginning through advanced techniques.
A practical guide to data mining using SQL and Excel Data Analysis Using SQL and Excel, 2nd Edition shows you how to leverage the two most popular tools for data query and analysis SQL and Excel to perform sophisticated data analysis without the need for complex and expensive data mining tools. Written by a leading expert on business data mining, this book shows you how to extract useful business information from relational databases. You'll learn the fundamental techniques before moving into the "where" and "why" of each analysis, and then learn how to design and perform these analyses using SQL and Excel. Examples include SQL and Excel code, and the appendix shows how non-standard constructs are implemented in other major databases, including Oracle and IBM DB2/UDB. The companion website includes datasets and Excel spreadsheets, and the book provides hints, warnings, and technical asides to help you every step of the way. Data Analysis Using SQL and Excel, 2nd Edition shows you how to perform a wide range of sophisticated analyses using these simple tools, sparing you the significant expense of proprietary data mining tools like SAS. * Understand core analytic techniques that work with SQL and Excel * Ensure your analytic approach gets you the results you need * Design and perform your analysis using SQL and Excel Data Analysis Using SQL and Excel, 2nd Edition shows you how to best use the tools you already know to achieve expert results.
Written in lucid language, this valuable textbook brings together fundamental concepts of data mining and data warehousing in a single volume. Important topics including information theory, decision tree, Naive Bayes classifier, distance metrics, partitioning clustering, associate mining, data marts and operational data store are discussed comprehensively. The textbook is written to cater to the needs of undergraduate students of computer science, engineering and information technology for a course on data mining and data warehousing. The text simplifies the understanding of the concepts through exercises and practical examples. Chapters such as classification, associate mining and cluster analysis are discussed in detail with their practical implementation using Weka and R language data mining tools. Advanced topics including big data analytics, relational data models and NoSQL are discussed in detail. Pedagogical features including unsolved problems and multiple-choice questions are interspersed throughout the book for better understanding.
Imagine spending a day with top analytical leaders and asking any question you want. In this book, Wayne Eckerson illustrates analytical best practices by weaving his perspective with commentary from seven directors of analytics who unveil their secrets of success. With an innovative flair, Eckerson tackles a complex subject with clarity and insight. Each of the books 20 chapters is a stand-alone essay on an analytical topic, yet collectively they form a concise methodology about how to implement a successful analytics program.
Let's step back to the year 1978. Sony introduces hip portable music with the Walkman, Illinois Bell Company releases the first mobile phone, Space Invaders kicks off the video game craze, and William Kent writes this book. We have made amazing progress in the last four decades in terms of portable music, mobile communication, and entertainment, making devices such as the original Sony Walkman and suitcase-sized mobile phones museum pieces today. Yet remarkably, the book Data and Reality is just as relevant to the field of data management today as it was in 1978. This book gracefully weaves the disciplines of psychology and philosophy with data management to create timeless takeaways on how we perceive and manage information. Although databases and related technology have come a long way since 1978, the process of eliciting business requirements and how we think about information remains constant. This book will provide valuable insights whether you are a 1970s data-processing expert or a modern-day business analyst, data modeller, database administrator, or data architect. This 3rd edition differs substantially from the first and second editions. Data modelling thought leader Steve Hoberman has updated many of the original examples and references and added his commentary throughout the book, including key points at the end of each chapter. The important takeaways in this book are rich with insight yet presented in a conversational and easy-to-grasp writing style. Here are just a few of the issues this book tackles: Has "business intelligence" replaced "artificial intelligence"? Why is a maps geographic landscape analogous to a data models information landscape? Where do forward and reverse engineering fit in our thought process? Why are we all becoming "data archaeologists"? What causes the communication chasm between the business professional and the information technology professional in most organisations, and how can the logical data model help bridge this chasm? Why do we invest in hardware and software to solve business problems before determining what the business problems are in the first place? What is the difference between oneness, sameness, and categories? Why does context play a role in every design decision? Why do the more important attributes become entities or relationships? Why do symbols speak louder than words? Whats the difference between a data modeller, a philosopher, and an artist? Why is the 1975 dream of mapping all attributes still a dream today? What influence does language have on our perception of reality? Can we distinguish between naming and describing?
Publisher's Note: Products purchased from Third Party sellers are not guaranteed by the publisher for quality, authenticity, or access to any online entitlements included with the product. Develop a custom, agile data warehousing and business intelligence architectureEmpower your users and drive better decision making across your enterprise with detailed instructions and best practices from an expert developer and trainer. The Data Warehouse Mentor: Practical Data Warehouse and Business Intelligence Insights shows how to plan, design, construct, and administer an integrated end-to-end DW/BI solution. Learn how to choose appropriate components, build an enterprise data model, configure data marts and data warehouses, establish data flow, and mitigate risk. Change management, data governance, and security are also covered in this comprehensive guide. Understand the components of BI and data warehouse systems Establish project goals and implement an effective deployment plan Build accurate logical and physical enterprise data models Gain insight into your company's transactions with data mining Input, cleanse, and normalize data using ETL (Extract, Transform, and Load) techniques Use structured input files to define data requirements Employ top-down, bottom-up, and hybrid design methodologies Handle security and optimize performance using data governance tools Robert Laberge is the founder of several Internet ventures and a principle consultant for the IBM Industry Models and Assets Lab, which has a focus on data warehousing and business intelligence solutions.
The final edition of the incomparable data warehousing and business intelligence reference, updated and expanded The Kimball Group Reader, Remastered Collection is the essential reference for data warehouse and business intelligence design, packed with best practices, design tips, and valuable insight from industry pioneer Ralph Kimball and the Kimball Group. This Remastered Collection represents decades of expert advice and mentoring in data warehousing and business intelligence, and is the final work to be published by the Kimball Group. Organized for quick navigation and easy reference, this book contains nearly 20 years of experience on more than 300 topics, all fully up-to-date and expanded with 65 new articles. The discussion covers the complete data warehouse/business intelligence lifecycle, including project planning, requirements gathering, system architecture, dimensional modeling, ETL, and business intelligence analytics, with each group of articles prefaced by original commentaries explaining their role in the overall Kimball Group methodology. Data warehousing/business intelligence industry's current multi-billion dollar value is due in no small part to the contributions of Ralph Kimball and the Kimball Group. Their publications are the standards on which the industry is built, and nearly all data warehouse hardware and software vendors have adopted their methods in one form or another. This book is a compendium of Kimball Group expertise, and an essential reference for anyone in the field. * Learn data warehousing and business intelligence from the field's pioneers * Get up to date on best practices and essential design tips * Gain valuable knowledge on every stage of the project lifecycle * Dig into the Kimball Group methodology with hands-on guidance Ralph Kimball and the Kimball Group have continued to refine their methods and techniques based on thousands of hours of consulting and training. This Remastered Collection of The Kimball Group Reader represents their final body of knowledge, and is nothing less than a vital reference for anyone involved in the field.
Do you enjoy completing puzzles? Perhaps one of the most challenging (yet rewarding) puzzles is delivering a successful data warehouse suitable for data mining and analytics. The Analytical Puzzle describes an unbiased, practical, and comprehensive approach to building a data warehouse which will lead to an increased level of business intelligence within your organisation. New technologies continuously impact this approach and therefore this book explains how to leverage big data, cloud computing, data warehouse appliances, data mining, predictive analytics, data visualisation and mobile devices. This book describes an unbiased, practical, and comprehensive approach to building a data warehouse which will lead to an increased level of business intelligence within your organisation. New technologies continuously impact this approach and therefore this book explains how to leverage big data, cloud computing, data warehouse appliances, data mining, predictive analytics, data visualisation and mobile devices.
Best practices and invaluable advice from world-renowned data warehouse experts
In this book, leading data warehouse experts from the Kimball Group share best practices for using the upcoming "Business Intelligence release" of SQL Server, referred to as SQL Server 2008 R2. In this new edition, the authors explain how SQL Server 2008 R2 provides a collection of powerful new tools that extend the power of its BI toolset to Excel and SharePoint users and they show how to use SQL Server to build a successful data warehouse that supports the business intelligence requirements that are common to most organizations. Covering the complete suite of data warehousing and BI tools that are part of SQL Server 2008 R2, as well as Microsoft Office, the authors walk you through a full project lifecycle, including design, development, deployment and maintenance.Features more than 50 percent new and revised material that covers the rich new feature set of the SQL Server 2008 R2 release, as well as the Office 2010 releaseIncludes brand new content that focuses on PowerPivot for Excel and SharePoint, Master Data Services, and discusses updated capabilities of SQL Server Analysis, Integration, and Reporting ServicesShares detailed case examples that clearly illustrate how to best apply the techniques described in the bookThe accompanying Web site contains all code samples as well as the sample database used throughout the case studies
"The Microsoft Data Warehouse Toolkit, Second Edition" provides you with the knowledge of how and when to use BI tools such as Analysis Services and Integration Services to accomplish your most essential data warehousing tasks.
Using Power Query, you can import, reshape, and cleanse any data from a simple interface, so you can mine that data for all of its hidden insights. Power Query is embedded in Excel, Power BI, and other Microsoft products, and leading Power Query expert Gil Raviv will help you make the most of it. Discover how to eliminate time-consuming manual data preparation, solve common problems, avoid pitfalls, and more. Then, walk through several complete analytics challenges, and integrate all your skills in a realistic chapter-length final project. By the time you're finished, you'll be ready to wrangle any data-and transform it into actionable knowledge. Prepare and analyze your data the easy way, with Power Query - Quickly prepare data for analysis with Power Query in Excel (also known as Get & Transform) and in Power BI - Solve common data preparation problems with a few mouse clicks and simple formula edits - Combine data from multiple sources, multiple queries, and mismatched tables - Master basic and advanced techniques for unpivoting tables - Customize transformations and build flexible data mashups with the M formula language - Address collaboration challenges with Power Query - Gain crucial insights into text feeds - Streamline complex social network analytics so you can do it yourself For all information workers, analysts, and any Excel user who wants to solve their own business intelligence problems.
Learn essential techniques from data warehouse legend Bill Inmon on how to build the reporting environment your business needs now! Answers for many valuable business questions hide in text. How well can your existing reporting environment extract the necessary text from email, spreadsheets, and documents, and put it in a useful format for analytics and reporting? Transforming the traditional data warehouse into an efficient unstructured data warehouse requires additional skills from the analyst, architect, designer, and developer. This book will prepare you to successfully implement an unstructured data warehouse and, through clear explanations, examples, and case studies, you will learn new techniques and tips to successfully obtain and analyse text. Master these ten objectives: Build an unstructured data warehouse using the 11-step approach; Integrate text and describe it in terms of homogeneity, relevance, medium, volume, and structure; Overcome challenges including blather, the Tower of Babel, and lack of natural relationships; Avoid the Data Junkyard and combat the "Spiders Web"; Reuse techniques perfected in the traditional data warehouse and Data Warehouse 2.0 , including iterative development; Apply essential techniques for textual Extract, Transform, and Load (ETL) such as phrase recognition, stop word filtering, and synonym replacement; Design the Document Inventory system and link unstructured text to structured data; Leverage indexes for efficient text analysis and taxonomies for useful external categorisation; Manage large volumes of data using advanced techniques such as backward pointers; Evaluate technology choices suitable for unstructured data processing, such as data warehouse appliances.
Publisher's Note: Products purchased from Third Party sellers are not guaranteed by the publisher for quality, authenticity, or access to any online entitlements included with the product. Foreword by Mark Stephen LaRow, Vice President of Products, MicroStrategy "A unique and authoritative book that blends recent research developments with industry-level practices for researchers, students, and industry practitioners."Il-Yeol Song, Professor, College of Information Science and Technology, Drexel University
Provides the fundamentals, technologies, and best practices in designing, constructing and managing mission critical, energy efficient data centers Organizations in need of high-speed connectivity and nonstop systems operations depend upon data centers for a range of deployment solutions. A data center is a facility used to house computer systems and associated components, such as telecommunications and storage systems. It generally includes multiple power sources, redundant data communications connections, environmental controls (e.g., air conditioning, fire suppression) and security devices. With contributions from an international list of experts, The Data Center Handbook instructs readers to: * Prepare strategic plan that includes location plan, site selection, roadmap and capacity planning * Design and build "green" data centers, with mission critical and energy-efficient infrastructure * Apply best practices to reduce energy consumption and carbon emissions * Apply IT technologies such as cloud and virtualization * Manage data centers in order to sustain operations with minimum costs * Prepare and practice disaster reovery and business continuity plan The book imparts essential knowledge needed to implement data center design and construction, apply IT technologies, and continually improve data center operations.
Are you struggling with the formal design of your organisation's data resource? Do you find yourself forced into generic data architectures and universal data models? Do you find yourself warping the business to fit a purchased application? Do you find yourself pushed into developing physical databases without formal logical design? Do you find disparate data throughout the organisation? If the answer to any of these questions is Yes, then you need to read Data Resource Design to help guide you through a formal design process that produces a high quality data resource within a single common data architecture. Most public and private sector organisations do not consistently follow a formal data resource design process that begins with the organisation's perception of the business world, proceeds through logical data design, through physical data design, and into implementation. Most organisations charge ahead with physical database implementation, physical package implementation, and other brute-force-physical approaches. The result is a data resource that becomes disparate and does not fully support the organisation in its business endeavours. This book describes how to formally design an organisation's data resource to meet its current and future business information demand. It builds on "Data Resource Simplexity", which described how to stop the burgeoning data disparity, and on "Data Resource Integration", which described how to understand and resolve an organisation's disparate data resource. It describes the concepts, principles, and techniques for building a high quality data resource based on an organisation's perception of the business world in which they operate. Like "Data Resource Simplexity" and "Data Resource Integration", Michael Brackett draws on five decades of data management experience building and managing data resources, and resolving disparate data in both public and private sector organisations. He leverages theories, concepts, principles, and techniques from a wide variety of disciplines, such as human dynamics, mathematics, physics, chemistry, philosophy, and biology, and applies them to properly designing data as a critical resource of an organisation. He shows how to understand the business environment where an organisation operates and design a data resource that supports the organisation in that business environment.
With this textbook, Vaisman and Zimanyi deliver excellent coverage of data warehousing and business intelligence technologies ranging from the most basic principles to recent findings and applications. To this end, their work is structured into three parts. Part I describes Fundamental Concepts including multi-dimensional models; conceptual and logical data warehouse design and MDX and SQL/OLAP. Subsequently, Part II details Implementation and Deployment, which includes physical data warehouse design; data extraction, transformation, and loading (ETL) and data analytics. Lastly, Part III covers Advanced Topics such as spatial data warehouses; trajectory data warehouses; semantic technologies in data warehouses and novel technologies like Map Reduce, column-store databases and in-memory databases.
As a key characteristic of the book, most of the topics are presented and illustrated using application tools. Specifically, a case study based on the well-known Northwind database illustrates how the concepts presented in the book can be implemented using Microsoft Analysis Services and Pentaho Business Analytics. All chapters are summarized using review questions and exercises to support comprehensive student learning. Supplemental material to assist instructors using this book as a course text is available at http: //cs.ulb.ac.be/DWSDIbook/, including electronic versions of the figures, solutions to all exercises, and a set of slides accompanying each chapter.
Overall, students, practitioners and researchers alike will find this book the most comprehensive reference work on data warehouses, with key topics described in a clear and educational style."
Direct from Microsoft, this Exam Ref is the official study guide for the new Microsoft 70-767 Implementing a SQL Data Warehouse certification exam. Exam Ref 70-767 Implementing a SQL Data Warehouse offers professional-level preparation that helps candidates maximize their exam performance and sharpen their skills on the job. It focuses on the specific areas of expertise modern IT professionals need to successfully build modern data warehouses to support advanced business intelligence solutions. Coverage includes: Designing and implementing an effective data warehouse, including dimension tables, fact tables, indexes, storage, partitioning, and more Establishing successful processes for extracting, transforming, and loading data (ETL) with SQL Server Integration Services (SSIS) and Transact-SQL (T-SQL) Integrating solutions that encompass cloud data and big data, using Polybase Integrate, Hadoop Integrate, the Azure Blob service, and related tools Ensuring high levels of data quality with a Data Quality Services (DQS) knowledge base and a Master Data Services (MDS) model Microsoft Exam Ref publications stand apart from third-party study guides because they: Provide guidance from Microsoft, the creator of Microsoft certification exams Target IT professional-level exam candidates with content focused on their needs, not "one-size-fits-all" content Streamline study by organizing material according to the exam's objective domain (OD), covering one functional group and its objectives in each chapter Feature Thought Experiments to guide candidates through a set of "what if?" scenarios, and prepare them more effectively for Pro-level style exam questions Explore big picture thinking around the planning and design aspects of the IT pro's job role For more information on Exam 70-767 and the MCSA: SQL 2016 Business Intelligence Development credential, visit microsoft.com/learning.
You may like...
Data Warehousing Fundamentals for IT…
Paulraj Ponniah Hardcover
Business unIntelligence - Insight…
Barry Devlin Paperback
Building The Data Warehouse
W.H. Inmon Paperback
Software-Defined Data Infrastructure…
Greg Schulz Hardcover R1,775 Discovery Miles 17 750
Data Resource Integration…
Michael H. Brackett Paperback
Testen Von Data-Warehouse- Und…
Oracle SOA Suite 12c Handbook
Lucas Jellema Paperback
Ontology-Based Data Access Leveraging…
Gerardo I. Simari, Cristian Molinaro, … Paperback
The Absolute Guide to Dashboarding and…
Kasper De Jonge Paperback
Data Deduplication for Data Optimization…
Daehee Kim, Sejun Song, … Hardcover