Your cart is empty
The increasing volume of data in modern business and science calls for more complex and sophisticated tools. Although advances in data mining technology have made extensive data collection much easier, it s still always evolving and there is a constant need for new techniques and tools that can help us transform this data into useful information and knowledge.
Since the previous edition s publication, great advances have
been made in the field of data mining. Not only does the third of
edition of "Data Mining: Concepts and Techniques" continue the
tradition of equipping you with an understanding and application of
the theory and practice of discovering patterns hidden in large
data sets, it also focuses on new, important topics in the field:
data warehouses and data cube technology, mining stream, mining
social networks, and mining spatial, multimedia and other complex
data. Each chapter is a stand-alone guide to a critical topic,
presenting proven algorithms and sound implementations ready to be
used directly or with strategic modification against live data.
This is the resource you need if you want to apply today s most
powerful data mining techniques to meet real business
* Presents dozens of algorithms and implementation examples, all in pseudo-code and suitable for use in real-world, large-scale data mining projects. * Addresses advanced topics such as mining object-relational databases, spatial databases, multimedia databases, time-series databases, text databases, the World Wide Web, and applications in several fields. *Provides a comprehensive, practical look at the concepts and techniques you need to get the most out of your data"
Information Modeling and Relational Databases, second edition,
provides an introduction to ORM (Object-Role Modeling)and much
more. In fact, it is the only book to go beyond introductory
coverage and provide all of the in-depth instruction you need to
transform knowledge from domain experts into a sound database
design. This book is intended for anyone with a stake in the
accuracy and efficacy of databases: systems analysts, information
modelers, database designers and administrators, and programmers.
"Data Mining: Practical Machine Learning Tools and Techniques" offers a thorough grounding in machine learning concepts as well as practical advice on applying machine learning tools and techniques in real-world data mining situations. This highly anticipated third edition of the most acclaimed work on data mining and machine learning will teach you everything you need to know about preparing inputs, interpreting outputs, evaluating results, and the algorithmic methods at the heart of successful data mining.
Thorough updates reflect the technical changes and
modernizations that have taken place in the field since the last
edition, including new material on Data Transformations, Ensemble
Learning, Massive Data Sets, Multi-instance Learning, plus a new
version of the popular Weka machine learning software developed by
the authors. Witten, Frank, and Hall include both tried-and-true
techniques of today as well as methods at the leading edge of
*Provides a thorough grounding in machine learning concepts as well as practical advice on applying the tools and techniques to your data mining projects *Offers concrete tips and techniques for performance improvement that work by transforming the input or output in machine learning methods *Includes downloadable Weka software toolkit, a collection of machine learning algorithms for data mining tasks-in an updated, interactive interface. Algorithms in toolkit cover: data pre-processing, classification, regression, clustering, association rules, visualization
"Data Modeling Essentials, Third Edition" provides expert tutelage
for data modelers, business analysts and systems designers at all
levels. Beginning with the basics, this book provides a thorough
grounding in theory before guiding the reader through the various
stages of applied data modeling and database design. Later chapters
address advanced subjects, including business rules, data
warehousing, enterprise-wide modeling and data management.
The key to client/server computing.
The authors provide detailed explanations of why various
problems occur as
This book brings all of the elements of database design together in
a single volume, saving the reader the time and expense of making
multiple purchases. It consolidates both introductory and advanced
topics, thereby covering the gamut of database design methodology ?
from ER and UML techniques, to conceptual data modeling and table
transformation, to storing XML and querying moving objects
The demand for SQL information and training continues to grow
with the need for a database behind every website capable of
offering web-based information queries. SQL is the de facto
standard for database retrieval, and if you need to access, update,
or utilize data in a modern database management system, youwill
need SQL to do it. TheSecond Editionof "Joe Celko's Trees and
Hierarchies in SQL for Smarties" covers two new sets of extensions
over three entirelynew chapters and expounds upon the changes that
have occurred in SQL standards since the previous edition's
publication. Benefit from mastering the challenging aspects of
these database applications in SQL as taught by Joe Celko, one of
the most-read SQL authors in the world.
*Expert advice from a noted SQL authority and award-winning columnist who has given 10 years of service to the ANSI SQL standards committee
*Teaches scores of advanced techniques that can be used with any product, in any SQL environment
*Offers graph theory and programming techniques for working around deficiencies and gives insight into real-world challenges"
A multitude of problems is likely to arise when developing data models. With dozens of attributes and millions of rows, data modelers are always in danger of inconsistency and inaccuracy. The development of the data model itself could result in difficulties presenting accurate data. The need to improve data models begins with getting it right in the first place.
Using real-world examples, "Developing High Quality Data Models"
walks the reader through identifying a number of data modeling
principles and analysis techniques that enable the development of
data models that both meet business requirements and have a
consistent basis. The reader is presented with a variety of generic
data model patterns that both exemplify the principles and
techniques discussed and build upon one another to give a powerful
and integrated generic data model. This model has wide
applicability across many disciplines in government and industry,
including but not limited to energy exploration, healthcare,
telecommunications, transportation, military defense,
transportation, and more.
* Uses a number of common data model patterns to explain how to develop data models over a wide scope in a way that is consistent and of high quality
*Offers generic data model templates that are reusable in many applications and are fundamental for developing more specific templates
*Develops ideas for creating consistent approaches to high quality data models
Would the basics of SQL querying be useful to you, but you don't
want to start at the very beginning? Do you wish you had an easy
way to ramp up quickly to get a basic understanding of key features
and capability? Then you need this book! Without a ton of
conceptual information or general programming basics you already
know, this book is a quick guide for computing professionals and
programmers to learn the basics--and more--in an easily digestible
"How to Build a Digital Library" is the only book that offers
all the knowledge and tools needed to construct and maintain a
digital library, regardless of the size or purpose. It is the
perfectly self-contained resource for individuals, agencies, and
institutions wishing to put this powerful tool to work in their
burgeoning information treasuries. The Second Edition reflects new
developments in the field as well as in the Greenstone Digital
Library open source software. In Part I, the authors have added an
entire new chapter on user groups, user support, collaborative
browsing, user contributions, and so on. There is also new material
on content-based queries, map-based queries, cross-media queries.
There is an increased emphasis placed on multimedia by adding a
"digitizing" section to each major media type. A new chapter has
also been added on "internationalization," which will address
Unicode standards, multi-language interfaces and collections, and
issues with non-European languages (Chinese, Hindi, etc.). Part II,
the software tools section, has been completely rewritten to
reflect the new developments in Greenstone Digital Library
Software, an internationally popular open source software tool with
a comprehensive graphical facility for creating and maintaining
digital libraries. As with the First Edition, a web site,
implemented as a digital library, will accompany the book and
provide access to color versions of all figures, two online
appendices, a full-text sentence-level index, and an automatically
generated glossary of acronyms and their definitions. In addition,
demonstration digital library collections will be included to
demonstrate particular points in the book. to access the online
content please visit, http: //www.greenstone.org/howto
*Outlines the history of libraries-- both traditional and digital-- and their impact on present practices and future directions. *Written for both technical and non-technical audiences and covers the entire spectrum of media, including text, images, audio, video, and related XML standards. *Web-enhanced with software documentation, color illustrations, full-text index, source code, and more."
Joe Celkos SQL for Smarties: Advanced SQL Programming offers tips and techniques in advanced programming. This book is the fourth edition and it consists of 39 chapters, starting with a comparison between databases and file systems. It covers transactions and currency control, schema level objects, locating data and schema numbers, base tables, and auxiliary tables. Furthermore, procedural, semi-procedural, and declarative programming are explored in this book. The book also presents the different normal forms in database normalization, including the first, second, third, fourth, fifth, elementary key, domain-key, and Boyce-Codd normal forms. It also offers practical hints for normalization and denormalization. The book discusses different data types, such as the numeric, temporal and character data types; the different predicates; and the simple and advanced SELECT statements. In addition, the book presents virtual tables, and it discusses data partitions in queries; grouping operations; simple aggregate functions; and descriptive statistics, matrices and graphs in SQL. The book concludes with a discussion about optimizing SQL. It will be of great value to SQL programmers. KEY FEATURES * Expert advice from a noted SQL authority and award-winning columnist who has given ten years service to the ANSI SQL standards committee * Teaches scores of advanced techniques that can be used with any product, in any SQL environment, whether it is an SQL 92 or SQL 2008 environment * Offers tips for working around deficiencies and gives insight into real-world challenges
Perfectly intelligent programmers often struggle when forced to
work with SQL. Why? Joe Celko believes the problem lies with their
procedural programming mindset, which keeps them from taking full
advantage of the power of declarative languages. The result is
overly complex and inefficient code, not to mention lost
In this complete revision and expansion of his first SQL Puzzles
book, Joe Celko challenges you with his trickiest puzzles and then
helps solve them with a variety of solutions and explanations. Joe
demonstrates the thought processes that are involved in attacking a
problem from an SQL perspective to help advanced database
programmers solve the puzzles you frequently face. These techniques
not only help with the puzzle at hand, but help develop the mindset
needed to solve the many difficult SQL puzzles you face every day.
Of course, part of the fun is to see whether or not you can write
better solutions than Joe s.
Principles of Transaction Processing is a comprehensive guide to developing applications, designing systems, and evaluating engineering products. The book provides detailed discussions of the internal workings of transaction processing systems, and it discusses how these systems work and how best to utilize them. It covers the architecture of Web Application Servers and transactional communication paradigms. The book is divided into 11 chapters, which cover the following: Overview of transaction processing application and system structureSoftware abstractions found in transaction processing systemsArchitecture of multitier applications and the functions of transactional middleware and database serversQueued transaction processing and its internals, with IBM's Websphere MQ and Oracle's Stream AQ as examplesBusiness process management and its mechanismsDescription of the two-phase locking function, B-tree locking and multigranularity locking used in SQL database systems and nested transaction lockingSystem recovery and its failuresTwo-phase commit protocolComparison between the tradeoffs of replicating servers versus replication resourcesTransactional middleware products and standardsFuture trends, such as cloud computing platforms, composing scalable systems using distributed computing components, the use of flash storage to replace disks and data streams from sensor devices as a source of transaction requests. The text meets the needs of systems professionals, such as IT application programmers who construct TP applications, application analysts, and product developers. The book will also be invaluable to students and novices in application programming.
Data analysis for database design is a subject of great practical
value to systems analysts and designers. This classic text has been
updated to include chapters on distributed database systems, query
optimisation and object-orientation.The SQL content now includes
features of SQL92 and SQL 99.
XML has become the lingua franca for representing business data,
for exchanging information between business partners and
applications, and for adding structure
Making the move from an IT technician or team member to management is one of the most difficult career steps you ll face. Help from management and targeted training can be hard to come by - and your success depends on your ability to adapt to your new role almost overnight. You might have years of experience in the trenches, but you ll quickly find that managing a team, setting budgets, and creating a winning strategy for the first time can be daunting tasks.
Now in its third edition, "IT Manager s Handbook" provides a
practical reference that you will return to again and again in an
ever-changing corporate environment where the demands on IT
continue to increase. Make your first 100 days really count with
the fundamental principles and core concepts critical to your
success as a new IT Manager. The book also includes discusses how
to develop an overall IT strategy as well as demonstrate the value
of IT to the company.
In this book, you ll learn how to: Manage your enterprise s new level of connectivity with a NEW chapter covering social media, handheld devices, and moreImplement and optimize cloud services to provide a better experience for your mobile and virtual workforce at a lower cost to your bottom lineIntegrate mobile applications into your company s strategyManage the money, including topics such as department budgets and leasing versus buyingWork with your "customers," whomever those might be for your IT shopHire, train, and manage your team and their projects so that you come in on time and budgetSecure your systems to face some of today's most challenging security challenges"
How do you approach answering queries when your data is stored in multiple databases that were designed independently by different people? This is first comprehensive book on data integration and is written by three of the most respected experts in the field.
This book provides an extensive introduction to the theory and concepts underlying today's data integration techniques, with detailed, instruction for their application using concrete examples throughout to explain the concepts. Data integration is the problem of answering queries that span multiple data sources (e.g., databases, web pages). Data integration problems surface in multiple contexts, including enterprise information integration, query processing on the Web, coordination between government agencies and collaboration between scientists. In some cases, data integration is the key bottleneck to making progress in a field.
The authors provide a working knowledge of data integration
concepts and techniques, giving you the tools you need to develop a
complete and concise package of algorithms and applications.
*Offers a range of data integration solutions enabling you to focus on what is most relevant to the problem at hand.
*Enables you to build your own algorithms and implement your own data integration applications
*Companion website with numerous project-based exercises and solutions and slides. Links to commercially available software allowing readers to build their own algorithms and implement their own data integration applications. Facebook page for reader input during and after publication.
Managing Time in Relational Databases: How to Design, Update and Query Temporal Data introduces basic concepts that will enable businesses to develop their own framework for managing temporal data. It discusses the management of uni-temporal and bi-temporal data in relational databases, so that they can be seamlessly accessed together with current data; the encapsulation of temporal data structures and processes; ways to implement temporal data management as an enterprise solution; and the internalization of pipeline datasets. The book is organized into three parts. Part 1 traces the history of temporal data management and presents a taxonomy of bi-temporal data management methods. Part 2 provides an introduction to Asserted Versioning, covering the origins of Asserted Versioning; core concepts of Asserted Versioning; the schema common to all asserted version tables, as well as the various diagrams and notations used in the rest of the book; and how the basic scenario works when the target of that activity is an asserted version table. Part 3 deals with designing, maintaining, and querying asserted version databases. It discusses the design of Asserted Versioning databases; temporal transactions; deferred assertions and other pipeline datasets; Allen relationships; and optimizing Asserted Versioning databases.
Joe Celko has looked deep into the code of SQL programmers and
found a consistent and troubling pattern - a frightening lack of
consistency between their individual encoding schemes and those of
the industries in which they operate. This translates into a series
of incompatible databases, each one an island unto itself that is
unable to share information with others in an age of
internationalization and business interdependence. Such
incompatibility severely hinders information flow and the quality
of company data.
The rapidly increasing volume of information contained in
relational databases places a strain on databases, performance, and
maintainability: DBAs are under greater pressure than ever to
optimize database structure for system performance and
Are you an SQL programmer that, like many, came to SQL after
learning and writing procedural or object-oriented code? Or have
switched jobs to where a different brand of SQL is being used, or
maybe even been told to learn SQL yourself?
Fuzzy Modeling and Genetic Algorithms for Data Mining and
Exploration is a handbook for analysts, engineers, and managers
involved in developing data mining models in business and
government. As you'll discover, fuzzy systems are extraordinarily
valuable tools for representing and manipulating all kinds of data,
and genetic algorithms and evolutionary programming techniques
drawn from biology provide the most effective means for designing
and tuning these systems.
Research in Information Systems helps supervisors and their
students get the most out of the PhD experience. It can be used as
a basis of courses for supervisors and their research students.
Life science data integration and interoperability is one of the
most challenging problems facing bioinformatics today. In the
current age of the life sciences, investigators have to interpret
many types of information from a variety of sources: lab
instruments, public databases, gene expression profiles, raw
sequence traces, single nucleotide polymorphisms, chemical
screening data, proteomic data, putative metabolic pathway models,
and many others. Unfortunately, scientists are not currently able
to easily identify and access this information because of the
variety of semantics, interfaces, and data formats used by the
underlying data sources.
You may like...
Designing Data-Intensive Web…
Stefano Ceri, Piero Fraternali, … Paperback R2,137 Discovery Miles 21 370
Advanced SQL:1999 - Understanding…
Jim Melton Paperback R2,007 Discovery Miles 20 070
Database Tuning - Principles…
Dennis Shasha, Philippe Bonnet Paperback R1,986 Discovery Miles 19 860
JDBC - Practical Guide for Java…
Gregory Speegle Paperback R992 Discovery Miles 9 920
The Object Data Standard - ODMG 3.0
R. Cattell, Douglas K. Barry, … Paperback R1,795 Discovery Miles 17 950
Rdb - A Comprehensive Guide
Lilian Hobbs, Ian Smith Paperback R2,101 Discovery Miles 21 010
Joe Celko's Data and Databases…
Joe Celko Paperback R1,791 Discovery Miles 17 910
A Complete Guide to DB2 Universal…
Don Chamberlin Paperback R2,409 Discovery Miles 24 090
Principles of Database Query Processing…
Clement T. Yu, Weiyi Meng Hardcover R2,682 Discovery Miles 26 820
Predictive Data Mining - A Practical…
Sholom M. Weiss, Nitin Indurkhya Paperback R1,684 Discovery Miles 16 840