1 / 62

Data Quality and Data Cleaning: An Overview – Continued

Data Quality and Data Cleaning: An Overview – Continued. Theodore Johnson johnsont@research.att.com AT&T Labs – Research (Lecture notes for CS541, 02/12/2004). Database Tools. Most DBMS ’ s provide many data consistency tools Transactions Data types

lmax
Télécharger la présentation

Data Quality and Data Cleaning: An Overview – Continued

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Quality and Data Cleaning: An Overview – Continued Theodore Johnson johnsont@research.att.com AT&T Labs – Research (Lecture notes for CS541, 02/12/2004)

  2. Database Tools • Most DBMS’s provide many data consistency tools • Transactions • Data types • Domains (restricted set of field values) • Constraints • Column Constraints • Not Null, Unique, Restriction of values • Table constraints • Primary and foreign key constraints • Powerful query language • Triggers • Timestamps, temporal DBMS

  3. Then why is every DB dirty? • Consistency constraints are often not used • Cost of enforcing the constraint • E.g., foreign key constraints, triggers. • Loss of flexibility • Constraints not understood • E.g., large, complex databases with rapidly changing requirements • DBA does not know / does not care. • Garbage in • Merged, federated, web-scraped DBs. • Undetectable problems • Incorrect values, missing data • Metadata not maintained • Database is too complex to understand

  4. Too complex to understand … • Recall Lecture 2 : ER diagrams • Modeling even toy problems gets complicated • Unintended consequences • Best example: cascading deletes to enforce participation constraints • Consider salesforce table and sales table. Participation constraint of salesforce in sales. Then you fire a salesman … • Real life is complicated. Hard to anticipate special situations • Textbook example of functional dependencies: zip code determines state. Except for a few zip codes in sparsely populated regions that straddle states.

  5. Tools • Extraction, Transformation, Loading • Approximate joins • Duplicate finding • Database exploration

  6. Data Loading • Extraction, Transformation, Loading (ETL) • The data might be derived from a questionable source. • Federated database, Merged databases • Text files, log records • Web scraping • The source database might admit a limited set of queries • The data might need restructuring • Field value transformation • Transform tables (e.g. denormalize, pivot, fold)

  7. Customer Part Sales Bob bolt 32 Bob nail 112 Bob rivet 44 Sue glue 12 Sue nail 8 Pete bolt 421 Pete glue 6 Customer bolt nail rivet glue Bob 32 112 44 0 Sue 0 8 0 12 Pete 421 0 0 6 (example of pivot) unpivot pivot

  8. ETL • Provides tools to • Access data (DB drivers, web page fetch, parse tools) • Validate data (ensure constraints) • Transform data (e.g. addresses, phone numbers) • Load data • Design automation • Schema mapping • Queries to data sets with limited query interfaces (web queries)

  9. (Example of schema mapping [MHH00]) Address ID Addr Mapping 1 Personnel Professor Name Sal ID Name Sal Student Name GPA Yr PayRate Mapping 2 Rank HrRate WorksOn Name Proj Hrs ProjRank

  10. Web Scraping • Lots of data in the web, but its mixed up with a lot of junk. • Problems: • Limited query interfaces • Fill in forms • “Free text” fields • E.g. addresses • Inconsistent output • I.e., html tags which mark interesting fields might be different on different pages. • Rapid change without notice.

  11. Tools • Automated generation of web scrapers • Excel will load html tables • Automatic translation of queries • Given a description of allowable queries on a particular source • Monitor results to detect quality deterioration • Extraction of data from free-form text • E.g. addresses, names, phone numbers • Auto-detect field domain

  12. Approximate Matching • Relate tuples whose fields are “close” • Approximate string matching • Generally, based on edit distance. • Fast SQL expression using a q-gram index • Approximate tree matching • For XML • Much more expensive than string matching • Recent research in fast approximations • Feature vector matching • Similarity search • Many techniques discussed in the data mining literature. • Ad-hoc matching • Look for a clever trick.

  13. Approximate Joins and Duplicate Elimination • Perform joins based on incomplete or corrupted information. • Approximate join : between two different tables • Duplicate elimination : within the same table • More general than approximate matching. • Semantics : Need to use special transforms and scoring functions. • Correlating information : verification from other sources, e.g. usage correlates with billing. • Missing data : Need to use several orthogonal search and scoring criteria. • But approximate matching is a valuable tool …

  14. Algorithm • Partition data set • By hash on computed key • By sort order on computed key • By similarity search / approximate match on computed key • Perform scoring within the partition • Hash : all pairs • Sort order, similarity search : target record to retrieved records • Record pairs with high scores are matches • Use multiple computed keys / hash functions • Duplicate elimination : duplicate records form an equivalence class.

  15. (Approximate Join Example) Sales “Gen” bucket Provisioning Provisioning Sales Genrl. Eclectic General Magic Gensys Genomic Research Genrl. Electric Genomic Research Gensys Inc. Match Genrl. Eclectic Genomic Research Gensys Genrl. Electric Genomic Research Gensys Inc.

  16. Database Exploration • Tools for finding problems in a database • Opposite of ETL • Similar to data quality mining • Simple queries are effective:Select Field, count(*) as Cnt from Table Group by Field Order by Cnt Desc • Hidden NULL values at the head of the list, typos at the end of the list • Just look at a sample of the data in the table.

  17. Database Profiling • Systematically collect summaries of the data in the database • Number of rows in each table • Number of unique, null values of each field • Skewness of distribution of field values • Data type, length of the field • Use free-text field extraction to guess field types (address, name, zip code, etc.) • Functional dependencies, keys • Join paths • Does the database contain what you think it contains? • Usually not.

  18. Finding Keys and Functional Dependencies • Key: set of fields whose value is unique in every row • Functional Dependency: A set of fields which determine the value of another field • E.g., ZipCode determines the value of State • But not really … • Problems: keys not identified, uniqueness not enforced, hidden keys and functional dependencies. • Key finding is expensive: O(fk) Count Distinct queries to find all keys of up to k fields. • Fortunately, we can prune a lot of this search space if we search only for minimal keys and FDs • Approximate keys : almost but not quite unique. • Approximate FD : similar idea

  19. Effective Algorithm • Eliminate “bad” fields • Float data type, mostly NULL, etc. • Collect an in-memory sample • Perhaps storing a hash of the field value • Compute count distinct on the sample • High count : verify by count distinct on database table. • Use Tane style level-wise pruning • Stop after examining 3-way or 4-way keys • False keys with enough attributes.

  20. Finding Join Paths • How do I correlate this information? • In large databases, hundreds of tables, thousands of fields. • Our experience: field names are very unreliable. • Natural join does not exist outside the laboratory. • Use data types and field characterization to narrow the search space.

  21. Min Hash Sampling • Special type of sampling which can estimate the resemblance of two sets • Size of intersection / size of union • Apply to set of values in a field, store the min hash sample in a database • Use an SQL query to find all fields with high resemblance to a given field • Small sample sizes suffice. • Problem: fields which join after a small data transformation • E.g “SS123-45-6789” vs. “123-45-6789” • Solution: collect min hash samples on the qgrams of a field • Alternative: collect sketches of qgram frequency vectors

  22. Domain Expertise • Data quality gurus: “We found these peculiar records in your database after running sophisticated algorithms!” Domain Experts: “Oh, those apples - we put them in the same baskets as oranges because there are too few apples to bother. Not a big deal. We knew that already.”

  23. Why Domain Expertise? • DE is important for understanding the data, the problem and interpreting the results • “The counter resets to 0 if the number of calls exceeds N”. • “The missing values are represented by 0, but the default billed amount is 0 too.” • Insufficient DE is a primary cause of poor DQ – data are unusable • DE should be documented as metadata

  24. Where is the Domain Expertise? • Usually in people’s heads – seldom documented • Fragmented across organizations • Often experts don’t agree. Force consensus. • Lost during personnel and project transitions • If undocumented, deteriorates and becomes fuzzy over time

  25. Metadata • Data about the data • Data types, domains, and constraints help, but are often not enough • Interpretation of values • Scale, units of measurement, meaning of labels • Interpretation of tables • Frequency of refresh, associations, view definitions • Most work done for scientific databases • Metadata can include programs for interpreting the data set.

  26. XML • Data interchange format, based on SGML • Tree structured • Multiple field values, complex structure, etc. • “Self-describing” : schema is part of the record • Field attributes • DTD : minimal schema in an XML record. <tutorial> <title> Data Quality and Data Cleaning: An Overview <\title> <Conference area=“database”> SIGMOD <\Conference> <author> T. Dasu <bio> Statistician <\bio> <\author> <author> T. Johnson <institution> AT&T Labs <\institution> <\author> <\tutorial>

  27. What’s Missing? • Most metadata relates to static properties • Database schema • Field interpretation • Data use and interpretation requires dynamic properties as well • What is the business process? • 80-20 rule

  28. Lineage Tracing • Record the processing used to create data • Coarse grained: record processing of a table • Fine grained: record processing of a record • Record graph of data transformation steps. • Used for analysis, debugging, feedback loops

  29. Case Study

  30. Case Study • Provisioning inventory database • Identify equipment needed to satisfy customer order. • False positives : provisioning delay • False negatives : decline the order, purchase unnecessary equipment • The initiative • Validate the corporate inventory • Build a database of record. • Has top management support.

  31. Task Description • OPED : operations database • Components available in each local warehouse • IOWA : information warehouse • Machine descriptions, owner descriptions • SAPDB : Sales and provisioning database • Used by sales force when dealing with clients. • Data flow OPED  IOWA  SAPDB

  32. Data Audits • Analyze databases and data flow to verify metadata / find problems • Documented metadata was insufficient • OPED.warehouseid is corrupted, workaround process used • 70 machine types in OPED, only 10 defined. • SAPDB contains only 15% of the records in OPED or IOWA • “Satellite” databases at local sites not integrated with main databases • Numerous workaround processes.

  33. Data Improvements • Satellite databases integrated into main databases. • Address mismatches cleaned up. • And so was the process which caused the mismatches • Static and dynamic data constraints defined. • Automated auditing process • Regular checks and cleanups

  34. What did we learn? • Take nothing for granted • Metadata is always wrong, every bad thing happens. • Manual entry and intervention causes problems • Automate processes. • Remove the need for manual intervention. • Make the regular process reflect practice. • Defining data quality metrics is key • Defines and measures the problem. • Creates metadata. • Organization-wide data quality • Data steward for the end-to-end process. • Data publishing to establish feedback loops.

  35. Research Directions

  36. Challenges in Data Quality • Multifaceted nature • Problems are introduced at all stages of the process. • but especially at organization boundaries. • Many types of data and applications. • Highly complex and context-dependent • The processes and entities are complex. • Many problems in many forms. • No silver bullet • Need an array of tools. • And the discipline to use them.

  37. Data Quality Research • Burning issues • Data quality mining • Advanced browsing / exploratory data mining • Reducing complexity • Data quality metrics

  38. “Interesting” Data Quality Research • Recent research that I think is interesting and important for an aspect of data quality. • CAVEAT • This list is meant to be an example, it is not exhaustive. • It contains research that I’ve read recently. • I’m not listing many interesting papers, including yours.

  39. Bellman • T. Dasu, T. Johnson, S. Muthukrishnan, V. Shkapenyuk, Mining database structure; or, how to build a data quality browser, SIGMOD 2002 pg 240-251 • “Data quality” browser. • Perform profiling on the database • Counts, keys, join paths, substring associations • Use to explore large databases. • Extract missing metadata.

  40. DBXplorer • S. Agrawal, S. Chaudhuri, G. Das, DBXplorer: A System for Keyword-Based Search over Relational Databases, ICDE 2002. • Keyword search in a relational database, independent of the schema. • Pre-processing to build inverted list indices (profiling). • Build join queries for multiple keyword search.

  41. Potters Wheel • V. Raman, J.M. Hellerstein, Potter's Wheel: An Interactive Data Cleaning System, VLDB 2001 pg. 381-390 • ETL tool, especially for web scraped data. • Two interesting features: • Scalable spreadsheet : interactive view of the results of applying a data transformation. • Field domain determination • Apply domain patterns to fields, see which ones fit best. • Report exceptions.

  42. OLAP Exploration • S. Sarawagi, G. Sathe, i3: Intelligent, Interactive Investigation of OLAP data cubes, SIGMOD 2000 pg. 589 • Suite of tools (operators) to automate the browsing of a data cube. • Find “interesting” regions

  43. Data Quality Mining Contaminated Data • Pearson, R. (2001) “Data Mining in the Face of Contaminated and Incomplete Records”, tutorial at SDM 2002 • Outliers in process modeling and identificationPearson, R.K.; Control Systems Technology, IEEE Transactions on , Volume: 10 Issue: 1 , Jan 2002 Page(s): 55 -63 • Methods • identifying outliers (Hampel limits), • missing value imputation, • compare results of fixed analysis on similar data subsets • others

  44. Data Quality Mining : Deviants • H.V. Jagadish, N. Koudas, S. Muthukrishnan, Mining Deviants in a Time Series Database, VLDB 1999 102-112. • Deviants : points in a time series which, when removed, yield best accuracy improvement in a histogram. • Use deviants to find glitches in time series data.

  45. Data Quality Mining • F. Korn, S. Muthukrishnan, Y. Zhu, Monitoring Data Quality Problems in Network Databases, VLDB 2003 • Define probably approximately correct constraints for a data feed (network performance data) • Range, smoothness, balance, functional dependence, unique keys • Automation of constraint selection and threshold setting • Raise alarm when constraints fail above tolerable level.

  46. Data Quality Mining: Depth Contours • S. Krishnan, N. Mustafa, S. Venkatasubramanian, Hardware-Assisted Computation of Depth Contours. SODA 2002 558-567. • Parallel computation of depth contours using graphics card hardware. • Cheap parallel processor • Depth contours : • Multidimensional analog of the median • Used for nonparametric statistics

  47. Depth Contours Points

  48. Approximate Matching • L. Gravano, P.G. Ipeirotis, N. Koudas, D. Srivastava, Text Joins in a RDBMS for Web Data Integration, WWW2003 • Approximate string matching using IR techniques • Weight edit distance by inverse frequency of differing tokens (words or q-grams) • If “Corp.” appears often, its presence or absence carries little weight. “IBM Corp.” close to “IBM”, far from “AT&T Corp.” • Define an SQL-queryable index

  49. Exploratory Data Mining • J.D. Becher, P. Berkhin, E. Freeman, Automating Exploratory Data Analysis for Efficient Data Mining, KDD 2000 • Use data mining and analysis tools to determine appropriate data models. • In this paper, attribute selection for classification.

  50. Exploratory Data Mining • R.T. Ng, L.V.S. Lakshmanan, J. Han, A. Pang, Exploratory Mining and Pruning Optimizations of Constrained Association Rules, SIGMOD 1998 pg 13-24 • Interactive exploration of data mining (association rule) results through constraint specification.

More Related