1 / 31

IBM Information Server

IBM Information Server. Simplifying the Creation of the Data Warehouse. The New Role of the Data Warehouse. The data warehouse is becoming a more active and integrated participant in enterprise architectures A source of the best information in the business Active source of analytics

gregoryboss
Télécharger la présentation

IBM Information Server

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. IBM Information Server Simplifying the Creation of the Data Warehouse

  2. The New Role of the Data Warehouse • The data warehouse is becoming a more active and integrated participant in enterprise architectures • A source of the best information in the business • Active source of analytics • Because of this, the data warehouse has new requirements • Must be more flexible and adaptable to change • Must have trustworthy, auditable information • Must represent the business view • Must be capable of scaling to meet ever-growing information volumes

  3. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  4. The IBM Solution: IBM Information ServerDelivering information you can trust IBM Information Server Unified Deployment Transform Deliver Understand Cleanse Discover, model, and govern information structure and content Standardize, merge, and correct information Combine and restructure information for new uses Synchronize, virtualize and move information for in-line delivery Unified Metadata Management Parallel Processing Rich Connectivity to Applications, Data, and Content

  5. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  6. Subject Matter Experts, Data Stewards Architects Data Admin Implementers Analysts Collaboration for Data Warehouse Design Collaboration Rational Data Architect IBM Business Glossary IBM Information Analyzer IBM QualityStage IBM DataStage Data-driven analysis, reporting, monitoring, data rule and integration specification Database application and transformation development Metadata and data-driven data modeling and management Business definition & ontology mapped to physical data IBM Metadata Server • Facilitate change management & reuse • Increase compliance to standards • Increase trust and confidence in information • Simplify integration

  7. Collaborative Metadata: From Analysis to Build Collaboration • Common metamodel provides seamless flow of metadata • Analysis activities populate information into data flow design • DataStage users can see the table metadata from Information Analyzer • Analysis Results and notes visible • Provides insight into quality of source • Provides guidance on how flow should be defined • Notes allow free-form collaboration across roles – ensuring knowledge is completed transferred from analysis to build

  8. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  9. Easily Embed Data Quality with Unified Design Auditable Data Quality • One design experience • Speeds Development time • Extended User orientation in a simplified design environment • Performance Oriented

  10. Measure Data Quality Over Time Using Baseline Reporting Auditable Data Quality • Compare quality results to a baseline to understand quality changes over time • Embed profiling tasks into sequencer to take before and after snapshots of data quality and rules adherence

  11. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  12. In-tool Metadata Visibility Metadata-Driven Design Impact Analysis: • Find dependencies …What does this item depend on? -Find where used …Where is this item used? Results shown using the Advanced Find window

  13. Job Difference – Integrated report Metadata-Driven Design Difference report displayed in Designer - jobs opened automatically from report hot links • Options available to: • Print report • Save report as HTML

  14. Slowly Changing Dimension Design Acceleration Metadata-Driven Design • New engine capabilities • Surrogate Key management • Updatable in-memory lookups • New & enhanced stages • Surrogate Key Generator • Slowly Changing Dimension • Single stage per Dimension • Quick setup and definition • Easy single point of maintenance

  15. Rapid Connectivity: Common Connectors Metadata-Driven Design Connection objects allow properties to be dropped onto stage Test the connection instantly Diagram lets you select the link to edit as though your on the canvas Parameter button on every field Graphical ODBC specific SQL builder Warning sign tells you which fields are mandatory

  16. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  17. Reuse: Find It Reuse • Find item in Repository tree • In-place find • Find by Name (Full or Partial) • Wild card support • Find next… • Filter on type

  18. Find – Advanced Search Criteria Reuse • Search on following criteria: • Object type • Job, Table Definition, Stage etc. • Creation • Date/Time • By User • Last Modification • Date/Time • By User • Where Used • What other objects use this object? • Dependencies of • What does this object use? • Options • Case • Match on “name & description” or “name or description”

  19. Reuse: Connection Objects Reuse • Allows saving of a re-usable connection path to a specific source or target • Username, password, db name etc. • Can be used for: • Stage connection properties • Loading onto a stage in the stage editor • Drag ‘n’ drop from Repository tree • Meta data import from that source or target • Drag ‘n’ drop table imported from that source or target onto canvas to create a pre-configured stage instance

  20. Reuse: Job Parameter Sets Reuse • New object in repository that contains the names and values of job parameters. • A Job Parameter set can be referenced by one or more jobs enabling easier deployment of jobs across machines and also enabling easy propagation of a changed job parameter value

  21. Reuse: Simply Deploy Data Flows as Shared Services Reuse • Automates the creation of information integration services including federation • Provides fundamental infrastructure services (security, logging, monitoring) • Provisions to leading web services JMS, EJB and SOAP over HTTP • Provides load balancing & fault tolerance for requests across multiple Service providers

  22. Critical Success Factors for Data Warehousing Metadata-Driven Design Acceleration & Automation Auditable Data Quality • Ensure quality embedded in data flows • Understand quality changes over time • Provide proof of quality and lineage Collaboration • Automate connection between design and build tasks • Provide in-tool metadata visibility • Easily connect to any data source • Seamless flow of metadata across roles • Shared understanding between business & IT • Team development Scalability Reuse • Seamless expansion of capacity • Resource estimation • Accurate performance analysis and balancing • Integrated object search • Object reuse optimization • Reuse of data flows through shared services Data Warehouse

  23. Job Performance Analysis Scalability A new visualization tool which: • Provides deeper insight into runtime job behavior. • Offers several categories of visualizations, including: • Record Throughput • CPU Utilization • Job Timing • Job Memory Utilization • Physical Machine Utilization • Hides runtime complexity by emphasizing the stages the customer placed on the designer canvas.

  24. Record Throughput Scalability • Breakdown of records read and records written per second. • Filters to show one line for each link drawn on the canvas initially. • Names used to refer to each dataset are the actual stage names on the canvas. • Advanced users can turn off filters and see every runtime dataset, including the inner operators of composites, and inserted operator datasets. • One tab for each partition, as well as the ability to show an overlay view including every partition for smaller jobs.

  25. CPU Utilization Scalability Total CPU and System Time • Visualizes the time in CPU of each operator. • Shows what operators were dominating the CPU at different points during the run. • Percentage view shows what percentage of the CPU load of the job each stage on the canvas was responsible for. • Inserted operators and Composite sub-operators automatically get bundled up in these results. • Advanced users can see combination, which will change this chart to reflect each process and the stages contained within. Percentage CPU Pie Chart

  26. Physical Machine Utilization Scalability Average Process Distribution Disk Throughput Free Memory Whisker Box Percent CPU Utilization

  27. Resource Estimation Scalability • Provides estimates for required disk space and CPU utilization. Helps with: • Job design –detect bottlenecks and optimize transformation logic to improve performance • Error protection – run with a range of data of particular interest for a better protection from job aborts due to bad data formats or insufficient null-handling • Resource allocation – determine allocation of scratch space and disk space to protect the job from aborts due to lack of space • Two Statistical Models • Static – provides worst case disk space estimates based on schema and job design. • Dynamic – Runs job and statistically samples actual resource usage. Then provides calculated estimates per node

  28. Resource Estimation Tool Layout Scalability

  29. Migration Path • Seamless upgrade for WebSphere DataStage users into the IBM Information Server • All DataStage jobs along with all other objects will migrate into the IBM Information Server • Upgrade for WebSphere QualityStage into the IBM Information Server • Existing QualityStage projects will migrate into the IBM Information Server • Conversion utilities for Standardize, Match, and Survive stages • All other stages will continue to execute

  30. Metadata-Driven Design Acceleration & Automation Collaboration Auditable Data Quality Scalability Reuse Data Warehouse Summary • Data Warehouses are becoming a tier one operational system in many companies • They must be able to adapt to change more quickly, must have authoritative information, and must be scalable • Platforms for building data warehouses must support metadata-driven design, collaboration, reuse, and auditable data quality, and must be able to scale to support growing data volumes • The IBM Information Server provides all of this in a unified platform IBM Information Server

More Related