SlideShare ist ein Scribd-Unternehmen logo
1 von 57
This document is the confidential property of BP plc.
All rights are reserved. Copyright © 2006.
Data Modelling as a Service (DMaaS) at Bp
DAMA International, San Diego, March 2008
Christopher Bradley & Ken Dunn
Contents
1. Information Architecture challenges at Bp
2. Our solution
a) Self service user administration & provisioning for BP
users
b) Automated Model publishing
c) Detailed reporting of ER/Studio and Repository usage for
user
tracking (and chargeback)
d) Judicious automation
e) Community of Interest
3. Next steps
1. Information Architecture challenges at Bp
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 4
BP is an oil, gas, petrochemicals & renewables company
We employ nearly 100,000 people
…operations on 6 continents and in over 100 countries
…market capitalisation of $250 billion
…revenues of $270 billion in 2006
…over 25,000 service stations worldwide
BP Overview
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 5
Areas of BP Business
Exploration &
Production
Gas
Refining
Alternative
Energy
Chemicals
LubricantsFuels Marketing
& Retail
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 6
DCT (IT) Landscape
Indicative Digital and Communication Technology statistics:
• 250 Data Centers, moving to 3 Mega Data Centers
• 80,000 Desktops – mainly Microsoft XP
• 6,000 Servers – Windows & Unix, some Linux
• 7,000 Applications - target to reduce significantly
• 33 instances of SAP (strategic ERP solution)
• 30 petabytes of spinning disk
• 26 major “data warehouses” (18 SAP BW, 3 Kalido)
• 150 applications independently maintaining Customer data
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
7
Characteristics:
• Well-integrated, enterprise-wide global data where appropriate
• A single view of customer and product master data Key Attributes
• Real-time straight-through processing in areas of need
• Overt focus on Data Quality
• Business insights through greater data visibility
• Business ownership with Single Point of Accountability for data
• DCT role in providing leadership, coordination and verification
VISION
Data and Information are effectively and efficiently managed as a shared corporate
asset that is easily accessible.
Information Architecture vision
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
8
Data
Types
Master Data MI/BI Data
Transaction
Data
Structured
Technical
Data
Digital
Document
Structure
Models / Taxonomy Catalog / Meta data
Information Architecture Framework
Integration
and Access
Quality
Lifecycle
Management
Process
Governance Planning People
Goals
Principles Purpose
ER/Studio
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
9
Challenges: Delivery environment
1. Decentralized management
− IM tools optional including ER/Studio (need to be persuasive)
− no single subject area model (linking entities important)
− few standards but many “guidelines” (modeling guidelines)
2. Project focused
− documentation gets lost in the project repository (drive to put models in
ER/Studio corporate repository)
− continuity of resources is difficult (strong community of interest)
− much project work out-sourced (looking to an accreditation program for
partners)
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
10
Challenges: Application environment
3. SAP
− SAP teams may believe that they only need to configure the application, thus
overlooking the importance of modeling
− gaining value from the modeling that is done (cultural change to get SAP team
to actually use models)
3. SOA
− demand for XML model management (ER/Studio used)
− looking for a quick way to turn data into services (using Composite EII
software)
3. Plus 5,000 other applications
− many different overlapping physical models (important to map to logical)
− much integration and ETL work (looking to establish canonical models)
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
11
Challenges: Modeling environment
6. Process modeling (ARIS)
− integration with data models important (have completed the logical
mapping of entities, still looking at best way to do integration)
− ensuring that “process modelers” don’t also develop the data models
(quality has shown to be variable!)
6. Architectural modeling (System Architect)
− need to integrate with data and process models (work in early stages)
− confusing for modelers as to which tool to use
− confusing for project teams as to where to find information
This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006.
12
Challenges: Publishing
8. PDF
− critical for inclusion in project documentation
− still major communication format
8. SharePoint
− official repository for most projects and architectural documentation
− have automated publication of all models so that they are available to all
project team members
− need good way to publish ER/Studio including zooming in and out!!
8. Wiki environment
− starting to be popular especially for gathering definitions
− need an easy way to keep definitions synchronized with models
13
Prior to 2006
2006 position:2006 position:
• Data modelling undertaken to different degrees in different Segments & Functions.
• Very wide variety of tools & techniques used to define DATA models
− ARIS, ERWin, System Architect, KMDM, Enterprise Architect, Power Designer,
Q_Designer, Rational, PowerPoint, Visio, …… others?
• Most commonly used tool in BP for Data Modelling is PowerPoint / Visio
• Projects encounter common cross Business data concepts, but still create their own
models & definitions.
• No repository of Data Models, nor Governance.
2006 position:2006 position:
• Data modelling undertaken to different degrees in different Segments & Functions.
• Very wide variety of tools & techniques used to define DATA models
− ARIS, ERWin, System Architect, KMDM, Enterprise Architect, Power Designer,
Q_Designer, Rational, PowerPoint, Visio, …… others?
• Most commonly used tool in BP for Data Modelling is PowerPoint / Visio
• Projects encounter common cross Business data concepts, but still create their own
models & definitions.
• No repository of Data Models, nor Governance.
Q4
2005
• Cross BP Data Modelling study – representation from all Segments + Functions.
• Developed agreed requirements statement for data modelling @ BP
• Comprehensive evaluation study
• Established x-BP licence agreements, MSLA & PSA.
• Cross BP Data Modelling study – representation from all Segments + Functions.
• Developed agreed requirements statement for data modelling @ BP
• Comprehensive evaluation study
• Established x-BP licence agreements, MSLA & PSA.
Data quality problems
Inconsistent Data definitions
Duplicated Data
Difficulty in reconciling MI
Models & knowledge lost after each project
2. Our solution
DMaaS portal
A Service not simply tools!
235 models
50,529 entities
Standards & Guidelines
“How to” guides
Web based
Step by step guides
BP Courses
Online & classroom
Several Video guides
Active FAQ & discussion board
Productivity, quality &
standards macros
Macros wish list
Active COI. Highly
attended & rated
a) Self service user administration & provisioning of users
• Self service user administration & provisioning
for BP users to:
− register for ER/Studio
− gain repository permissions
− repository password change
− licence server access
− view registered users / managers (&
members) of teams can see who’s
registered
a) Self Service
− View registered users / managers (&
members) of teams can see who’s
registered
Self Service
• Lets managers know who has registered (or who has not) on their team
• Lets users verify they are registered correctly
• Lets users see other members of the data modelling community at BP
Self Service – Example: View users
− Register new user
Self Service
Register
Self Service – Example: Register New User
1. New user request
submitted from
SharePoint
2. Request received
and validated
against BP Active
Directory
3. User created in
database
4. User created in
Repository
5. User given default
permissions
6. Welcome email sent
= Embarcadero
components
b) Automated publishing of models to SharePoint
b) Model Publishing
• Publishing of models from repository to BP Data Modelling Environment
SharePoint
− Completely automatic generation of models in HTML (no need to produce
ER/Studio report settings files)
− Usual approach is to utilise report wizard
− Approach would be unworkable for BP’s large # of models
− Automatically generate report settings files
− Customise generated reports
− Layouts, title etc
− Automatic uploading to SharePoint
− Uploading of 1000’s of files to SharePoint is very problematic
− Restart built into our upload jobs
− Report home page in SharePoint mimics repository structure
− Highlights when repository models and SharePoint reports not synchronised
− Publishing meta data to inform users of status
Model Publishing - Example
1. Query for updated
models
2. Generate settings
file
3. Generate HTML
version of model
4. Upload HTML to
SharePoint
5. Generate and
update repository
page
Model publishing
Model publishing
Model publishing
Zoom inside the browser!
c) User & usage reporting
c) User & Usage Reporting
• Detailed reporting of ER/Studio and Repository usage for user
tracking (and chargeback)
• Custom solution
• Database of users
− User department & contact details
− MAC address
− Repository id
• Licence server usage
− Peak number of concurrent users (are we approaching licence limit?)
− Number of unique users registered and using DME (monitor take-up)
User & Usage Reporting
Concurrent License Usage
0
5
10
15
20
25
30
35
08May
22May
05Jun
19Jun
03Jul
17Jul
31Jul
14Aug
28Aug
11Sep
25Sep
09Oct
23Oct
06Nov
20Nov
04Dec
18Dec
01Jan
Max Usage
Unique Users
• Log files are copied from the server and parsed
• Usage graph shows peak concurrent license usage and number of unique users for a
given day
• Allows license purchasing decisions to be based on actual usage
• Allows Data Modelling Environment take-up to be monitored
d) Judicious automation
Generic Import/Export
Makes changes
to the model, e.g.
add entites and
attributes
Search Repository
Double-click to
get diagram then
view entity
Copy Entity For Re-Use
Select entities
and run the
Copy Entity
macro
Copy Entity For Re-Use
Run the Paste
Entity macro in
a new diagram
Copy Entity For Re-Use
Run the Entity Re-use
Report macro to see the
list of re-used entities and
their differences
Entity Mapping
Define a mapping concept
then check diagram into
Repository – this allows
entities to be mapped to
Entity Mapping
Define a mapping concept
then check diagram into
Repository – this allows
entities to be mapped to
Entity Mapping
Reference the mapping
concept from the Manage
Mapping Concepts macro
– this creates list
attachments to represent
the mapping
Entity Mapping
Generate a Mapping
Report, lists entities (or
submodels) and where
they are linked to
Render Stylesheet
Generate a Mapping
Report, lists entities (or
submodels) and where
they are linked to
Render Stylesheet
Apply the simple
stylesheet –
everything
becomes white
Render Stylesheet
Change
stylesheet, all
entities with an
‘EDM Business
Domain’
attachment
become red
Render Stylesheet
Change stylesheet
again, fill colour is
based on
attachment value;
Customers
become blue,
Commodities
become green
Validate Data Model
− Data modelling standards and guidelines have been developed.
− Large number of users are utilising ER/Studio (>300).
− No formal process or organisational function to check quality of data
models.
− An automated process (macro) provides a first level assessment of
model quality (i.e. conformance to standards & good practices).
− This does NOT provide any assessment of contentcontent quality – this can only
be accomplished by data model
domain expert review of model.
− Automated populates
the “Validation State” within the
model status block.
− Option to run “statistics only”
report on models in specific
project folders.
BP Model Status
Status: Approved
Type: Project
Validation State: Validated 25/12/2007 73%
Reviewed by: Chris Bradley (BRADC6)
Approved by: Ken Dunn (DUNNKB)
e) Community of interest
51
Community of Interest (COI)
• Purpose:
− This CoI is to share business cases, issues, best practices, guidance, project experiences, and propose domain
directives for Data Modelling at BP.
• Why:
− Data Modelling is undertaken at different levels across BP (Enterprise, Conceptual, Logical, Physical, Message).
− ER/Studio is an accepted & supported tool that BP has adopted across the Enterprise
− Several projects are using ER/Studio at BP today and even more in the future
− Avoid project islands, re-inventing the wheel, gather project synergies
• Share “best practices”
• Charter:
https://wss2.bp.com/DCT/EA/teams/EAPublic/GIA/DME/Admin/Community%20of%20Interest/Data%20Modelling%20CO
• Membership:
− The Data Modelling COI is open to all interested BP staff
− Third parties such as consultants and offshore providers may also participate by invitation. Any consultants /
contractors or other 3rd parties participating will have a current NDA with BP.
− Primarily driven by technical demands
• Involvement of Embarcadero:
− Input from Embarcadero
− COI can influence Embarcadero product development though our involvement in PAC
PAC 4th
– 7th
Feb. Key product requests to Christopher.Bradley2@uk.bp.com
• Meeting Frequency and length:
− Monthly – last Tuesday of the month; 90 minutes / online & “real” meeting
• Agenda items:
− Product & DME news, “how to” sessions, user experiences, hot-topic issues.
52
StonglyAgree
Agree
Disagree
StronglyDisagree
79%
77%
70%
55%
60%
4%
0%
10%
20%
30%
40%
50%
60%
70%
80%
User Survey:
What benefits are you gaining from the Data service?
We are not
obtaining any
benefits
We are obtaining
benefit through use of
a common modellingcommon modelling
tooltool
We are obtaining benefit
through utilisation of a
common repositorycommon repository
We are obtaining
benefit through use of
common standards,common standards,
guidelines &guidelines &
processesprocesses
We are obtaining
benefit through re-usere-use
of models &of models &
artefactsartefacts We are obtaining benefit
through provision of
central support & helpcentral support & help
2006 & 2007 - evangelise
53
Governance & management
Best
practices
DM
Tools
Notation
DM
Repository
Common (core)
set of data definitions
e.g. Master DataImplementation
guidelines 200+ users; 8000+ viewers
BP Enterprise model
Conceptual models
Logical models
Physical models
Industry standard models
Template models
235 models
50,529 entities
Top 10 BP reasons for developing data model
1. Capturing Business Requirements
2. Promotes Reuse, Consistency, Quality
3. Bridge Between Business and Technology
Personnel
4. Assessing Fit of Package Solutions
5. Identify and Manage Redundant Data
6. Sets Context for Project within the Enterprise
7. Interaction Analysis: Compliments Process Model
8. Pictures Communicate Better than Words
9. Avoid Late Discovery of Missed Requirements
10. Critical in Managing Integration Between Systems
GET STARTED
Register for ER/Studio license
Training
List of users
Sign up to newsletter
Change repository permissions
Community of Interest
Productivity Macros
Web publication of models
3. Next steps
Challenges
• SAP Architects
− “We don’t need to do Data Modelling”
• Selling / promoting purpose of Data Modelling
− It’s NOT just for bespoke database developments!
• Expanding online community of interest
• Certification of internal AND supplier staff
− An “approved” supplier doesn’t necessarily mean they know Data
Modelling!
• Interactive training
• Web portal to interrogate repository
− Develop & promote Business Data Dictionary
• Drive re-use
− Linking model artefacts to drive re-use (e.g. Entities from Master Data
Models)
56
Next steps: 2008 & onwards
SOA:
Important in an SoA World.
Definition of data & consequently calls to / results from
services is vital.
Straight through processing can exacerbate the issue
what does the data mean?
which definition of X (e.g. “cost of goods”)?
need to utilise the logical model and ERP models
definitions
Data Lineage:
Repository based Data migration design - Consistency
Source to target mapping
Reverse engineer & generate Informatica ETL
Impact analysis
ERP:
Model Data requirements – aid configuration / fit for
purpose evaluation
Data Integration
Legacy Data take on
Master Data integration
BI / DW:
Model Data requirements in Dimensional Model
Reverse engineer BW Info Cubes, BO Universes,
etc
Generate Star / Snowflake / Starflake schemas
Message modelling:
Hierarchic view of data model
Canonicals
Utilise “Sub-models” for each XML
message
Generate XSD
Import WSDL
Customise XSD via ER/Studio macros
Very powerful XML features in new V7.5
Approved status of models by ….
Enterprise, Segment, Function
Model validation service
Promotion of “approved” e.g. master data models
Promotion of Industry standard models (e.g. PODS)
Drive quality model culture
Cross domain Governance
Modelling (Data lineage) befits for SOX compliance
Reward re-use
Demonstrate benefits of reuse
Make re-use the default behaviour
Share BP benefits success stories (e.g. GOIL)
57
Questions?
Contact details
Chris Bradley
Head Of Information Management Practice
Chris.Bradley@ipl.com
+44 1225 475000
Ken Dunn
Head of Information Architecture
Ken,Dunn@bp.com
+1 630 836 7805

Weitere ähnliche Inhalte

Was ist angesagt?

Enabling a Data Mesh Architecture with Data Virtualization
Enabling a Data Mesh Architecture with Data VirtualizationEnabling a Data Mesh Architecture with Data Virtualization
Enabling a Data Mesh Architecture with Data Virtualization
Denodo
 
Wallchart - Data Warehouse Documentation Roadmap
Wallchart - Data Warehouse Documentation RoadmapWallchart - Data Warehouse Documentation Roadmap
Wallchart - Data Warehouse Documentation Roadmap
David Walker
 
Data Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to MeshData Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to Mesh
Jeffrey T. Pollock
 

Was ist angesagt? (20)

Data modelling 101
Data modelling 101Data modelling 101
Data modelling 101
 
Data Profiling, Data Catalogs and Metadata Harmonisation
Data Profiling, Data Catalogs and Metadata HarmonisationData Profiling, Data Catalogs and Metadata Harmonisation
Data Profiling, Data Catalogs and Metadata Harmonisation
 
Modernizing to a Cloud Data Architecture
Modernizing to a Cloud Data ArchitectureModernizing to a Cloud Data Architecture
Modernizing to a Cloud Data Architecture
 
Enabling a Data Mesh Architecture with Data Virtualization
Enabling a Data Mesh Architecture with Data VirtualizationEnabling a Data Mesh Architecture with Data Virtualization
Enabling a Data Mesh Architecture with Data Virtualization
 
The Business Glossary, Data Dictionary, Data Catalog Trifecta
The Business Glossary, Data Dictionary, Data Catalog TrifectaThe Business Glossary, Data Dictionary, Data Catalog Trifecta
The Business Glossary, Data Dictionary, Data Catalog Trifecta
 
How to Build & Sustain a Data Governance Operating Model
How to Build & Sustain a Data Governance Operating Model How to Build & Sustain a Data Governance Operating Model
How to Build & Sustain a Data Governance Operating Model
 
Chapter 4: Data Architecture Management
Chapter 4: Data Architecture ManagementChapter 4: Data Architecture Management
Chapter 4: Data Architecture Management
 
Improving Data Literacy Around Data Architecture
Improving Data Literacy Around Data ArchitectureImproving Data Literacy Around Data Architecture
Improving Data Literacy Around Data Architecture
 
Master Data Management - Aligning Data, Process, and Governance
Master Data Management - Aligning Data, Process, and GovernanceMaster Data Management - Aligning Data, Process, and Governance
Master Data Management - Aligning Data, Process, and Governance
 
‏‏‏‏‏‏‏‏Chapter 11: Meta-data Management
‏‏‏‏‏‏‏‏Chapter 11: Meta-data Management‏‏‏‏‏‏‏‏Chapter 11: Meta-data Management
‏‏‏‏‏‏‏‏Chapter 11: Meta-data Management
 
Data Lakehouse, Data Mesh, and Data Fabric (r1)
Data Lakehouse, Data Mesh, and Data Fabric (r1)Data Lakehouse, Data Mesh, and Data Fabric (r1)
Data Lakehouse, Data Mesh, and Data Fabric (r1)
 
Data Governance and Metadata Management
Data Governance and Metadata ManagementData Governance and Metadata Management
Data Governance and Metadata Management
 
Building Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics PrimerBuilding Lakehouses on Delta Lake with SQL Analytics Primer
Building Lakehouses on Delta Lake with SQL Analytics Primer
 
How to identify the correct Master Data subject areas & tooling for your MDM...
How to identify the correct Master Data subject areas & tooling for your MDM...How to identify the correct Master Data subject areas & tooling for your MDM...
How to identify the correct Master Data subject areas & tooling for your MDM...
 
Example data specifications and info requirements framework OVERVIEW
Example data specifications and info requirements framework OVERVIEWExample data specifications and info requirements framework OVERVIEW
Example data specifications and info requirements framework OVERVIEW
 
Data Governance Takes a Village (So Why is Everyone Hiding?)
Data Governance Takes a Village (So Why is Everyone Hiding?)Data Governance Takes a Village (So Why is Everyone Hiding?)
Data Governance Takes a Village (So Why is Everyone Hiding?)
 
Data Catalog as a Business Enabler
Data Catalog as a Business EnablerData Catalog as a Business Enabler
Data Catalog as a Business Enabler
 
Best Practices in Metadata Management
Best Practices in Metadata ManagementBest Practices in Metadata Management
Best Practices in Metadata Management
 
Wallchart - Data Warehouse Documentation Roadmap
Wallchart - Data Warehouse Documentation RoadmapWallchart - Data Warehouse Documentation Roadmap
Wallchart - Data Warehouse Documentation Roadmap
 
Data Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to MeshData Mesh Part 4 Monolith to Mesh
Data Mesh Part 4 Monolith to Mesh
 

Andere mochten auch

Selecting Data Management Tools - A practical approach
Selecting Data Management Tools - A practical approachSelecting Data Management Tools - A practical approach
Selecting Data Management Tools - A practical approach
Christopher Bradley
 
Data Governance by stealth v0.0.2
Data Governance by stealth v0.0.2Data Governance by stealth v0.0.2
Data Governance by stealth v0.0.2
Christopher Bradley
 

Andere mochten auch (20)

CDMP preparation workshop EDW2016
CDMP preparation workshop EDW2016CDMP preparation workshop EDW2016
CDMP preparation workshop EDW2016
 
Chief Data Officer: Evolution to the Chief Analytics Officer and Data Science
Chief Data Officer: Evolution to the Chief Analytics Officer and Data ScienceChief Data Officer: Evolution to the Chief Analytics Officer and Data Science
Chief Data Officer: Evolution to the Chief Analytics Officer and Data Science
 
You're the New CDO, Now What?
You're the New CDO, Now What?You're the New CDO, Now What?
You're the New CDO, Now What?
 
Information Management Training Options
Information Management Training OptionsInformation Management Training Options
Information Management Training Options
 
Information Management Capabilities, Competencies & Staff Maturity Assessment
Information Management Capabilities, Competencies & Staff Maturity AssessmentInformation Management Capabilities, Competencies & Staff Maturity Assessment
Information Management Capabilities, Competencies & Staff Maturity Assessment
 
Fate of the Chief Data Officer
Fate of the Chief Data OfficerFate of the Chief Data Officer
Fate of the Chief Data Officer
 
Information Management Training & Certification
Information Management Training & CertificationInformation Management Training & Certification
Information Management Training & Certification
 
Information Management training courses in Dubai
Information Management training courses in DubaiInformation Management training courses in Dubai
Information Management training courses in Dubai
 
Big Data Readiness Assessment
Big Data Readiness AssessmentBig Data Readiness Assessment
Big Data Readiness Assessment
 
Information Management Fundamentals DAMA DMBoK training course synopsis
Information Management Fundamentals DAMA DMBoK training course synopsisInformation Management Fundamentals DAMA DMBoK training course synopsis
Information Management Fundamentals DAMA DMBoK training course synopsis
 
Chief Data Officers At Work
Chief Data Officers At WorkChief Data Officers At Work
Chief Data Officers At Work
 
DAMA CDMP exam cram
DAMA CDMP exam cramDAMA CDMP exam cram
DAMA CDMP exam cram
 
Data Modelling and WITSML
Data Modelling and WITSMLData Modelling and WITSML
Data Modelling and WITSML
 
CDMP Overview Professional Information Management Certification
CDMP Overview Professional Information Management CertificationCDMP Overview Professional Information Management Certification
CDMP Overview Professional Information Management Certification
 
Data modeling for the business
Data modeling for the businessData modeling for the business
Data modeling for the business
 
Selecting Data Management Tools - A practical approach
Selecting Data Management Tools - A practical approachSelecting Data Management Tools - A practical approach
Selecting Data Management Tools - A practical approach
 
Data Governance by stealth v0.0.2
Data Governance by stealth v0.0.2Data Governance by stealth v0.0.2
Data Governance by stealth v0.0.2
 
DAMA BCS Chris Bradley Information is at the Heart of ALL architectures 18_06...
DAMA BCS Chris Bradley Information is at the Heart of ALL architectures 18_06...DAMA BCS Chris Bradley Information is at the Heart of ALL architectures 18_06...
DAMA BCS Chris Bradley Information is at the Heart of ALL architectures 18_06...
 
Master Data Management
Master Data ManagementMaster Data Management
Master Data Management
 
Pellustro Overview
Pellustro OverviewPellustro Overview
Pellustro Overview
 

Ähnlich wie BP Data Modelling as a Service (DMaaS)

Feature Store as a Data Foundation for Machine Learning
Feature Store as a Data Foundation for Machine LearningFeature Store as a Data Foundation for Machine Learning
Feature Store as a Data Foundation for Machine Learning
Provectus
 
HarshaKore-HCLTechnologies
HarshaKore-HCLTechnologiesHarshaKore-HCLTechnologies
HarshaKore-HCLTechnologies
harsha kore
 

Ähnlich wie BP Data Modelling as a Service (DMaaS) (20)

Sanjay Lakhanpal 2015
Sanjay Lakhanpal 2015Sanjay Lakhanpal 2015
Sanjay Lakhanpal 2015
 
Nw2008 tips tricks_edw_v10
Nw2008 tips tricks_edw_v10Nw2008 tips tricks_edw_v10
Nw2008 tips tricks_edw_v10
 
Amit_Kumar_CV
Amit_Kumar_CVAmit_Kumar_CV
Amit_Kumar_CV
 
How to transport PeopleSoft Crystal to BIP via automation_M.... (1).pptx
How to transport PeopleSoft Crystal to BIP via automation_M.... (1).pptxHow to transport PeopleSoft Crystal to BIP via automation_M.... (1).pptx
How to transport PeopleSoft Crystal to BIP via automation_M.... (1).pptx
 
rough-work.pptx
rough-work.pptxrough-work.pptx
rough-work.pptx
 
Power BI Governance and Development Best Practices - Presentation at #MSBIFI ...
Power BI Governance and Development Best Practices - Presentation at #MSBIFI ...Power BI Governance and Development Best Practices - Presentation at #MSBIFI ...
Power BI Governance and Development Best Practices - Presentation at #MSBIFI ...
 
Feature Store as a Data Foundation for Machine Learning
Feature Store as a Data Foundation for Machine LearningFeature Store as a Data Foundation for Machine Learning
Feature Store as a Data Foundation for Machine Learning
 
AhmedWasfi2015
AhmedWasfi2015AhmedWasfi2015
AhmedWasfi2015
 
Data Governance challenges in a major Energy Company
Data Governance challenges in a major Energy CompanyData Governance challenges in a major Energy Company
Data Governance challenges in a major Energy Company
 
Self-Service Analytics with Guard Rails
Self-Service Analytics with Guard RailsSelf-Service Analytics with Guard Rails
Self-Service Analytics with Guard Rails
 
Data-Ed: Data Architecture Requirements
Data-Ed: Data Architecture Requirements Data-Ed: Data Architecture Requirements
Data-Ed: Data Architecture Requirements
 
Data-Ed Online: Data Architecture Requirements
Data-Ed Online: Data Architecture RequirementsData-Ed Online: Data Architecture Requirements
Data-Ed Online: Data Architecture Requirements
 
Dw bi
Dw biDw bi
Dw bi
 
Abdul ETL Resume
Abdul ETL ResumeAbdul ETL Resume
Abdul ETL Resume
 
HarshaKore-HCLTechnologies
HarshaKore-HCLTechnologiesHarshaKore-HCLTechnologies
HarshaKore-HCLTechnologies
 
Murali tummala resume in SAP BO/BI
Murali tummala resume in SAP BO/BIMurali tummala resume in SAP BO/BI
Murali tummala resume in SAP BO/BI
 
Sami patel full_resume
Sami patel full_resumeSami patel full_resume
Sami patel full_resume
 
Resume Pallavi Mishra as of 2017 Feb
Resume Pallavi Mishra as of 2017 FebResume Pallavi Mishra as of 2017 Feb
Resume Pallavi Mishra as of 2017 Feb
 
Data-Ed Webinar: Data Architecture Requirements
Data-Ed Webinar: Data Architecture RequirementsData-Ed Webinar: Data Architecture Requirements
Data-Ed Webinar: Data Architecture Requirements
 
Data-Ed: Data Architecture Requirements
Data-Ed: Data Architecture Requirements  Data-Ed: Data Architecture Requirements
Data-Ed: Data Architecture Requirements
 

Mehr von Christopher Bradley

Information Management best_practice_guide
Information Management best_practice_guideInformation Management best_practice_guide
Information Management best_practice_guide
Christopher Bradley
 

Mehr von Christopher Bradley (14)

Data is NOT the new oil - the Data Asset IS different
Data is NOT the new oil - the Data Asset IS differentData is NOT the new oil - the Data Asset IS different
Data is NOT the new oil - the Data Asset IS different
 
Information Management Training Courses & Certification
Information Management Training Courses & CertificationInformation Management Training Courses & Certification
Information Management Training Courses & Certification
 
Is the Data asset really different?
Is the Data asset really different?Is the Data asset really different?
Is the Data asset really different?
 
Information Management best_practice_guide
Information Management best_practice_guideInformation Management best_practice_guide
Information Management best_practice_guide
 
Big data Readiness white paper
Big data  Readiness white paperBig data  Readiness white paper
Big data Readiness white paper
 
Information is at the heart of all architecture disciplines
Information is at the heart of all architecture disciplinesInformation is at the heart of all architecture disciplines
Information is at the heart of all architecture disciplines
 
Advanced Data Modelling course 3 day synopsis
Advanced Data Modelling course 3 day synopsisAdvanced Data Modelling course 3 day synopsis
Advanced Data Modelling course 3 day synopsis
 
Data Modelling Fundamentals course 3 day synopsis
Data Modelling Fundamentals course 3 day synopsisData Modelling Fundamentals course 3 day synopsis
Data Modelling Fundamentals course 3 day synopsis
 
Data Management Capabilities for the Oil & Gas Industry 17-19 March, Dubai
Data Management Capabilities for the Oil & Gas Industry  17-19 March, DubaiData Management Capabilities for the Oil & Gas Industry  17-19 March, Dubai
Data Management Capabilities for the Oil & Gas Industry 17-19 March, Dubai
 
DMBOK 2.0 and other frameworks including TOGAF & COBIT - keynote from DAMA Au...
DMBOK 2.0 and other frameworks including TOGAF & COBIT - keynote from DAMA Au...DMBOK 2.0 and other frameworks including TOGAF & COBIT - keynote from DAMA Au...
DMBOK 2.0 and other frameworks including TOGAF & COBIT - keynote from DAMA Au...
 
Information is at the heart of all architecture disciplines & why Conceptual ...
Information is at the heart of all architecture disciplines & why Conceptual ...Information is at the heart of all architecture disciplines & why Conceptual ...
Information is at the heart of all architecture disciplines & why Conceptual ...
 
BDA 2012 Big data why the big fuss?
BDA 2012 Big data why the big fuss?BDA 2012 Big data why the big fuss?
BDA 2012 Big data why the big fuss?
 
Data Modelling is NOT just for RDBMS's
Data Modelling is NOT just for RDBMS'sData Modelling is NOT just for RDBMS's
Data Modelling is NOT just for RDBMS's
 
Implementing Effective Data Governance
Implementing Effective Data GovernanceImplementing Effective Data Governance
Implementing Effective Data Governance
 

Kürzlich hochgeladen

Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
Joaquim Jorge
 

Kürzlich hochgeladen (20)

Automating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps ScriptAutomating Google Workspace (GWS) & more with Apps Script
Automating Google Workspace (GWS) & more with Apps Script
 
Scaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organizationScaling API-first – The story of a global engineering organization
Scaling API-first – The story of a global engineering organization
 
Artificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and MythsArtificial Intelligence: Facts and Myths
Artificial Intelligence: Facts and Myths
 
Exploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone ProcessorsExploring the Future Potential of AI-Enabled Smartphone Processors
Exploring the Future Potential of AI-Enabled Smartphone Processors
 
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
Mastering MySQL Database Architecture: Deep Dive into MySQL Shell and MySQL R...
 
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, AdobeApidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
Apidays New York 2024 - Scaling API-first by Ian Reasor and Radu Cotescu, Adobe
 
Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...Driving Behavioral Change for Information Management through Data-Driven Gree...
Driving Behavioral Change for Information Management through Data-Driven Gree...
 
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdfUnderstanding Discord NSFW Servers A Guide for Responsible Users.pdf
Understanding Discord NSFW Servers A Guide for Responsible Users.pdf
 
Boost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivityBoost PC performance: How more available memory can improve productivity
Boost PC performance: How more available memory can improve productivity
 
Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)Powerful Google developer tools for immediate impact! (2023-24 C)
Powerful Google developer tools for immediate impact! (2023-24 C)
 
Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024Partners Life - Insurer Innovation Award 2024
Partners Life - Insurer Innovation Award 2024
 
GenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day PresentationGenCyber Cyber Security Day Presentation
GenCyber Cyber Security Day Presentation
 
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law DevelopmentsTrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
TrustArc Webinar - Stay Ahead of US State Data Privacy Law Developments
 
Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024Finology Group – Insurtech Innovation Award 2024
Finology Group – Insurtech Innovation Award 2024
 
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
Connector Corner: Accelerate revenue generation using UiPath API-centric busi...
 
🐬 The future of MySQL is Postgres 🐘
🐬  The future of MySQL is Postgres   🐘🐬  The future of MySQL is Postgres   🐘
🐬 The future of MySQL is Postgres 🐘
 
2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...2024: Domino Containers - The Next Step. News from the Domino Container commu...
2024: Domino Containers - The Next Step. News from the Domino Container commu...
 
Strategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a FresherStrategies for Landing an Oracle DBA Job as a Fresher
Strategies for Landing an Oracle DBA Job as a Fresher
 
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time AutomationFrom Event to Action: Accelerate Your Decision Making with Real-Time Automation
From Event to Action: Accelerate Your Decision Making with Real-Time Automation
 
GenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdfGenAI Risks & Security Meetup 01052024.pdf
GenAI Risks & Security Meetup 01052024.pdf
 

BP Data Modelling as a Service (DMaaS)

  • 1. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. Data Modelling as a Service (DMaaS) at Bp DAMA International, San Diego, March 2008 Christopher Bradley & Ken Dunn
  • 2. Contents 1. Information Architecture challenges at Bp 2. Our solution a) Self service user administration & provisioning for BP users b) Automated Model publishing c) Detailed reporting of ER/Studio and Repository usage for user tracking (and chargeback) d) Judicious automation e) Community of Interest 3. Next steps
  • 3. 1. Information Architecture challenges at Bp
  • 4. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 4 BP is an oil, gas, petrochemicals & renewables company We employ nearly 100,000 people …operations on 6 continents and in over 100 countries …market capitalisation of $250 billion …revenues of $270 billion in 2006 …over 25,000 service stations worldwide BP Overview
  • 5. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 5 Areas of BP Business Exploration & Production Gas Refining Alternative Energy Chemicals LubricantsFuels Marketing & Retail
  • 6. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2007. 6 DCT (IT) Landscape Indicative Digital and Communication Technology statistics: • 250 Data Centers, moving to 3 Mega Data Centers • 80,000 Desktops – mainly Microsoft XP • 6,000 Servers – Windows & Unix, some Linux • 7,000 Applications - target to reduce significantly • 33 instances of SAP (strategic ERP solution) • 30 petabytes of spinning disk • 26 major “data warehouses” (18 SAP BW, 3 Kalido) • 150 applications independently maintaining Customer data
  • 7. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 7 Characteristics: • Well-integrated, enterprise-wide global data where appropriate • A single view of customer and product master data Key Attributes • Real-time straight-through processing in areas of need • Overt focus on Data Quality • Business insights through greater data visibility • Business ownership with Single Point of Accountability for data • DCT role in providing leadership, coordination and verification VISION Data and Information are effectively and efficiently managed as a shared corporate asset that is easily accessible. Information Architecture vision
  • 8. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 8 Data Types Master Data MI/BI Data Transaction Data Structured Technical Data Digital Document Structure Models / Taxonomy Catalog / Meta data Information Architecture Framework Integration and Access Quality Lifecycle Management Process Governance Planning People Goals Principles Purpose ER/Studio
  • 9. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 9 Challenges: Delivery environment 1. Decentralized management − IM tools optional including ER/Studio (need to be persuasive) − no single subject area model (linking entities important) − few standards but many “guidelines” (modeling guidelines) 2. Project focused − documentation gets lost in the project repository (drive to put models in ER/Studio corporate repository) − continuity of resources is difficult (strong community of interest) − much project work out-sourced (looking to an accreditation program for partners)
  • 10. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 10 Challenges: Application environment 3. SAP − SAP teams may believe that they only need to configure the application, thus overlooking the importance of modeling − gaining value from the modeling that is done (cultural change to get SAP team to actually use models) 3. SOA − demand for XML model management (ER/Studio used) − looking for a quick way to turn data into services (using Composite EII software) 3. Plus 5,000 other applications − many different overlapping physical models (important to map to logical) − much integration and ETL work (looking to establish canonical models)
  • 11. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 11 Challenges: Modeling environment 6. Process modeling (ARIS) − integration with data models important (have completed the logical mapping of entities, still looking at best way to do integration) − ensuring that “process modelers” don’t also develop the data models (quality has shown to be variable!) 6. Architectural modeling (System Architect) − need to integrate with data and process models (work in early stages) − confusing for modelers as to which tool to use − confusing for project teams as to where to find information
  • 12. This document is the confidential property of BP plc. All rights are reserved. Copyright © 2006. 12 Challenges: Publishing 8. PDF − critical for inclusion in project documentation − still major communication format 8. SharePoint − official repository for most projects and architectural documentation − have automated publication of all models so that they are available to all project team members − need good way to publish ER/Studio including zooming in and out!! 8. Wiki environment − starting to be popular especially for gathering definitions − need an easy way to keep definitions synchronized with models
  • 13. 13 Prior to 2006 2006 position:2006 position: • Data modelling undertaken to different degrees in different Segments & Functions. • Very wide variety of tools & techniques used to define DATA models − ARIS, ERWin, System Architect, KMDM, Enterprise Architect, Power Designer, Q_Designer, Rational, PowerPoint, Visio, …… others? • Most commonly used tool in BP for Data Modelling is PowerPoint / Visio • Projects encounter common cross Business data concepts, but still create their own models & definitions. • No repository of Data Models, nor Governance. 2006 position:2006 position: • Data modelling undertaken to different degrees in different Segments & Functions. • Very wide variety of tools & techniques used to define DATA models − ARIS, ERWin, System Architect, KMDM, Enterprise Architect, Power Designer, Q_Designer, Rational, PowerPoint, Visio, …… others? • Most commonly used tool in BP for Data Modelling is PowerPoint / Visio • Projects encounter common cross Business data concepts, but still create their own models & definitions. • No repository of Data Models, nor Governance. Q4 2005 • Cross BP Data Modelling study – representation from all Segments + Functions. • Developed agreed requirements statement for data modelling @ BP • Comprehensive evaluation study • Established x-BP licence agreements, MSLA & PSA. • Cross BP Data Modelling study – representation from all Segments + Functions. • Developed agreed requirements statement for data modelling @ BP • Comprehensive evaluation study • Established x-BP licence agreements, MSLA & PSA. Data quality problems Inconsistent Data definitions Duplicated Data Difficulty in reconciling MI Models & knowledge lost after each project
  • 16. A Service not simply tools! 235 models 50,529 entities Standards & Guidelines “How to” guides Web based Step by step guides BP Courses Online & classroom Several Video guides Active FAQ & discussion board Productivity, quality & standards macros Macros wish list Active COI. Highly attended & rated
  • 17. a) Self service user administration & provisioning of users
  • 18. • Self service user administration & provisioning for BP users to: − register for ER/Studio − gain repository permissions − repository password change − licence server access − view registered users / managers (& members) of teams can see who’s registered a) Self Service
  • 19. − View registered users / managers (& members) of teams can see who’s registered Self Service
  • 20. • Lets managers know who has registered (or who has not) on their team • Lets users verify they are registered correctly • Lets users see other members of the data modelling community at BP Self Service – Example: View users
  • 21. − Register new user Self Service
  • 23. Self Service – Example: Register New User 1. New user request submitted from SharePoint 2. Request received and validated against BP Active Directory 3. User created in database 4. User created in Repository 5. User given default permissions 6. Welcome email sent = Embarcadero components
  • 24.
  • 25. b) Automated publishing of models to SharePoint
  • 26. b) Model Publishing • Publishing of models from repository to BP Data Modelling Environment SharePoint − Completely automatic generation of models in HTML (no need to produce ER/Studio report settings files) − Usual approach is to utilise report wizard − Approach would be unworkable for BP’s large # of models − Automatically generate report settings files − Customise generated reports − Layouts, title etc − Automatic uploading to SharePoint − Uploading of 1000’s of files to SharePoint is very problematic − Restart built into our upload jobs − Report home page in SharePoint mimics repository structure − Highlights when repository models and SharePoint reports not synchronised − Publishing meta data to inform users of status
  • 27. Model Publishing - Example 1. Query for updated models 2. Generate settings file 3. Generate HTML version of model 4. Upload HTML to SharePoint 5. Generate and update repository page
  • 31. c) User & usage reporting
  • 32. c) User & Usage Reporting • Detailed reporting of ER/Studio and Repository usage for user tracking (and chargeback) • Custom solution • Database of users − User department & contact details − MAC address − Repository id • Licence server usage − Peak number of concurrent users (are we approaching licence limit?) − Number of unique users registered and using DME (monitor take-up)
  • 33. User & Usage Reporting Concurrent License Usage 0 5 10 15 20 25 30 35 08May 22May 05Jun 19Jun 03Jul 17Jul 31Jul 14Aug 28Aug 11Sep 25Sep 09Oct 23Oct 06Nov 20Nov 04Dec 18Dec 01Jan Max Usage Unique Users • Log files are copied from the server and parsed • Usage graph shows peak concurrent license usage and number of unique users for a given day • Allows license purchasing decisions to be based on actual usage • Allows Data Modelling Environment take-up to be monitored
  • 35. Generic Import/Export Makes changes to the model, e.g. add entites and attributes
  • 36. Search Repository Double-click to get diagram then view entity
  • 37. Copy Entity For Re-Use Select entities and run the Copy Entity macro
  • 38. Copy Entity For Re-Use Run the Paste Entity macro in a new diagram
  • 39. Copy Entity For Re-Use Run the Entity Re-use Report macro to see the list of re-used entities and their differences
  • 40. Entity Mapping Define a mapping concept then check diagram into Repository – this allows entities to be mapped to
  • 41. Entity Mapping Define a mapping concept then check diagram into Repository – this allows entities to be mapped to
  • 42. Entity Mapping Reference the mapping concept from the Manage Mapping Concepts macro – this creates list attachments to represent the mapping
  • 43. Entity Mapping Generate a Mapping Report, lists entities (or submodels) and where they are linked to
  • 44. Render Stylesheet Generate a Mapping Report, lists entities (or submodels) and where they are linked to
  • 45. Render Stylesheet Apply the simple stylesheet – everything becomes white
  • 46. Render Stylesheet Change stylesheet, all entities with an ‘EDM Business Domain’ attachment become red
  • 47. Render Stylesheet Change stylesheet again, fill colour is based on attachment value; Customers become blue, Commodities become green
  • 48.
  • 49. Validate Data Model − Data modelling standards and guidelines have been developed. − Large number of users are utilising ER/Studio (>300). − No formal process or organisational function to check quality of data models. − An automated process (macro) provides a first level assessment of model quality (i.e. conformance to standards & good practices). − This does NOT provide any assessment of contentcontent quality – this can only be accomplished by data model domain expert review of model. − Automated populates the “Validation State” within the model status block. − Option to run “statistics only” report on models in specific project folders. BP Model Status Status: Approved Type: Project Validation State: Validated 25/12/2007 73% Reviewed by: Chris Bradley (BRADC6) Approved by: Ken Dunn (DUNNKB)
  • 50. e) Community of interest
  • 51. 51 Community of Interest (COI) • Purpose: − This CoI is to share business cases, issues, best practices, guidance, project experiences, and propose domain directives for Data Modelling at BP. • Why: − Data Modelling is undertaken at different levels across BP (Enterprise, Conceptual, Logical, Physical, Message). − ER/Studio is an accepted & supported tool that BP has adopted across the Enterprise − Several projects are using ER/Studio at BP today and even more in the future − Avoid project islands, re-inventing the wheel, gather project synergies • Share “best practices” • Charter: https://wss2.bp.com/DCT/EA/teams/EAPublic/GIA/DME/Admin/Community%20of%20Interest/Data%20Modelling%20CO • Membership: − The Data Modelling COI is open to all interested BP staff − Third parties such as consultants and offshore providers may also participate by invitation. Any consultants / contractors or other 3rd parties participating will have a current NDA with BP. − Primarily driven by technical demands • Involvement of Embarcadero: − Input from Embarcadero − COI can influence Embarcadero product development though our involvement in PAC PAC 4th – 7th Feb. Key product requests to Christopher.Bradley2@uk.bp.com • Meeting Frequency and length: − Monthly – last Tuesday of the month; 90 minutes / online & “real” meeting • Agenda items: − Product & DME news, “how to” sessions, user experiences, hot-topic issues.
  • 52. 52 StonglyAgree Agree Disagree StronglyDisagree 79% 77% 70% 55% 60% 4% 0% 10% 20% 30% 40% 50% 60% 70% 80% User Survey: What benefits are you gaining from the Data service? We are not obtaining any benefits We are obtaining benefit through use of a common modellingcommon modelling tooltool We are obtaining benefit through utilisation of a common repositorycommon repository We are obtaining benefit through use of common standards,common standards, guidelines &guidelines & processesprocesses We are obtaining benefit through re-usere-use of models &of models & artefactsartefacts We are obtaining benefit through provision of central support & helpcentral support & help
  • 53. 2006 & 2007 - evangelise 53 Governance & management Best practices DM Tools Notation DM Repository Common (core) set of data definitions e.g. Master DataImplementation guidelines 200+ users; 8000+ viewers BP Enterprise model Conceptual models Logical models Physical models Industry standard models Template models 235 models 50,529 entities Top 10 BP reasons for developing data model 1. Capturing Business Requirements 2. Promotes Reuse, Consistency, Quality 3. Bridge Between Business and Technology Personnel 4. Assessing Fit of Package Solutions 5. Identify and Manage Redundant Data 6. Sets Context for Project within the Enterprise 7. Interaction Analysis: Compliments Process Model 8. Pictures Communicate Better than Words 9. Avoid Late Discovery of Missed Requirements 10. Critical in Managing Integration Between Systems GET STARTED Register for ER/Studio license Training List of users Sign up to newsletter Change repository permissions Community of Interest Productivity Macros Web publication of models
  • 55. Challenges • SAP Architects − “We don’t need to do Data Modelling” • Selling / promoting purpose of Data Modelling − It’s NOT just for bespoke database developments! • Expanding online community of interest • Certification of internal AND supplier staff − An “approved” supplier doesn’t necessarily mean they know Data Modelling! • Interactive training • Web portal to interrogate repository − Develop & promote Business Data Dictionary • Drive re-use − Linking model artefacts to drive re-use (e.g. Entities from Master Data Models)
  • 56. 56 Next steps: 2008 & onwards SOA: Important in an SoA World. Definition of data & consequently calls to / results from services is vital. Straight through processing can exacerbate the issue what does the data mean? which definition of X (e.g. “cost of goods”)? need to utilise the logical model and ERP models definitions Data Lineage: Repository based Data migration design - Consistency Source to target mapping Reverse engineer & generate Informatica ETL Impact analysis ERP: Model Data requirements – aid configuration / fit for purpose evaluation Data Integration Legacy Data take on Master Data integration BI / DW: Model Data requirements in Dimensional Model Reverse engineer BW Info Cubes, BO Universes, etc Generate Star / Snowflake / Starflake schemas Message modelling: Hierarchic view of data model Canonicals Utilise “Sub-models” for each XML message Generate XSD Import WSDL Customise XSD via ER/Studio macros Very powerful XML features in new V7.5 Approved status of models by …. Enterprise, Segment, Function Model validation service Promotion of “approved” e.g. master data models Promotion of Industry standard models (e.g. PODS) Drive quality model culture Cross domain Governance Modelling (Data lineage) befits for SOX compliance Reward re-use Demonstrate benefits of reuse Make re-use the default behaviour Share BP benefits success stories (e.g. GOIL)
  • 57. 57 Questions? Contact details Chris Bradley Head Of Information Management Practice Chris.Bradley@ipl.com +44 1225 475000 Ken Dunn Head of Information Architecture Ken,Dunn@bp.com +1 630 836 7805