noaa documentation improvement ted habermann how do we measure and visualize improvements in noaa...

17
NOAA Documentation Improvement Ted Habermann do we measure and visualize improvements in NOAA Documentatio Record Count Completeness (Rubric Scores) Less Metadata More Metadata More Complete Less Complete Good Not So Good

Upload: edgar-white

Post on 17-Dec-2015

222 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

NOAA Documentation ImprovementTed HabermannHow do we measure and visualize improvements in NOAA Documentation?

Record Count

Com

plet

enes

s (R

ubric

Sco

res)

Less Metadata More Metadata

MoreComplete

LessComplete

Good

Not SoGood

Page 2: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

NGDC Solar Metadata History

Late in 2011 Bill Denig, the Chief of the Solar and Terrestrial Physics Division at NGDC, decided that, in order to understand this “metadata thing “ he had to actually work on some metadata. He started working with one record on his desktop using Oxygen (the dreaded XML editor) and the NGDC rubric. He got a very high score and was able to extend his experience to several other records.

He achieved high scores with a small number of records.

Bill’s experience and confidence increased and two months later he extended his collection to include records translated from FGDC to ISO.

This increased the number of records and decreased the scores (FGDC translations generally yield scores between 16 and 20).

Since then Bill has improved existing records and added new, high-quality records to

steadily increase his average score.

Page 3: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Geophysics Components

Components are re-useable pieces of documentation that allow “normalization” of information in metadata collections.

One Month

Page 4: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Acce

ssib

ility

% Records with Data Access Service Link

Historic Trend T

%

NOAA Documentation Dashboard

Metadata Completeness Scores

NOAA 46%

Line Ofc 1 25%

Line Ofc 2 80%

Line Ofc 3 50%

Mean σ Min Max

NOAA

Line Ofc 1

Line Ofc 2

Data Access Service Types Offered

#

WMS NoneWCS DAP Esri WFS

Doc

umen

tatio

n Metadata Dialects Used

% o

f Rec

ords

FreeText

ISO FGDC OBIS DC None

Managers

Scientists andData Managers

Page 5: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Documentation Metric Services

Collector

Web Accessible Folders

Scores Database

Calculate Metrics

ValidationBad LinksBad xLinksRubric Scores

Bad XMLUnique ContactsComponent Report

Page 6: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Collection Characteristics

FGDC

Legacy Recent

Completeness

Most Metadata Are Old

Sporadic Work

Periods

Metadata Revision Date

Discovery

Understanding

Goal: Improve metadata when you revise it:

Page 7: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

2. Assess

3. Create and Improve

4. Publish/Preserve

Standards ExpertsData Collectors/Providers Data UsersData Stewards

DiscoveryUseUnderstanding

CollectionsDatasetsServices

Data Stewardship Teams

Line Office / Program Process

1. Identify Expertise

Initial Evaluation

On-Going Evaluation

SpiralsRubrics

On-Going Input

Wiki

Catalogs WAFs

Consultation / Guidance Use Cases / Needs

Page 8: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Documentation Capabilities

WAFUnified access to metadata with multiple viewsAutomated custom processing and metadata quality checksTranslation from many dialects to consistent international standardHarvest targets for Geospatial One Stop and data.gov

Spiral TrackerConsistent rubric score calculationScore distributions help identify improvement stepsDatabase provides history and access to scores / recordsSupports stewardship teams and managers

WikiProvides community guidance, examples, successesInformation available and shared with national and international partners

Page 9: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Required NOAA-Wide Support NOAA Line/Staff Offices and Programs must build on a strong foundation of support across all of NOAA

1. Develop and implement common metadata management tools

2. Use rubrics to establish a baseline and monitor progress3. Promote and highlight good examples4. Support training specifically targeted at improving NOAA’s

data documentation5. Initiate teams to work on “special documentation problems”

that cross Line and Staff Offices6. Encourage and support participation in the ISO and Open

Geospatial Consortium (OGC)

Page 10: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Questions?

Page 11: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

NOAA Documentation Improvement Components

DirectiveVision and goals, identify standards, describe common nomenclature, identify responsibilities

Efficiently provide consistent guidance implementations, evaluations, and metadata management capabilities for all Line Offices and ProgramsTools

Plans

Community guidance, training, examples, best practices, broad input , Line Office and Program plans

Heterogeneous

Homogeneous

Page 12: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Metric Calculation System

FGDCRecord Scores

Existing XSLT (NCDDC) Rubric XSLT (TBD)ISO

Collector

Web Accessible Folders

Record Scores

Rubric XSLT (TBD)THREDDS

Desktop EditorsMETAVIST , CatMDEdit,

ArcCatalog, XMLSpy, Oxygen

Web ToolsMERMaid,

inPort, NMMR, Geonetwork

Page 13: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Current Test Cases

Collector

CoRIS1642

Records

NOS737

Records

NESDIS/OAR1521

Records

UAF552

Records

FGDC

GOSIC~400

Records

ISO NcML DIF

NMFS~400

Records

inPort

Page 14: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Spiral 1:Initial ContentSpiral 1:Initial Content

Metadata ContentIndependent of standardMetadata ContentIndependent of standard

StandardGuidance /

Implementation

StandardGuidance /

Implementation

Check BackWith UsersCheck BackWith Users

Spiral 2-N: Scientific QuestionsNew RequirementsNew Use Cases

Check BackWith Data Collectors/

Providers

Check BackWith Data Collectors/

Providers

Spiral Development / Training

Page 15: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

IdentificationIdTitleAbstractResource DateTopic CategoryTheme KeywordMetadata ContactScience Contact

IdentificationIdTitleAbstractResource DateTopic CategoryTheme KeywordMetadata ContactScience Contact

ExtentGeospatial Bounding BoxTemporal Start/EndVertical Min/MaxPlace Keywords

ExtentGeospatial Bounding BoxTemporal Start/EndVertical Min/MaxPlace Keywords

DistributionDistributor ContactOnline ResourceDistribution FormatData Center KeywordsBrowse Graphic

DistributionDistributor ContactOnline ResourceDistribution FormatData Center KeywordsBrowse Graphic

Text SearchesPurposeExtent DescriptionLineage StatementProject Keywords

Text SearchesPurposeExtent DescriptionLineage StatementProject Keywords

Quality/LineageSourcesProcess StepsQuality Reports / Coverages

Quality/LineageSourcesProcess StepsQuality Reports / Coverages

Acquisition InformationInstrumentPlatformInstrument KeywordsPlatform Keywords

Acquisition InformationInstrumentPlatformInstrument KeywordsPlatform Keywords

Content InformationAttribute TypeAttribute NamesAttribute DefinitionsAttribute Units

Content InformationAttribute TypeAttribute NamesAttribute DefinitionsAttribute Units

DiscoveryDiscovery UnderstandingUnderstanding

ConnectionOnlineResource:Linkage (URL)NameDescriptionFunction

ConnectionOnlineResource:Linkage (URL)NameDescriptionFunction

Spiral Development / Training: Potential Spirals

Page 17: NOAA Documentation Improvement Ted Habermann How do we measure and visualize improvements in NOAA Documentation? Record Count Completeness (Rubric Scores)

Spiral Tracker

WAF

ScoreHistory Score

Distribution

SpecificRecords

http://www.ngdc.noaa.gov/idb/struts/results?t=103068&s=20&d=25