1 / 56

Data Citation: Key to Discovery, Reuse, and Tracking Impact

Data Citation: Key to Discovery, Reuse, and Tracking Impact. Curating and Managing Research Data for Reuse ICPSR Summer Program August 2, 2013 Elizabeth Moss, MSLIS eammoss@umich.edu. Today’s talk. A tour of the ICPSR Bibliography of Data-related Literature

dior
Télécharger la présentation

Data Citation: Key to Discovery, Reuse, and Tracking Impact

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Data Citation: Key to Discovery,Reuse, and Tracking Impact Curating and Managing Research Data for Reuse ICPSR Summer Program August 2, 2013 Elizabeth Moss, MSLIS eammoss@umich.edu

  2. Today’s talk • A tour of the ICPSRBibliography of Data-related Literature • The challenges of tracking data reuse (you have to be able to discern data use before you can track data reuse) • Efforts to improve citing standards and practices, leading to sharing and impact

  3. Who uses these shared data?How are they used?With what impact?

  4. Link research data to scholarly literature about it • Increase likelihood of discovery and reuse • Aid students, instructors, researchers, and funders The ICPSRBibliography of Data-related Literature

  5. It’s really a searchable database . . .. . . containing over 65,000 citations of known published and unpublished works resulting from analyses of data archived at ICPSR . . . that resides in Oracle, with an internal UI for database management. . . that can generate study bibliographieslinking each study with the literature about it, and out to the full text

  6. It’s useful to all stakeholders Instructors direct students to begin data-related research projects by reading some of the major works based on the data Advanced researchers also use it to conduct a focused literature review before deciding to use a dataset Reporters and policymakers looking for processed statistics look for reports explaining studies Principal investigators and funding agencies want to track how data are used after they are deposited

  7. But challenging to provide

  8. Provide PIs and data users with citations (since 1990) and DOIs (since 2008) for all study-level data

  9. Explicit citation,in the references, with the DOI “The use of DOI names for the citing of data sets would make their provenance trackable and citable and therefore allow interoperability with existing reference services like Thomson Reuters “Web of Science . . .” From: http://www.codata.org/taskgroups/TGdatacitation/index.html doi:10.3886/ICPSR21240

  10. The state of data citation in the social science literature

  11. Data “Sighting” (implicit) vs. Data Citing (explicit)

  12. Typical “sightings” • Sample described, not named, no author information, no access information, only a publication cited • Data named in text, with some attribution, but no access information • Cited in reference section, but with no permanent, unique identifier, so difficult for indexing scripts to find to automate tracking

  13. Challenges in database search infrastructure • Journal databases fielded for journal article discovery are not ideal for finding data “sightation” • No field searching on methods sections • Full-text search brings back too many bad hits • Limiting to abstract misses too many good hits

  14. Challenges in tracking many studies • Tension between highly curating a manageable collection and minimally maintaining a broad collection • Too many publications for efficient collection by humans, so we must make it easy for scripts to do it reliably

  15. Challenges of completeness • Data use that is too difficult/costly to find cannot be counted • A selective sample, difficult to draw accurate conclusions in broad analyses of reuse

  16. Challenges in lack of data management planning • Publishing sequence prevents citation creation before publication • Potential for change by educating the PI/mentor; graduate directors; liaisonlibrarians • Consciousness raising starting to occur due to funders’ requirements

  17. Poorly described and cited data + Excessive human search effort = Too costly, too questionable for confident measure of impact

  18. Citing data with a DOI + Minimal human search effort = High hit accuracy for the cost, and better confidence of impact measures

  19. Building a culture of viable data citation to improve measures of impact

  20. From: CODATA Data Citation Standards and Practices Task Group. 2012. Task Group Data Citation and Attribution Bibliography http://www.codata.org/taskgroups/TGdatacitation/docs/CODATA_DDCTG_BestPracticesBib_FINAL_17June2012.pdf

  21. http://www.datacite.org/

  22. The tool enables users to search the DataCite Metadata Store for their works, and subsequently to add (or claim) those research outputs – including datasets, software, and other types – to their ORCID profile. This should increase the visibility of these research outputs, and will make it easier to use these data citations in applications that connect to the ORCID Registry – ImpactStory is one of several services already doing this. http://odin-project.eu/2013/05/13/new-orcid-integrated-data-citation-tool/ http://odin-project.eu/

  23. Finding data with simple search fields Integration with Web of Knowledge All Databases: Research data is equal to research literature

  24. Articles linked to underlying data. Increased data discovery. Reward for data citation. Potential for automated tracking. What audience does this have? Anecdotally, no large group of adopters yet. Converting journal search infrastructure to meet the needs of data, but synching metadata still a work in progress.

  25. http://iassistdata.org/

  26. “CODATA, the Committee on Data for Science and Technology, is an interdisciplinary Scientific Committee of the International Council for Science (ICSU), was established 40 years ago. CODATA works to improve the quality, reliability, management and accessibility of data of importance to all fields of science and technology.” From: http://www.codata.org/about/who.html http://www.codata.org/taskgroups/TGdatacitation/index.html

  27. “The move to encourage wider access to the results of publicly-funded research will have limited impact without the associated tools, networks and standards that are needed for sharing and mining of data. The Research Data Alliance aims to provide them.” https://rd-alliance.org/

  28. Data-PASS partners work to change publishing practice

  29. Altmetrics are an attempt to augment or replace the inadequate ways we now use to determine relevant and significant sources of knowledge: 1. peer review 2. citation counting 3. journal impact factors • In-text links, • blogs, • tweets, • bookmarks, • likes, • data downloads . . . Altmetrics.org/manifesto

More Related