DataCite will hold its second Summer Meeting on August 24th and 25th at the historic Shattuck Plaza Hotel in Berkeley, California. The Summer Meeting will be a 1.5 day event and you can register at: http://datacite2011.eventbrite.com/ . The Summer Meeting brings together people from research organisations, data centers, government, and information service providers to hear about the latest developments in data science, data citation, discovery, and reuse. It also provides opportunities to exchange experience and influence the next generation of data citation services. This year’s program will include sessions on data citation, data publishing, and discussions on the new challenges that come with increased accessRead More →

DDI Workshop: Managing Metadata for Longitudinal Data – Best Practices September, 19-23, 2011 Leibniz Center for Informatics, Schloss Dagstuhl, Wadern, Germany Goals This symposium-style workshop will bring together representatives from major longitudinal data collection efforts to share expertise and to explore the use of the DDI metadata standard as a means of managing and structuring longitudinal study documentation. Participants will work collaboratively to create best practices for documenting longitudinal data in its various forms, including panel data and repeated cross-sections. Description of the workshop Longitudinal survey data carry special challenges related to documenting and managing data over time, over geography, and across multiple languages. ThisRead More →

At the end of April, arXiv posted an update on their sustainability initiative. This and all arXiv sustainability work should be mandatory reading for all who are working on large, collaborative digital initiatives. Recent updates include the 2011 projected budget and the full support documentation are also available.Read More →

The Data Documentation Initiative 3 (DDI 3) standard is a simply fabulous and full standard for metadata (data about data) as well as for the data contents, making it a full payload standard. DDI 3 is such an exciting standard because it allows for the possibility of true and full computational support for data harmonization and for really working with longitudinal data. It’s the type of data standard I’d been waiting for because it gets it. Data standards need to be able to support documenting, containing, expressing, and computing (analysis, harmonization, limitations on disclosure, everything we now do with less than ideal systems and methods).Read More →

Awesome news from CDL, so reposting below. The original is here. Prototype interface released for searching archival authority records CDL’s Digital Special Collections program is pleased to announce the public release of a draft prototype historical access system for the Social Networks and Archival Context Project (SNAC). SNAC is a two-year research project, funded by the National Endowment for the Humanities, that is creating a set of authority records by extracting information from archival finding aids and enhancing it with other sources.  The project uses the new standard Encoded Archival Context—Corporate bodies, Persons, and Families (EAC-CPF).  Data for the research is being provided by theRead More →

A MADS/RDF ontology developed at the Library of Congress is available for a public review period until Jan. 14, 2011.  The MADS/RDF (Metadata Authority Description Schema in RDF) vocabulary is a data model for authority and vocabulary data used within the library and information science (LIS) community, which is inclusive of museums, archives, and other cultural institutions. It is presented as an OWL ontology. Documentation and the ontology are available at: http://www.loc.gov/standards/mads/rdf/ Based on the MADS/XML schema, MADS/RDF provides a means to record data from the Machine Readable Cataloging (MARC) Authorities format in RDF for use in semantic applications and Linked Data projects. MADS/RDF isRead More →

SobekCM – the system powering the UF Digital Collections, the Digital Library of the Caribbean, and many other rich collections – will soon have advanced support for finding guides in EAD. This has been in process as a complete solution for the full workflow and it’s nearing completion. Check out the EADs we’re testing with here. The benefits from fully supporting EAD within the same digital library system supporting digital objects is enormous: Finding guides can be displayed, searched, and used within the same system as the digital objects they reference (increases usability from consistent navigation, ease of searching a single system, additional benefits fromRead More →

BWF MetaEdit is a free, open source tool that supports embedding, validating, and exporting of metadata in Broadcast WAVE Format (BWF) files. BWF MetaEdit is available for download at SourceForge: http://sourceforge.net/projects/bwfmetaedit/. BWF MetaEdit was developed by the Federal Agencies Digitization Guidelines Initiative to support its guideline for embedded metadata in the bext and INFO chunks (http://www.digitizationguidelines.gov/audio-visual/documents/wave_metadata.html).  The application was developed by AudioVisual Preservation Solutions (http://www.avpreserve.com/). Users of BWF MetaEdit can: Import, edit, embed, and export specified metadata elements in WAVE audio files Export technical metadata from Format Chunks and minimal metadata from bext and INFO chunks as comma-separated values and/or XML, across a set ofRead More →

The most recent Code4Lib Journal issue has an excellent article that should be mandatory reading for anyone working or with a library. The article is “Interpreting MARC: Where’s the Bibliographic Data?” by Jason Thomale. In it, he explains in extremely clear terms exactly what MARC is not. He begins by explaining that MARC pre-dated relational databases. That means everything we think about for computers, digital processing, data structures, and logic doesn’t apply for MARC. The title of this blog post is from one of the article’s notes: There is also the statement about working with MARC data purportedly made by Google engineer Leonid Taycher thatRead More →

From the SNAC website: Leveraging the new standard Encoded Archival Context-Corporate Bodies, Persons, and Families (EAC-CPF), the SNAC Project will use digital technology to “unlock” descriptions of people from finding aids and link them together in exciting new ways. We will: Create efficient open-source tools that allow archivists to separate the process of describing people from that of records. Create a prototype integrated historical resource and access system that will link descriptions of people to one another and to descriptions of resources in archives, libraries and museums; online biographical and historical databases; and other diverse resources.Read More →