Changes

Jump to: navigation, search

2015 Prepared Talk Proposals

65,597 bytes added, 01:03, 8 November 2014
m
Protected "2015 Prepared Talk Proposals": Proposal deadline hit ([Edit=Allow only administrators] (indefinite) [Move=Allow only administrators] (indefinite))
Proposals can be submitted through Friday, November 7, 2014 at 5pm PST (GMT−8). Voting will start on November 11, 2014 and continue through November 25, 2014. The URL to submit votes will be announced on the Code4Lib website and mailing list and will require an active code4lib.org account to participate. The final list of presentations will be announced in early- to mid-December.
 
'''Proposals for Prepared Talks:'''
'''Talk Proposals'''
== Zines + Gamification = Awesomest Metadata Literacy Outreach Event Ever! ==
* [http://www.JenniferHecker.info Jennifer Hecker], jenniferraehecker@gmail.com, [http://www.lib.utexas.edu/subject/zines University of Texas Libraries] & [http://www.AustinFanzineProject.org Austin Fanzine Project]
* [http://anomalily.net/ Lillian Karabaic], librarian@iprc.org, [http://www.iprc.org/ Independent Publishing Resource Center] (Portland)
In academic libraries, and elsewhere, the popularity of zine (a magazine produced for love, not profit) collections is on the rise. At the same time, metadata literacy is becoming an increasingly important skill, helping people navigate and understand digital environments and interactions. We have found a way to teach metadata literacy to the general public that isn’t super-boring – in fact, we’ve made it downright fun!
 
First, volunteer zine librarian Lillian Karabaic of Portland’s Independent Publishing Resource Center facilitated the creation of a gamified cataloging interface for the IPRC’s annual Raiders of the Lost Archives backlog-busting 24-hour volunteer cataloging event.
 
Then, archivist Jennifer Hecker facilitated the adaptation of the IPRC’s game for use in a similar, but also very different context – promoting UT Libraries newly-acquired zine collections. The main goal of the academic-library-based event was increasing excitement around the collections, but with the side goal of building metadata literacy, and introducing an understanding of library cataloging issues.
 
The Texas modification also conforms to the xZINECOREx metadata schema developed by the national [http://zinelibraries.info/ Zine Librarians Interest Group], and triggered interesting conversations with the Libraries’s cataloging department about evolving metadata standards and how to incorporate the products of crowd-sourcing projects into existing workflows.
 
Both games will be demoed.
 
We have never presented at Code4lib.
 
== Do the Semantic FRBRoo ==
* Rosie Le Faive, rlefaive@upei.ca, University of Prince Edward Island
 
[http://www.islandora.ca Islandora] is great for creating repositories of any data type, but how can you model meaningful relationships between digital objects and use them to tell a story?
 
At UPEI, I’m assembling an ethnography of Prince Edward Island’s traditional fiddle music that includes musical clips, video clips, oral histories, musical notation, images, and ethnographic commentaries. In order to present an exhibition-style site, I’m tying these digital objects together via the people, places, events, tunes and topics that they share or describe.
 
To describe the relationships, I’m extending Islandora to use [http://www.cidoc-crm.org/frbr_inro.html FRBRoo], a vocabulary that combines the FRBR model with CIDOC-CRM, the the object-oriented museum documentation ontology. These modules being developed will allow other researchers to create a structured, navigable digital repository of diverse object types, that uses Islandora as an exhibition platform.
 
== Our $50,000 Problem: Why Library School? ==
* Jennie Rose Halperin, jhalperin@mozilla.com, Mozilla Corporation
 
57 library schools in the United States are churning out approximately 100 graduates per year, many with debt upwards of $50,000. According to ONet, [http://www.inthelibrarywiththeleadpipe.org/2011/is-the-united-states-training-too-many-librarians-or-too-few-part-1/ 84% of library jobs in the US require an MLS.] The library profession is [http://dpeaflcio.org/programs-publications/issue-fact-sheets/library-workers-facts-figures/) 92% white and 82% female and entry-level librarians can expect to make $32,500 per year.]
 
Contrasted with developers, who are almost [http://www.ncwit.org/blog/did-you-know-demographics-technical-women 90% male] and can expect to make [http://www.forbes.com/sites/jennagoudreau/2011/06/01/best-entry-level-jobs/ $70,000 in an entry-level position,] these numbers are dismal.
 
According to a recent survey, the top skill that outgoing library students want to know is “programming” and yet many MLS programs still consider Microsoft Word an essential technology skill.
 
What is going on here? Why do we accept this fate, where mostly female debt-burdened professionals continue to be thrown onto the work force without the education their expensive degrees promised?
 
As a community we need to come together to stop this cycle. We need to provide better support and mentorship to diversify and keep the profession relevant and help librarianship move into the future it deserves.
 
This talk will walk through the challenges of navigating a hostile employment environment as well as present models for better development and future state imagining.
 
== No cataloging software? Need more than Dublin Core? No problem!: Experiences with CollectiveAccess ==
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University
* Rachel Wittmann, rwittma@clemson.edu, Clemson University
 
Clemson University Libraries has implemented the open-source software CollectiveAccess for customized digital collection needs. CollectiveAccess is an open-source project with the goal of providing a flexible way to manage and publish museum and archival collections. There are several applications associated with the projects; most used are: Providence (for cataloging and entering metadata) and Pawtucket (for displaying objects in a collection for the public). It has many profiles readily available for installing with existing library standards, such as Dublin Core, and there is a robust syntax for creating your own profiles to fit custom tailored metadata schemas. Plus, the user interface allows you to modify the metadata profile quickly and easily.
 
In this talk, we will discuss:
* Our experiences with installing Providence and creating an installation profile that satisfies the needs of many of the Clemson Libraries digital archiving processes.
* The stumbling blocks experienced in that process and how they were resolved.
* The available plugins sourcing widely used authorities, such as Library of Congress thesauri and GeoNames.org, and how they have been used by our projects.
* A brief overview of the export and import functions and also current workflow practices within Providence.
* Future plans & the role of CollectiveAccess at Clemson University Libraries
 
== Getting ContentDM and Wordpress to Play Together ==
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University
* John Mignault, jmignault@metro.org, Empire State Digital Network
With the growth in the number of hubs providing metadata to the Digital Public Library of America, many of them are using REPOX, a tool originally created for the Europeana project, to aggregate disparate metadata feeds and transform them into formats suitable for ingest into DPLA. The Empire State Digital Network, the forthcoming DPLA service hub for NY state, is using it to prepare for our first ingest into DPLA in early 2015. We'll take a look at REPOX and its capabilities and how it can be useful for ingesting and transforming metadata, and also discuss some things we've learned in massaging widely varied metadata feeds.
== Beyond Open Source ==
Checksum technology is used all over the place, from git commits to authenticating Linux packages. It is most commonly used in the digital preservation field to monitor materials in storage for changes that will occur over time or used in the transmission of files during duplication. But do you even checksum, bro? I want this talk to move checksums from a position of mysterious macho jargon to something everyone can understand and want to use. I think a lot of people have heard of checksum but don’t know where to begin when it comes to actually using it at their institution. And cryptography is hella intimidating! This talk will cover what checksums are, how they can be integrated into a library or archival workflow, protecting collections requiring additional levels of security, algorithms used to verify file fixity and how they are different, and other aspects of cryptographic technology. Oh, and please note that all points in this talk will be emphasized or lightly performed through Taylor Swift lyrics. Seriously, this talk will consist of at least 50% Taylor Swift. Can you, like, even?
 
== Level Up Your Coding with Code Club (yes, you can talk about it) ==
 
* Coral Sheldon-Hess, coral@sheldon-hess.org
 
Reading code is a necessary part of becoming a better developer. It gives you more experience and more insight into How Things Are (or Aren't) Done; it builds your intuition about how to solve problems with code; and it increases your confidence that you, too, can tackle whatever technological problems you're facing.
 
But you don't have to read code alone! (Which is good. It's really not fun to read code alone.)
 
In late 2014, a group of librarians formed two Code Clubs, inspired by [http://bloggytoons.com/code-club/ this talk by Saron] (of Bloggytoons fame). I'd like to tell you about how we've structured our Code Clubs, what has gone well, what we've learned, and what you need to do to form your own Code Club. I'll share a list of the codebases we've looked at, too, to help you get your own Code Club off the ground!
 
== The Growth of a Programmer ==
* [[User:jgo | Joshua Gomez]], Getty Research Institute, jgomez@getty.edu
 
Just like other creative endeavors, software developers can experience periods of great productivity or find themselves in a rut. After contemplating the alternating periods in my own career I've noticed several factors that have effected my own professional growth and happiness, including: mentorship, structure, community, teamwork, environment, formal education, etc. Not all of the factors need to be present at all times; but some mixture of them is critical for continued growth. In this talk, I will articulate these factors, discuss how they can effect a developer's career, and how they can be sought out when missing. This talk is aimed at both new developers looking to strike their own path as well as the veterans that lead or mentor them.
 
== Developing a Fedora 4.0 Content Model for Disk Images ==
 
* Matthew Farrell, matthew.j.farrell@duke.edu, Duke University Libraries
* Alexandra Chassanoff, achass@email.unc.edu, BitCurator Access Project Manager
 
As the acquisition of born-digital materials grows, institutions are seeking methods to facilitate easy ingest into their repositories and provide access to disk images and files derived or extracted from disk images. In this session, we describe our development of a Fedora 4.0 Content model for disk images, including acceptable image file formats and the rationale behind those choices. We will also discuss efforts to integrate the disk image content model into the BitCurator Access environment. Unlike generalized, format-agnostic content models which might treat the disk image as a generic bitstream, a content model designed for disk images enables expression of relationships among associated content in the collection such as files extracted from images and other born-digital and digitized material associated with the same creator. It also enables capture of file-system attributes such as file paths, timestamps, whether files are allocated/deleted, etc. Further, a disk image content model suggests further steps repositories can take in order to transform and re-use associated metadata generated during the creation and forensic analysis of the disk image.
 
== Data acquisition and publishing tools in R ==
 
* Scott Chamberlain, scott@ropensci.org, rOpenSci/UC Berkeley - first-time presenter
 
R is an open source programming environment that is widely used among researchers in many fields. R is powerful because it's free, increasingly robust, and facilitates reproducible research, an increasingly sought after goal in academia. Although tools for data manipulation/visualization/analysis are well developed in R, data acquisition and publishing tools are not. rOpenSci is a collaborative effort to create the tools necessary to complete the reproducible research workflow. This presentation discusses the need for these tools, including examples, including interacting with the repositories Mendeley, Dryad, DataONE, and Figshare. In addition, we are building tools for searching scholarly metadata and acuiring full text of open access articles in a standarized way across metadata providers (e.g., Crossref, DataCite, DPLA) and publishers (e.g., PLOS, PeerJ, BMC, Pubmed). Last, we are building out tools for data reading and writing in Ecologial Metadata Language (EML).
 
== SPLUNK: Log File Analysis ==
 
* Jim LeFager, jlefager@depaul.edu, DePaul University Library
DePaul University Library recently took over monitoring and maintaining of the library EZproxy servers this past year and using Splunk, a machine data analysis tool, we are able to gather information and statistics on our electronic resource usage in addition to monitoring the servers. Splunk is a tool that can collect, analyze, and visualize log files and other machine data in real time and this has allowed for gathering realtime usage statistics for our electronic resources allowing us to filter by multiple facets including IP Range, Group Membership (student, faculty), so that we can see who is accessing our resources and from where. Splunk allows our library to query our data and create rich custom dashboards as well as create alerts that can be triggered when certain conditions are met, such as error codes, which can send an email alert to a group of users. We will be leveraging Splunk to monitor all library web applications going forward. This talk will review setting up Splunk and best practices in using the available features and customizations available including creating queries, alerts, and custom dashboards.
 
== Your code does not exist in a vacuum ==
* Becky Yoose, yoosebec at grinnell dot edu, Grinnell College (Done a lightning talk, MC duties, but have not presented a prepared talk)
 
“If you have something to say, then say it in code…” - Sebastian Hammer, code4lib 2009
 
In its 10 year run, code4lib has covered the spectrum of libtech development, from search to repositories to interfaces. However, during this time there has been little discussion about this one little fact about development - code does not exist in a vacuum.
 
Like the comment above, code has something to say. A person’s or organization’s culture and beliefs influences code in all steps of the development cycle. What development method you use, tools, programming languages, licenses - everything is interconnected with and influenced by the philosophies, economics, social structures, and cultural beliefs of the developer and their organization/community.
 
This talk will discuss these interconnections and influences when one develops code for libraries, focusing on several development practices (such as “Fail Fast, Fail Often” and Agile) and licensing choices (such as open source) that libtech has either tried to model or incorporate into mainstream libtech practices. It’ll only scratch the surface of the many influences present in libtech development, but it will give folks a starting point to further investigate these connections at their own organizations and as a community as a whole.
 
tl;dr - this will be a messy theoretical talk about technology and libraries. No shiny code slides, no live demos. You might come out of this talk feeling uncomfortable. Your code does not exist in a vacuum. Then again, you don’t exist in a vacuum either.
 
== The Metadata Hopper: Mapping and Merging Metadata Standards for Simple, User-Friendly Access ==
 
* Tracy Seneca, tjseneca@uic.edu, University of Illinois at Chicago
* Esther Verreau: verreau1@uic.edu, University of Illinois at Chicago
 
The Chicago Collections Consortium: 15 institutions and growing! 8 distinct EAD standards! At least 3 permutations of MARC, and we lost count of the varieties of custom CONTENTdm image collections. Not to mention the 14,730 unique subject terms, nearly all of which lead our poor end-users to exactly one organization's content.
 
All large content aggregation projects have faced this challenge, and there are a few emerging tools to help us wrangle disparate metadata into new contexts. The Metadata Hopper is one such tool. The Metadata Hopper enables archivists to map their local metadata standards to standardized deposit records, and tags those materials using a shared vocabulary, integrating them into a user-friendly portal without disrupting local practices. In last year's Code4Lib lightning talk we described the challenges that the Chicago Collections Consortium faces in creating shared, in-depth access to archival and digital collections about Chicago history and culture across CCC member organizations. This year, thanks to the Andrew W. Mellon Foundation, we have a working Django application to demonstrate. In this talk we'll discuss the design that enables multiple layers of flexibility, from the ability to accept a variety of metadata standards to designing for an open source audience.
 
http://chicagocollectionsconsortium.org
 
== Programmers are not projects: lessons learned from managing humans ==
 
* Erin White, erwhite@vcu.edu, Virginia Commonwealth University - first-time presenter
 
Managing projects is one thing, but managing people is another. Whether we’re hired as managers or grow “organically” into management roles, sometimes technical people end up leading technical teams (gasp!). I’ll talk about lessons I’ve learned about hiring, retaining, and working long-term and day-to-day with highly tech-competent humans. I’ll also talk about navigating the politics of libraryland, juggling different types of projects, and working with constrained budgets to make good things and keep talented people engaged.
 
== Practical Strategies for Picking Low-Hanging Fruits to Improve Your Library's Web Usability and UX ==
 
* Bohyun Kim, bkim@hshsl.umaryland.edu, University of Maryland, Baltimore
 
Have you ever tried to fix an obvious (to you at least!) problem in Web usability or UX (user experience) only to face strong resistance from the library staff? Are you a strong advocate for making library resources, systems, services, and space as usable as possible, but do you often find yourself struggling to get the point across and/or obtain the crucial buy-in from colleagues and administrators?
 
There is no shortage of Web usability and UX guidelines. But applying them to a library and implementing desired changes often involve a long and slow process. To tackle this issue, this talk will focus on how to utilize the 'expert review' process (aka 'heuristic evaluation') as a preliminary or even preparatory step before embarking on more time-and-labor-intensive usability testing and user research. Several examples from simple fixes to more nuanced usability and UX issues in libraries will be discussed to your heart's content. The goal of this talk is to provide practical strategies for picking as many low-hanging fruits as possible to make a real (albeit small) difference to your library's Web usability and UX effectively and efficiently.
 
== A Semantic Makeover for CMS Data ==
 
* Bill Levay, wjlevay@gmail.com, Linked Jazz Project
 
How can we take semi-structured but messy metadata from a repository like CONTENTdm and transform it into rich linked data? Working with metadata from Tulane’s Hogan Jazz Archive Photography Collection, the Linked Jazz Project used Open Refine and Python scripts to tease out proper names, match them with name authority URIs, and specify FOAF relationships between musicians who appear together in photographs. Additional RDF triples were created for any dates associated with the photos, and for those images with place information we employed GeoNames URIs. Historical images and data that were siloed can now interact with other datasets, like Linked Jazz’s rich set of names and personal relationships, and can be visualized [[http://linkedjazz.org/tulane/ see prototype visualization]] or otherwise presented on the web in any number of ways. I have not previously presented at a Code4Lib conference.
 
== Taking User Experience (UX) to new heights ==
* Kayne Richens, kayne.richens@deakin.edu.au, Deakin University
 
User Experience, or "UX", is for more than just websites. At Deakin University Library we're exploring ways to improve the user experience inside our campus library spaces, by putting new technologies front and centre in the overall experience for our students. How are we doing this? We’re collaborating with the University's IT department and exploring the following Library-changing opportunities:
 
- Augmented Reality for Way-finding: We’re tackling that infamous thing that all Libraries can't get right – way-finding. We're enhancing library tour information and way-finding experiences by introducing augmented reality solutions.
- Heat mapping the library with wi-fi: We’re using our existing wi-fi infrastructure to present "heat maps" of library space utilisation, allowing our users to easily locate the space that best suits their needs, whether it be busy spaces to collaborate, or quiet spaces to study. And by overlaying computer usage and group study room bookings, users can quickly locate the space they need.
- Video chat library service: We’re piloting video-conferencing facilities in our group study rooms and spaces, connecting users and librarians and other professionals.
This talk will look at how these different technologies will be brought together to provide improved user experiences, as well some of the evidence and reasons that helped us to identify our needs, so you can too.
 
==How to Hack it as a Working Parent: or, Should Your Face be Bathed in the Blue Glow of a Phone at 2 AM?==
 
*Margaret Heller, Loyola University Chicago, mheller1@luc.edu
*Christina Salazar, California State University Channel Islands, christina.salazar@csuci.edu
*May Yan, Ryerson University, may.yan@ryerson.ca
 
Modern technology has made it easier than ever for parents employed in technical environments to keep up with work at all hours and in all locations. This makes it possible to work a flexible schedule, but also may lead to problems with work/life balance and furthering unreasonable expectations about working hours. Add to that shifting gender roles and limited paid parental leave in the United States and you have potential for burnout and a certainty for anxiety. It raises the additioal question of whether the “always connected” mindset puts up a barrier to some populations who otherwise might be better represented in open source and library technology communities.
 
This presentation will address tools that are useful for working parents in technical library positions, and share some lessons learned about using these tools while maintaining a reasonable work/life balance. We will consider a question that Karen Coyle raised back in 1996:
“What if the thousands of hours of graveyard shift amateur hacking wasn't really the best way to get the job done? That would be unthinkable.”
 
For those who are able to take an extended parental leave, we will present strategies for minimizing the impact to your career and your employer. For those (particularly in the United States) who are only able to take a short leave will require different strategies. Despite different levels of preparation, all are useful exercises in succession planning and making a stronger workplace and future ability to work a flexible schedule through reviewing workloads, cross-training personnel, hiring contract replacements, and creative divisions of labor. Such preparation makes work better for everyone, kids or no kids or caretakers of any kind.
 
==Making your digital objects embeddable around the web==
* Jessie Keck, jkeck@stanford.edu, Stanford University Libraries
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries
 
With more and more content from our digital repositories making their way into our discovery environments we quickly realize that we’re repeatedly re-inventing the wheel when it comes to creating “Viewers” for these digital objects. With various different types of viewers necessary (books, images, audio, video, geospatial data, etc) the burden of getting these viewers into various environments (topic guides, blogs, catalogs, etc) becomes exponential.
 
In this talk we’ll discuss how Stanford University Libraries implemented an oEmbed service to create an extensible viewer framework for all of its digital content. Using this service we’ve been able to easily integrate viewers into various discovery applications as well as make it easy for end users who discover our objects to easily embed customized versions into their own websites and blogs.
 
==So you want to make your geospatial data discoverable==
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries
 
Finding data for research or coursework can be one of the most time intensive tasks for a scholar or student. We introduce GeoBlacklight, an open source, multi-institutional software project focused on solving these common challenges at institutions across the world. GeoBlacklight prioritizes user experience, integrates with many GIS tools, and streamlines the use and organization of geospatial data. This talk will provide an introduction to the software, demonstrate current functionality, and provide a road map for future work.
 
== Clueless-Driven Development: How I learned to migrate to Fedora 4 ==
 
* Adam Wead, awead@psu.edu, Penn State University
 
Recently I was tasked with migrating the content from our Fedora3 repository to the new Fedora4 repository architecture.
Despite a wealth of community support, I had no idea how to approach, or even begin to solve this problem. I knew I
wanted to follow best practices and use test-driven development to build my solution, but had no idea where to start.
Despite this initial setback, I was able to start writing tests with only a vague understanding of the problem. As my
tests exposed where my understanding of the problem was flawed, my code evolved, and within a week I had arrived at a
working solution that exhibited all the hallmarks of good testing and software design.
 
This talk recounts the process I went through from starting with practically nothing, to arriving at a working solution.
You can follow the rules of test-driven development, but you can write tests in an expressive way to describe the
problem instead of just describing what the code should do. It was also essential to begin testing from an integration
viewpoint as opposed to a unit one, because at the outset the units were unknown and were later realized through further
development. For the presentation, I will be demonstrating using RSpec and Ruby. All the code examples will be related
to the Hydra software stack; however, I hope to show that the processes at work will be applicable in any context.
 
 
== Designing and Leading a Kick A** Tech Team ==
* Sibyl Schaefer, sschaefer@rockarch.org, Rockefeller Archive Center
 
New managers are often promoted without receiving management training, yet management is not something you just figure out. The experience of being expected to know how to manage, yet not being trained to do so often results in new managers feeling isolated and unsure how to move from making to managing. In this talk I’ll focus on my own managerial experience of designing and leading an archival tech team in a small independent archives. Topics covered will include hiring, delegating, creating a team culture, and leading people whose specialized knowledge exceeds your own. The talk take-aways should be applicable to managers and employees at large and small institutions alike.
 
 
==American (Archives) Horror Story: LTO Failure and Data Loss ==
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH
* Casey Davis, casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH
 
Here’s a story to send shivers down archival spines: when transferring video files off LTO for the American Archive project, WGBH got an initial failure rate of 57%. After repeat tries, the rates improved; still, an unnervingly large percentage of files were never able to be transferred successfully. Even more unnerving, going public with our horror story got a big response from other archives using LTO -- it seems like many institutions are having similarly scary results. What are the real risks with LTO tape? Are there steps that archives should be taking to better circumvent those risks? This presentation will share information about LTO storage failures across archives world and discuss the process of investigating the problem at WGBH by testing different methods of data retrieval from LTO (direct and networked downloads, individual file retrieval and bulk data dump, use of LTO 4 and LTO 6 decks) and using checksum comparisons and file analysis and characterization tools such as ffprobe, mediainfo and exiftool to analyze failed files. We'll also present whatever results we’ve managed to turn up by the time of Code4Lib!
 
== PBCore in Action: Three Words, Not Two! ==
* Casey E. Davis, casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH
* Andrew (Drew) Myers, andrew_myers@wgbh.org, Supervising Developer, WGBH
 
In 2001, public media representatives developed the PBCore XML schema to establish a common language for managing metadata about their analog and digital audio and video. Since then, PBCore has been adopted by a number of organizations and archivists in the moving image archival community. The schema has also undergone a few revisions, but on more than one occasion it was left orphaned and with little to no support.
Times have changed. You may have heard the news that PBCore is back in action as part of the American Archive of Public Broadcasting initiative and via the Association of Moving Image Archivists (AMIA) PBCore Advisory Subcommittee. A group of archivists, public media stakeholders, and engaged users have come together to provide necessary, sustaining support for the standard and to see to its further development.
At this session, we'll discuss the scope and uses of PBCore in digital preservation and access, report on the progress and goals of the PBCore Advisory Subcommittee, and share how the group (by the time of the conference) will have transformed the XML schema into an RDF ontology, bringing PBCore into the second decade of the 21st century. #PBHardcore
 
==Collaborating to Avert the Digital Graveyard==
 
* Harish Nayak, hnayak@library.rochester.edu, University of Rochester Libraries
* Sean Morris, smorris@library.rochester.edu, University of Rochester Libraries
 
In 1995, the Robbins Library at the University of Rochester created a digital collection of Arthurian texts, images, and bibliographies. Together with medieval scholars, we recently completed the redesign and development of an interface for this collection. Using FRBR concepts, we re-conceptualized organization and editing workflow from the ground up in a mobile-first Drupal-based project.
 
In this talk we will describe the project as well as how we utilized the techniques of work practice study and user centered design to maintain engagement with reluctant stakeholders, nontechnical scholars, and VERY meticulous graduate students. Neither of us have previously presented at a Code4Lib conference.
 
==Docker? VMs? EC2? Yes! With Packer.io==
 
* Kevin S. Clarke, ksclarke@gmail.com, Digital Library Programmer, UCLA
 
There are a lot of exciting ways to deploy a software stack nowadays. Many of our library systems are fully virtualized. Docker is a compelling alternative, and there are also cloud options like Amazon's EC2. This talk will introduce Packer.io, a tool for creating identical machine images for multiple platforms (e.g., Docker, VMWare, VirtualBox, EC2, GCE, OpenStack, et al.) all from a single source configuration. It works well with Ansible, Chef, Puppet, Salt, and plain old Bash scripts. And, it's designed to be scriptable so that builds can be automated. This presentation will show how easy it is to use Packer.io to bring up a set of related services like Fedora 4, Grinder (for stress testing), and Graphite (for charting metrics). As an added value, all the buzzwords in this proposal will be defined and explained!
 
== Technology on your Wrist: Cross-platform Smartwatch Development for Libraries ==
* [[User:sanderson|Steven Carl Anderson]], sanderson@bpl.org, [http://www.bpl.org Boston Public Library] (no previously accepted prepared talks but have done lightning talks in the past)
 
I'll be the first to admit: smartwatches are unlikely to completely revolutionize how a library provides online services. But I believe they still represent an opportunity to further enhance existing library services and resources in a unique way.
 
At the Boston Public Library (BPL), we're in the initial phases of designing a modest smartwatch app to provide notifications for circulation availability and checked-out-material due-date alerts by the end of current year. We're starting small, but we plan to evolve the concept over time as we see what (if any) traction such an application gets with potential users. For example, we plan to explore the possibility of adding "nearest branch to my current location" functionality to this app.
 
Despite the "development phase" of this application as of this writing, this talk is not being given by a novice. As a technology enthusiast, I've released [http://www.phdgaming.com/smartwatch_projects/ five smartwatch applications] and have had two of those be finalists in a [http://www.phdgaming.com/samsung_challenge/ Samsung sponsored development challenge]. This experience not only will allow for the BPL to avoid many beginner mistakes in its smartwatch app development but also gives a much more complete understanding of the smartwatch development ecosystem.
 
This talk will explore the following questions:
 
* What kinds of online library services could potentially be transformed or translated into the smartwatch/wearable domain? What kinds of services are better left alone? These questions are currently being explored and I'll talk about our plans and experiences. Included will be any statistical information from our application launch along with statistics from my personal development.
 
* How to support all the different operating systems these devices run without painful modifications to your codebase. (There's Tizen that is used by Samsung's Gear 2 and Gear S, Android Wear that is used by most other non-Apple manufacturers, then there is Apple's upcoming smartwatch itself, etc.)
 
* How to support different screen resolutions on such a small device. From round to rectangular to perfectly square, smartwatches come in all different shapes these days.
 
* What are the app stores like on these platforms? As I support multiple applications through different distribution networks, a guide to navigating how to distribute one's app is included and I'll reveal how these systems work “behind the curtain.”
 
* What are common issues and pitfalls to avoid when doing development? Tips on broken APIs and how to cope or optimizing your code will be included.
 
==Seeing the Forest From the Trees: The Art of Creating Workflows for Digital Projects ==
* Jen LaBarbera, j.labarbera@neu.edu, NDSR Resident, Northeastern University
* Joey Heinen, joseph_heinen@harvard.edu, NDSR Resident, Harvard University
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH
* Tricia Patterson, triciap@mit.edu, NDSR Resident, MIT
 
We have to "turn projects into programs" in order to create a solid and sustainable digital preservation initiative...but what the heck does that even mean? What does that look like?
 
In this talk, members of the inaugural Boston cohort of the National Digital Stewardship Residency will discuss one piece of our digital preservation test kitchen: our stabs at creating digital workflows that will (hopefully) help our institutions turn digital preservation projects into programs. Specifically, we will talk about how difficult it is to create a general and overarching workflow for digital preservation tasks (e.g. ingest into repositories, format migrations, etc.) that incorporates various technical tools while also taking into account the myriad and unending list of possible exceptions or special scenarios. Turning these complicated, specific processes into a simplified and generalized workflow is an art. We haven't necessarily perfected that art yet, but in this talk, we'll share what has worked for us -- and what hasn't. We’ll also touch on the importance of documentation, and achieving that delicate balance of adequately thorough documentation that doesn’t pose the risk of information avalanche. These processes often create more questions than answers, but we'll share the answers that we (and our mentors) have found along the way!
 
== Annotations as Linked Data with Fedora4 and Triannon (a Real Use Case for RDF!) ==
 
* Rob Sanderson, azaroth@stanford.edu, Stanford University Libraries
* Naomi Dushay, ndushay@stanford.edu, Stanford University Libraries
 
Annotations on content resources allow users to contribute knowledge within the digital repository space. W3C Open Annotation provides a comprehensive model for web annotation on all types of content, using Linked Data as a fundamental framework. Annotation clients generate instances of this model, typically using a JSON serialization, but need to store that data somewhere using a standard interaction pattern so that best of breed clients, servers, and data can be mixed and matched.
 
Stanford is using Fedora4 for managing Open Annotations, via a middleware component called Triannon. Triannon receives the JSON data from the annotation client, and uses the Linked Data Platform API implementation in Fedora4 to create, retrieve, update and delete the constituent resources. Triannon could be easily modified to use other LDP implementations, or could be modified to work with linked data other than annotations.
 
== Hydras in the Wild: A survey of current projects ==
* Mark Bussey, mark@curationexperts.com, Data Curation Experts
 
You've seen the tutorials, but [https://github.com/projecthydra/hydra/wiki/Dive-into-Hydra Dive Into Hydra] seems to leave something wanting. What can you really do using the Hydra Framework? This talks looks at a number of current Hydra projects and highlights the design and functional features unique to each. Compare and contrast UX, design and functional capabilites from a range of hydra-based repositories including:
* Avalon for media discovery and distribution (Indiana and Northwestern Univiersities)
* HydraDam for media archive management(WGBH)
* HyHull for general Institutional Repository needs (University of Hull)
* T-DIL for slide library functions (Tufts University)
* Sufia & Scholarsphere as a bundled self-deposit IR solution (Pennsylvania State University)
* Curate & Worthwhile as general purpose repository platforms (Multiple Insititutions)
This will be a whirlwind tour aimed at providing ideas and inspiration for your own repository development project.
 
== Hydra Makeovers! ==
 
* Alicia Cozine, alicia@curationexperts.com, Data Curation Experts
* Patrick Feeley, pgf8@case.edu, Case Western Reserve University
 
Compare two Hydra-based applications with the systems they replaced. Marvel at the Before and After snapshots of functionality, speed, and look & feel.
* '''Digital Case 2.0''' is an institutional repository, complete with administrative tools, derivatives transcoding, flexible XML metadata storage, embargo and lease capability, faceted searching, and content viewers for texts/TEI, images, audio recordings, and videos. Digital Case 2.0 is based on worthwhile, an open-source IR starter gem.
* '''The Tufts Digital Image Library''' is a specialized tool for art and art history resources, offering image collections with user access controls, image ordering, collection nesting, drag-and-drop organization, slideshows, and export capability.
Both new systems are built on hydra, the open-source Ruby-on-Rails repository solution that incorporates Fedora for storage, Solr for indexing, and Blacklight for search optimization. Their beauty is not just skin-deep!
 
 
 
== Helping Google (and scholars, researchers, educators, & the public) find archival audio ==
 
* Anne Wootton, anne@popuparchive.org, Pop Up Archive (www.popuparchive.org)
 
Culturally significant digital audio collections are hard to discover on the web. There are major barriers keeping this valuable media from scholars, researchers, and the general public:
 
Audio is opaque: you can’t picture sound, or skim the words in a recording.
Audio is hard to share: there’s no text to interact with.
Audio is not text: but since text is the medium of the web, there’s no path for audiences to find content-rich audio.
Audio metadata is inconsistent and incomplete.
 
At Pop Up Archive, we're helping solve this problem making the spoken word searchable. We began as a UC-Berkeley School of Information Master's thesis to provide better access to recorded sound for audio producers, journalists, and historians. Today, Pop Up Archive processes thousands of hours of sound from all over the web to create automatic, timestamped transcripts and keywords, working with media companies and institutions like NPR, KQED, HuffPost Live, Princeton, and Stanford. We're building collections of sound from journalists, media organizations, and oral history archives from around the world. Pop Up Archive is supported by the John S. and James L. Knight Foundation, the National Endowment for the Humanities, and 500 Startups.
 
== Digital Content Integrated with ILS Data for User Discovery: Lessons Learned ==
 
* Naomi Dushay, ndushay@stanford.edu, Stanford University Libraries
* Laney McGlohon, laneymcg@stanford.edu, Stanford University Libraries (first-time presenter)
 
So you want to expose your digital content in your discovery interface, integrated with the data from your ILS? How do you make the best information user searchable? How do you present complete, up to date search results with a minimum of duplicate entries?
 
At Stanford, we have these cases and more:
* digital content with no metadata in ILS
* digital content for metadata in ILS
* digital content with its own metadata derived from ILS metadata.
 
We will describe our efforts to accommodate multiple updatable metadata sources for materials in the ILS and our Digital Object Repository while presenting users with reduced duplication in SearchWorks. Included will be some failures, some successes, and an honest assessment of where we are now.
 
== Show All the Things: Kanban for Libraries ==
 
* Mike Hagedon, mhagedon@email.arizona.edu, University of Arizona Libraries (first-time presenter)
 
The web developers at the University of Arizona Libraries had a problem: we were working on a major website rebuild project with no clear way to prioritize it against our other work. We knew we wanted to follow Agile principles and initially chose Scrum to organize and communicate about our work. But we found that certain core pieces of Scrum did not work for our team. Then we discovered Kanban, an Agile meta-process for organizing work (team or individual) that treats the work more as a flow than as a series of fixed time boxes. I’ll be talking about our journey toward finding a process that works for our team and how we’ve applied the principles of Kanban to better get our work done. Specifically, I'll discuss principles like how to visualize all your work, how to limit how much you’re doing (to get more done!), and how to optimize the flow of your work.
 
== DIY Catalog ==
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]
 
With so many discovery layers being made available that didn’t quite fit user needs but still came with substantial costs, Hennepin County Library determined it would develop one that could handle its customizations. So they built their own ILS API, Web services, and an interface that provides all the features (and existing custom features) of the old catalog, with new features planned, all within a responsive layout. Get a look at the system architecture, how the front end communicates through the layers back to the databases, and how to manage the infrastructure.
 
== OverDrive: Full Integration ==
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]
 
How do you increase usage of your ebooks? Seamless integration in the catalog certainly helps. Users can search for digital titles, limit to just those that are available, place a request, borrow, and manage all their digital titles from the library website alongside all the other physical titles the library has. Get a quick demo and learn how this was all made possible.
 
The [//www.hclib.org Hennepin County Library] is ranked one of [//www.thedigitalshift.com/2014/10/ebooks/overdrive-api-usage-indicates-growth-ebook-checkouts-via-opac/ the top users of OverDrive and OverDrive’s API] and is looking to more fully integrate 3M Cloud, OneClickDigital, and other digital resources.
 
== Dynamic Indexing: a Tragic Solr Story ==
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]
 
Loading data from an ILS into Solr isn’t so hard, unless it needs to be dynamic, fast, and hold more data than what can be found in 1.5 million MARC records. Some additional information we’ve incorporated are from Syndetics, ILS circulation, and OverDrive. We’ll share the nitty gritty details and what we learned about dynamic Solr indexing, including how to get good performance, how to deal with indexing failures, how to schedule it all to keep the data up-to-date, and some things you can do with that data such as popularity ratings.
 
== Fedora 4, Survey of Core Capabilities and External Modules: ==
* Andrew Woods, awoods@duraspace.org, DuraSpace
 
Now that the production release of Fedora 4.0 is out, the time is right to take a step back and review the core capabilities offered by Fedora:
* Basic CRUD
* Versioning
*- Transactions
* etc
 
...and take a closer look at the supported external modules:
* Authorization
* Solr integration
* Triplestore integration
* Camel integration
* etc
 
In addition to clarifying specific implementation details, this is an opportunity to surface community requirements which may have not been addressed in the initial release.
 
== Bleeding edge beacons: redesign of a library tour with new technology ==
 
* Neal Henshaw, Virginia Tech
* Somiah Lattimore, Straight Up Creative
* Keith Gilbertson, keith.gilbertson@vt.edu, Virginia Tech
 
This group has not yet presented at Code4Lib
 
A technology liaison, a designer, and a developer are working together to update a mobile audio tour of an academic library.
 
The walking tour was originally conceived of and created by the technology liaison to provide an orientation experience for incoming students. In its current version, patrons visit several stations in the library and activate an audio description of each station by scanning a QR barcode with a mobile device.
 
We are building a prototype to test with focus groups so that we can launch a new version of the tour in early 2015. The tour has been reimagined with a professional and communicative user interface that presents students with learning goals for each location in the library. The new version of the tour is a mobile application with integrated support for micro-location technology provided through Bluetooth low energy beacon devices, known informally as iBeacons, installed throughout the building. As a touring student walks through the library, her location is noted by the app so that an appropriate video segment is automatically played according to the current location.
We will discuss perspectives on the design of the project, including
* Designing with technology to achieve learning goals
* Designing graphical elements, interaction, and user experience
* Designing to maintain compatibility with older technologies
 
We’ll explain our efforts to make the application usable by those with hearing impairments and mobility impairments, reactions from our first users, and challenges in working with the relatively new beacon technology.
 
== Distributed Remediation: Small tools for big problems: ==
* Matt Miller, matthewmiller@nypl.org, New York Public Library, NYPL Labs
 
Remediation of legacy data can be automated only so much. Certain essential cleanup tasks, such as aligning a name with the correct authorized version, is very difficult for a computer yet trivial for a person. While it is these types remediations that will allow an institution to take advantage and participate in the web of Linked Open Data, a wholly manual approach is unrealistic. However, by augmenting automated remediation with a light human touch we can quickly and efficiently reach our goals. This talk will look at tools and methods being developed at NYPL Labs to empower library staff and the public to help clean up our legacy metadata through collaborative remediation.
 
==VuFind + WorldCat: Open Source Discovery Meets Big Library Data==
 
* Karen A. Coombs, coombsk@oclc.org, OCLC
* Demian Katz, demian.katz@villanova.edu, Villanova University
 
Good collaboration is crucial to any integration. Our project, to integrate the open source discovery tool, VuFind, with results from the new WorldCat Discovery API, is no different. We want to exploit the flexibility of VuFind and add the depth and breadth of WorldCat and central index content. Plus, the project has the potential to eliminate the hurdle of exporting and indexing MARC records for 32 VuFind libraries.
 
While we are really excited about the UI enhancements, we are also curious to see how we can share our domain expertise between our two organizations to get the project done quickly. How will different perspectives shape our existing code bases? How will we adjust to working on the same code base simultaneously? And how can we best incorporate what we learn along the way?
 
There is also new territory to explore using the API: integrating bibliographic and article data into a single results set. And the overarching question: will we be able to take advantage of Linked Data in the WorldCat Discovery API to create some “glue” between records in existing VuFind indexes and third-party data providers?
 
This presentation will discuss our adventures in the OCLC WorldShare Platform/VuFind collaboration: the opportunities, challenges and results. You’ll learn what worked, what didn’t and how you can improve your own discovery interface integration project—no matter what provider or APIs you use.
 
== Book Reader Bingo: Which Page-Turner Should I Use? ==
* Eben English, eenglish [at] bpl.org, Boston Public Library
 
Another day, another library reinventing the book-viewer wheel. When will the madness end? This talk will explore the current landscape of book-viewer/page-turner applications for digital library systems with an eye towards helping you make the right decision for your project. We'll look at some the major players on the market (such as Internet Archive BookReader, Wellcome Player, Mirador, and WDL-Viewer to name a few) and compare them based on a number of criteria: feature sets, mobile/tablet friendliness, ease of integration, code health, test coverage, "market share" (number of implementers), and other important factors. We'll look at the results of ACTUAL USABILITY TESTS to see what features users REALLY want in a book-viewer, and how each app measures up.
 
We'll also discuss important recent trends (such as the IIIF Presentation API, ReadersFirst, and NYPL's Library Simplified initiative) that have the potential to shape the book-viewer development landscape in the immediate future. Which page-turner applications are best poised to adopt/integrate/leverage these emerging standards? Which will become obsolete? This talk has the answers you need.
 
== Rich Citations ==
* Adam Becker, abecker@plos.org, Public Library of Science
* Erik Hetzner, ehetzner@plos.org, Public Library of Science
 
Citations should connect you to the research you need. They should
link directly to the relevant papers and data. They should tell you
everything you need to know about how and where two pieces of research
are connected. They should be easy to use and they should never get in
the way of the reading experience. In short, citations should not
clumsily point to a mere chunk of plain undifferentiated text sitting
in a static list at the end of the paper. But with almost no
exceptions, citations in scientific papers do just that.
 
PLOS Labs has been working on a project to capture extra information
about citations. To this end, we have designed a metadata structure
describing the context of a paper's citations, called rich citations.
This structured information includes complete bibliographic
information for the cited items, location and context of the in-text
citation, co-citations, reference license and status (updated,
retracted, etc.), and more.
 
We have processed the complete PLOS corpus to extract this rich
citation metadata, and made this data available in an API located at
http://api.richcitations.org/.
 
We have also developed a JavaScript overlay to enhance a paper's view
using rich citations (http://alpha.richcitations.org). This overlay
uses the underlying rich citation metadata to allow the user to view
information about an in-text citation, quickly navigate between
citations, and sort and filter the paper's reference list.
 
In this talk, we will describe the rich citation metadata that we are
capturing. We will demonstrate how this metadata can enhance a
reader's experience of an article and how it can be used by
researchers to better understand how citations are used in the
scientific community. We will also discuss our ongoing plans to extend
this project to the wider literature beyond PLOS, and how interested
members of the community can help.
 
== Scriptaloging with ArchivesSpace and NodeJS ==
 
* Brian Hoffman, brianjhoffman@gmail.com
 
ArchivesSpace is a new web application for managing archival collections. It has a browser-based interface for entering and editing metadata, and can import data serialized as EAD, MARC, and several other formats. But there may be situations where neither of these are quite what you want. For instance, you may have a large folder of images that each need a digital object record; or you may want to export an EAD for every collection in your repository; or calculate the total extent of your collection; or execute a global search and replace; or batch-update barcodes, etc. You could write a plugin using ArchivesSpace’s plugin API, but that requires facility with Ruby as well as access to the environment where the application is running. A more lightweight approach is to access your data through ArchivesSpace’s powerful REST API, and process it using whatever scripting language you prefer. This talk will present some simple “scriptaloging” solutions that a moderately skilled programmer can use to automate data entry or import tasks using an extendable command line tool written in NodeJS (https://www.npmjs.org/package/as-cli) and loosely inspired by Drupal’s drush utilitly.
 
== Consuming Big Linked Open Data in Practice: Authority Shifts and Identifier Drift ==
* Kathryn Stine, katstine@berkeley.edu, UC Berkeley (first-time presenter)
* Stephanie Collett, stephanie.collett@ucop.edu, California Digital Library, UC (c4l 2012 presenter)
 
Increasingly, authoritative datasets of interest to libraries (subjects, names, classifications, etc.) are are available in bulk, exposed as linked open data. Unfettered access can allow libraries to aggregate, connect, and augment data in new ways that will benefit users. This talk will describe our exploratory experience integrating bulk data from the Virtual International Authority File (VIAF) into HathiTrust metadata to improve discovery and collection management.
 
Authoritative data is not static - datasets change with new contributions and re-clustering, resulting in new identifier relationships. We will describe the challenges this presents with accessing, processing, and syncing our metadata with a massive, complex linked dataset. We will talk about our technical approach to navigating an ecosystem of identifiers and mitigating cached identifier drift between systems as authority data shifts. We aim to spark conversation about data accessibility and the relationships between local, consortial, and authoritative metadata as the library community moves beyond “Hello, world” linked data examples to integrating this data at scale into existing systems.
 
== Your Chocolate is in My Peanut Butter! Mixing up Content and Presentation Layers to Build Smarter Books in Browsers with RDFa, Schema.org, and Linked Data Topics ==
 
* Jason A. Clark [@jaclark], Head, Library Informatics & Computing, Montana State University, jaclark@montana.edu
* Scott W. H. Young [@hei_scott], Digital Initiatives Librarian, Montana State University, swyoung@montana.edu
 
Common methods of publishing book content have focused on various implementations of existing technologies such as LaTeX, Markdown, and .epub. A common theme within this development has been the separation of presentation layers and content layers. What if there was another way? In responding to that question, we’ll look at our local @msulibrary prototype software funded by an IMLS Sparks! Innovation grant for presenting books inside of web browsers (https://github.com/jasonclark/bib-template). Our talk will focus on the tools and technologies of open web publishing. We’ll consider the strange and wonderful benefits of integrating the presentation layer and content layer using semantic RDFa HTML5 markup, and we’ll demonstrate how describing and displaying books within an open web model impacts discovery, eBook production, and machine-readability.
 
Our session will include:
* The benefits of using RDFa, Schema.org, and linked data models for book production
* How structured data models for book content can turn your webpage into your API
* Analyze the effects of this practice for machine-understanding, SEO, and UX
* Discuss the advantages and disadvantages of this model as it applies to a range of book genres, including web book prototypes for fiction and poetry.
 
Chocolate/Peanut Butter... RDFa/HTML5... Linked Data/Dbpedia Topics... "Great tastes that go great together."
 
== Integration/Collaboration: Playing Well With Others ==
* Sarah Romkey, sromkey@artefactual.com, Artefactual Systems
* Justin Simpson, jsimpson@artefactual.com, Artefactual Systems
 
There are many different software applications and systems being used in collecting institutions to handle all aspects of managing their digital collections. Within a single institution, it is possible to have several tools with overlapping functionality. Sometimes the biggest gains in functionality and productivity come from a focus on connecting these systems together.
 
In this talk we will present several examples of integration and collaboration that we have been involved in while working on the Archivematica project. Examples of integrations with both open source and proprietary systems, as well as examples of collaborative development processes.
 
== BYOB: Build Your Own Bootstrap ==
 
* Axa Mei Liauw, aliauw@princeton.edu, Princeton University Library
* Kevin Reiss, kr2@princeton.edu, Princeton University Library
 
Bootstrap is awesome! It allows for quick and easy front-end development without demanding too much time with the design of the user interface. A framework like Bootstrap can also help libraries address one of our top usability problems: wild inconsistency of styles and layouts across our own in-house applications and websites.
 
However, what if you do not want to compromise your creativity and make your project visually stand out? With enough Sass or Less knowledge it is possible to highly customize Bootstrap to fit your needs, but in the end you still end up with your design and code buying into Bootstrap's markup and design conventions, which are not always what you want. Sometimes you simply want to start with a clean slate and not fight the framework and use only the components and features you need.
 
In that case, build your own Bootstrap! In this talk we will discuss techniques for using some of the emerging tools from the front-end development and design communities to create a maintainable, modular in-house Sass toolkit that we are implementing within a variety of tools including Drupal, Rails applications, in house PHP applications, as well as within vendor solutions like Libguides and ILLiad. Some specific tools we will discuss are Grunt, Bower, and Sass tools like Singularity, Breakpoint, and Bourbon.
 
== Plz to let me download this ebook: an idea for better leisure reading access ==
 
* Lauren Ajamie, lajamie@nd.edu, University of Notre Dame
 
(I haven't presented at Code4Lib before)
 
The landscape of finding and getting fiction (and non-fiction for leisure reading) has changed dramatically in the past few years, and one of the most disruptive new developments could turn out to be Oyster, a "Netflix for books" ebook subscription. While somewhat expensive and with a developing selection of materials, the platform is amazingly easy to use, and makes finding and reading ebooks a two step process: find a book, start reading. Compare this to getting ebooks (or even print books) from your library, a process which could take multiple user names and up to a dozen steps. For both the future of libraries, and my own selfish desires, I want to make this better! This speculative talk will discuss an idea I have (that I will need help with) to make finding and borrowing leisure reading materials less frustrating, and will hopefully result in a library-wide conversation about the discovery of, and access to, leisure reading materials.
 
== Formats For All! The Future of the Archivematica Format Policy Registry ==
* Misty De Meo, mdemeo@artefactual.com, Artefactual Systems (first-time talk submitter)
 
Archivematica's Format Policy Registry (FPR) is an open-source preservation planning tool. The FPR is designed to abstract away many of the challenges involved in identifying file formats, as well as picking the right tools to use to perform tasks like metadata extraction, format transformation and normalization.
 
The first part of this talk will focus on the FPR's structure and its capabilities, but this isn't just a vendor pitch. The current version of the registry is only designed around Archivematica's usecases, but we want to bring the FPR to the world and make it work great for other software too. Share your format policies between Archivematica, Islandora, and Hydra! Share your file IDs between applications so you never misidentify a file again! Build new software without having to worry about the details of how to run external tools on files! The second part of the talk will focus on possible future developments, how the FPR might change, and what a community effort to bring the FPR forward might look like.
 
== PeerLibrary – Facilitating the global conversation on academic literature ==
 
* Timothy Quach, UC Berkeley
* Rachel Wolfsohn, UC Berkeley, rachel at peerlibrary.org
* Mitar Milutinovic, UC Berkeley, mitar.code4lib at peerlibrary.org
* Not presented or attended code4lib before
 
We present [https://peerlibrary.org/ PeerLibrary] as an example of how to intertwine various tools and methods to spread the conversation about academic publications. PeerLibrary is unique because of our collaborative annotation and discussions layer integrated with the in-browser PDF viewer. PeerLibrary provides tools to annotate and comment directly on the works. Users can highlight important information and take notes in the margins online, just as how one would mark up a physical research paper.
 
Our vision is a virtual space spans the horizontal of all academic knowledge, where individuals across the globe can connect and learn things only another human can help teach. PeerLibrary can be used as a system for depositing original work, archive navigation, to help organize conferences and journals, as a reference manager, a social network to follow what others are doing, for proposing/recommending connections, collaboration, reading, commenting, organizing or sharing papers. PeerLibrary is a platform that synthesizes an online repository and those who utilize it. Our longer-term goal is for PeerLibrary to become a web-first standards-based publication platform. We believe that all stages of research should be collaborative, from the idea, to experiments, conclusions and publication (constant feedback).
 
To achieve this vision we are working with partners. Such as [http://www.contentmine.org/ ContentMine], to push facts they mine in academic publications into commons available through PeerLibrary. We use Internet Archive to permanently store open access publications and public annotations for future generations. We are using code developed at projects [http://hypothes.is/ Hypothes.is] and [http://annotateit.org/ Annotator] to ensure compatibility with other annotation projects and platforms. We are also using the [http://meteor.com/ Meteor framework], [http://lens.elifesciences.org/ eLife Lens], [https://github.com/guardian/scribe The Guardian Scribe] and [https://github.com/mozilla/pdf.js Mozilla PDF.js]. Our presentation will be an exploration of how these organizations are contributing to the Open Scholarship initiative.
 
With no subscription or registration needed, anybody can easily access the annotations in PeerLibrary. By eliminating the need to print files, the research process is contained in one online medium, creating a greater sense of organization. Furthermore, note-taking tools encourage active analysis, a recording of insights and questions while reading. Users can choose to keep their highlights and annotations private, or they can make them public to engage in an open online discussion. We believe that the collection of tools provided in PeerLibrary will make education more inclusive.
 
See it live here: https://peerlibrary.org/
 
Project presentation: https://vimeo.com/93085636
 
Screencast: https://vimeo.com/109787685

Navigation menu