<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>https://wiki.code4lib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=MatthewMiller</id>
		<title>Code4Lib - User contributions [en]</title>
		<link rel="self" type="application/atom+xml" href="https://wiki.code4lib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=MatthewMiller"/>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/Special:Contributions/MatthewMiller"/>
		<updated>2026-04-09T06:31:04Z</updated>
		<subtitle>User contributions</subtitle>
		<generator>MediaWiki 1.26.2</generator>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2015_Preconference_Proposals&amp;diff=42012</id>
		<title>2015 Preconference Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2015_Preconference_Proposals&amp;diff=42012"/>
				<updated>2014-11-07T22:41:53Z</updated>
		
		<summary type="html">&lt;p&gt;MatthewMiller: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;== Instructions ==&lt;br /&gt;
Thank you for considering proposing a pre-conference! Here are a few details:&lt;br /&gt;
&lt;br /&gt;
* We will be taking pre-conference proposals until '''November 7, 2014'''&lt;br /&gt;
* If you cannot or do not want to edit this wiki directly, you can email your proposals to cmh2166@columbia.edu or collie@msu.edu&lt;br /&gt;
* Examples from the 2014 pre-conference proposals can be found at [[2014 preconference proposals|http://wiki.code4lib.org/2014_preconference_proposals]]&lt;br /&gt;
* If you are interested in ''attending'' a particular pre-conference, please append your name below that proposal (indicating interest in more than one proposal is fine!)&lt;br /&gt;
* If you have an idea for a pre-conference, but cannot facilitate yourself please post the idea below and email cmh2116@columbia.edu or collie@msu.edu&lt;br /&gt;
* '''NOTE:''' Pre-conferences are NOT included in the Code4Lib Conference price and will be held on Monday, February 9, 2015 as either full day or half day sessions&lt;br /&gt;
* Please use the template for proposals provided in the pre-formatted block below&lt;br /&gt;
&lt;br /&gt;
=== Proposal formatting guidelines: ===&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
=== Preconference Title: ===&lt;br /&gt;
 &lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Full Day&amp;quot;''' &lt;br /&gt;
'''&amp;quot;Half Day [Morning]&amp;quot;''' &lt;br /&gt;
'''&amp;quot;Half Day [Afternoon]&amp;quot;''' &lt;br /&gt;
&lt;br /&gt;
* Facilitator's name, affiliation, and email address&lt;br /&gt;
* Second facilitator's name, affiliation, email address, if second speaker&lt;br /&gt;
&lt;br /&gt;
Abstract&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
 &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Pre-conference Proposals ==&lt;br /&gt;
&lt;br /&gt;
Post your ideas here!&lt;br /&gt;
&lt;br /&gt;
=== Delivering and Preserving GIS Data ===&lt;br /&gt;
 &lt;br /&gt;
'''Half Day [Morning]'''&lt;br /&gt;
&lt;br /&gt;
* Darren Hardy, Stanford University, drh@stanford.edu&lt;br /&gt;
* Jack Reed, Stanford University, pjreed@stanford.edu&lt;br /&gt;
&lt;br /&gt;
We will discuss how to set up a spatial data infrastructure (SDI) to deliver GIS data, to manage GIS content in a Fedora repository for preservation, and to establish metadata requirements for good spatial discovery. By the end of the workshop you will have a working SDI! This workshop is a compliment to the GeoBlacklight workshop in the afternoon.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# [[User:Ssimpkin|Sarah Simpkin]]&lt;br /&gt;
# Vicky Steeves&lt;br /&gt;
# Andrew Battista&lt;br /&gt;
# Peggy Griesinger&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== A hands-on introduction to GeoBlacklight ===&lt;br /&gt;
&lt;br /&gt;
'''Half Day [Afternoon]'''&lt;br /&gt;
&lt;br /&gt;
* Darren Hardy, Stanford University, drh@stanford.edu&lt;br /&gt;
* Jack Reed, Stanford University, pjreed@stanford.edu&lt;br /&gt;
&lt;br /&gt;
GeoBlacklight is a discovery solution for geospatial data that builds on the successful Blacklight platform. Many libraries have collections of GIS data that aren’t easily discoverable. This will be a hands-on workshop, focused on installing and running GeoBlacklight which builds on the morning workshop &amp;quot;Delivering and Preserving GIS Data&amp;quot;.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# [[User:Ssimpkin|Sarah Simpkin]]&lt;br /&gt;
# Vicky Steeves&lt;br /&gt;
# Andrew Battista&lt;br /&gt;
# Peggy Griesinger&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
===RailsBridge: Intro to programming in Ruby on Rails===&lt;br /&gt;
'''&amp;quot;Half-Day&amp;quot; [morning]'''&lt;br /&gt;
* Contact Carolyn Cole, Penn State University, carolyn@psu.edu&lt;br /&gt;
* Additional instructors welcome&lt;br /&gt;
&lt;br /&gt;
Interested in learning how to program? Want to build your own web application? Never written a line of code before and are a little intimidated? There's no need to be! [http://www.railsbridge.org/ RailsBridge] is a friendly place to get together and learn how to write some code.&lt;br /&gt;
&lt;br /&gt;
RailsBridge is a great workshop that opens the doors to projects like [http://projectblacklight.org/ Blacklight] and [http://projecthydra.org/ Hydra] and [https://github.com/traject-project/traject Traject].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Maura Carbone&lt;br /&gt;
#Vicky Steeves&lt;br /&gt;
# Peggy Griesinger&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Replace yourself with a painfully complex bash script...or try Ansible ===&lt;br /&gt;
'''Half Day [Morning]'''&lt;br /&gt;
&lt;br /&gt;
* Chad Nelson, chad dot nelson @ lyrasis dot org&lt;br /&gt;
* Blake Carver, Blake dot carver @lyrasis dot org&lt;br /&gt;
&lt;br /&gt;
Abstract: &lt;br /&gt;
&lt;br /&gt;
[http://www.ansible.com Ansible] is an open source automation and [http://en.wikipedia.org/wiki/Configuration_management configuration management] tool that focuses on simplicity to help make your life as a developer, or a sysadmin, or even a full on devops-er, easier. This workshop will cover the basic building blocks used in Ansible as well as some best practices for maintaining your Ansible code. We will start by working through a simple example together, and then participants will be given time to work on their own projects with instructors providing guidance and troubleshooting along the way. By the end of the session, participants will have a working knowledge of Ansible and be able to write a working [http://docs.ansible.com/playbooks.html playbook] to meet local needs.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
 &lt;br /&gt;
=== Intro to Docker ===&lt;br /&gt;
'''Half Day [Whenever]'''&lt;br /&gt;
&lt;br /&gt;
* John Fink, McMaster University, john dot fink at gmail dot com&lt;br /&gt;
* Francis Kayiwa, Kayiwa Consulting , francis dot kayiwa at gmail dot com&lt;br /&gt;
&lt;br /&gt;
Abstract:&lt;br /&gt;
&lt;br /&gt;
[http://docker.io Docker] ([http://journal.code4lib.org/articles/9669 jbfink code4lib journal article]) is an open source Linux operating system-level virtualization framework that has seen great uptake over the past year. This workshop will take you through the basic features of Docker, including setup, importing of containers, development workflows and deploying. Knowing when Docker is useful and when it isn't will also be covered. Ideally, every attendee will have ample experience creating and running their own Docker instances by the end.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
1. Jim Hahn&lt;br /&gt;
&lt;br /&gt;
=== Code Retreat ===&lt;br /&gt;
'''Full Day'''&lt;br /&gt;
&lt;br /&gt;
* Jeremy Friesen, University of Notre Dame, jfriesen at nd dot edu&lt;br /&gt;
* Additional facilitators welcome; Especially if you have CodeRetreat experience.&lt;br /&gt;
&lt;br /&gt;
Abstract:&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Coderetreat is a day-long, intensive practice event, focusing on the fundamentals of software development and design.&lt;br /&gt;
By providing developers the opportunity to take part in focused practice, away from the pressures of 'getting things done', the coderetreat format has proven itself to be a highly effective means of skill improvement.&lt;br /&gt;
Practicing the basic principles of modular and object-oriented design, developers can improve their ability to write code that minimizes the cost of change over time.&amp;quot; [http://coderetreat.org/about About Code Retreat]&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
=== Presentations workshop ===&lt;br /&gt;
 &lt;br /&gt;
'''&amp;quot;Half Day [Afternoon]&amp;quot;'''  (but could be expanded based on interest)&lt;br /&gt;
&lt;br /&gt;
* Chris Beer, Stanford University, cabeer@stanford.edu&lt;br /&gt;
* Additional facilitators welcome.&lt;br /&gt;
&lt;br /&gt;
This is a preconference session intended for first time Code4Lib speakers, habitual procrastinators, experienced speakers, those thinking about offering lightning talks, etc. If you're preparing a talk for this year's Code4Lib, this workshop is an opportunity to rehearse your presentation, get feedback from peers, get familiar with the presentation technology, etc.&lt;br /&gt;
 &lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#Vicky Steeves&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Dive into Hydra  ===&lt;br /&gt;
 &lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Half Day [Afternoon]&amp;quot;''' &lt;br /&gt;
&lt;br /&gt;
* Justin Coyne, Data Curation Experts, justin@curationexperts.com&lt;br /&gt;
* Bess Sadler, Stanford University, bess@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Hydra is a collaboration of over 30 educational institutions who work together to solve their repository needs by building open-source software.   Dive into Hydra is a course that bootstraps you into the Hydra software framework.  We'll start at the basics and walk you through the various layers of the Hydra stack.   We'll conclude by installing the Worthwhile gem, enabling every participant to walk away with their own Institutional Repository.  Participants who have prior exposure to web programming will get the most out of this course.  It's recommended (but not required) that you attend &amp;quot;RailsBridge&amp;quot; prior to this workshop.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Maura Carbone&lt;br /&gt;
# Peggy Griesinger&lt;br /&gt;
#&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
=== code4lib/Write The Docs barcamp ===&lt;br /&gt;
 &lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Full Day&amp;quot;''', with options for jumping in for half a day&lt;br /&gt;
&lt;br /&gt;
* code4lib wrangler: Becky Yoose, yoosebec at grinnell dot edu&lt;br /&gt;
* Write the Docs contacts: TBA&lt;br /&gt;
&lt;br /&gt;
Abstract&lt;br /&gt;
&lt;br /&gt;
Documentation. We all know that we need it for things we develop, but most of us either keep putting it off or write documentation that is not maintained, clear, concise, and so on. We're all guilty! So what's stopping us from doing better docs? Luckily, Portland is also the home to the NA Write the Docs conference, and is home for many folks who live and breathe documentation. This barcamp is open to both code4lib and non-code4lib conference attendees and is intended to provide a space where code4libbers can find practices and tools in creating better documentation for all as well as documentation wonks can find out ways in which the library wonks can help with better documentation access and organization. &lt;br /&gt;
&lt;br /&gt;
Remember, like metadata, documentation is a love note to the future.&lt;br /&gt;
&lt;br /&gt;
More information about Write the Docs at http://conf.writethedocs.org/&lt;br /&gt;
&lt;br /&gt;
There will be a nominal fee (t/b/d) for non-Code4LibCon attendees (subject to organizer approval). &lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
'''Full day'''&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
'''Morning'''&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
'''Afternoon'''&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Linked Data Workshop ===&lt;br /&gt;
&lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Half Day [morning]&amp;quot;''' &lt;br /&gt;
&lt;br /&gt;
* Karen Estlund, University of Oregon, kestlund@uoregon.edu&lt;br /&gt;
* Tom Johnson, DPLA, tom@dp.la&lt;br /&gt;
&lt;br /&gt;
Abstract:&lt;br /&gt;
&lt;br /&gt;
Developer and metadata experts-focused linked data workshop. Topics covered will include: linked open data principles, converting existing data, and modeling linked data in DAMS.&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Logan Cox&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Code4Arc ===&lt;br /&gt;
&lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Full Day&amp;quot;''' (with options for half day participation)&lt;br /&gt;
&lt;br /&gt;
* Sarah Romkey, Artefactual Systems, sromkey@artefactual.com&lt;br /&gt;
* Justin Simpson, Artefactual Systems, jsimpson@artefactual.com&lt;br /&gt;
* Chris Fitzpatrick, ArchivesSpace, chris.fitzpatrick@lyrasis.org&lt;br /&gt;
* Alexandra Chassanoff, BitCurator Access, bitcurator@gmail.com&lt;br /&gt;
&lt;br /&gt;
Abstract:&lt;br /&gt;
&lt;br /&gt;
What does it mean to Code for Archives? Is it different than coding for libraries, and if so, how? &lt;br /&gt;
&lt;br /&gt;
Code4Lib is a wonderful and successful model (you must agree or you wouldn't be reading this). This workshop is an attempt to create a space to replicate the model in an Archival context. A space to talk about development for archives, and the particular challenges of developing archival systems.  Topics to discuss include Integration between different Archival software tools, and between Archival tools/workflows and larger institutional tools like institutional repositories, discovery and access systems.&lt;br /&gt;
&lt;br /&gt;
The schedule may include the following:&lt;br /&gt;
&lt;br /&gt;
* Panel type conversations about the State of Art in Archives &lt;br /&gt;
* Case Studies - discussion of workflows at specific institutions, including gaps in tools and how those are being addressed or could be addressed &lt;br /&gt;
* Tool Demos - access to demos of some of the open source tools used in an Archival Context (examples include ArchivesSpace, Archivematica, BitCurator, AtoM)&lt;br /&gt;
&lt;br /&gt;
Artefactual will provide demos running Archivematica and AtoM, Lyrasis will do so for ArchivesSpace, BitCurator will for BitCurator.  We encourage others to chime in here to expand the list of tools available to touch and play with. &lt;br /&gt;
&lt;br /&gt;
When signing up, please indicate if you are an end-user or a developer.&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Laney McGlohon - developer&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Fail4Lib 2015 ===&lt;br /&gt;
'''Half Day [TBD, probably afternoon]'''&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, akorphan (at) ncsu.edu&lt;br /&gt;
* Jason Casden, jmcasden (at) ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Abstract:&lt;br /&gt;
&lt;br /&gt;
Failure. Failure never changes. Since failure is an inescapable part of our professional work, it's important to be familiar with it, to acknowledge it, and to grow from it -- and, in contravention to longstanding tradition, to accept it as a fact of development life. At Fail4Lib, we'll talk about our own experiences with projects gone wrong, explore some famous design failures in the real world, and talk about how we can come to terms with the reality of failure, to make it part of our creative process -- rather than something to be shunned. Let's train ourselves to understand and embrace failure, encourage enlightened risk-taking, and seek out opportunities to fail and learn. This way, when we do what we do -- and fail at what we do -- we'll do so with grace and without fear.&lt;br /&gt;
&lt;br /&gt;
This year's preconference will include new case studies and an improved discussion format. Repeat customers are welcome! (Fail early, fail often.)&lt;br /&gt;
&lt;br /&gt;
The schedule may include the following:&lt;br /&gt;
&lt;br /&gt;
* Case studies. Avoid our own mistakes by bearing witness to the failures of others.&lt;br /&gt;
* Confessionals, for those willing to share. Let's learn from our own (and each others') failures.&lt;br /&gt;
* Group therapy. Vent about your own experiences in a judgment-free setting. Explore how we can make our organizations less risk-averse and more failure-tolerant.&lt;br /&gt;
&lt;br /&gt;
''Interested in attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Coding Custom Solutions for Every Department in the Library with File Analyzer ===&lt;br /&gt;
 &lt;br /&gt;
One of:&lt;br /&gt;
'''&amp;quot;Half Day [Morning]&amp;quot;''' &lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Library, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
Abstract&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Library has shared an application called the [http://georgetown-university-libraries.github.io/File-Analyzer/ File Analyzer] that has allowed us to build custom solutions for nearly every department in the library.&lt;br /&gt;
&lt;br /&gt;
* Analyzing Marc Records for the Cataloging department&lt;br /&gt;
* Transferring ILS invoices for the University Account System for the Acquisitions department &lt;br /&gt;
* Delivering patron fines to the Bursar’s office for the Access Service department&lt;br /&gt;
* Summarizing student worker timesheet data for the Finance department&lt;br /&gt;
* Validating counter compliant reports for the Electronic Resources department&lt;br /&gt;
* Preparing ingest packages for the Digital Services department&lt;br /&gt;
* Validating checksums for the Preservation department&lt;br /&gt;
&lt;br /&gt;
This hands on workshop will step through the components of the application framework.  Workshop participants will install and develop custom File Analyzer tasks in this session.&lt;br /&gt;
&lt;br /&gt;
The workshop agenda will loosely follow the [https://github.com/Georgetown-University-Libraries/File-Analyzer/wiki/File-Analyzer-Training----Code4Lib-2014 pre-conference agenda from Code4Lib 2014].&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Confessions of the (Accidental) Code Hoarder: How to make your Code Sharable: ===&lt;br /&gt;
 &lt;br /&gt;
'''Half Day [Whenever]'''&lt;br /&gt;
&lt;br /&gt;
* Karen A. Coombs, OCLC, coombsk@oclc.org&lt;br /&gt;
&lt;br /&gt;
Abstract&lt;br /&gt;
Have you built something cool and useful that you want to share with others? This preconference session will discuss techniques and tools for sharing code. Using our own OCLC Developer Network PHP authentication code libraries as an example, we will discuss a set of recommended best practices for how to share your code.&lt;br /&gt;
 &lt;br /&gt;
We’ll start with coding standards and test writing so you can be confident of the quality of your code. Next we'll discuss inline documentation as a tool for developers and how auto-generating documentation will save you time and effort. Lastly we'll provide an overview of the tricky areas of dependency and package management, and distribution tools. Along the way, we'll cover PHP coding standards, testing, and popular PHP tools including PHPDoc for documentation, Composer for smooth installations, and using GitHub and Packagist to manage distribution, updates and community feedback.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Peggy Griesinger&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
=== UXtravaganza ===&lt;br /&gt;
'''&amp;quot;Half or Full Day [Based on Interest?, Morning/Afternoon Doesn’t Matter]&amp;quot;'''&lt;br /&gt;
&lt;br /&gt;
* William Hicks, University of North Texas, William.hicks@unt.edu&lt;br /&gt;
* Volunteers?&lt;br /&gt;
&lt;br /&gt;
Abstract&lt;br /&gt;
&lt;br /&gt;
I’m envisioning a 1/2 of full day for front-end developers, content strategy people, and other misfits with an interest in user experience, where we can talk about our shared problems, use cases, the state of current research, and play with each other’s sites. A half day seems doable, but if there’s significant enough interest we could push for a full?  Here are a few of the things I think might be interesting to see happen:&lt;br /&gt;
&lt;br /&gt;
* '''Analytics Share-fest:''' A few volunteers demonstrate data about their websites, catalogs, archival/digital collections. Most of us know our own sites but it would be interesting/validating to share this data with others so we can start to see commonalities between institutions, in certain kinds of systems, etc. For anyone using event tracking, or using click- or heat-maps, this would be a great opportunity to show off what people are seeing.&lt;br /&gt;
&lt;br /&gt;
* '''UX Best Practices Catch Up:''' This spring I had the opportunity to attend a few days worth of usability workshops from the Nielsen-Norman Group, most of which was focused on mobile. I could distill down a lot of the information into an short presentation.  Since this is a constantly moving area of research it would be nice to see a few people do other similar short presentations on some current trends/findings relevant to libraries, search, etc.&lt;br /&gt;
&lt;br /&gt;
* '''Mobile Dev Lab:''' The UNT Libraries has been collecting a small set of smartphones and tablets for testing and development. Basically an [http://labup.org Open Device Lab].  We have about a dozen devices now of varying sizes, OS, OS Versions, + Google Glass. I’ll bring the devices, you can bring yours, and assuming we can get the wifi up and running we can test our sites/services with our big sausage fingers rather than pretending to do so through emulators and the one or two devices we each usually have on hand. If anyone is game they can do a tutorial on Browser-based Inspector Tools, Browser-Cams, or other testing services.&lt;br /&gt;
&lt;br /&gt;
* '''The Eye’s Have It.''' The UNT Libraries is also in the process of acquiring an eye tracker and software for usability and other gaze-based research studies. We’ll take possession of it shortly after this pre-conference proposal is due and will have a couple of months to play with them before the conference.  Assuming we can get our act together learning the device and can get past the technical hurdles of setting it up at the pre-conference, we could try to do some live demos on each other’s sites; i.e. You nominate a site/service, someone in the audience volunteers to wear the device, and we all watch them struggle do the tasks you request on a projector. Rinse. Lather. Repeat. It would hardly be scientific, but it sure would be fun. As a backup, if we have some sites nominated beforehand, I can run a few students at my library through some tasks here and we can show off the results to the crowd.&lt;br /&gt;
&lt;br /&gt;
For those of you wanting to attend and help out, I’d really like to see some discussion on typography, writing for the web, “dealing with business/administrative requirements from on-high&amp;quot;, maybe do some prototyping exercises, etc. Similarly if anyone is interested in doing some tutorials on bootstrap or how-to’s on running a usability test, that would be rad. But we need you to step up and steer part of the time for most of this to work, so if you are interested in some aspect, and especially if you want to volunteer to lead a bit of the time, contact me.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your&lt;br /&gt;
name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Intro to Git &amp;amp; possibly beyond ===&lt;br /&gt;
 &lt;br /&gt;
'''Half Day [Whenever]'''&lt;br /&gt;
&lt;br /&gt;
* Erin Fahy, Stanford University, efahy@stanford.edu&lt;br /&gt;
* Shaun Trujillo, Mount Holyoke College, strujill@mtholyoke.edu&lt;br /&gt;
&lt;br /&gt;
We can start with the basics of Git and discuss ways in which it can help you version control just about any file, not just code. Points we can go over:&lt;br /&gt;
&lt;br /&gt;
* What is a Distributed Version Control System?&lt;br /&gt;
* What's the difference between Git and Github.com?&lt;br /&gt;
* How to initialize new Git projects locally and on a remote server/Github&lt;br /&gt;
* Cloning/Forking existing projects and keeping up to date&lt;br /&gt;
* The wonderful world of Git branches&lt;br /&gt;
* Interactive rebasing&lt;br /&gt;
* Contributing code to existing projects &amp;amp; what pull requests are&lt;br /&gt;
* How to handle merge conflicts&lt;br /&gt;
* Overview of workflows and branch best practices&lt;br /&gt;
* (time allowing) Advanced git: pre/post hooks, submodules, anything else?&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
&lt;br /&gt;
=== Visualizing Library Data ===&lt;br /&gt;
 &lt;br /&gt;
'''&amp;quot;Half Day [Morning||Afternoon]&amp;quot;''' &lt;br /&gt;
&lt;br /&gt;
* Matt Miller, matthewmiller@nypl.org, New York Public Library, NYPL Labs&lt;br /&gt;
&lt;br /&gt;
Visualizing your institution’s data can give new insight about your holding’s strengths, weaknesses and outliers. They can also provide potential new avenues for discovery and access. This half day session will focus on programmatically visualizing library metadata. Emphasis will be on creating web-based visualizations utilizing libraries such as d3.js but attention paid towards visualizing large datasets while keeping them web accessible. By then end of the session participants will have template, sample code and methodologies enabling them to start producing visualization with their own data.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
#&lt;br /&gt;
 &lt;br /&gt;
&lt;br /&gt;
[[Category:Code4Lib2015]]&lt;/div&gt;</summary>
		<author><name>MatthewMiller</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2015_Prepared_Talk_Proposals&amp;diff=42006</id>
		<title>2015 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2015_Prepared_Talk_Proposals&amp;diff=42006"/>
				<updated>2014-11-07T20:51:22Z</updated>
		
		<summary type="html">&lt;p&gt;MatthewMiller: /* Distributed Remediation: Small tools for big problems: */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Code4lib 2015 is a loosely-structured conference that provides people working at the intersection of libraries/archives/museums/cultural heritage and technology with a chance to share ideas, be inspired, and forge collaborations. For more information about the Code4lib community, please visit http://code4lib.org/about/. &lt;br /&gt;
The conference will be held at the Portland Hilton &amp;amp; Executive Tower in Portland, Oregon, from February 9-12, 2015.&lt;br /&gt;
&lt;br /&gt;
'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
We encourage everyone to propose a talk.&lt;br /&gt;
 &lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
* Projects you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* Tools and technologies – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* Technical issues - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* Relevant non-technical issues – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
Proposals can be submitted through Friday, November 7, 2014 at 5pm PST (GMT−8). Voting will start on November 11, 2014 and continue through November 25, 2014. The URL to submit votes will be announced on the Code4Lib website and mailing list and will require an active code4lib.org account to participate. The final list of presentations will be announced in early- to mid-December.&lt;br /&gt;
&lt;br /&gt;
'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Log in to the Code4lib wiki and edit this wiki page using the prescribed format. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist voters in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
Please follow the formatting guidelines:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Talk Title: ==&lt;br /&gt;
 &lt;br /&gt;
* Speaker's name,  email address, and (optional) affiliation&lt;br /&gt;
* Second speaker's name, email address, and affiliation, if second speaker&lt;br /&gt;
&lt;br /&gt;
Abstract of no more than 500 words.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
== Zines + Gamification = Awesomest Metadata Literacy Outreach Event Ever! ==&lt;br /&gt;
 &lt;br /&gt;
* [http://www.JenniferHecker.info Jennifer Hecker], jenniferraehecker@gmail.com, [http://www.lib.utexas.edu/subject/zines University of Texas Libraries] &amp;amp; [http://www.AustinFanzineProject.org Austin Fanzine Project]&lt;br /&gt;
* [http://anomalily.net/ Lillian Karabaic], librarian@iprc.org, [http://www.iprc.org/ Independent Publishing Resource Center] (Portland)&lt;br /&gt;
 &lt;br /&gt;
In academic libraries, and elsewhere, the popularity of zine (a magazine produced for love, not profit) collections is on the rise. At the same time, metadata literacy is becoming an increasingly important skill, helping people navigate and understand digital environments and interactions. We have found a way to teach metadata literacy to the general public that isn’t super-boring – in fact, we’ve made it downright fun!&lt;br /&gt;
&lt;br /&gt;
First, volunteer zine librarian Lillian Karabaic of Portland’s Independent Publishing Resource Center facilitated the creation of a gamified cataloging interface for the IPRC’s annual Raiders of the Lost Archives backlog-busting 24-hour volunteer cataloging event.&lt;br /&gt;
&lt;br /&gt;
Then, archivist Jennifer Hecker facilitated the adaptation of the IPRC’s game for use in a similar, but also very different context – promoting UT Libraries newly-acquired zine collections. The main goal of the academic-library-based event was increasing excitement around the collections, but with the side goal of building metadata literacy, and introducing an understanding of library cataloging issues.&lt;br /&gt;
&lt;br /&gt;
The Texas modification also conforms to the xZINECOREx metadata schema developed by the national [http://zinelibraries.info/ Zine Librarians Interest Group], and triggered interesting conversations with the Libraries’s cataloging department about evolving metadata standards and how to incorporate the products of crowd-sourcing projects into existing workflows.&lt;br /&gt;
&lt;br /&gt;
Both games will be demoed.&lt;br /&gt;
&lt;br /&gt;
We have never presented at Code4lib.&lt;br /&gt;
&lt;br /&gt;
== Do the Semantic FRBRoo ==&lt;br /&gt;
* Rosie Le Faive, rlefaive@upei.ca, University of Prince Edward Island&lt;br /&gt;
&lt;br /&gt;
[http://www.islandora.ca Islandora] is great for creating repositories of any data type, but how can you model meaningful relationships between digital objects and use them to tell a story?&lt;br /&gt;
&lt;br /&gt;
At UPEI, I’m assembling an ethnography of Prince Edward Island’s traditional fiddle music that includes musical clips, video clips, oral histories, musical notation, images, and ethnographic commentaries. In order to present an exhibition-style site, I’m tying these digital objects together via the people, places, events, tunes and topics that they share or describe. &lt;br /&gt;
&lt;br /&gt;
To describe the relationships, I’m extending Islandora to use [http://www.cidoc-crm.org/frbr_inro.html FRBRoo], a vocabulary that combines the FRBR model with CIDOC-CRM, the the object-oriented museum documentation ontology. These modules being developed will allow other researchers to create a structured, navigable digital repository of diverse object types, that uses Islandora as an exhibition platform. &lt;br /&gt;
&lt;br /&gt;
== Our $50,000 Problem: Why Library School? ==&lt;br /&gt;
* Jennie Rose Halperin, jhalperin@mozilla.com, Mozilla Corporation&lt;br /&gt;
&lt;br /&gt;
57 library schools in the United States are churning out approximately 100 graduates per year, many with debt upwards of $50,000.  According to ONet, [http://www.inthelibrarywiththeleadpipe.org/2011/is-the-united-states-training-too-many-librarians-or-too-few-part-1/ 84% of library jobs in the US require an MLS.] The library profession is [http://dpeaflcio.org/programs-publications/issue-fact-sheets/library-workers-facts-figures/) 92% white and 82% female and entry-level librarians can expect to make $32,500 per year.]&lt;br /&gt;
&lt;br /&gt;
Contrasted with developers, who are almost [http://www.ncwit.org/blog/did-you-know-demographics-technical-women 90% male] and can expect to make [http://www.forbes.com/sites/jennagoudreau/2011/06/01/best-entry-level-jobs/ $70,000 in an entry-level position,] these numbers are dismal.&lt;br /&gt;
&lt;br /&gt;
According to a recent survey, the top skill that outgoing library students want to know is “programming” and yet many MLS programs still consider Microsoft Word an essential technology skill.&lt;br /&gt;
&lt;br /&gt;
What is going on here? Why do we accept this fate, where mostly female debt-burdened professionals continue to be thrown onto the work force without the education their expensive degrees promised?&lt;br /&gt;
&lt;br /&gt;
As a community we need to come together to stop this cycle. We need to provide better support and mentorship to diversify and keep the profession relevant and help librarianship move into the future it deserves.&lt;br /&gt;
&lt;br /&gt;
This talk will walk through the challenges of navigating a hostile employment environment as well as present models for better development and future state imagining.&lt;br /&gt;
&lt;br /&gt;
== No cataloging software? Need more than Dublin Core? No problem!: Experiences with CollectiveAccess ==&lt;br /&gt;
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University&lt;br /&gt;
* Rachel Wittmann, rwittma@clemson.edu, Clemson University&lt;br /&gt;
&lt;br /&gt;
Clemson University Libraries has implemented the open-source software CollectiveAccess for customized digital collection needs. CollectiveAccess is an open-source project with the goal of providing a flexible way to manage and publish museum and archival collections. There are several applications associated with the projects; most used are: Providence (for cataloging and entering metadata) and Pawtucket (for displaying objects in a collection for the public). It has many profiles readily available for installing with existing library standards, such as Dublin Core, and there is a robust syntax for creating your own profiles to fit custom tailored metadata schemas. Plus, the user interface allows you to modify the metadata profile quickly and easily.&lt;br /&gt;
&lt;br /&gt;
In this talk, we will discuss:&lt;br /&gt;
* Our experiences with installing Providence and creating an installation profile that satisfies the needs of many of the Clemson Libraries digital archiving processes. &lt;br /&gt;
* The stumbling blocks experienced in that process and how they were resolved.&lt;br /&gt;
* The available plugins sourcing widely used authorities, such as Library of Congress thesauri and GeoNames.org, and how they have been used by our projects. &lt;br /&gt;
* A brief overview of the export and import functions and also current workflow practices within Providence.&lt;br /&gt;
* Future plans &amp;amp; the role of CollectiveAccess at Clemson University Libraries&lt;br /&gt;
&lt;br /&gt;
== Getting ContentDM and Wordpress to Play Together ==&lt;br /&gt;
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University&lt;br /&gt;
&lt;br /&gt;
Clemson University Libraries has a very strong program for digitizing and archiving photographs, and the Digital Imaging team processes many hundreds of photographs every month. These images are managed using different methods, including ContentDM, a digital collection manager.&lt;br /&gt;
&lt;br /&gt;
ContentDM provides various methods for searching and displaying photographs, along with their metadata. However, recent initiatives have resulted in the need to leverage those collections into exhibits displayed on other library-related websites, such as our Special Collections unit. The Clemson Libraries has invested heavily in Wordpress as our content management system of choice, and it seemed most efficient not to have to export and import images into our Wordpress sites in order to provide exhibited images.&lt;br /&gt;
&lt;br /&gt;
Fortunately, ContentDM has provided an API to many of their functions, allowing the extraction of metadata and even rescaled images through URLs. This project has been developing a plugin for Wordpress that integrates with ContentDM through shortcodes that Wordpress editors can easily include in their content. These shortcodes allow editors to choose how many images, which images from which collections, thumbnail sizes, etc. to display in different gallery styles. Plans are for it to allow integration with different plugins such as Fancybox and Masonry.&lt;br /&gt;
&lt;br /&gt;
In this presentation, I will demonstrate the current state of the plugin and discuss future plans. &lt;br /&gt;
&lt;br /&gt;
==Refinery — An open source locally deployable web platform for the analysis of large document collections==&lt;br /&gt;
 &lt;br /&gt;
* [[User:DaeilKim|Daeil Kim]], The New York Times, daeil.kim@nytimes.com&lt;br /&gt;
&lt;br /&gt;
Refinery is an open source web platform for the analysis of large unstructured document collections. It extracts meaningful semantic themes within documents also known as &amp;quot;topics&amp;quot; which can be thought of as word clouds composed of terms that highly co-occur with one another. Once this semantic index is formed, one can extract relevant documents related to these topics and further refine their contents through a summarization process that allows users to search for phrases that are relevant to them within the corpus. The goal of Refinery is to make this whole process easier and to provide some of the latest scalable versions of these learning algorithms in an intuitive web-based interface. Refinery is also meant to be run locally, thus bypassing the need for securing document collections over the internet. The talk will go through some of the technologies involved and a demo of the app.&lt;br /&gt;
&lt;br /&gt;
For more info check out http://www.docrefinery.org.&lt;br /&gt;
&lt;br /&gt;
==Drupal 8 — Evolution &amp;amp; Revolution==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Drupal 8 is in beta and nearing release. Among its many features, it notably has become more developer friendly through its adoption of the Symfony PHP framework along with Symfony's outstanding set of libraries (like Guzzle) and tools (like Composer). And, in implementing the Twig theming system, it is can begin to escape PHPtemplate. These moves also make it easier to create headless systems that uses Angular.js and other systems for presentation, or even forgo presentation entirely.&lt;br /&gt;
&lt;br /&gt;
From the site-builder's perspective, Drupal 8 provides a much smother experience and makes it easier to build and implement site recipes.&lt;br /&gt;
&lt;br /&gt;
==Using GameSalad to Build a Gamified Information Literacy Mobile App for Higher Education==&lt;br /&gt;
 &lt;br /&gt;
* [[User:StanBogdanov|Stanislav 'Stan' Bogdanov]],  stan@stanrb.com, Adelphi University and [http://bogliollc.com Boglio LLC]&lt;br /&gt;
&lt;br /&gt;
GameSalad is a popular tool for developing mobile and desktop games with little actual programming. In this presentation, Stan Bogdanov breaks down the development process he followed while building [https://github.com/stanrb/mobiLit mobiLit], a mobile app with the goal of being the first open-source gamified information literacy app to be used as part of a college-level information literacy curriculum. He will go through the basics of using GameSalad to create an app that can be easily customized by non-programmers and the instructional principles used to teach the material in a mobile medium. Stan will also go through two qualitative design studies he did on the app and discuss their results and the lessons learned from building mobiLit. The session will conclude with an overview of the next steps for the [https://github.com/stanrb/mobiLit mobiLit project].&lt;br /&gt;
&lt;br /&gt;
==The Impossible Search: Pulling data from multiple unknown sources==&lt;br /&gt;
 &lt;br /&gt;
* Riley Childs, no official affiliation (currently a Senior in High School at Charlotte United Christian Academy), rchilds (AT) cucawarriors.com &lt;br /&gt;
&lt;br /&gt;
It's easy to search data you know the structure of, but what if you need to pull in data from sources that don't have a standard structure. The ability to search community events along with your standard catalog search results is an example, but often the only way to pull these events is through XML, JSON, (Insert structured format here), or even just raw html. But how do you get that structure? That simple question is what makes this impossible. The process to define and process this structure takes a lot of manual labor, especially if the data you are pulling is just HTML, and then every time you add data to the index you have to run all the data through a script to pull in data in a format Solr or an other index can use. This talk will focus on Solr, but the principles explained will apply to many other indexes.&lt;br /&gt;
&lt;br /&gt;
==What! You're Not Using Docker?==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Boring part: Docker[1] is a container system that provides benefits similar to virtualization with only a fraction of the overhead. Scintillating part: Docker can host between four to six times the number of service instances than systems such as Xen or VMWare on a given piece of hardware. But thats not all! Docker also makes it simple(r) to create transportable instances, so you can spin up development servers on your laptop.&lt;br /&gt;
&lt;br /&gt;
*[1]https://www.docker.com/&lt;br /&gt;
&lt;br /&gt;
== Video Accessibility, WebVTT, and Timed Text Track Tricks ==&lt;br /&gt;
&lt;br /&gt;
* Jason Ronallo, jronallo@gmail.com, NCSU Libraries&lt;br /&gt;
&lt;br /&gt;
Video on the Web presents new challenges and opportunities. How do you make your video more accessible to those with various disabilities and needs? I'll show you how. This presentation will focus on how to write and deliver captions, subtitles, audio descriptions, and timed metadata tracks for Web video using the WebVTT W3C standard. Encoding timed text tracks in this way opens up opportunities for new functionality on your websites beyond accessibility. The presentation will show some examples of the potential for using timed text tracks in creative ways. I'll cover all the HTML and JavaScript you will need to know as well as some of the CSS and other bits you could probably do without but are too fun to pass up.&lt;br /&gt;
&lt;br /&gt;
== Categorizing Records with Random Forests ==&lt;br /&gt;
 &lt;br /&gt;
* Geoffrey Boushey, geoffrey.boushey@ucsf.edu, UCSF Library&lt;br /&gt;
Academic libraries are increasingly responsible for providing ingest, search, discovery, and analysis for data sets.  Emerging techniques from data science and machine learning can provide librarians and developers with an opportunity to generate new insights and services from these document collections.  This presentation will provide a brief overview of common machine learning classification techniques, then dive into a more detailed example using a random forest to assign keywords to research data sets.  The talk will emphasize the insight that can be gained from machine learning rather than the inner workings of the algorithms.  The overall goal of this presentation is to provide librarians and developers with the context to recognize an opportunity to apply machine learning categorization techniques at their home campuses and organizations.  &lt;br /&gt;
&lt;br /&gt;
== Data Science in Libraries ==&lt;br /&gt;
 &lt;br /&gt;
* Devon Smith, smithde@oclc.org, OCLC&lt;br /&gt;
&lt;br /&gt;
Data Science is increasing in buzz and hype. I'll go over what it is, what it isn't, and how it fits in libraries.&lt;br /&gt;
&lt;br /&gt;
== PDF metadata extraction for academic literature == &lt;br /&gt;
&lt;br /&gt;
* Kevin Savage, kevin.savage at mendeley.com, Mendeley&lt;br /&gt;
* Joyce Stack, joyce.stack at mendeley.com, Mendeley&lt;br /&gt;
&lt;br /&gt;
Mendeley recently added a, &amp;quot;document from file,&amp;quot; endpoint to its API which attempts to extract metadata such as title and authors directly from PDF files. This talk will describe at a high level the machine learning methods we used including how we measured and tuned our model. We will then delve more deeply into our stack, the tools we used, some of the things that didn't work and why PDFs are the worst thing ever to compute over.&lt;br /&gt;
&lt;br /&gt;
== Giving Users What They Want: Record Grouping in VuFind ==&lt;br /&gt;
 &lt;br /&gt;
* Mark Noble,  mark@marmot.org, [//www.marmot.org Marmot Library Network]&lt;br /&gt;
&lt;br /&gt;
In 2013, Marmot did extensive usability studies with patrons to determine what was difficult in the catalog.  Many patrons had problems sifting through all of the various formats and editions of a title.  In 2014 we developed a method for [//mercury.marmot.org/Union/Search?lookfor=divergent grouping records] so only a single work is shown in search results and all formats and editions are listed under that work.  We will discuss our definition of a 'work' based on FRBR principles; combining meta data from MARC records with metadata from other sources like OverDrive; the technical details of Record Grouping; the design decisions made during implementation; and the reaction from users and staff.&lt;br /&gt;
&lt;br /&gt;
== Topic Space: a mobile augmented reality recommendation app ==&lt;br /&gt;
&lt;br /&gt;
* Jim Hahn, jimhahn@illinois.edu, University of Illinois at Urbana-Champaign&lt;br /&gt;
&lt;br /&gt;
The Topic Space module (http://minrvaproject.org/modules_topicspace.php ) was developed with an IMLS Sparks! Grant to investigate augmented reality technologies for in-library recommendations. The funding allowed for sustained university community collaboration by the University Library, the Graduate School of Library and Information Science, as well as graduate student programmers sourced from the Department of Computer Science. Collaborators designed app functionality and identified relevant open source libraries that could power optical character recognition (OCR) functionality from within the mobile phone.&lt;br /&gt;
&lt;br /&gt;
Topic space allows a user to take a picture of an item's call number in the book stacks. The module will show the user other books that are relevant but that are not shelved nearby. It can also show users books that are normally shelved here but that are currently checked out. Recommendations are based on Library of Congress subject headings and ILS circulation data which indicate recommendation candidates based on total check-outs. &lt;br /&gt;
&lt;br /&gt;
Research questions included development of back end (server-side) pattern matching algorithms for recommendations, and a rapid formative evaluation of interface design that would provide optimal user experience for navigation of the book stacks as a context to recommendations.&lt;br /&gt;
&lt;br /&gt;
Along with the Topic Space native app, grant collaborators prototyped web based recommendations which could serve as a new way of providing readers advisory and “more like this” recommendations from discovery interfaces accessed through desktop browsers. Outcomes of the grant include the availability of the [https://play.google.com/store/apps/details?id=edu.illinois.ugl.minrva Topic Spaces module within Minrva app on the Android Play store] and an experimental [http://backbonejs.org/ Backbone.js] based [http://minrva-dev.library.illinois.edu Topic Space web app].&lt;br /&gt;
&lt;br /&gt;
== Leveling Up Your Git Workflow ==&lt;br /&gt;
&lt;br /&gt;
* Megan Kudzia, moneill@albion.edu, Albion College Library&lt;br /&gt;
* Kate Sears, eks11@albion.edu, Albion College Library&lt;br /&gt;
&lt;br /&gt;
Have you started experimenting with Git on your own, but now you need to include others in your projects? Learn from our mistakes! Transitioning from a one-person git workflow and repo structure, to a structure that includes multiple people (including student workers), is not for the faint of heart. We'll talk about why we decided to work this way, our path to developing a git culture amongst ourselves, conceptual and technical difficulties we've faced, what we learned, and where we are now. Also with pretty pictures (aka workflow drawings).&lt;br /&gt;
&lt;br /&gt;
== Drone Loaning Program: Because Laptops are so last century ==&lt;br /&gt;
&lt;br /&gt;
 * Uche Enwesi, uenwesi@umd.edu, University of Maryland Libraries&lt;br /&gt;
 * Francis Kayiwa, fkayiwa@umd.edu, University of Maryland Libraries&lt;br /&gt;
&lt;br /&gt;
At Univ. Maryland we are in the very early stages of looking into allowing our student body get their hands on a drone. Yes that's right we will let students take out a drone for n amount of hours to work on projects of their choosing. The talk will talk about the logistics of getting a program of this sort from concept to &amp;quot;Is the drone available?&amp;quot;. If people sign waivers we will also promise not to crash the drone into code4lib attendees.&lt;br /&gt;
&lt;br /&gt;
== Got Git? Getting More Out of Your GitHub Repositories ==&lt;br /&gt;
&lt;br /&gt;
 * Terry Brady, twb27@georgetown.edu, Georgetown University Library&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss how librarians, developers, and system administrators at Georgetown University are maximizing their use of the public and private GitHub repositories. &lt;br /&gt;
&lt;br /&gt;
In additional to all of the great benefits of using Git for code management, the GitHub interface provides a powerful set of tools to showcase a project and to keep your users informed of developments to your project.  These tools can assist with marketing and outreach - turning your code repository into a focus of conversation!&lt;br /&gt;
&lt;br /&gt;
* [http://georgetown-university-libraries.github.io/File-Analyzer/ Style-able Project Pages]&lt;br /&gt;
* [https://github.com/Georgetown-University-Libraries/File-Analyzer/wiki Project Wikis]&lt;br /&gt;
* [https://github.com/Georgetown-University-Libraries/Georgetown-University-Libraries-Code/releases Project Release Notes/Portfolios]&lt;br /&gt;
* [https://rawgit.com/Georgetown-University-Libraries/Georgetown-University-Libraries-Code/master/samples/GoogleSpreadsheetFilter.html Web Resources That Can Be Directly Requested]&lt;br /&gt;
* Gists for code sharing&lt;br /&gt;
* Private Repositories and Organizational Groups&lt;br /&gt;
* Pull Request Conversation Tracking&lt;br /&gt;
* Customized Issue management&lt;br /&gt;
&lt;br /&gt;
== Quick Wins for Every Department in the Library - File Analyzer! ==&lt;br /&gt;
&lt;br /&gt;
 * Terry Brady, twb27@georgetown.edu, Georgetown University Library&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Library has customized workflows for nearly every department in our library with a single code base.&lt;br /&gt;
* Analyzing Marc Records for the Cataloging department&lt;br /&gt;
* Transferring ILS invoices for the University Account System for the Acquisitions department &lt;br /&gt;
* Delivering patron fines to the Bursar’s office for the Access Service department&lt;br /&gt;
* Summarizing student worker timesheet data for the Finance department&lt;br /&gt;
* Validating COUNTER compliant reports for the Electronic Resources department&lt;br /&gt;
* Generating ingest packages for the Digital Services department&lt;br /&gt;
* Validating checksums for the Preservation department&lt;br /&gt;
&lt;br /&gt;
Learn how you can customize the [http://georgetown-university-libraries.github.io/File-Analyzer/ File Analyzer] to become a hero in your library!&lt;br /&gt;
&lt;br /&gt;
==The Geospatial World is Moving from Maps *on* the Web to Maps *of* the web. Libraries can too==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Copystar|Mita Williams]], mita@uwindsor.ca, User Experience Librarian, University of Windsor&lt;br /&gt;
&lt;br /&gt;
The transition from paper maps to digital ones changed much more than the maps themselves; it changed the very foundation of how we work and how we find each other. Now maps are transforming again.  The Geospatial World is moving from GIS systems that are institutionally-focused, expensive, feature-burdened, and binds data into a complicated and demanding user-hostile interface. From this transition from digital to web-based digital geospatial tools has come growth and development in new forms of map-based investigative journalism, activism, scholarship, and business ventures. This talk will highlight the conditions and strategies that made these changes possible as a means to draw a path by which librarians through our own work may follow, dragons notwithstanding. &lt;br /&gt;
&lt;br /&gt;
== Building Your Own Federated Search ==&lt;br /&gt;
&lt;br /&gt;
* Rich Trott, Richard.Trott@ucsf.edu, UC San Francisco&lt;br /&gt;
&lt;br /&gt;
Advances in modern browsers have created some interesting possibilities for federated search. This presentation will cover common techniques and pitfalls in building a federated search. We will discuss what principles guided our decisions when implementing our own federated search. We will show tools we've built and our findings from building and using experimental prototypes.&lt;br /&gt;
&lt;br /&gt;
Your higher education institution likely offers dozens of online resources for educators, students, researchers, and the public. And each of these online resources likely has its own search tool. But users can't be expected to search in dozens of different interfaces to find what they're looking for. A typical solution for this issue is federated search. &lt;br /&gt;
&lt;br /&gt;
==  Indexing Linked Data with LDPath ==&lt;br /&gt;
&lt;br /&gt;
* Chris Beer, cabeer@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
LDPath [1] is a simple query language for indexing linked open data, with support for caching, content negotiation, and integration with non-RDF endpoints. This talk will demonstrate the features and potential of the language and framework to index a resource with links into id.loc.gov, viaf.org, geonames.org, etc to build an application-ready document.&lt;br /&gt;
&lt;br /&gt;
[1] http://marmotta.apache.org/ldpath/language.html&lt;br /&gt;
&lt;br /&gt;
== Show Me the Money: Integrating an LMS with Payment Providers ==&lt;br /&gt;
 &lt;br /&gt;
* Josh Weisman,  Josh.Weisman@exlibrisgroup.com, Development Director-Resources Management, Ex Libris Group&lt;br /&gt;
&lt;br /&gt;
In order to provide an easy and convenient way for patrons to pay fines, we are exploring ways to integrate the library management system with online payment providers such as PayPal. With many LMS systems being designed and developed for the cloud, we should be able to provide the frictionless user experience our patrons have come to expect from online transactions. In this session we'll discuss strategies for integration and review a sample application which uses REST APIs from a library management system to integrate with PayPal.&lt;br /&gt;
&lt;br /&gt;
== Shibboleth Federated Authentication for Library Applications: ==&lt;br /&gt;
&lt;br /&gt;
* Scott Fisher, scott.fisher@ucop.edu, California Digital Library&lt;br /&gt;
* Ken Weiss, ken.weiss@ucop.edu, California Digital Library&lt;br /&gt;
&lt;br /&gt;
Shibboleth is the most widely-used method to provide single-sign-on authentication to academic applications where users come from many different institutions. Shibboleth, the InCommon education and research trust framework, and the SAML protocol comprise a very powerful - but very complicated - solution to this very complicated problem. Scott and Ken have implemented Shibboleth for multiple library applications. They will share their understanding of the good, the bad, and the underlying spaghetti that makes it all work. Ken will discuss some of the technical aspects of the solution, touching on optimal and non-optimal use cases, administrative challenges, and authorization concerns. Scott will describe the implementation pattern for multi-institution single-sign-on that the California Digital Library has evolved, using the recently released Dash application (http://dash.cdlib.org) as an example.&lt;br /&gt;
&lt;br /&gt;
==Scientific Data: A Needs Assessment Journey==&lt;br /&gt;
 &lt;br /&gt;
*[[User:VickySteeves| Vicky Steeves]], vsteeves@amnh.org, American Museum of Natural History&lt;br /&gt;
&lt;br /&gt;
While surveying digital research and collections data in the research science divisions at the American Museum of Natural History in NYC (as a part of my [http://ndsr.nycdigital.org/ National Digital Stewardship Residency] project), I have come across the big data hogs (genome sequencing and CT scanning) and the little pieces of data (images, publications), all equally important to not only scientific discovery, but as nodes in the history of science. &lt;br /&gt;
&lt;br /&gt;
In this session, I will discuss the development of my needs assessment surveys for scientific datasets and the interview process with Museum curators and researchers as background, seguing into an explanation of the results. I will then combine my findings into preliminary selection criteria to choose tools for digital preservation and management unique to scientific datasets. This will brooke a discussion on emerging standards, tools, and technologies in big data, specific to research science. &lt;br /&gt;
&lt;br /&gt;
I will conclude with preliminary findings on emerging technology that can be used to answer concerns surrounding the management and digital preservation of these data. I am hoping the Q&amp;amp;A session can be used to both answer questions about my project, and function as a way for you (the larger tech-savy library community)  to discuss the tools I’ve touched on in this talk. &lt;br /&gt;
&lt;br /&gt;
== Feminist Human Computer Interaction (HCI) in Library Software ==&lt;br /&gt;
 &lt;br /&gt;
* Bess Sadler,  bess@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Libraries are not neutral repositories of knowledge. Library classification systems and search technologies tend to reflect the inequalities, biases, ethnocentrism, and power imbalances of the societies in which they are built [1]. How might we better resist these tendencies in the library software we create? This talk will examine some qualities of feminist HCI (pluralism, self-disclosure, participation, ecology, advocacy, and embodiment) [2] through the lens of library software. &lt;br /&gt;
&lt;br /&gt;
[1] Olson, Hope A. (2002). The Power to Name: Locating the Limits of Subject Representation in Libraries. Dordrecht, The Netherlands: Kluwer Academic Publishers.&lt;br /&gt;
&lt;br /&gt;
[2] Bardzell, Shaowen. Feminist HCI: Taking Stock and Outlining an Agenda for Design. CHI 2010: HCI For All. http://dmrussell.net/CHI2010/docs/p1301.pdf&lt;br /&gt;
&lt;br /&gt;
== Heiðrún: DPLA's Metadata Harvesting, Mapping and Enhancement System ==&lt;br /&gt;
&lt;br /&gt;
* Audrey Altman, audrey at dp.la, Digital Public Library of America&lt;br /&gt;
* Gretchen Gueguen, gretchen at dp.la, Digital Public Library of America&lt;br /&gt;
* Mark Breedlove, mb at dp.la, Digital Public Library of America&lt;br /&gt;
&lt;br /&gt;
The Digital Public Library of America aggregates metadata for over 8 million objects from more than 24 direct partners, or Hubs, using its Metadata Application Profile (MAP), an RDF metadata application profile based on the Europeana Data Model. After working with the initial system for harvesting, mapping and enhancing our Hub’s metadata for a year, we realized that it was inadequate for working with data at this scale. There were architectural issues; it was opaque to non-developer and partner staff; there were inadequate tools for quality assurance and analysis; and the system was unaware that it was working with RDF data. As the network of Hubs expanded and we ingested more metadata, it became harder and harder to know when or why a harvest, a mapping task, or an enrichment went wrong because the tools for quality assurance were largely inadequate. &lt;br /&gt;
&lt;br /&gt;
The DPLA Content and Technology teams decided to develop a new system from the ground up to address those problems. Development of Heidrun, the internal version of the new system, started in October 2014. Heidrun’s goals are to make it easier for us to harvest and map metadata from various sources and in variety of schemas to the DPLA MAP, to better enrich that metadata using external data sources, and to actively involve our partners in the ingestion process through access to better QA tools. Heidrun and its componentry are built on Ruby on Rails, Blacklight, and ActiveTriples. Our presentation will give some background on our design principles and processes used during development, the architecture of the system, and its functionality. We plan to release a version of Heidrun and its components as a generalized metadata aggregation system for use by DPLA Hubs and others working to aggregate cultural heritage metadata.&lt;br /&gt;
&lt;br /&gt;
== OS or GTFO: Program or Perish ==&lt;br /&gt;
*Tessa Fallon, tessa.fallon@gmail.com&lt;br /&gt;
&lt;br /&gt;
Description TBD&lt;br /&gt;
&lt;br /&gt;
== Creating Dynamic— and Cheap!— Digital Displays with HTML 5 Authoring Software ==&lt;br /&gt;
* Chris Woodall, cmwoodall@salisbury.edu, Salisbury University Libraries&lt;br /&gt;
Would your library like to have large digital signage that displays dynamic information such as library hours, weather, room availability, and more? Have you looked into purchasing large digital signage, only to be turned off by the high price tag and lack of customization available with commercial solutions? Our library has developed a cheap and effective alternative to these systems using HTML 5 authoring software, a large TV, and freely-available APIs from Google, Springshare, and others. At this session, you’ll learn about the system that we have in place for displaying dynamic and easily-updatable information on our library’s large digital display, and how you can easily create something similar for your library.&lt;br /&gt;
&lt;br /&gt;
== REPOX: Metadata Blender ==&lt;br /&gt;
 &lt;br /&gt;
* John Mignault, jmignault@metro.org, Empire State Digital Network&lt;br /&gt;
&lt;br /&gt;
With the growth in the number of hubs providing metadata to the Digital Public Library of America, many of them are using REPOX, a tool originally created for the Europeana project, to aggregate disparate metadata feeds and transform them into formats suitable for ingest into DPLA. The Empire State Digital Network, the forthcoming DPLA service hub for NY state, is using it to prepare for our first ingest into DPLA in early 2015.  We'll take a look at REPOX and its capabilities and how it can be useful for ingesting and transforming metadata, and also discuss some things we've learned in massaging widely varied metadata feeds.&lt;br /&gt;
&lt;br /&gt;
== Beyond Open Source ==&lt;br /&gt;
&lt;br /&gt;
* Jason Casden, jmcasden@ncsu.edu, NCSU Libraries&lt;br /&gt;
* Bret Davidson, bddavids@ncsu.edu, NCSU Libraries&lt;br /&gt;
&lt;br /&gt;
The Code4Lib community has produced an increasingly impressive collection of open source software over the last decade, but much of this creative work remains out of reach for large portions of the library community. Do the relatively privileged institutions represented by a majority of Code4Lib participants have a professional responsibility to support the adoption of their innovations?&lt;br /&gt;
&lt;br /&gt;
Drawing from old and new software packaging and distribution approaches (from freeware to Docker), we will propose extending the open source software values of collaboration and transparency to include the wide and affordable distribution of software. We believe this will not only simplify the process of sharing our applications within the Code4Lib community, but also make it possible for less well resourced institutions to actually use our software. We will identify areas of need, present our experiences with the users of our own open source projects, discuss our attempts to go beyond open source, and make an argument for the internal value of supporting and encouraging a vibrant library ecosystem.&lt;br /&gt;
&lt;br /&gt;
[[Category:Code4Lib2015]] &lt;br /&gt;
[[Category:Talk Proposals]]&lt;br /&gt;
&lt;br /&gt;
== Making It Work: Problem Solving Using Open Source at a Small Academic Library ==&lt;br /&gt;
 &lt;br /&gt;
* Adam Strohm, astrohm@iit.edu, Illinois Institute of Technology&lt;br /&gt;
* Max King, mking9@iit.edu, Illinois Institute of Technology&lt;br /&gt;
&lt;br /&gt;
The Illinois Institute of Technology campus was added to the National Register of Historic Places in 2005, and contains a building, Mies van der Rohe's S.R. Crown Hall, that was named a National Historic Landmark in 2001. Creating a digital resource that can adequately showcase the campus and its architecture is challenge enough in and of itself, but doing so as a two-person team of relative newcomers, at a university library without dedicated programmers on staff, ups the ante considerably.&lt;br /&gt;
The challenges of technical know-how, staff time, and funding are nothing new to anyone working on digital projects at a university library, and are amplified when doing so at a smaller institution. This talk covers the conception, development, and design of the campus map site that was built, concentrating on the problem-solving strategies developed to cope with limited technical and financial resources.&lt;br /&gt;
We'll talk about our approach to development with Open Source software, including Omeka, along with the Neatline and Simile Timeline plugins. We'll also discuss the juggling act of designing for mobile mapping functionality without sacrificing desktop design, weighing the costs of increased functionality versus our ability to time-effectively include that functionality, and the challenge of building a site that could be developed iteratively, with an eye towards future enhancement and sustainability. Finally, we’ll provide recommendations for other librarians at smaller institutions for their own efforts at digital development.&lt;br /&gt;
&lt;br /&gt;
== Recording Digitization History: Metadata Options for the Process History of Audiovisual Materials ==&lt;br /&gt;
 &lt;br /&gt;
* Peggy Griesinger, peggy_griesinger@moma.org, Museum of Modern Art&lt;br /&gt;
&lt;br /&gt;
The Museum of Modern Art has amassed a large collection of audiovisual materials over its many decades of existence. In order to preserve these materials, much of the audiovisual collection has been digitized. This is a complex process involving numerous steps and devices, and the methods used for digitization can have an effect on the quality of the file that is preserved. Therefore, knowing exactly how something was digitized is critical for future stewards of these objects to be able to properly care for and preserve them. However, detailed technical information about the processes involved in the digitization of audiovisual materials is not defined explicitly in most metadata schemas used for audiovisual materials. In order to record process history using existing metadata standards, some level of creativity is required to allow existing standards to express this information.&lt;br /&gt;
&lt;br /&gt;
This talk will detail different metadata standards, including PBCore, PREMIS, and reVTMD, that can be implemented as methods of recording this information. Specifically, the talk will examine efforts to integrate this metadata into the Museum of Modern Art’s new digital repository, the DRMC. This talk will provide background on the DRMC as well as MoMA’s specific institutional needs for process history metadata, then discuss different metadata implementations we have considered to document process history.&lt;br /&gt;
&lt;br /&gt;
== Pig Kisses Elephant: Building Research Data Services for Web Archives ==&lt;br /&gt;
 &lt;br /&gt;
* Jefferson Bailey,  jefferson@archive.org, Internet Archive&lt;br /&gt;
* Vinay Goel, vinay@archive.org, Internet Archive&lt;br /&gt;
&lt;br /&gt;
More and more libraries and archives are creating web archiving programs.  For both new and established programs, these archives can consist of hundreds of thousands, if not millions, of born-digital resources within a single collection; as such, they are ideally suited for large-scale computational study and analysis. Yet current access methods for web archives consist largely of browsing the archived web in the same manner as browsing the live web and the size of these collections and complexity of the WARC format can make aggregate analysis difficult. This talk will describe a project to create new ways for users and researchers to access and study web archives by offering extracted and post-processed datasets derived from web collections. Working with the 325+ institutions and their 2600+ collections within the Archive-It service, the Internet Archive is building methods to deliver a variety of datasets culled from collections of web content, including extracted metadata packaged in JSON, longitudinal link graph data, named entities, and other types of data. The talk will cover the technical details of building dataset production pipelines with Apache Pig, Hadoop, and tools like Stanford NER, the programmatic aspects of building data services for archives and researchers, and ongoing work to create new ways to access and study web archives.&lt;br /&gt;
&lt;br /&gt;
== Awesome Pi, LOL! ==&lt;br /&gt;
&lt;br /&gt;
* Matt Connolly, mconnolly@cornell.edu, Cornell University Library&lt;br /&gt;
* Jennifer Colt, jrc88@cornell.edu, Cornell University Library&lt;br /&gt;
&lt;br /&gt;
Inspired by Harvard Library Lab’s “Awesome Box” project, Cornell’s Library Outside the Library (LOL) group is piloting a more automated approach to letting our users tell us which materials they find particularly stunning. Armed with a Raspberry Pi, a barcode scanner, and some bits of kit that flash and glow, we have ventured into the foreign world of hardware development. This talk will discuss what it’s like for software developers and designers to get their hands dirty, how patrons are reacting to the Awesomizer, and LOL’s not-afraid-to-fail philosophy of experimentation.&lt;br /&gt;
&lt;br /&gt;
== You Gotta Keep 'em Separated: The Case for &amp;quot;Bento Box&amp;quot; Discovery Interfaces ==&lt;br /&gt;
 &lt;br /&gt;
* Jason Thomale,  jason.thomale@unt.edu, University of North Texas Libraries&lt;br /&gt;
&lt;br /&gt;
I know, I know--proposing a talk about Resource Discovery is like, ''so'' 2010.&lt;br /&gt;
&lt;br /&gt;
The thing is, practically all of us--in academic libraries at least--have a similar set up for discovery, with just a few variations, and so talking about it still seems useful. Stop me if this sounds familiar. You've got a single search box on the library homepage as a starting point for discovery. And it's probably a tabbed affair, with an option for searching the catalog for books, an option for searching a discovery service for articles, an option for searching databases, and maybe a few others. Maybe you have an option to search everything at once--probably the default, if you have it. And, if you're a crazy hepcat, maybe you ''only'' have your one search that searches everything, with no tabs.&lt;br /&gt;
&lt;br /&gt;
Now, the question is, for your &amp;quot;everything&amp;quot; search, are you doing a combined list of results, or are you doing it bento-box style, with a short results list from each category displayed in its own compartment?&lt;br /&gt;
&lt;br /&gt;
At UNT, we've been holding off on implementing an &amp;quot;everything&amp;quot; search, for various reasons. One reason is that the evidence for either style hasn't been very clear. There's this persistent paradox that we just can't reconcile: users tell us, through word and action, that they prefer searching Google, yet, libraries aren't Google, and there are valid design reasons why we shouldn't try to oversimplify our discovery interfaces to be like Google. And there's user data that supports both sides.&lt;br /&gt;
&lt;br /&gt;
Holding off on making this decision has granted us 2 years of data on how people use our tabbed search interface that does ''not'' include an &amp;quot;everything&amp;quot; search. Recently I conducted a thorough analysis of this data--specifically the usage and query data for our catalog and discovery system (Summon). And I think it helps make the case for a bento box style discovery interface. To be clear, it isn't exactly the smoking gun that I was hoping for, but the picture it paints I think is telling. At the very least, it points away from a combined-results approach.&lt;br /&gt;
&lt;br /&gt;
I'm proposing a talk discussing the data we've collected, the trends we've seen, and what I think it all means--plus other reasons that we're jumping on the &amp;quot;bento box&amp;quot; discovery bandwagon and why I think &amp;quot;bento box&amp;quot; is at this point the path that least sells our souls.&lt;br /&gt;
&lt;br /&gt;
== Don’t know about you, but I’m feeling like SHA-2!: Checksumming with Taylor Swift ==&lt;br /&gt;
 &lt;br /&gt;
* Ashley Blewer!, ashley.blewer@gmail.com&lt;br /&gt;
&lt;br /&gt;
Checksum technology is used all over the place, from git commits to authenticating Linux packages. It is most commonly used in the digital preservation field to monitor materials in storage for changes that will occur over time or used in the transmission of files during duplication. But do you even checksum, bro? I want this talk to move checksums from a position of mysterious macho jargon to something everyone can understand and want to use. I think a lot of people have heard of checksum but don’t know where to begin when it comes to actually using it at their institution. And cryptography is hella intimidating! This talk will cover what checksums are, how they can be integrated into a library or archival workflow, protecting collections requiring additional levels of security, algorithms used to verify file fixity and how they are different, and other aspects of cryptographic technology. Oh, and please note that all points in this talk will be emphasized or lightly performed through Taylor Swift lyrics. Seriously, this talk will consist of at least 50% Taylor Swift. Can you, like, even?&lt;br /&gt;
&lt;br /&gt;
== Level Up Your Coding with Code Club (yes, you can talk about it) ==&lt;br /&gt;
&lt;br /&gt;
* Coral Sheldon-Hess, coral@sheldon-hess.org&lt;br /&gt;
&lt;br /&gt;
Reading code is a necessary part of becoming a better developer. It gives you more experience and more insight into How Things Are (or Aren't) Done; it builds your intuition about how to solve problems with code; and it increases your confidence that you, too, can tackle whatever technological problems you're facing.&lt;br /&gt;
&lt;br /&gt;
But you don't have to read code alone! (Which is good. It's really not fun to read code alone.) &lt;br /&gt;
&lt;br /&gt;
In late 2014, a group of librarians formed two Code Clubs, inspired by [http://bloggytoons.com/code-club/ this talk by Saron] (of Bloggytoons fame). I'd like to tell you about how we've structured our Code Clubs, what has gone well, what we've learned, and what you need to do to form your own Code Club. I'll share a list of the codebases we've looked at, too, to help you get your own Code Club off the ground! &lt;br /&gt;
&lt;br /&gt;
== The Growth of a Programmer ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:jgo | Joshua Gomez]], Getty Research Institute, jgomez@getty.edu&lt;br /&gt;
&lt;br /&gt;
Just like other creative endeavors, software developers can experience periods of great productivity or find themselves in a rut. After contemplating the alternating periods in my own career I've noticed several factors that have effected my own professional growth and happiness, including: mentorship, structure, community, teamwork, environment, formal education, etc. Not all of the factors need to be present at all times; but some mixture of them is critical for continued growth. In this talk, I will articulate these factors, discuss how they can effect a developer's career, and how they can be sought out when missing. This talk is aimed at both new developers looking to strike their own path as well as the veterans that lead or mentor them.&lt;br /&gt;
&lt;br /&gt;
== Developing a Fedora 4.0 Content Model for Disk Images ==&lt;br /&gt;
&lt;br /&gt;
* Matthew Farrell, matthew.j.farrell@duke.edu, Duke University Libraries&lt;br /&gt;
* Alexandra Chassanoff, achass@email.unc.edu, BitCurator Access Project Manager&lt;br /&gt;
&lt;br /&gt;
As the acquisition of born-digital materials grows, institutions are seeking methods to facilitate easy ingest into their repositories and provide access to disk images and files derived or extracted from disk images. In this session, we describe our development of a Fedora 4.0 Content model for disk images, including acceptable image file formats and the rationale behind those choices.  We will also discuss efforts to integrate the disk image content model into the BitCurator Access environment. Unlike generalized, format-agnostic content models which might treat the disk image as a generic bitstream, a content model designed for disk images enables expression of relationships among associated content in the collection such as files extracted from images and other born-digital and digitized material associated with the same creator.  It also enables capture of file-system attributes such as file paths, timestamps, whether files are allocated/deleted, etc.  Further, a disk image content model suggests further steps repositories can take in order to transform and re-use associated metadata generated during the creation and forensic analysis of the disk image.&lt;br /&gt;
&lt;br /&gt;
== Data acquisition and publishing tools in R ==&lt;br /&gt;
&lt;br /&gt;
* Scott Chamberlain,  scott@ropensci.org, rOpenSci/UC Berkeley - first-time presenter&lt;br /&gt;
&lt;br /&gt;
R is an open source programming environment that is widely used among researchers in many fields. R is powerful because it's free, increasingly robust, and facilitates reproducible research, an increasingly sought after goal in academia. Although tools for data manipulation/visualization/analysis are well developed in R, data acquisition and publishing tools are not. rOpenSci is a collaborative effort to create the tools necessary to complete the reproducible research workflow. This presentation discusses the need for these tools, including examples, including interacting with the repositories Mendeley, Dryad, DataONE, and Figshare. In addition, we are building tools for searching scholarly metadata and acuiring full text of open access articles in a standarized way across metadata providers (e.g., Crossref, DataCite, DPLA) and publishers (e.g., PLOS, PeerJ, BMC, Pubmed). Last, we are building out tools for data reading and writing in Ecologial Metadata Language (EML).&lt;br /&gt;
&lt;br /&gt;
== SPLUNK: Log File Analysis ==&lt;br /&gt;
&lt;br /&gt;
* Jim LeFager, jlefager@depaul.edu, DePaul University Library&lt;br /&gt;
DePaul University Library recently took over monitoring and maintaining of the library EZproxy servers this past year and using Splunk, a machine data analysis tool, we are able to gather information and statistics on our electronic resource usage in addition to monitoring the servers. Splunk is a tool that can collect, analyze, and visualize log files and other machine data in real time and this has allowed for gathering realtime usage statistics for our electronic resources allowing us to filter by multiple facets including IP Range, Group Membership (student, faculty), so that we can see who is accessing our resources and from where. Splunk allows our library to query our data and create rich custom dashboards as well as create alerts that can be triggered when certain conditions are met, such as error codes, which can send an email alert to a group of users. We will be leveraging Splunk to monitor all library web applications going forward. This talk will review setting up Splunk and best practices in using the available features and customizations available including creating queries, alerts, and custom dashboards.  &lt;br /&gt;
&lt;br /&gt;
== Your code does not exist in a vacuum ==&lt;br /&gt;
* Becky Yoose, yoosebec at grinnell dot edu, Grinnell College (Done a lightning talk, MC duties, but have not presented a prepared talk)&lt;br /&gt;
&lt;br /&gt;
“If you have something to say, then say it in code…” - Sebastian Hammer, code4lib 2009&lt;br /&gt;
&lt;br /&gt;
In its 10 year run, code4lib has covered the spectrum of libtech development, from search to repositories to interfaces. However, during this time there has been little discussion about this one little fact about development - code does not exist in a vacuum. &lt;br /&gt;
&lt;br /&gt;
Like the comment above, code has something to say. A person’s or organization’s culture and beliefs influences code in all steps of the development cycle. What development method you use, tools, programming languages, licenses - everything is interconnected with and influenced by the philosophies, economics, social structures, and cultural beliefs of the developer and their organization/community.&lt;br /&gt;
&lt;br /&gt;
This talk will discuss these interconnections and influences when one develops code for libraries, focusing on several development practices (such as “Fail Fast, Fail Often” and Agile)   and licensing choices (such as open source) that libtech has either tried to model or incorporate into mainstream libtech practices. It’ll only scratch the surface of the many influences present in libtech development, but it will give folks a starting point to further investigate these connections at their own organizations and as a community as a whole.&lt;br /&gt;
&lt;br /&gt;
tl;dr - this will be a messy theoretical talk about technology and libraries. No shiny code slides, no live demos. You might come out of this talk feeling uncomfortable. Your code does not exist in a vacuum. Then again, you don’t exist in a vacuum either.&lt;br /&gt;
&lt;br /&gt;
== The Metadata Hopper: Mapping and Merging Metadata Standards for Simple, User-Friendly Access ==&lt;br /&gt;
&lt;br /&gt;
* Tracy Seneca, tjseneca@uic.edu, University of Illinois at Chicago&lt;br /&gt;
* Esther Verreau: verreau1@uic.edu, University of Illinois at Chicago&lt;br /&gt;
&lt;br /&gt;
The Chicago Collections Consortium: 15 institutions and growing!  8 distinct EAD standards! At least 3 permutations of MARC, and we lost count of the varieties of custom CONTENTdm image collections.  Not to mention the 14,730 unique subject terms, nearly all of which lead our poor end-users to exactly one organization's content. &lt;br /&gt;
&lt;br /&gt;
All large content aggregation projects have faced this challenge, and there are a few emerging tools to help us wrangle disparate metadata into new contexts.  The Metadata Hopper is one such tool. The Metadata Hopper enables archivists to map their local metadata standards to standardized deposit records, and tags those materials using a shared vocabulary, integrating them into a user-friendly portal without disrupting local practices. In last year's Code4Lib lightning talk we described the challenges that the Chicago Collections Consortium faces in creating shared, in-depth access to archival and digital collections about Chicago history and culture across CCC member organizations. This year, thanks to the Andrew W. Mellon Foundation, we have a working Django application to demonstrate.  In this talk we'll discuss the design that enables multiple layers of flexibility, from the ability to accept a variety of metadata standards to designing for an open source audience.&lt;br /&gt;
&lt;br /&gt;
http://chicagocollectionsconsortium.org&lt;br /&gt;
&lt;br /&gt;
== Programmers are not projects: lessons learned from managing humans ==&lt;br /&gt;
&lt;br /&gt;
* Erin White, erwhite@vcu.edu, Virginia Commonwealth University - first-time presenter&lt;br /&gt;
&lt;br /&gt;
Managing projects is one thing, but managing people is another. Whether we’re hired as managers or grow “organically” into management roles, sometimes technical people end up leading technical teams (gasp!). I’ll talk about lessons I’ve learned about hiring, retaining, and working long-term and day-to-day with highly tech-competent humans. I’ll also talk about navigating the politics of libraryland, juggling different types of projects, and working with constrained budgets to make good things and keep talented people engaged.&lt;br /&gt;
&lt;br /&gt;
== Practical Strategies for Picking Low-Hanging Fruits to Improve Your Library's Web Usability and UX ==&lt;br /&gt;
&lt;br /&gt;
* Bohyun Kim, bkim@hshsl.umaryland.edu, University of Maryland, Baltimore&lt;br /&gt;
&lt;br /&gt;
Have you ever tried to fix an obvious (to you at least!) problem in Web usability or UX (user experience) only to face strong resistance from the library staff? Are you a strong advocate for making library resources, systems, services, and space as usable as possible, but do you often find yourself struggling to get the point across and/or obtain the crucial buy-in from colleagues and administrators? &lt;br /&gt;
&lt;br /&gt;
There is no shortage of Web usability and UX guidelines. But applying them to a library and implementing desired changes often involve a long and slow process. To tackle this issue, this talk will focus on how to utilize the 'expert review' process (aka 'heuristic evaluation') as a preliminary or even preparatory step before embarking on more time-and-labor-intensive usability testing and user research. Several examples from  simple fixes to more nuanced usability and UX issues in libraries will be discussed to your heart's content. The goal of this talk is to provide practical strategies for picking as many low-hanging fruits as possible to make a real (albeit small) difference to your library's Web usability and UX effectively and efficiently.&lt;br /&gt;
&lt;br /&gt;
== A Semantic Makeover for CMS Data ==&lt;br /&gt;
&lt;br /&gt;
* Bill Levay, wjlevay@gmail.com, Linked Jazz Project&lt;br /&gt;
&lt;br /&gt;
How can we take semi-structured but messy metadata from a repository like CONTENTdm and transform it into rich linked data? Working with metadata from Tulane’s Hogan Jazz Archive Photography Collection, the Linked Jazz Project used Open Refine and Python scripts to tease out proper names, match them with name authority URIs, and specify FOAF relationships between musicians who appear together in photographs. Additional RDF triples were created for any dates associated with the photos, and for those images with place information we employed GeoNames URIs. Historical images and data that were siloed can now interact with other datasets, like Linked Jazz’s rich set of names and personal relationships, and can be visualized [link to come] or otherwise presented on the web in any number of ways. I have not previously presented at a Code4Lib conference.&lt;br /&gt;
&lt;br /&gt;
== Taking User Experience (UX) to new heights ==&lt;br /&gt;
 &lt;br /&gt;
* Kayne Richens, kayne.richens@deakin.edu.au, Deakin University&lt;br /&gt;
&lt;br /&gt;
User Experience, or &amp;quot;UX&amp;quot;, is for more than just websites. At Deakin University Library we're exploring ways to improve the user experience inside our campus library spaces, by putting new technologies front and centre in the overall experience for our students. How are we doing this? We’re collaborating with the University's IT department and exploring the following Library-changing opportunities:&lt;br /&gt;
&lt;br /&gt;
- Augmented Reality for Way-finding: We’re tackling that infamous thing that all Libraries can't get right – way-finding. We're enhancing library tour information and way-finding experiences by introducing augmented reality solutions.&lt;br /&gt;
 &lt;br /&gt;
- Heat mapping the library with wi-fi: We’re using our existing wi-fi infrastructure to present &amp;quot;heat maps&amp;quot; of library space utilisation, allowing our users to easily locate the space that best suits their needs, whether it be busy spaces to collaborate, or quiet spaces to study. And by overlaying computer usage and group study room bookings, users can quickly locate the space they need.&lt;br /&gt;
 &lt;br /&gt;
- Video chat library service: We’re piloting video-conferencing facilities in our group study rooms and spaces, connecting users and librarians and other professionals.&lt;br /&gt;
         &lt;br /&gt;
This talk will look at how these different technologies will be brought together to provide improved user experiences, as well some of the evidence and reasons that helped us to identify our needs, so you can too.&lt;br /&gt;
&lt;br /&gt;
==How to Hack it as a Working Parent: or, Should Your Face be Bathed in the Blue Glow of a Phone at 2 AM?==&lt;br /&gt;
&lt;br /&gt;
*Margaret Heller, Loyola University Chicago, mheller1@luc.edu&lt;br /&gt;
*Christina Salazar, California State University Channel Islands, christina.salazar@csuci.edu&lt;br /&gt;
*May Yan, Ryerson University, may.yan@ryerson.ca&lt;br /&gt;
&lt;br /&gt;
Modern technology has made it easier than ever for parents employed in technical environments to keep up with work at all hours and in all locations. This makes it possible to work a flexible schedule, but also may lead to problems with work/life balance and furthering unreasonable expectations about working hours. Add to that shifting gender roles and limited paid parental leave in the United States and you have potential for burnout and a certainty for anxiety. It raises the additioal question of whether the “always connected” mindset puts up a barrier to some populations who otherwise might be better represented in open source and library technology communities. &lt;br /&gt;
&lt;br /&gt;
This presentation will address tools that are useful for working parents in technical library positions, and share some lessons learned about using these tools while maintaining a reasonable work/life balance. We will consider a question that Karen Coyle raised back in 1996: &lt;br /&gt;
“What if the thousands of hours of graveyard shift amateur hacking wasn't really the best way to get the job done? That would be unthinkable.” &lt;br /&gt;
&lt;br /&gt;
For those who are able to take an extended parental leave, we will present strategies for minimizing the impact to your career and your employer. For those (particularly in the United States) who are only able to take a short leave will require different strategies. Despite different levels of preparation, all are useful exercises in succession planning and making a stronger workplace and future ability to work a flexible schedule through reviewing workloads, cross-training personnel, hiring contract replacements, and creative divisions of labor. Such preparation makes work better for everyone, kids or no kids or caretakers of any kind.&lt;br /&gt;
&lt;br /&gt;
==Making your digital objects embeddable around the web==&lt;br /&gt;
 &lt;br /&gt;
* Jessie Keck, jkeck@stanford.edu, Stanford University Libraries&lt;br /&gt;
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
With more and more content from our digital repositories making their way into our discovery environments we quickly realize that we’re repeatedly re-inventing the wheel when it comes to creating “Viewers” for these digital objects.  With various different types of viewers necessary (books, images, audio, video, geospatial data, etc) the burden of getting these viewers into various environments (topic guides, blogs, catalogs, etc) becomes exponential.&lt;br /&gt;
&lt;br /&gt;
In this talk we’ll discuss how Stanford University Libraries implemented an oEmbed service to create an extensible viewer framework for all of its digital content. Using this service we’ve been able to easily integrate viewers into various discovery applications as well as make it easy for end users who discover our objects to easily embed customized versions into their own websites and blogs.&lt;br /&gt;
&lt;br /&gt;
==So you want to make your geospatial data discoverable==&lt;br /&gt;
 &lt;br /&gt;
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Finding data for research or coursework can be one of the most time intensive tasks for a scholar or student. We introduce GeoBlacklight, an open source, multi-institutional software project focused on solving these common challenges at institutions across the world. GeoBlacklight prioritizes user experience, integrates with many GIS tools, and streamlines the use and organization of geospatial data. This talk will provide an introduction to the software, demonstrate current functionality, and provide a road map for future work.&lt;br /&gt;
&lt;br /&gt;
== Clueless-Driven Development: How I learned to migrate to Fedora 4 ==&lt;br /&gt;
&lt;br /&gt;
* Adam Wead, awead@psu.edu, Penn State University&lt;br /&gt;
&lt;br /&gt;
Recently I was tasked with migrating the content from our Fedora3 repository to the new Fedora4 repository architecture.&lt;br /&gt;
Despite a wealth of community support, I had no idea how to approach, or even begin to solve this problem. I knew I&lt;br /&gt;
wanted to follow best practices and use test-driven  development to build my solution, but had no idea where to start.&lt;br /&gt;
Despite this initial setback, I was able to start writing tests with only a  vague understanding of the problem. As my&lt;br /&gt;
tests exposed where my understanding of the problem was flawed, my code evolved, and within a week I had arrived  at a&lt;br /&gt;
working solution that exhibited all the hallmarks of good testing and software design.&lt;br /&gt;
&lt;br /&gt;
This talk recounts the process I went through from starting with practically nothing, to arriving at a working solution.&lt;br /&gt;
You can follow the rules of  test-driven development, but you can write tests in an expressive way to describe the&lt;br /&gt;
problem instead of just describing what the code should do. It was also essential to begin testing from an integration&lt;br /&gt;
viewpoint as opposed to a unit one, because at the outset the units were unknown and were later realized through further&lt;br /&gt;
development. For the presentation, I will be demonstrating using RSpec and Ruby. All the code examples will be related&lt;br /&gt;
to the Hydra software stack; however, I hope to show  that the processes at work will be applicable in any context.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Designing and Leading a Kick A** Tech Team ==&lt;br /&gt;
 &lt;br /&gt;
* Sibyl Schaefer,  sschaefer@rockarch.org, Rockefeller Archive Center&lt;br /&gt;
&lt;br /&gt;
New managers are often promoted without receiving management training, yet management is not something you just figure out. The experience of being expected to know how to manage, yet not being trained to do so often results in new managers feeling isolated and unsure how to move from making to managing. In this talk I’ll focus on my own managerial experience of designing and leading an archival tech team in a small independent archives. Topics covered will include hiring, delegating, creating a team culture, and leading people whose specialized knowledge exceeds your own. The talk take-aways should be applicable to managers and employees at large and small institutions alike.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==American (Archives) Horror Story: LTO Failure and Data Loss ==&lt;br /&gt;
 &lt;br /&gt;
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH&lt;br /&gt;
* Casey Davis, casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH&lt;br /&gt;
&lt;br /&gt;
Here’s a story to send shivers down archival spines: when transferring video files off LTO for the American Archive project, WGBH got an initial failure rate of 57%.   After repeat tries, the rates improved; still, an unnervingly large percentage of files were never able to be transferred successfully.   Even more unnerving, going public with our horror story got a big response from other archives using LTO -- it seems like many institutions are having similarly scary results.   What are the real risks with LTO tape?  Are there steps that archives should be taking to better circumvent those risks?  This presentation will share information about LTO storage failures across archives world and discuss the process of investigating the problem at WGBH by testing different methods of data retrieval from LTO (direct and networked downloads, individual file retrieval and bulk data dump, use of LTO 4 and LTO 6 decks) and using checksum comparisons and file analysis and characterization tools such as ffprobe, mediainfo and exiftool to analyze failed files.  We'll also present whatever results we’ve managed to turn up by the time of Code4Lib!&lt;br /&gt;
&lt;br /&gt;
== PBCore in Action: Three Words, Not Two! ==&lt;br /&gt;
 &lt;br /&gt;
* Casey E. Davis,  casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH&lt;br /&gt;
* Andrew (Drew) Myers, andrew_myers@wgbh.org, Supervising Developer, WGBH&lt;br /&gt;
&lt;br /&gt;
In 2001, public media representatives developed the PBCore XML schema to establish a common language for managing metadata about their analog and digital audio and video. Since then, PBCore has been adopted by a number of organizations and archivists in the moving image archival community. The schema has also undergone a few revisions, but on more than one occasion it was left orphaned and with little to no support.&lt;br /&gt;
 &lt;br /&gt;
Times have changed. You may have heard the news that PBCore is back in action as part of the American Archive of Public Broadcasting initiative and via the Association of Moving Image Archivists (AMIA) PBCore Advisory Subcommittee. A group of archivists, public media stakeholders, and engaged users have come together to provide necessary, sustaining support for the standard and to see to its further development. &lt;br /&gt;
 &lt;br /&gt;
At this session, we'll discuss the scope and uses of PBCore in digital preservation and access, report on the progress and goals of the PBCore Advisory Subcommittee, and share how the group (by the time of the conference) will have transformed the XML schema into an RDF ontology, bringing PBCore into the second decade of the 21st century. #PBHardcore&lt;br /&gt;
&lt;br /&gt;
==Collaborating to Avert the Digital Graveyard==&lt;br /&gt;
&lt;br /&gt;
* Harish Nayak, hnayak@library.rochester.edu, University of Rochester Libraries &lt;br /&gt;
* Sean Morris, smorris@library.rochester.edu, University of Rochester Libraries &lt;br /&gt;
&lt;br /&gt;
In 1995, the Robbins Library at the University of Rochester created a digital collection of Arthurian texts, images, and bibliographies. Together with medieval scholars, we recently completed the redesign and development of an interface for this collection. Using FRBR concepts, we re-conceptualized organization and editing workflow from the ground up in a mobile-first Drupal-based project. &lt;br /&gt;
&lt;br /&gt;
In this talk we will describe the project as well as how we utilized the techniques of work practice study and user centered design to maintain engagement with reluctant stakeholders, nontechnical scholars, and VERY meticulous graduate students.  Neither of us have previously presented at a Code4Lib conference.&lt;br /&gt;
&lt;br /&gt;
==Docker? VMs? EC2? Yes! With Packer.io==&lt;br /&gt;
&lt;br /&gt;
* Kevin S. Clarke, ksclarke@gmail.com, Digital Library Programmer, UCLA&lt;br /&gt;
&lt;br /&gt;
There are a lot of exciting ways to deploy a software stack nowadays. Many of our library systems are fully virtualized. Docker is a compelling alternative, and there are also cloud options like Amazon's EC2. This talk will introduce Packer.io, a tool for creating identical machine images for multiple platforms (e.g., Docker, VMWare, VirtualBox, EC2, GCE, OpenStack, et al.) all from a single source configuration.  It works well with Ansible, Chef, Puppet, Salt, and plain old Bash scripts. And, it's designed to be scriptable so that builds can be automated. This presentation will show how easy it is to use Packer.io to bring up a set of related services like Fedora 4, Grinder (for stress testing), and Graphite (for charting metrics). As an added value, all the buzzwords in this proposal will be defined and explained!&lt;br /&gt;
&lt;br /&gt;
== Technology on your Wrist: Cross-platform Smartwatch Development for Libraries ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:sanderson|Steven Carl Anderson]], sanderson@bpl.org, [http://www.bpl.org Boston Public Library] (no previously accepted prepared talks but have done lightning talks in the past)&lt;br /&gt;
&lt;br /&gt;
I'll be the first to admit: smartwatches are unlikely to completely revolutionize how a library provides online services. But I believe they still represent an opportunity to further enhance existing library services and resources in a unique way.&lt;br /&gt;
&lt;br /&gt;
At the Boston Public Library (BPL), we're in the initial phases of designing a modest smartwatch app to provide notifications for circulation availability and checked-out-material due-date alerts by the end of current year. We're starting small, but we plan to evolve the concept over time as we see what (if any) traction such an application gets with potential users. For example, we plan to explore the possibility of adding &amp;quot;nearest branch to my current location&amp;quot; functionality to this app.&lt;br /&gt;
&lt;br /&gt;
Despite the &amp;quot;development phase&amp;quot; of this application as of this writing, this talk is not being given by a novice. As a technology enthusiast, I've released [http://www.phdgaming.com/smartwatch_projects/ five smartwatch applications] and have had two of those be finalists in a [http://www.phdgaming.com/samsung_challenge/ Samsung sponsored development challenge]. This experience not only will allow for the BPL to avoid many beginner mistakes in its smartwatch app development but also gives a much more complete understanding of the smartwatch development ecosystem.&lt;br /&gt;
&lt;br /&gt;
This talk will explore the following questions:&lt;br /&gt;
&lt;br /&gt;
* What kinds of online library services could potentially be transformed or translated into the smartwatch/wearable domain? What kinds of services are better left alone? These questions are currently being explored and I'll talk about our plans and experiences. Included will be any statistical information from our application launch along with statistics from my personal development.&lt;br /&gt;
&lt;br /&gt;
* How to support all the different operating systems these devices run without painful modifications to your codebase. (There's Tizen that is used by Samsung's Gear 2 and Gear S, Android Wear that is used by most other non-Apple manufacturers, then there is Apple's upcoming smartwatch itself, etc.)&lt;br /&gt;
&lt;br /&gt;
* How to support different screen resolutions on such a small device. From round to rectangular to perfectly square, smartwatches come in all different shapes these days.&lt;br /&gt;
&lt;br /&gt;
* What are the app stores like on these platforms? As I support multiple applications through different distribution networks, a guide to navigating how to distribute one's app is included and I'll reveal how these systems work “behind the curtain.”&lt;br /&gt;
&lt;br /&gt;
* What are common issues and pitfalls to avoid when doing development? Tips on broken APIs and how to cope or optimizing your code will be included.&lt;br /&gt;
&lt;br /&gt;
==Seeing the Forest From the Trees: The Art of Creating Workflows for Digital Projects ==&lt;br /&gt;
 &lt;br /&gt;
* Jen LaBarbera, j.labarbera@neu.edu, NDSR Resident, Northeastern University&lt;br /&gt;
* Joey Heinen, joseph_heinen@harvard.edu, NDSR Resident, Harvard University&lt;br /&gt;
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH&lt;br /&gt;
* Tricia Patterson, triciap@mit.edu, NDSR Resident, MIT&lt;br /&gt;
&lt;br /&gt;
We have to &amp;quot;turn projects into programs&amp;quot; in order to create a solid and sustainable digital preservation initiative...but what the heck does that even mean? What does that look like?&lt;br /&gt;
&lt;br /&gt;
In this talk, members of the inaugural Boston cohort of the National Digital Stewardship Residency will discuss one piece of our digital preservation test kitchen: our stabs at creating digital workflows that will (hopefully) help our institutions turn digital preservation projects into programs. Specifically, we will talk about how difficult it is to create a general and overarching workflow for digital preservation tasks (e.g. ingest into repositories, format migrations, etc.) that incorporates various technical tools while also taking into account the myriad and unending list of possible exceptions or special scenarios. Turning these complicated, specific processes into a simplified and generalized workflow is an art. We haven't necessarily perfected that art yet, but in this talk, we'll share what has worked for us -- and what hasn't. We’ll also touch on the importance of documentation, and achieving that delicate balance of adequately thorough documentation that doesn’t pose the risk of information avalanche. These processes often create more questions than answers, but we'll share the answers that we (and our mentors) have found along the way!&lt;br /&gt;
&lt;br /&gt;
== Annotations as Linked Data with Fedora4 and Triannon (a Real Use Case for RDF!) ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, azaroth@stanford.edu,  Stanford University Libraries&lt;br /&gt;
* Naomi Dushay, ndushay@stanford.edu,  Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Annotations on content resources allow users to contribute knowledge within the digital repository space.  W3C Open Annotation provides a comprehensive model for web annotation on all types of content, using Linked Data as a fundamental framework.  Annotation clients generate instances of this model, typically using a JSON serialization, but need to store that data somewhere using a standard interaction pattern so that best of breed clients, servers, and data can be mixed and matched.&lt;br /&gt;
&lt;br /&gt;
Stanford is using Fedora4 for managing Open Annotations, via a middleware component called Triannon.  Triannon receives the JSON data from the annotation client, and uses the Linked Data Platform API implementation in Fedora4 to create, retrieve, update and delete the constituent resources.  Triannon could be easily modified to use other LDP implementations, or could be modified to work with linked data other than annotations.&lt;br /&gt;
&lt;br /&gt;
== Hydras in the Wild: A survey of current projects ==&lt;br /&gt;
 &lt;br /&gt;
* Mark Bussey,  mark@curationexperts.com, Data Curation Experts&lt;br /&gt;
&lt;br /&gt;
You've seen the tutorials, but [https://github.com/projecthydra/hydra/wiki/Dive-into-Hydra Dive Into Hydra] seems to leave something wanting.  What can you really do using the Hydra Framework?  This talks looks at a number of current Hydra projects and highlights the design and functional features unique to each. Compare and contrast UX, design and functional capabilites from a range of hydra-based repositories including:&lt;br /&gt;
* Avalon for media discovery and distribution (Indiana and Northwestern Univiersities) &lt;br /&gt;
* HydraDam for media archive management(WGBH)&lt;br /&gt;
* HyHull for general Institutional Repository needs (University of Hull)&lt;br /&gt;
* T-DIL for slide library functions (Tufts University)&lt;br /&gt;
* Sufia &amp;amp; Scholarsphere as a bundled self-deposit IR solution (Pennsylvania State University)&lt;br /&gt;
* Curate &amp;amp; Worthwhile as general purpose repository platforms (Multiple Insititutions)&lt;br /&gt;
This will be a whirlwind tour aimed at providing ideas and inspiration for your own repository development project.&lt;br /&gt;
&lt;br /&gt;
== Hydra Makeovers! ==&lt;br /&gt;
&lt;br /&gt;
* Alicia Cozine, alicia@curationexperts.com, Data Curation Experts&lt;br /&gt;
* Patrick Feeley, pgf8@case.edu, Case Western Reserve University&lt;br /&gt;
&lt;br /&gt;
Compare two Hydra-based applications with the systems they replaced. Marvel at the Before and After snapshots of functionality, speed, and look &amp;amp; feel. &lt;br /&gt;
* '''Digital Case 2.0''' is an institutional repository, complete with administrative tools, derivatives transcoding, flexible XML metadata storage, embargo and lease capability, faceted searching, and content viewers for texts/TEI, images, audio recordings, and videos. Digital Case 2.0 is based on worthwhile, an open-source IR starter gem. &lt;br /&gt;
* '''The Tufts Digital Image Library''' is a specialized tool for art and art history resources, offering image collections with user access controls, image ordering, collection nesting, drag-and-drop organization, slideshows, and export capability. &lt;br /&gt;
Both new systems are built on hydra, the open-source Ruby-on-Rails repository solution that incorporates Fedora for storage, Solr for indexing, and Blacklight for search optimization. Their beauty is not just skin-deep!&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Helping Google (and scholars, researchers, educators, &amp;amp; the public) find archival audio ==&lt;br /&gt;
&lt;br /&gt;
* Anne Wootton, anne@popuparchive.org, Pop Up Archive (www.popuparchive.org)&lt;br /&gt;
&lt;br /&gt;
Culturally significant digital audio collections are hard to discover on the web. There are major barriers keeping this valuable media from scholars, researchers, and the general public:&lt;br /&gt;
&lt;br /&gt;
Audio is opaque: you can’t picture sound, or skim the words in a recording. &lt;br /&gt;
Audio is hard to share: there’s no text to interact with. &lt;br /&gt;
Audio is not text: but since text is the medium of the web, there’s no path for audiences to find content-rich audio.&lt;br /&gt;
Audio metadata is inconsistent and incomplete.&lt;br /&gt;
&lt;br /&gt;
At Pop Up Archive, we're helping solve this problem making the spoken word searchable. We began as a UC-Berkeley School of Information Master's thesis to provide better access to recorded sound for audio producers, journalists, and historians. Today, Pop Up Archive processes thousands of hours of sound from all over the web to create automatic, timestamped transcripts and keywords, working with media companies and institutions like NPR, KQED, HuffPost Live, Princeton, and Stanford. We're building collections of sound from journalists, media organizations, and oral history archives from around the world. Pop Up Archive is supported by the John S. and James L. Knight Foundation, the National Endowment for the Humanities, and 500 Startups.&lt;br /&gt;
&lt;br /&gt;
== Digital Content Integrated with ILS Data for User Discovery:  Lessons Learned ==&lt;br /&gt;
&lt;br /&gt;
* Naomi Dushay, ndushay@stanford.edu,  Stanford University Libraries&lt;br /&gt;
* Laney McGlohon, laneymcg@stanford.edu,  Stanford University Libraries (first-time presenter)&lt;br /&gt;
&lt;br /&gt;
So you want to expose your digital content in your discovery interface, integrated with the data from your ILS?  How do you make the best information user searchable?  How do you present complete, up to date search results with a minimum of duplicate entries?&lt;br /&gt;
&lt;br /&gt;
At Stanford, we have these cases and more:&lt;br /&gt;
* digital content with no metadata in ILS&lt;br /&gt;
* digital content for metadata in ILS&lt;br /&gt;
* digital content with its own metadata derived from ILS metadata.&lt;br /&gt;
&lt;br /&gt;
We will describe our efforts to accommodate multiple updatable metadata sources for materials in the ILS and our Digital Object Repository while presenting users with reduced duplication in SearchWorks.  Included will be some failures, some successes, and an honest assessment of where we are now.&lt;br /&gt;
&lt;br /&gt;
== Show All the Things: Kanban for Libraries == &lt;br /&gt;
&lt;br /&gt;
* Mike Hagedon, mhagedon@email.arizona.edu, University of Arizona Libraries (first-time presenter)&lt;br /&gt;
&lt;br /&gt;
The web developers at the University of Arizona Libraries had a problem: we were working on a major website rebuild project with no clear way to prioritize it against our other work. We knew we wanted to follow Agile principles and initially chose Scrum to organize and communicate about our work. But we found that certain core pieces of Scrum did not work for our team. Then we discovered Kanban, an Agile meta-process for organizing work (team or individual) that treats the work more as a flow than as a series of fixed time boxes. I’ll be talking about our journey toward finding a process that works for our team and how we’ve applied the principles of Kanban to better get our work done. Specifically, I'll discuss principles like how to visualize all your work, how to limit how much you’re doing (to get more done!), and how to optimize the flow of your work.&lt;br /&gt;
&lt;br /&gt;
== DIY Catalog ==&lt;br /&gt;
 &lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
With so many discovery layers being made available that didn’t quite fit user needs but still came with substantial costs, Hennepin County Library determined it would develop one that could handle its customizations.  So they built their own ILS API, Web services, and an interface that provides all the features (and existing custom features) of the old catalog, with new features planned, all within a responsive layout.  Get a look at the system architecture, how the front end communicates through the layers back to the databases, and how to manage the infrastructure.&lt;br /&gt;
&lt;br /&gt;
== OverDrive: Full Integration ==&lt;br /&gt;
 &lt;br /&gt;
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
How do you increase usage of your ebooks?  Seamless integration in the catalog certainly helps. Users can search for digital titles, limit to just those that are available, place a request, borrow, and manage all their digital titles from the library website alongside all the other physical titles the library has.  Get a quick demo and learn how this was all made possible.&lt;br /&gt;
&lt;br /&gt;
The [//www.hclib.org Hennepin County Library] is ranked one of [//www.thedigitalshift.com/2014/10/ebooks/overdrive-api-usage-indicates-growth-ebook-checkouts-via-opac/ the top users of OverDrive and OverDrive’s API] and is looking to more fully integrate 3M Cloud, OneClickDigital, and other digital resources.&lt;br /&gt;
&lt;br /&gt;
== Dynamic Indexing: a Tragic Solr Story ==&lt;br /&gt;
 &lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
Loading data from an ILS into Solr isn’t so hard, unless it needs to be dynamic, fast, and hold more data than what can be found in 1.5 million MARC records. Some additional information we’ve incorporated are from Syndetics, ILS circulation, and OverDrive.  We’ll share the nitty gritty details and what we learned about dynamic Solr indexing, including how to get good performance, how to deal with indexing failures, how to schedule it all to keep the data up-to-date, and some things you can do with that data such as popularity ratings.&lt;br /&gt;
&lt;br /&gt;
== Fedora 4, Survey of Core Capabilities and External Modules: == &lt;br /&gt;
* Andrew Woods, awoods@duraspace.org, DuraSpace&lt;br /&gt;
&lt;br /&gt;
Now that the production release of Fedora 4.0 is out, the time is right to take a step back and review the core capabilities offered by Fedora:&lt;br /&gt;
* Basic CRUD&lt;br /&gt;
* Versioning&lt;br /&gt;
*- Transactions&lt;br /&gt;
* etc&lt;br /&gt;
&lt;br /&gt;
...and take a closer look at the supported external modules:&lt;br /&gt;
* Authorization&lt;br /&gt;
* Solr integration&lt;br /&gt;
* Triplestore integration&lt;br /&gt;
* Camel integration&lt;br /&gt;
* etc&lt;br /&gt;
&lt;br /&gt;
In addition to clarifying specific implementation details, this is an opportunity to surface community requirements which may have not been addressed in the initial release.&lt;br /&gt;
&lt;br /&gt;
== Bleeding edge beacons: redesign of a library tour with new technology ==&lt;br /&gt;
&lt;br /&gt;
* Neal Henshaw, Virginia Tech &lt;br /&gt;
* Somiah Lattimore, Straight Up Creative&lt;br /&gt;
* Keith Gilbertson, keith.gilbertson@vt.edu, Virginia Tech&lt;br /&gt;
&lt;br /&gt;
This group has not yet presented at Code4Lib&lt;br /&gt;
&lt;br /&gt;
A technology liaison, a designer, and a developer are working together to update a mobile audio tour of an academic library.&lt;br /&gt;
&lt;br /&gt;
The walking tour was originally conceived of and created by the technology liaison to provide an orientation experience for incoming students. In its current version, patrons visit several stations in the library and activate an audio description of each station by scanning a QR barcode with a mobile device.&lt;br /&gt;
&lt;br /&gt;
We are building a prototype to test with focus groups so that we can launch a new version of the tour in early 2015. The tour has been reimagined with a professional and communicative user interface that presents students with learning goals for each location in the library.  The new version of the tour is a mobile application with integrated support for micro-location technology provided through Bluetooth low energy beacon devices, known informally as iBeacons, installed throughout the building. As a touring student walks through the library, her location is noted by the app so that an appropriate video segment is automatically played according to the current location.&lt;br /&gt;
 &lt;br /&gt;
We will discuss perspectives on the design of the project, including&lt;br /&gt;
 &lt;br /&gt;
* Designing with technology to achieve learning goals&lt;br /&gt;
* Designing graphical elements, interaction, and user experience&lt;br /&gt;
* Designing to maintain compatibility with older technologies&lt;br /&gt;
&lt;br /&gt;
We’ll explain our efforts to make the application usable by those with hearing impairments and mobility impairments, reactions from our first users, and challenges in working with the relatively new beacon technology.&lt;br /&gt;
&lt;br /&gt;
== Distributed Remediation: Small tools for big problems: ==&lt;br /&gt;
 &lt;br /&gt;
* Matt Miller,  matthewmiller@nypl.org, New York Public Library, NYPL Labs&lt;br /&gt;
&lt;br /&gt;
Remediation of legacy data can be automated only so much. Certain essential cleanup tasks, such as aligning a name with the correct authorized version, is very difficult for a computer yet trivial for a person. While it is these types remediations that will allow an institution to take advantage and participate in the web of Linked Open Data, a wholly manual approach is unrealistic. However, by augmenting automated remediation with a light human touch we can quickly and efficiently reach our goals. This talk will look at tools and methods being developed at NYPL Labs to empower library staff and the public to help clean up our legacy metadata through collaborative remediation.&lt;/div&gt;</summary>
		<author><name>MatthewMiller</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2015_Prepared_Talk_Proposals&amp;diff=42004</id>
		<title>2015 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2015_Prepared_Talk_Proposals&amp;diff=42004"/>
				<updated>2014-11-07T20:49:23Z</updated>
		
		<summary type="html">&lt;p&gt;MatthewMiller: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;Code4lib 2015 is a loosely-structured conference that provides people working at the intersection of libraries/archives/museums/cultural heritage and technology with a chance to share ideas, be inspired, and forge collaborations. For more information about the Code4lib community, please visit http://code4lib.org/about/. &lt;br /&gt;
The conference will be held at the Portland Hilton &amp;amp; Executive Tower in Portland, Oregon, from February 9-12, 2015.&lt;br /&gt;
&lt;br /&gt;
'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
We encourage everyone to propose a talk.&lt;br /&gt;
 &lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
* Projects you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* Tools and technologies – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* Technical issues - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* Relevant non-technical issues – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
Proposals can be submitted through Friday, November 7, 2014 at 5pm PST (GMT−8). Voting will start on November 11, 2014 and continue through November 25, 2014. The URL to submit votes will be announced on the Code4Lib website and mailing list and will require an active code4lib.org account to participate. The final list of presentations will be announced in early- to mid-December.&lt;br /&gt;
&lt;br /&gt;
'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Log in to the Code4lib wiki and edit this wiki page using the prescribed format. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist voters in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
Please follow the formatting guidelines:&lt;br /&gt;
&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Talk Title: ==&lt;br /&gt;
 &lt;br /&gt;
* Speaker's name,  email address, and (optional) affiliation&lt;br /&gt;
* Second speaker's name, email address, and affiliation, if second speaker&lt;br /&gt;
&lt;br /&gt;
Abstract of no more than 500 words.&lt;br /&gt;
&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
== Zines + Gamification = Awesomest Metadata Literacy Outreach Event Ever! ==&lt;br /&gt;
 &lt;br /&gt;
* [http://www.JenniferHecker.info Jennifer Hecker], jenniferraehecker@gmail.com, [http://www.lib.utexas.edu/subject/zines University of Texas Libraries] &amp;amp; [http://www.AustinFanzineProject.org Austin Fanzine Project]&lt;br /&gt;
* [http://anomalily.net/ Lillian Karabaic], librarian@iprc.org, [http://www.iprc.org/ Independent Publishing Resource Center] (Portland)&lt;br /&gt;
 &lt;br /&gt;
In academic libraries, and elsewhere, the popularity of zine (a magazine produced for love, not profit) collections is on the rise. At the same time, metadata literacy is becoming an increasingly important skill, helping people navigate and understand digital environments and interactions. We have found a way to teach metadata literacy to the general public that isn’t super-boring – in fact, we’ve made it downright fun!&lt;br /&gt;
&lt;br /&gt;
First, volunteer zine librarian Lillian Karabaic of Portland’s Independent Publishing Resource Center facilitated the creation of a gamified cataloging interface for the IPRC’s annual Raiders of the Lost Archives backlog-busting 24-hour volunteer cataloging event.&lt;br /&gt;
&lt;br /&gt;
Then, archivist Jennifer Hecker facilitated the adaptation of the IPRC’s game for use in a similar, but also very different context – promoting UT Libraries newly-acquired zine collections. The main goal of the academic-library-based event was increasing excitement around the collections, but with the side goal of building metadata literacy, and introducing an understanding of library cataloging issues.&lt;br /&gt;
&lt;br /&gt;
The Texas modification also conforms to the xZINECOREx metadata schema developed by the national [http://zinelibraries.info/ Zine Librarians Interest Group], and triggered interesting conversations with the Libraries’s cataloging department about evolving metadata standards and how to incorporate the products of crowd-sourcing projects into existing workflows.&lt;br /&gt;
&lt;br /&gt;
Both games will be demoed.&lt;br /&gt;
&lt;br /&gt;
We have never presented at Code4lib.&lt;br /&gt;
&lt;br /&gt;
== Do the Semantic FRBRoo ==&lt;br /&gt;
* Rosie Le Faive, rlefaive@upei.ca, University of Prince Edward Island&lt;br /&gt;
&lt;br /&gt;
[http://www.islandora.ca Islandora] is great for creating repositories of any data type, but how can you model meaningful relationships between digital objects and use them to tell a story?&lt;br /&gt;
&lt;br /&gt;
At UPEI, I’m assembling an ethnography of Prince Edward Island’s traditional fiddle music that includes musical clips, video clips, oral histories, musical notation, images, and ethnographic commentaries. In order to present an exhibition-style site, I’m tying these digital objects together via the people, places, events, tunes and topics that they share or describe. &lt;br /&gt;
&lt;br /&gt;
To describe the relationships, I’m extending Islandora to use [http://www.cidoc-crm.org/frbr_inro.html FRBRoo], a vocabulary that combines the FRBR model with CIDOC-CRM, the the object-oriented museum documentation ontology. These modules being developed will allow other researchers to create a structured, navigable digital repository of diverse object types, that uses Islandora as an exhibition platform. &lt;br /&gt;
&lt;br /&gt;
== Our $50,000 Problem: Why Library School? ==&lt;br /&gt;
* Jennie Rose Halperin, jhalperin@mozilla.com, Mozilla Corporation&lt;br /&gt;
&lt;br /&gt;
57 library schools in the United States are churning out approximately 100 graduates per year, many with debt upwards of $50,000.  According to ONet, [http://www.inthelibrarywiththeleadpipe.org/2011/is-the-united-states-training-too-many-librarians-or-too-few-part-1/ 84% of library jobs in the US require an MLS.] The library profession is [http://dpeaflcio.org/programs-publications/issue-fact-sheets/library-workers-facts-figures/) 92% white and 82% female and entry-level librarians can expect to make $32,500 per year.]&lt;br /&gt;
&lt;br /&gt;
Contrasted with developers, who are almost [http://www.ncwit.org/blog/did-you-know-demographics-technical-women 90% male] and can expect to make [http://www.forbes.com/sites/jennagoudreau/2011/06/01/best-entry-level-jobs/ $70,000 in an entry-level position,] these numbers are dismal.&lt;br /&gt;
&lt;br /&gt;
According to a recent survey, the top skill that outgoing library students want to know is “programming” and yet many MLS programs still consider Microsoft Word an essential technology skill.&lt;br /&gt;
&lt;br /&gt;
What is going on here? Why do we accept this fate, where mostly female debt-burdened professionals continue to be thrown onto the work force without the education their expensive degrees promised?&lt;br /&gt;
&lt;br /&gt;
As a community we need to come together to stop this cycle. We need to provide better support and mentorship to diversify and keep the profession relevant and help librarianship move into the future it deserves.&lt;br /&gt;
&lt;br /&gt;
This talk will walk through the challenges of navigating a hostile employment environment as well as present models for better development and future state imagining.&lt;br /&gt;
&lt;br /&gt;
== No cataloging software? Need more than Dublin Core? No problem!: Experiences with CollectiveAccess ==&lt;br /&gt;
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University&lt;br /&gt;
* Rachel Wittmann, rwittma@clemson.edu, Clemson University&lt;br /&gt;
&lt;br /&gt;
Clemson University Libraries has implemented the open-source software CollectiveAccess for customized digital collection needs. CollectiveAccess is an open-source project with the goal of providing a flexible way to manage and publish museum and archival collections. There are several applications associated with the projects; most used are: Providence (for cataloging and entering metadata) and Pawtucket (for displaying objects in a collection for the public). It has many profiles readily available for installing with existing library standards, such as Dublin Core, and there is a robust syntax for creating your own profiles to fit custom tailored metadata schemas. Plus, the user interface allows you to modify the metadata profile quickly and easily.&lt;br /&gt;
&lt;br /&gt;
In this talk, we will discuss:&lt;br /&gt;
* Our experiences with installing Providence and creating an installation profile that satisfies the needs of many of the Clemson Libraries digital archiving processes. &lt;br /&gt;
* The stumbling blocks experienced in that process and how they were resolved.&lt;br /&gt;
* The available plugins sourcing widely used authorities, such as Library of Congress thesauri and GeoNames.org, and how they have been used by our projects. &lt;br /&gt;
* A brief overview of the export and import functions and also current workflow practices within Providence.&lt;br /&gt;
* Future plans &amp;amp; the role of CollectiveAccess at Clemson University Libraries&lt;br /&gt;
&lt;br /&gt;
== Getting ContentDM and Wordpress to Play Together ==&lt;br /&gt;
* [[User:SeanHendricks|Sean Q. Hendricks]], sqhendr@clemson.edu, Clemson University&lt;br /&gt;
&lt;br /&gt;
Clemson University Libraries has a very strong program for digitizing and archiving photographs, and the Digital Imaging team processes many hundreds of photographs every month. These images are managed using different methods, including ContentDM, a digital collection manager.&lt;br /&gt;
&lt;br /&gt;
ContentDM provides various methods for searching and displaying photographs, along with their metadata. However, recent initiatives have resulted in the need to leverage those collections into exhibits displayed on other library-related websites, such as our Special Collections unit. The Clemson Libraries has invested heavily in Wordpress as our content management system of choice, and it seemed most efficient not to have to export and import images into our Wordpress sites in order to provide exhibited images.&lt;br /&gt;
&lt;br /&gt;
Fortunately, ContentDM has provided an API to many of their functions, allowing the extraction of metadata and even rescaled images through URLs. This project has been developing a plugin for Wordpress that integrates with ContentDM through shortcodes that Wordpress editors can easily include in their content. These shortcodes allow editors to choose how many images, which images from which collections, thumbnail sizes, etc. to display in different gallery styles. Plans are for it to allow integration with different plugins such as Fancybox and Masonry.&lt;br /&gt;
&lt;br /&gt;
In this presentation, I will demonstrate the current state of the plugin and discuss future plans. &lt;br /&gt;
&lt;br /&gt;
==Refinery — An open source locally deployable web platform for the analysis of large document collections==&lt;br /&gt;
 &lt;br /&gt;
* [[User:DaeilKim|Daeil Kim]], The New York Times, daeil.kim@nytimes.com&lt;br /&gt;
&lt;br /&gt;
Refinery is an open source web platform for the analysis of large unstructured document collections. It extracts meaningful semantic themes within documents also known as &amp;quot;topics&amp;quot; which can be thought of as word clouds composed of terms that highly co-occur with one another. Once this semantic index is formed, one can extract relevant documents related to these topics and further refine their contents through a summarization process that allows users to search for phrases that are relevant to them within the corpus. The goal of Refinery is to make this whole process easier and to provide some of the latest scalable versions of these learning algorithms in an intuitive web-based interface. Refinery is also meant to be run locally, thus bypassing the need for securing document collections over the internet. The talk will go through some of the technologies involved and a demo of the app.&lt;br /&gt;
&lt;br /&gt;
For more info check out http://www.docrefinery.org.&lt;br /&gt;
&lt;br /&gt;
==Drupal 8 — Evolution &amp;amp; Revolution==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Drupal 8 is in beta and nearing release. Among its many features, it notably has become more developer friendly through its adoption of the Symfony PHP framework along with Symfony's outstanding set of libraries (like Guzzle) and tools (like Composer). And, in implementing the Twig theming system, it is can begin to escape PHPtemplate. These moves also make it easier to create headless systems that uses Angular.js and other systems for presentation, or even forgo presentation entirely.&lt;br /&gt;
&lt;br /&gt;
From the site-builder's perspective, Drupal 8 provides a much smother experience and makes it easier to build and implement site recipes.&lt;br /&gt;
&lt;br /&gt;
==Using GameSalad to Build a Gamified Information Literacy Mobile App for Higher Education==&lt;br /&gt;
 &lt;br /&gt;
* [[User:StanBogdanov|Stanislav 'Stan' Bogdanov]],  stan@stanrb.com, Adelphi University and [http://bogliollc.com Boglio LLC]&lt;br /&gt;
&lt;br /&gt;
GameSalad is a popular tool for developing mobile and desktop games with little actual programming. In this presentation, Stan Bogdanov breaks down the development process he followed while building [https://github.com/stanrb/mobiLit mobiLit], a mobile app with the goal of being the first open-source gamified information literacy app to be used as part of a college-level information literacy curriculum. He will go through the basics of using GameSalad to create an app that can be easily customized by non-programmers and the instructional principles used to teach the material in a mobile medium. Stan will also go through two qualitative design studies he did on the app and discuss their results and the lessons learned from building mobiLit. The session will conclude with an overview of the next steps for the [https://github.com/stanrb/mobiLit mobiLit project].&lt;br /&gt;
&lt;br /&gt;
==The Impossible Search: Pulling data from multiple unknown sources==&lt;br /&gt;
 &lt;br /&gt;
* Riley Childs, no official affiliation (currently a Senior in High School at Charlotte United Christian Academy), rchilds (AT) cucawarriors.com &lt;br /&gt;
&lt;br /&gt;
It's easy to search data you know the structure of, but what if you need to pull in data from sources that don't have a standard structure. The ability to search community events along with your standard catalog search results is an example, but often the only way to pull these events is through XML, JSON, (Insert structured format here), or even just raw html. But how do you get that structure? That simple question is what makes this impossible. The process to define and process this structure takes a lot of manual labor, especially if the data you are pulling is just HTML, and then every time you add data to the index you have to run all the data through a script to pull in data in a format Solr or an other index can use. This talk will focus on Solr, but the principles explained will apply to many other indexes.&lt;br /&gt;
&lt;br /&gt;
==What! You're Not Using Docker?==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Boring part: Docker[1] is a container system that provides benefits similar to virtualization with only a fraction of the overhead. Scintillating part: Docker can host between four to six times the number of service instances than systems such as Xen or VMWare on a given piece of hardware. But thats not all! Docker also makes it simple(r) to create transportable instances, so you can spin up development servers on your laptop.&lt;br /&gt;
&lt;br /&gt;
*[1]https://www.docker.com/&lt;br /&gt;
&lt;br /&gt;
== Video Accessibility, WebVTT, and Timed Text Track Tricks ==&lt;br /&gt;
&lt;br /&gt;
* Jason Ronallo, jronallo@gmail.com, NCSU Libraries&lt;br /&gt;
&lt;br /&gt;
Video on the Web presents new challenges and opportunities. How do you make your video more accessible to those with various disabilities and needs? I'll show you how. This presentation will focus on how to write and deliver captions, subtitles, audio descriptions, and timed metadata tracks for Web video using the WebVTT W3C standard. Encoding timed text tracks in this way opens up opportunities for new functionality on your websites beyond accessibility. The presentation will show some examples of the potential for using timed text tracks in creative ways. I'll cover all the HTML and JavaScript you will need to know as well as some of the CSS and other bits you could probably do without but are too fun to pass up.&lt;br /&gt;
&lt;br /&gt;
== Categorizing Records with Random Forests ==&lt;br /&gt;
 &lt;br /&gt;
* Geoffrey Boushey, geoffrey.boushey@ucsf.edu, UCSF Library&lt;br /&gt;
Academic libraries are increasingly responsible for providing ingest, search, discovery, and analysis for data sets.  Emerging techniques from data science and machine learning can provide librarians and developers with an opportunity to generate new insights and services from these document collections.  This presentation will provide a brief overview of common machine learning classification techniques, then dive into a more detailed example using a random forest to assign keywords to research data sets.  The talk will emphasize the insight that can be gained from machine learning rather than the inner workings of the algorithms.  The overall goal of this presentation is to provide librarians and developers with the context to recognize an opportunity to apply machine learning categorization techniques at their home campuses and organizations.  &lt;br /&gt;
&lt;br /&gt;
== Data Science in Libraries ==&lt;br /&gt;
 &lt;br /&gt;
* Devon Smith, smithde@oclc.org, OCLC&lt;br /&gt;
&lt;br /&gt;
Data Science is increasing in buzz and hype. I'll go over what it is, what it isn't, and how it fits in libraries.&lt;br /&gt;
&lt;br /&gt;
== PDF metadata extraction for academic literature == &lt;br /&gt;
&lt;br /&gt;
* Kevin Savage, kevin.savage at mendeley.com, Mendeley&lt;br /&gt;
* Joyce Stack, joyce.stack at mendeley.com, Mendeley&lt;br /&gt;
&lt;br /&gt;
Mendeley recently added a, &amp;quot;document from file,&amp;quot; endpoint to its API which attempts to extract metadata such as title and authors directly from PDF files. This talk will describe at a high level the machine learning methods we used including how we measured and tuned our model. We will then delve more deeply into our stack, the tools we used, some of the things that didn't work and why PDFs are the worst thing ever to compute over.&lt;br /&gt;
&lt;br /&gt;
== Giving Users What They Want: Record Grouping in VuFind ==&lt;br /&gt;
 &lt;br /&gt;
* Mark Noble,  mark@marmot.org, [//www.marmot.org Marmot Library Network]&lt;br /&gt;
&lt;br /&gt;
In 2013, Marmot did extensive usability studies with patrons to determine what was difficult in the catalog.  Many patrons had problems sifting through all of the various formats and editions of a title.  In 2014 we developed a method for [//mercury.marmot.org/Union/Search?lookfor=divergent grouping records] so only a single work is shown in search results and all formats and editions are listed under that work.  We will discuss our definition of a 'work' based on FRBR principles; combining meta data from MARC records with metadata from other sources like OverDrive; the technical details of Record Grouping; the design decisions made during implementation; and the reaction from users and staff.&lt;br /&gt;
&lt;br /&gt;
== Topic Space: a mobile augmented reality recommendation app ==&lt;br /&gt;
&lt;br /&gt;
* Jim Hahn, jimhahn@illinois.edu, University of Illinois at Urbana-Champaign&lt;br /&gt;
&lt;br /&gt;
The Topic Space module (http://minrvaproject.org/modules_topicspace.php ) was developed with an IMLS Sparks! Grant to investigate augmented reality technologies for in-library recommendations. The funding allowed for sustained university community collaboration by the University Library, the Graduate School of Library and Information Science, as well as graduate student programmers sourced from the Department of Computer Science. Collaborators designed app functionality and identified relevant open source libraries that could power optical character recognition (OCR) functionality from within the mobile phone.&lt;br /&gt;
&lt;br /&gt;
Topic space allows a user to take a picture of an item's call number in the book stacks. The module will show the user other books that are relevant but that are not shelved nearby. It can also show users books that are normally shelved here but that are currently checked out. Recommendations are based on Library of Congress subject headings and ILS circulation data which indicate recommendation candidates based on total check-outs. &lt;br /&gt;
&lt;br /&gt;
Research questions included development of back end (server-side) pattern matching algorithms for recommendations, and a rapid formative evaluation of interface design that would provide optimal user experience for navigation of the book stacks as a context to recommendations.&lt;br /&gt;
&lt;br /&gt;
Along with the Topic Space native app, grant collaborators prototyped web based recommendations which could serve as a new way of providing readers advisory and “more like this” recommendations from discovery interfaces accessed through desktop browsers. Outcomes of the grant include the availability of the [https://play.google.com/store/apps/details?id=edu.illinois.ugl.minrva Topic Spaces module within Minrva app on the Android Play store] and an experimental [http://backbonejs.org/ Backbone.js] based [http://minrva-dev.library.illinois.edu Topic Space web app].&lt;br /&gt;
&lt;br /&gt;
== Leveling Up Your Git Workflow ==&lt;br /&gt;
&lt;br /&gt;
* Megan Kudzia, moneill@albion.edu, Albion College Library&lt;br /&gt;
* Kate Sears, eks11@albion.edu, Albion College Library&lt;br /&gt;
&lt;br /&gt;
Have you started experimenting with Git on your own, but now you need to include others in your projects? Learn from our mistakes! Transitioning from a one-person git workflow and repo structure, to a structure that includes multiple people (including student workers), is not for the faint of heart. We'll talk about why we decided to work this way, our path to developing a git culture amongst ourselves, conceptual and technical difficulties we've faced, what we learned, and where we are now. Also with pretty pictures (aka workflow drawings).&lt;br /&gt;
&lt;br /&gt;
== Drone Loaning Program: Because Laptops are so last century ==&lt;br /&gt;
&lt;br /&gt;
 * Uche Enwesi, uenwesi@umd.edu, University of Maryland Libraries&lt;br /&gt;
 * Francis Kayiwa, fkayiwa@umd.edu, University of Maryland Libraries&lt;br /&gt;
&lt;br /&gt;
At Univ. Maryland we are in the very early stages of looking into allowing our student body get their hands on a drone. Yes that's right we will let students take out a drone for n amount of hours to work on projects of their choosing. The talk will talk about the logistics of getting a program of this sort from concept to &amp;quot;Is the drone available?&amp;quot;. If people sign waivers we will also promise not to crash the drone into code4lib attendees.&lt;br /&gt;
&lt;br /&gt;
== Got Git? Getting More Out of Your GitHub Repositories ==&lt;br /&gt;
&lt;br /&gt;
 * Terry Brady, twb27@georgetown.edu, Georgetown University Library&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss how librarians, developers, and system administrators at Georgetown University are maximizing their use of the public and private GitHub repositories. &lt;br /&gt;
&lt;br /&gt;
In additional to all of the great benefits of using Git for code management, the GitHub interface provides a powerful set of tools to showcase a project and to keep your users informed of developments to your project.  These tools can assist with marketing and outreach - turning your code repository into a focus of conversation!&lt;br /&gt;
&lt;br /&gt;
* [http://georgetown-university-libraries.github.io/File-Analyzer/ Style-able Project Pages]&lt;br /&gt;
* [https://github.com/Georgetown-University-Libraries/File-Analyzer/wiki Project Wikis]&lt;br /&gt;
* [https://github.com/Georgetown-University-Libraries/Georgetown-University-Libraries-Code/releases Project Release Notes/Portfolios]&lt;br /&gt;
* [https://rawgit.com/Georgetown-University-Libraries/Georgetown-University-Libraries-Code/master/samples/GoogleSpreadsheetFilter.html Web Resources That Can Be Directly Requested]&lt;br /&gt;
* Gists for code sharing&lt;br /&gt;
* Private Repositories and Organizational Groups&lt;br /&gt;
* Pull Request Conversation Tracking&lt;br /&gt;
* Customized Issue management&lt;br /&gt;
&lt;br /&gt;
== Quick Wins for Every Department in the Library - File Analyzer! ==&lt;br /&gt;
&lt;br /&gt;
 * Terry Brady, twb27@georgetown.edu, Georgetown University Library&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Library has customized workflows for nearly every department in our library with a single code base.&lt;br /&gt;
* Analyzing Marc Records for the Cataloging department&lt;br /&gt;
* Transferring ILS invoices for the University Account System for the Acquisitions department &lt;br /&gt;
* Delivering patron fines to the Bursar’s office for the Access Service department&lt;br /&gt;
* Summarizing student worker timesheet data for the Finance department&lt;br /&gt;
* Validating COUNTER compliant reports for the Electronic Resources department&lt;br /&gt;
* Generating ingest packages for the Digital Services department&lt;br /&gt;
* Validating checksums for the Preservation department&lt;br /&gt;
&lt;br /&gt;
Learn how you can customize the [http://georgetown-university-libraries.github.io/File-Analyzer/ File Analyzer] to become a hero in your library!&lt;br /&gt;
&lt;br /&gt;
==The Geospatial World is Moving from Maps *on* the Web to Maps *of* the web. Libraries can too==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Copystar|Mita Williams]], mita@uwindsor.ca, User Experience Librarian, University of Windsor&lt;br /&gt;
&lt;br /&gt;
The transition from paper maps to digital ones changed much more than the maps themselves; it changed the very foundation of how we work and how we find each other. Now maps are transforming again.  The Geospatial World is moving from GIS systems that are institutionally-focused, expensive, feature-burdened, and binds data into a complicated and demanding user-hostile interface. From this transition from digital to web-based digital geospatial tools has come growth and development in new forms of map-based investigative journalism, activism, scholarship, and business ventures. This talk will highlight the conditions and strategies that made these changes possible as a means to draw a path by which librarians through our own work may follow, dragons notwithstanding. &lt;br /&gt;
&lt;br /&gt;
== Building Your Own Federated Search ==&lt;br /&gt;
&lt;br /&gt;
* Rich Trott, Richard.Trott@ucsf.edu, UC San Francisco&lt;br /&gt;
&lt;br /&gt;
Advances in modern browsers have created some interesting possibilities for federated search. This presentation will cover common techniques and pitfalls in building a federated search. We will discuss what principles guided our decisions when implementing our own federated search. We will show tools we've built and our findings from building and using experimental prototypes.&lt;br /&gt;
&lt;br /&gt;
Your higher education institution likely offers dozens of online resources for educators, students, researchers, and the public. And each of these online resources likely has its own search tool. But users can't be expected to search in dozens of different interfaces to find what they're looking for. A typical solution for this issue is federated search. &lt;br /&gt;
&lt;br /&gt;
==  Indexing Linked Data with LDPath ==&lt;br /&gt;
&lt;br /&gt;
* Chris Beer, cabeer@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
LDPath [1] is a simple query language for indexing linked open data, with support for caching, content negotiation, and integration with non-RDF endpoints. This talk will demonstrate the features and potential of the language and framework to index a resource with links into id.loc.gov, viaf.org, geonames.org, etc to build an application-ready document.&lt;br /&gt;
&lt;br /&gt;
[1] http://marmotta.apache.org/ldpath/language.html&lt;br /&gt;
&lt;br /&gt;
== Show Me the Money: Integrating an LMS with Payment Providers ==&lt;br /&gt;
 &lt;br /&gt;
* Josh Weisman,  Josh.Weisman@exlibrisgroup.com, Development Director-Resources Management, Ex Libris Group&lt;br /&gt;
&lt;br /&gt;
In order to provide an easy and convenient way for patrons to pay fines, we are exploring ways to integrate the library management system with online payment providers such as PayPal. With many LMS systems being designed and developed for the cloud, we should be able to provide the frictionless user experience our patrons have come to expect from online transactions. In this session we'll discuss strategies for integration and review a sample application which uses REST APIs from a library management system to integrate with PayPal.&lt;br /&gt;
&lt;br /&gt;
== Shibboleth Federated Authentication for Library Applications: ==&lt;br /&gt;
&lt;br /&gt;
* Scott Fisher, scott.fisher@ucop.edu, California Digital Library&lt;br /&gt;
* Ken Weiss, ken.weiss@ucop.edu, California Digital Library&lt;br /&gt;
&lt;br /&gt;
Shibboleth is the most widely-used method to provide single-sign-on authentication to academic applications where users come from many different institutions. Shibboleth, the InCommon education and research trust framework, and the SAML protocol comprise a very powerful - but very complicated - solution to this very complicated problem. Scott and Ken have implemented Shibboleth for multiple library applications. They will share their understanding of the good, the bad, and the underlying spaghetti that makes it all work. Ken will discuss some of the technical aspects of the solution, touching on optimal and non-optimal use cases, administrative challenges, and authorization concerns. Scott will describe the implementation pattern for multi-institution single-sign-on that the California Digital Library has evolved, using the recently released Dash application (http://dash.cdlib.org) as an example.&lt;br /&gt;
&lt;br /&gt;
==Scientific Data: A Needs Assessment Journey==&lt;br /&gt;
 &lt;br /&gt;
*[[User:VickySteeves| Vicky Steeves]], vsteeves@amnh.org, American Museum of Natural History&lt;br /&gt;
&lt;br /&gt;
While surveying digital research and collections data in the research science divisions at the American Museum of Natural History in NYC (as a part of my [http://ndsr.nycdigital.org/ National Digital Stewardship Residency] project), I have come across the big data hogs (genome sequencing and CT scanning) and the little pieces of data (images, publications), all equally important to not only scientific discovery, but as nodes in the history of science. &lt;br /&gt;
&lt;br /&gt;
In this session, I will discuss the development of my needs assessment surveys for scientific datasets and the interview process with Museum curators and researchers as background, seguing into an explanation of the results. I will then combine my findings into preliminary selection criteria to choose tools for digital preservation and management unique to scientific datasets. This will brooke a discussion on emerging standards, tools, and technologies in big data, specific to research science. &lt;br /&gt;
&lt;br /&gt;
I will conclude with preliminary findings on emerging technology that can be used to answer concerns surrounding the management and digital preservation of these data. I am hoping the Q&amp;amp;A session can be used to both answer questions about my project, and function as a way for you (the larger tech-savy library community)  to discuss the tools I’ve touched on in this talk. &lt;br /&gt;
&lt;br /&gt;
== Feminist Human Computer Interaction (HCI) in Library Software ==&lt;br /&gt;
 &lt;br /&gt;
* Bess Sadler,  bess@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Libraries are not neutral repositories of knowledge. Library classification systems and search technologies tend to reflect the inequalities, biases, ethnocentrism, and power imbalances of the societies in which they are built [1]. How might we better resist these tendencies in the library software we create? This talk will examine some qualities of feminist HCI (pluralism, self-disclosure, participation, ecology, advocacy, and embodiment) [2] through the lens of library software. &lt;br /&gt;
&lt;br /&gt;
[1] Olson, Hope A. (2002). The Power to Name: Locating the Limits of Subject Representation in Libraries. Dordrecht, The Netherlands: Kluwer Academic Publishers.&lt;br /&gt;
&lt;br /&gt;
[2] Bardzell, Shaowen. Feminist HCI: Taking Stock and Outlining an Agenda for Design. CHI 2010: HCI For All. http://dmrussell.net/CHI2010/docs/p1301.pdf&lt;br /&gt;
&lt;br /&gt;
== Heiðrún: DPLA's Metadata Harvesting, Mapping and Enhancement System ==&lt;br /&gt;
&lt;br /&gt;
* Audrey Altman, audrey at dp.la, Digital Public Library of America&lt;br /&gt;
* Gretchen Gueguen, gretchen at dp.la, Digital Public Library of America&lt;br /&gt;
* Mark Breedlove, mb at dp.la, Digital Public Library of America&lt;br /&gt;
&lt;br /&gt;
The Digital Public Library of America aggregates metadata for over 8 million objects from more than 24 direct partners, or Hubs, using its Metadata Application Profile (MAP), an RDF metadata application profile based on the Europeana Data Model. After working with the initial system for harvesting, mapping and enhancing our Hub’s metadata for a year, we realized that it was inadequate for working with data at this scale. There were architectural issues; it was opaque to non-developer and partner staff; there were inadequate tools for quality assurance and analysis; and the system was unaware that it was working with RDF data. As the network of Hubs expanded and we ingested more metadata, it became harder and harder to know when or why a harvest, a mapping task, or an enrichment went wrong because the tools for quality assurance were largely inadequate. &lt;br /&gt;
&lt;br /&gt;
The DPLA Content and Technology teams decided to develop a new system from the ground up to address those problems. Development of Heidrun, the internal version of the new system, started in October 2014. Heidrun’s goals are to make it easier for us to harvest and map metadata from various sources and in variety of schemas to the DPLA MAP, to better enrich that metadata using external data sources, and to actively involve our partners in the ingestion process through access to better QA tools. Heidrun and its componentry are built on Ruby on Rails, Blacklight, and ActiveTriples. Our presentation will give some background on our design principles and processes used during development, the architecture of the system, and its functionality. We plan to release a version of Heidrun and its components as a generalized metadata aggregation system for use by DPLA Hubs and others working to aggregate cultural heritage metadata.&lt;br /&gt;
&lt;br /&gt;
== OS or GTFO: Program or Perish ==&lt;br /&gt;
*Tessa Fallon, tessa.fallon@gmail.com&lt;br /&gt;
&lt;br /&gt;
Description TBD&lt;br /&gt;
&lt;br /&gt;
== Creating Dynamic— and Cheap!— Digital Displays with HTML 5 Authoring Software ==&lt;br /&gt;
* Chris Woodall, cmwoodall@salisbury.edu, Salisbury University Libraries&lt;br /&gt;
Would your library like to have large digital signage that displays dynamic information such as library hours, weather, room availability, and more? Have you looked into purchasing large digital signage, only to be turned off by the high price tag and lack of customization available with commercial solutions? Our library has developed a cheap and effective alternative to these systems using HTML 5 authoring software, a large TV, and freely-available APIs from Google, Springshare, and others. At this session, you’ll learn about the system that we have in place for displaying dynamic and easily-updatable information on our library’s large digital display, and how you can easily create something similar for your library.&lt;br /&gt;
&lt;br /&gt;
== REPOX: Metadata Blender ==&lt;br /&gt;
 &lt;br /&gt;
* John Mignault, jmignault@metro.org, Empire State Digital Network&lt;br /&gt;
&lt;br /&gt;
With the growth in the number of hubs providing metadata to the Digital Public Library of America, many of them are using REPOX, a tool originally created for the Europeana project, to aggregate disparate metadata feeds and transform them into formats suitable for ingest into DPLA. The Empire State Digital Network, the forthcoming DPLA service hub for NY state, is using it to prepare for our first ingest into DPLA in early 2015.  We'll take a look at REPOX and its capabilities and how it can be useful for ingesting and transforming metadata, and also discuss some things we've learned in massaging widely varied metadata feeds.&lt;br /&gt;
&lt;br /&gt;
== Beyond Open Source ==&lt;br /&gt;
&lt;br /&gt;
* Jason Casden, jmcasden@ncsu.edu, NCSU Libraries&lt;br /&gt;
* Bret Davidson, bddavids@ncsu.edu, NCSU Libraries&lt;br /&gt;
&lt;br /&gt;
The Code4Lib community has produced an increasingly impressive collection of open source software over the last decade, but much of this creative work remains out of reach for large portions of the library community. Do the relatively privileged institutions represented by a majority of Code4Lib participants have a professional responsibility to support the adoption of their innovations?&lt;br /&gt;
&lt;br /&gt;
Drawing from old and new software packaging and distribution approaches (from freeware to Docker), we will propose extending the open source software values of collaboration and transparency to include the wide and affordable distribution of software. We believe this will not only simplify the process of sharing our applications within the Code4Lib community, but also make it possible for less well resourced institutions to actually use our software. We will identify areas of need, present our experiences with the users of our own open source projects, discuss our attempts to go beyond open source, and make an argument for the internal value of supporting and encouraging a vibrant library ecosystem.&lt;br /&gt;
&lt;br /&gt;
[[Category:Code4Lib2015]] &lt;br /&gt;
[[Category:Talk Proposals]]&lt;br /&gt;
&lt;br /&gt;
== Making It Work: Problem Solving Using Open Source at a Small Academic Library ==&lt;br /&gt;
 &lt;br /&gt;
* Adam Strohm, astrohm@iit.edu, Illinois Institute of Technology&lt;br /&gt;
* Max King, mking9@iit.edu, Illinois Institute of Technology&lt;br /&gt;
&lt;br /&gt;
The Illinois Institute of Technology campus was added to the National Register of Historic Places in 2005, and contains a building, Mies van der Rohe's S.R. Crown Hall, that was named a National Historic Landmark in 2001. Creating a digital resource that can adequately showcase the campus and its architecture is challenge enough in and of itself, but doing so as a two-person team of relative newcomers, at a university library without dedicated programmers on staff, ups the ante considerably.&lt;br /&gt;
The challenges of technical know-how, staff time, and funding are nothing new to anyone working on digital projects at a university library, and are amplified when doing so at a smaller institution. This talk covers the conception, development, and design of the campus map site that was built, concentrating on the problem-solving strategies developed to cope with limited technical and financial resources.&lt;br /&gt;
We'll talk about our approach to development with Open Source software, including Omeka, along with the Neatline and Simile Timeline plugins. We'll also discuss the juggling act of designing for mobile mapping functionality without sacrificing desktop design, weighing the costs of increased functionality versus our ability to time-effectively include that functionality, and the challenge of building a site that could be developed iteratively, with an eye towards future enhancement and sustainability. Finally, we’ll provide recommendations for other librarians at smaller institutions for their own efforts at digital development.&lt;br /&gt;
&lt;br /&gt;
== Recording Digitization History: Metadata Options for the Process History of Audiovisual Materials ==&lt;br /&gt;
 &lt;br /&gt;
* Peggy Griesinger, peggy_griesinger@moma.org, Museum of Modern Art&lt;br /&gt;
&lt;br /&gt;
The Museum of Modern Art has amassed a large collection of audiovisual materials over its many decades of existence. In order to preserve these materials, much of the audiovisual collection has been digitized. This is a complex process involving numerous steps and devices, and the methods used for digitization can have an effect on the quality of the file that is preserved. Therefore, knowing exactly how something was digitized is critical for future stewards of these objects to be able to properly care for and preserve them. However, detailed technical information about the processes involved in the digitization of audiovisual materials is not defined explicitly in most metadata schemas used for audiovisual materials. In order to record process history using existing metadata standards, some level of creativity is required to allow existing standards to express this information.&lt;br /&gt;
&lt;br /&gt;
This talk will detail different metadata standards, including PBCore, PREMIS, and reVTMD, that can be implemented as methods of recording this information. Specifically, the talk will examine efforts to integrate this metadata into the Museum of Modern Art’s new digital repository, the DRMC. This talk will provide background on the DRMC as well as MoMA’s specific institutional needs for process history metadata, then discuss different metadata implementations we have considered to document process history.&lt;br /&gt;
&lt;br /&gt;
== Pig Kisses Elephant: Building Research Data Services for Web Archives ==&lt;br /&gt;
 &lt;br /&gt;
* Jefferson Bailey,  jefferson@archive.org, Internet Archive&lt;br /&gt;
* Vinay Goel, vinay@archive.org, Internet Archive&lt;br /&gt;
&lt;br /&gt;
More and more libraries and archives are creating web archiving programs.  For both new and established programs, these archives can consist of hundreds of thousands, if not millions, of born-digital resources within a single collection; as such, they are ideally suited for large-scale computational study and analysis. Yet current access methods for web archives consist largely of browsing the archived web in the same manner as browsing the live web and the size of these collections and complexity of the WARC format can make aggregate analysis difficult. This talk will describe a project to create new ways for users and researchers to access and study web archives by offering extracted and post-processed datasets derived from web collections. Working with the 325+ institutions and their 2600+ collections within the Archive-It service, the Internet Archive is building methods to deliver a variety of datasets culled from collections of web content, including extracted metadata packaged in JSON, longitudinal link graph data, named entities, and other types of data. The talk will cover the technical details of building dataset production pipelines with Apache Pig, Hadoop, and tools like Stanford NER, the programmatic aspects of building data services for archives and researchers, and ongoing work to create new ways to access and study web archives.&lt;br /&gt;
&lt;br /&gt;
== Awesome Pi, LOL! ==&lt;br /&gt;
&lt;br /&gt;
* Matt Connolly, mconnolly@cornell.edu, Cornell University Library&lt;br /&gt;
* Jennifer Colt, jrc88@cornell.edu, Cornell University Library&lt;br /&gt;
&lt;br /&gt;
Inspired by Harvard Library Lab’s “Awesome Box” project, Cornell’s Library Outside the Library (LOL) group is piloting a more automated approach to letting our users tell us which materials they find particularly stunning. Armed with a Raspberry Pi, a barcode scanner, and some bits of kit that flash and glow, we have ventured into the foreign world of hardware development. This talk will discuss what it’s like for software developers and designers to get their hands dirty, how patrons are reacting to the Awesomizer, and LOL’s not-afraid-to-fail philosophy of experimentation.&lt;br /&gt;
&lt;br /&gt;
== You Gotta Keep 'em Separated: The Case for &amp;quot;Bento Box&amp;quot; Discovery Interfaces ==&lt;br /&gt;
 &lt;br /&gt;
* Jason Thomale,  jason.thomale@unt.edu, University of North Texas Libraries&lt;br /&gt;
&lt;br /&gt;
I know, I know--proposing a talk about Resource Discovery is like, ''so'' 2010.&lt;br /&gt;
&lt;br /&gt;
The thing is, practically all of us--in academic libraries at least--have a similar set up for discovery, with just a few variations, and so talking about it still seems useful. Stop me if this sounds familiar. You've got a single search box on the library homepage as a starting point for discovery. And it's probably a tabbed affair, with an option for searching the catalog for books, an option for searching a discovery service for articles, an option for searching databases, and maybe a few others. Maybe you have an option to search everything at once--probably the default, if you have it. And, if you're a crazy hepcat, maybe you ''only'' have your one search that searches everything, with no tabs.&lt;br /&gt;
&lt;br /&gt;
Now, the question is, for your &amp;quot;everything&amp;quot; search, are you doing a combined list of results, or are you doing it bento-box style, with a short results list from each category displayed in its own compartment?&lt;br /&gt;
&lt;br /&gt;
At UNT, we've been holding off on implementing an &amp;quot;everything&amp;quot; search, for various reasons. One reason is that the evidence for either style hasn't been very clear. There's this persistent paradox that we just can't reconcile: users tell us, through word and action, that they prefer searching Google, yet, libraries aren't Google, and there are valid design reasons why we shouldn't try to oversimplify our discovery interfaces to be like Google. And there's user data that supports both sides.&lt;br /&gt;
&lt;br /&gt;
Holding off on making this decision has granted us 2 years of data on how people use our tabbed search interface that does ''not'' include an &amp;quot;everything&amp;quot; search. Recently I conducted a thorough analysis of this data--specifically the usage and query data for our catalog and discovery system (Summon). And I think it helps make the case for a bento box style discovery interface. To be clear, it isn't exactly the smoking gun that I was hoping for, but the picture it paints I think is telling. At the very least, it points away from a combined-results approach.&lt;br /&gt;
&lt;br /&gt;
I'm proposing a talk discussing the data we've collected, the trends we've seen, and what I think it all means--plus other reasons that we're jumping on the &amp;quot;bento box&amp;quot; discovery bandwagon and why I think &amp;quot;bento box&amp;quot; is at this point the path that least sells our souls.&lt;br /&gt;
&lt;br /&gt;
== Don’t know about you, but I’m feeling like SHA-2!: Checksumming with Taylor Swift ==&lt;br /&gt;
 &lt;br /&gt;
* Ashley Blewer!, ashley.blewer@gmail.com&lt;br /&gt;
&lt;br /&gt;
Checksum technology is used all over the place, from git commits to authenticating Linux packages. It is most commonly used in the digital preservation field to monitor materials in storage for changes that will occur over time or used in the transmission of files during duplication. But do you even checksum, bro? I want this talk to move checksums from a position of mysterious macho jargon to something everyone can understand and want to use. I think a lot of people have heard of checksum but don’t know where to begin when it comes to actually using it at their institution. And cryptography is hella intimidating! This talk will cover what checksums are, how they can be integrated into a library or archival workflow, protecting collections requiring additional levels of security, algorithms used to verify file fixity and how they are different, and other aspects of cryptographic technology. Oh, and please note that all points in this talk will be emphasized or lightly performed through Taylor Swift lyrics. Seriously, this talk will consist of at least 50% Taylor Swift. Can you, like, even?&lt;br /&gt;
&lt;br /&gt;
== Level Up Your Coding with Code Club (yes, you can talk about it) ==&lt;br /&gt;
&lt;br /&gt;
* Coral Sheldon-Hess, coral@sheldon-hess.org&lt;br /&gt;
&lt;br /&gt;
Reading code is a necessary part of becoming a better developer. It gives you more experience and more insight into How Things Are (or Aren't) Done; it builds your intuition about how to solve problems with code; and it increases your confidence that you, too, can tackle whatever technological problems you're facing.&lt;br /&gt;
&lt;br /&gt;
But you don't have to read code alone! (Which is good. It's really not fun to read code alone.) &lt;br /&gt;
&lt;br /&gt;
In late 2014, a group of librarians formed two Code Clubs, inspired by [http://bloggytoons.com/code-club/ this talk by Saron] (of Bloggytoons fame). I'd like to tell you about how we've structured our Code Clubs, what has gone well, what we've learned, and what you need to do to form your own Code Club. I'll share a list of the codebases we've looked at, too, to help you get your own Code Club off the ground! &lt;br /&gt;
&lt;br /&gt;
== The Growth of a Programmer ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:jgo | Joshua Gomez]], Getty Research Institute, jgomez@getty.edu&lt;br /&gt;
&lt;br /&gt;
Just like other creative endeavors, software developers can experience periods of great productivity or find themselves in a rut. After contemplating the alternating periods in my own career I've noticed several factors that have effected my own professional growth and happiness, including: mentorship, structure, community, teamwork, environment, formal education, etc. Not all of the factors need to be present at all times; but some mixture of them is critical for continued growth. In this talk, I will articulate these factors, discuss how they can effect a developer's career, and how they can be sought out when missing. This talk is aimed at both new developers looking to strike their own path as well as the veterans that lead or mentor them.&lt;br /&gt;
&lt;br /&gt;
== Developing a Fedora 4.0 Content Model for Disk Images ==&lt;br /&gt;
&lt;br /&gt;
* Matthew Farrell, matthew.j.farrell@duke.edu, Duke University Libraries&lt;br /&gt;
* Alexandra Chassanoff, achass@email.unc.edu, BitCurator Access Project Manager&lt;br /&gt;
&lt;br /&gt;
As the acquisition of born-digital materials grows, institutions are seeking methods to facilitate easy ingest into their repositories and provide access to disk images and files derived or extracted from disk images. In this session, we describe our development of a Fedora 4.0 Content model for disk images, including acceptable image file formats and the rationale behind those choices.  We will also discuss efforts to integrate the disk image content model into the BitCurator Access environment. Unlike generalized, format-agnostic content models which might treat the disk image as a generic bitstream, a content model designed for disk images enables expression of relationships among associated content in the collection such as files extracted from images and other born-digital and digitized material associated with the same creator.  It also enables capture of file-system attributes such as file paths, timestamps, whether files are allocated/deleted, etc.  Further, a disk image content model suggests further steps repositories can take in order to transform and re-use associated metadata generated during the creation and forensic analysis of the disk image.&lt;br /&gt;
&lt;br /&gt;
== Data acquisition and publishing tools in R ==&lt;br /&gt;
&lt;br /&gt;
* Scott Chamberlain,  scott@ropensci.org, rOpenSci/UC Berkeley - first-time presenter&lt;br /&gt;
&lt;br /&gt;
R is an open source programming environment that is widely used among researchers in many fields. R is powerful because it's free, increasingly robust, and facilitates reproducible research, an increasingly sought after goal in academia. Although tools for data manipulation/visualization/analysis are well developed in R, data acquisition and publishing tools are not. rOpenSci is a collaborative effort to create the tools necessary to complete the reproducible research workflow. This presentation discusses the need for these tools, including examples, including interacting with the repositories Mendeley, Dryad, DataONE, and Figshare. In addition, we are building tools for searching scholarly metadata and acuiring full text of open access articles in a standarized way across metadata providers (e.g., Crossref, DataCite, DPLA) and publishers (e.g., PLOS, PeerJ, BMC, Pubmed). Last, we are building out tools for data reading and writing in Ecologial Metadata Language (EML).&lt;br /&gt;
&lt;br /&gt;
== SPLUNK: Log File Analysis ==&lt;br /&gt;
&lt;br /&gt;
* Jim LeFager, jlefager@depaul.edu, DePaul University Library&lt;br /&gt;
DePaul University Library recently took over monitoring and maintaining of the library EZproxy servers this past year and using Splunk, a machine data analysis tool, we are able to gather information and statistics on our electronic resource usage in addition to monitoring the servers. Splunk is a tool that can collect, analyze, and visualize log files and other machine data in real time and this has allowed for gathering realtime usage statistics for our electronic resources allowing us to filter by multiple facets including IP Range, Group Membership (student, faculty), so that we can see who is accessing our resources and from where. Splunk allows our library to query our data and create rich custom dashboards as well as create alerts that can be triggered when certain conditions are met, such as error codes, which can send an email alert to a group of users. We will be leveraging Splunk to monitor all library web applications going forward. This talk will review setting up Splunk and best practices in using the available features and customizations available including creating queries, alerts, and custom dashboards.  &lt;br /&gt;
&lt;br /&gt;
== Your code does not exist in a vacuum ==&lt;br /&gt;
* Becky Yoose, yoosebec at grinnell dot edu, Grinnell College (Done a lightning talk, MC duties, but have not presented a prepared talk)&lt;br /&gt;
&lt;br /&gt;
“If you have something to say, then say it in code…” - Sebastian Hammer, code4lib 2009&lt;br /&gt;
&lt;br /&gt;
In its 10 year run, code4lib has covered the spectrum of libtech development, from search to repositories to interfaces. However, during this time there has been little discussion about this one little fact about development - code does not exist in a vacuum. &lt;br /&gt;
&lt;br /&gt;
Like the comment above, code has something to say. A person’s or organization’s culture and beliefs influences code in all steps of the development cycle. What development method you use, tools, programming languages, licenses - everything is interconnected with and influenced by the philosophies, economics, social structures, and cultural beliefs of the developer and their organization/community.&lt;br /&gt;
&lt;br /&gt;
This talk will discuss these interconnections and influences when one develops code for libraries, focusing on several development practices (such as “Fail Fast, Fail Often” and Agile)   and licensing choices (such as open source) that libtech has either tried to model or incorporate into mainstream libtech practices. It’ll only scratch the surface of the many influences present in libtech development, but it will give folks a starting point to further investigate these connections at their own organizations and as a community as a whole.&lt;br /&gt;
&lt;br /&gt;
tl;dr - this will be a messy theoretical talk about technology and libraries. No shiny code slides, no live demos. You might come out of this talk feeling uncomfortable. Your code does not exist in a vacuum. Then again, you don’t exist in a vacuum either.&lt;br /&gt;
&lt;br /&gt;
== The Metadata Hopper: Mapping and Merging Metadata Standards for Simple, User-Friendly Access ==&lt;br /&gt;
&lt;br /&gt;
* Tracy Seneca, tjseneca@uic.edu, University of Illinois at Chicago&lt;br /&gt;
* Esther Verreau: verreau1@uic.edu, University of Illinois at Chicago&lt;br /&gt;
&lt;br /&gt;
The Chicago Collections Consortium: 15 institutions and growing!  8 distinct EAD standards! At least 3 permutations of MARC, and we lost count of the varieties of custom CONTENTdm image collections.  Not to mention the 14,730 unique subject terms, nearly all of which lead our poor end-users to exactly one organization's content. &lt;br /&gt;
&lt;br /&gt;
All large content aggregation projects have faced this challenge, and there are a few emerging tools to help us wrangle disparate metadata into new contexts.  The Metadata Hopper is one such tool. The Metadata Hopper enables archivists to map their local metadata standards to standardized deposit records, and tags those materials using a shared vocabulary, integrating them into a user-friendly portal without disrupting local practices. In last year's Code4Lib lightning talk we described the challenges that the Chicago Collections Consortium faces in creating shared, in-depth access to archival and digital collections about Chicago history and culture across CCC member organizations. This year, thanks to the Andrew W. Mellon Foundation, we have a working Django application to demonstrate.  In this talk we'll discuss the design that enables multiple layers of flexibility, from the ability to accept a variety of metadata standards to designing for an open source audience.&lt;br /&gt;
&lt;br /&gt;
http://chicagocollectionsconsortium.org&lt;br /&gt;
&lt;br /&gt;
== Programmers are not projects: lessons learned from managing humans ==&lt;br /&gt;
&lt;br /&gt;
* Erin White, erwhite@vcu.edu, Virginia Commonwealth University - first-time presenter&lt;br /&gt;
&lt;br /&gt;
Managing projects is one thing, but managing people is another. Whether we’re hired as managers or grow “organically” into management roles, sometimes technical people end up leading technical teams (gasp!). I’ll talk about lessons I’ve learned about hiring, retaining, and working long-term and day-to-day with highly tech-competent humans. I’ll also talk about navigating the politics of libraryland, juggling different types of projects, and working with constrained budgets to make good things and keep talented people engaged.&lt;br /&gt;
&lt;br /&gt;
== Practical Strategies for Picking Low-Hanging Fruits to Improve Your Library's Web Usability and UX ==&lt;br /&gt;
&lt;br /&gt;
* Bohyun Kim, bkim@hshsl.umaryland.edu, University of Maryland, Baltimore&lt;br /&gt;
&lt;br /&gt;
Have you ever tried to fix an obvious (to you at least!) problem in Web usability or UX (user experience) only to face strong resistance from the library staff? Are you a strong advocate for making library resources, systems, services, and space as usable as possible, but do you often find yourself struggling to get the point across and/or obtain the crucial buy-in from colleagues and administrators? &lt;br /&gt;
&lt;br /&gt;
There is no shortage of Web usability and UX guidelines. But applying them to a library and implementing desired changes often involve a long and slow process. To tackle this issue, this talk will focus on how to utilize the 'expert review' process (aka 'heuristic evaluation') as a preliminary or even preparatory step before embarking on more time-and-labor-intensive usability testing and user research. Several examples from  simple fixes to more nuanced usability and UX issues in libraries will be discussed to your heart's content. The goal of this talk is to provide practical strategies for picking as many low-hanging fruits as possible to make a real (albeit small) difference to your library's Web usability and UX effectively and efficiently.&lt;br /&gt;
&lt;br /&gt;
== A Semantic Makeover for CMS Data ==&lt;br /&gt;
&lt;br /&gt;
* Bill Levay, wjlevay@gmail.com, Linked Jazz Project&lt;br /&gt;
&lt;br /&gt;
How can we take semi-structured but messy metadata from a repository like CONTENTdm and transform it into rich linked data? Working with metadata from Tulane’s Hogan Jazz Archive Photography Collection, the Linked Jazz Project used Open Refine and Python scripts to tease out proper names, match them with name authority URIs, and specify FOAF relationships between musicians who appear together in photographs. Additional RDF triples were created for any dates associated with the photos, and for those images with place information we employed GeoNames URIs. Historical images and data that were siloed can now interact with other datasets, like Linked Jazz’s rich set of names and personal relationships, and can be visualized [link to come] or otherwise presented on the web in any number of ways. I have not previously presented at a Code4Lib conference.&lt;br /&gt;
&lt;br /&gt;
== Taking User Experience (UX) to new heights ==&lt;br /&gt;
 &lt;br /&gt;
* Kayne Richens, kayne.richens@deakin.edu.au, Deakin University&lt;br /&gt;
&lt;br /&gt;
User Experience, or &amp;quot;UX&amp;quot;, is for more than just websites. At Deakin University Library we're exploring ways to improve the user experience inside our campus library spaces, by putting new technologies front and centre in the overall experience for our students. How are we doing this? We’re collaborating with the University's IT department and exploring the following Library-changing opportunities:&lt;br /&gt;
&lt;br /&gt;
- Augmented Reality for Way-finding: We’re tackling that infamous thing that all Libraries can't get right – way-finding. We're enhancing library tour information and way-finding experiences by introducing augmented reality solutions.&lt;br /&gt;
 &lt;br /&gt;
- Heat mapping the library with wi-fi: We’re using our existing wi-fi infrastructure to present &amp;quot;heat maps&amp;quot; of library space utilisation, allowing our users to easily locate the space that best suits their needs, whether it be busy spaces to collaborate, or quiet spaces to study. And by overlaying computer usage and group study room bookings, users can quickly locate the space they need.&lt;br /&gt;
 &lt;br /&gt;
- Video chat library service: We’re piloting video-conferencing facilities in our group study rooms and spaces, connecting users and librarians and other professionals.&lt;br /&gt;
         &lt;br /&gt;
This talk will look at how these different technologies will be brought together to provide improved user experiences, as well some of the evidence and reasons that helped us to identify our needs, so you can too.&lt;br /&gt;
&lt;br /&gt;
==How to Hack it as a Working Parent: or, Should Your Face be Bathed in the Blue Glow of a Phone at 2 AM?==&lt;br /&gt;
&lt;br /&gt;
*Margaret Heller, Loyola University Chicago, mheller1@luc.edu&lt;br /&gt;
*Christina Salazar, California State University Channel Islands, christina.salazar@csuci.edu&lt;br /&gt;
*May Yan, Ryerson University, may.yan@ryerson.ca&lt;br /&gt;
&lt;br /&gt;
Modern technology has made it easier than ever for parents employed in technical environments to keep up with work at all hours and in all locations. This makes it possible to work a flexible schedule, but also may lead to problems with work/life balance and furthering unreasonable expectations about working hours. Add to that shifting gender roles and limited paid parental leave in the United States and you have potential for burnout and a certainty for anxiety. It raises the additioal question of whether the “always connected” mindset puts up a barrier to some populations who otherwise might be better represented in open source and library technology communities. &lt;br /&gt;
&lt;br /&gt;
This presentation will address tools that are useful for working parents in technical library positions, and share some lessons learned about using these tools while maintaining a reasonable work/life balance. We will consider a question that Karen Coyle raised back in 1996: &lt;br /&gt;
“What if the thousands of hours of graveyard shift amateur hacking wasn't really the best way to get the job done? That would be unthinkable.” &lt;br /&gt;
&lt;br /&gt;
For those who are able to take an extended parental leave, we will present strategies for minimizing the impact to your career and your employer. For those (particularly in the United States) who are only able to take a short leave will require different strategies. Despite different levels of preparation, all are useful exercises in succession planning and making a stronger workplace and future ability to work a flexible schedule through reviewing workloads, cross-training personnel, hiring contract replacements, and creative divisions of labor. Such preparation makes work better for everyone, kids or no kids or caretakers of any kind.&lt;br /&gt;
&lt;br /&gt;
==Making your digital objects embeddable around the web==&lt;br /&gt;
 &lt;br /&gt;
* Jessie Keck, jkeck@stanford.edu, Stanford University Libraries&lt;br /&gt;
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
With more and more content from our digital repositories making their way into our discovery environments we quickly realize that we’re repeatedly re-inventing the wheel when it comes to creating “Viewers” for these digital objects.  With various different types of viewers necessary (books, images, audio, video, geospatial data, etc) the burden of getting these viewers into various environments (topic guides, blogs, catalogs, etc) becomes exponential.&lt;br /&gt;
&lt;br /&gt;
In this talk we’ll discuss how Stanford University Libraries implemented an oEmbed service to create an extensible viewer framework for all of its digital content. Using this service we’ve been able to easily integrate viewers into various discovery applications as well as make it easy for end users who discover our objects to easily embed customized versions into their own websites and blogs.&lt;br /&gt;
&lt;br /&gt;
==So you want to make your geospatial data discoverable==&lt;br /&gt;
 &lt;br /&gt;
* Jack Reed, pjreed@stanford.edu, Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Finding data for research or coursework can be one of the most time intensive tasks for a scholar or student. We introduce GeoBlacklight, an open source, multi-institutional software project focused on solving these common challenges at institutions across the world. GeoBlacklight prioritizes user experience, integrates with many GIS tools, and streamlines the use and organization of geospatial data. This talk will provide an introduction to the software, demonstrate current functionality, and provide a road map for future work.&lt;br /&gt;
&lt;br /&gt;
== Clueless-Driven Development: How I learned to migrate to Fedora 4 ==&lt;br /&gt;
&lt;br /&gt;
* Adam Wead, awead@psu.edu, Penn State University&lt;br /&gt;
&lt;br /&gt;
Recently I was tasked with migrating the content from our Fedora3 repository to the new Fedora4 repository architecture.&lt;br /&gt;
Despite a wealth of community support, I had no idea how to approach, or even begin to solve this problem. I knew I&lt;br /&gt;
wanted to follow best practices and use test-driven  development to build my solution, but had no idea where to start.&lt;br /&gt;
Despite this initial setback, I was able to start writing tests with only a  vague understanding of the problem. As my&lt;br /&gt;
tests exposed where my understanding of the problem was flawed, my code evolved, and within a week I had arrived  at a&lt;br /&gt;
working solution that exhibited all the hallmarks of good testing and software design.&lt;br /&gt;
&lt;br /&gt;
This talk recounts the process I went through from starting with practically nothing, to arriving at a working solution.&lt;br /&gt;
You can follow the rules of  test-driven development, but you can write tests in an expressive way to describe the&lt;br /&gt;
problem instead of just describing what the code should do. It was also essential to begin testing from an integration&lt;br /&gt;
viewpoint as opposed to a unit one, because at the outset the units were unknown and were later realized through further&lt;br /&gt;
development. For the presentation, I will be demonstrating using RSpec and Ruby. All the code examples will be related&lt;br /&gt;
to the Hydra software stack; however, I hope to show  that the processes at work will be applicable in any context.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Designing and Leading a Kick A** Tech Team ==&lt;br /&gt;
 &lt;br /&gt;
* Sibyl Schaefer,  sschaefer@rockarch.org, Rockefeller Archive Center&lt;br /&gt;
&lt;br /&gt;
New managers are often promoted without receiving management training, yet management is not something you just figure out. The experience of being expected to know how to manage, yet not being trained to do so often results in new managers feeling isolated and unsure how to move from making to managing. In this talk I’ll focus on my own managerial experience of designing and leading an archival tech team in a small independent archives. Topics covered will include hiring, delegating, creating a team culture, and leading people whose specialized knowledge exceeds your own. The talk take-aways should be applicable to managers and employees at large and small institutions alike.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==American (Archives) Horror Story: LTO Failure and Data Loss ==&lt;br /&gt;
 &lt;br /&gt;
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH&lt;br /&gt;
* Casey Davis, casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH&lt;br /&gt;
&lt;br /&gt;
Here’s a story to send shivers down archival spines: when transferring video files off LTO for the American Archive project, WGBH got an initial failure rate of 57%.   After repeat tries, the rates improved; still, an unnervingly large percentage of files were never able to be transferred successfully.   Even more unnerving, going public with our horror story got a big response from other archives using LTO -- it seems like many institutions are having similarly scary results.   What are the real risks with LTO tape?  Are there steps that archives should be taking to better circumvent those risks?  This presentation will share information about LTO storage failures across archives world and discuss the process of investigating the problem at WGBH by testing different methods of data retrieval from LTO (direct and networked downloads, individual file retrieval and bulk data dump, use of LTO 4 and LTO 6 decks) and using checksum comparisons and file analysis and characterization tools such as ffprobe, mediainfo and exiftool to analyze failed files.  We'll also present whatever results we’ve managed to turn up by the time of Code4Lib!&lt;br /&gt;
&lt;br /&gt;
== PBCore in Action: Three Words, Not Two! ==&lt;br /&gt;
 &lt;br /&gt;
* Casey E. Davis,  casey_davis@wgbh.org, Project Manager, American Archive of Public Broadcasting, WGBH&lt;br /&gt;
* Andrew (Drew) Myers, andrew_myers@wgbh.org, Supervising Developer, WGBH&lt;br /&gt;
&lt;br /&gt;
In 2001, public media representatives developed the PBCore XML schema to establish a common language for managing metadata about their analog and digital audio and video. Since then, PBCore has been adopted by a number of organizations and archivists in the moving image archival community. The schema has also undergone a few revisions, but on more than one occasion it was left orphaned and with little to no support.&lt;br /&gt;
 &lt;br /&gt;
Times have changed. You may have heard the news that PBCore is back in action as part of the American Archive of Public Broadcasting initiative and via the Association of Moving Image Archivists (AMIA) PBCore Advisory Subcommittee. A group of archivists, public media stakeholders, and engaged users have come together to provide necessary, sustaining support for the standard and to see to its further development. &lt;br /&gt;
 &lt;br /&gt;
At this session, we'll discuss the scope and uses of PBCore in digital preservation and access, report on the progress and goals of the PBCore Advisory Subcommittee, and share how the group (by the time of the conference) will have transformed the XML schema into an RDF ontology, bringing PBCore into the second decade of the 21st century. #PBHardcore&lt;br /&gt;
&lt;br /&gt;
==Collaborating to Avert the Digital Graveyard==&lt;br /&gt;
&lt;br /&gt;
* Harish Nayak, hnayak@library.rochester.edu, University of Rochester Libraries &lt;br /&gt;
* Sean Morris, smorris@library.rochester.edu, University of Rochester Libraries &lt;br /&gt;
&lt;br /&gt;
In 1995, the Robbins Library at the University of Rochester created a digital collection of Arthurian texts, images, and bibliographies. Together with medieval scholars, we recently completed the redesign and development of an interface for this collection. Using FRBR concepts, we re-conceptualized organization and editing workflow from the ground up in a mobile-first Drupal-based project. &lt;br /&gt;
&lt;br /&gt;
In this talk we will describe the project as well as how we utilized the techniques of work practice study and user centered design to maintain engagement with reluctant stakeholders, nontechnical scholars, and VERY meticulous graduate students.  Neither of us have previously presented at a Code4Lib conference.&lt;br /&gt;
&lt;br /&gt;
==Docker? VMs? EC2? Yes! With Packer.io==&lt;br /&gt;
&lt;br /&gt;
* Kevin S. Clarke, ksclarke@gmail.com, Digital Library Programmer, UCLA&lt;br /&gt;
&lt;br /&gt;
There are a lot of exciting ways to deploy a software stack nowadays. Many of our library systems are fully virtualized. Docker is a compelling alternative, and there are also cloud options like Amazon's EC2. This talk will introduce Packer.io, a tool for creating identical machine images for multiple platforms (e.g., Docker, VMWare, VirtualBox, EC2, GCE, OpenStack, et al.) all from a single source configuration.  It works well with Ansible, Chef, Puppet, Salt, and plain old Bash scripts. And, it's designed to be scriptable so that builds can be automated. This presentation will show how easy it is to use Packer.io to bring up a set of related services like Fedora 4, Grinder (for stress testing), and Graphite (for charting metrics). As an added value, all the buzzwords in this proposal will be defined and explained!&lt;br /&gt;
&lt;br /&gt;
== Technology on your Wrist: Cross-platform Smartwatch Development for Libraries ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:sanderson|Steven Carl Anderson]], sanderson@bpl.org, Boston Public Library (no previously accepted prepared talks but have done lightning talks in the past)&lt;br /&gt;
&lt;br /&gt;
I'll be the first to admit: smartwatches are unlikely to completely revolutionize how a library provides online services. But I believe they still represent an opportunity to further enhance existing library services and resources in a unique way.&lt;br /&gt;
&lt;br /&gt;
At the Boston Public Library (BPL), we're in the initial phases of designing a modest smartwatch app to provide notifications for circulation availability and checked-out-material due-date alerts by the end of current year. We're starting small, but we plan to evolve the concept over time as we see what (if any) traction such an application gets with potential users. For example, we plan to explore the possibility of adding &amp;quot;nearest branch to my current location&amp;quot; functionality to this app.&lt;br /&gt;
&lt;br /&gt;
Despite the &amp;quot;development phase&amp;quot; of this application as of this writing, this talk is not being given by a novice. As a technology enthusiast, I've released [http://www.phdgaming.com/smartwatch_projects/ five smartwatch applications] and have had two of those be finalists in a [http://www.phdgaming.com/samsung_challenge/ Samsung sponsored development challenge]. This experience not only will allow for the BPL to avoid many beginner mistakes in its smartwatch app development but also gives a much more complete understanding of the smartwatch development ecosystem.&lt;br /&gt;
&lt;br /&gt;
This talk will explore the following questions:&lt;br /&gt;
&lt;br /&gt;
* What kinds of online library services could potentially be transformed or translated into the smartwatch/wearable domain? What kinds of services are better left alone? These questions are currently being explored and I'll talk about our plans and experiences. Included will be any statistical information from our application launch along with statistics from my personal development.&lt;br /&gt;
&lt;br /&gt;
* How to support all the different operating systems these devices run without painful modifications to your codebase. (There's Tizen that is used by Samsung's Gear 2 and Gear S, Android Wear that is used by most other non-Apple manufacturers, then there is Apple's upcoming smartwatch itself, etc.)&lt;br /&gt;
&lt;br /&gt;
* How to support different screen resolutions on such a small device. From round to rectangular to perfectly square, smartwatches come in all different shapes these days.&lt;br /&gt;
&lt;br /&gt;
* What are the app stores like on these platforms? As I support multiple applications through different distribution networks, a guide to navigating how to distribute one's app is included and I'll reveal how these systems work “behind the curtain.”&lt;br /&gt;
&lt;br /&gt;
* What are common issues and pitfalls to avoid when doing development? Tips on broken APIs and how to cope or optimizing your code will be included.&lt;br /&gt;
&lt;br /&gt;
==Seeing the Forest From the Trees: The Art of Creating Workflows for Digital Projects ==&lt;br /&gt;
 &lt;br /&gt;
* Jen LaBarbera, j.labarbera@neu.edu, NDSR Resident, Northeastern University&lt;br /&gt;
* Joey Heinen, joseph_heinen@harvard.edu, NDSR Resident, Harvard University&lt;br /&gt;
* Rebecca Fraimow, rebecca_fraimow@wgbh.org, NDSR Resident, WGBH&lt;br /&gt;
* Tricia Patterson, triciap@mit.edu, NDSR Resident, MIT&lt;br /&gt;
&lt;br /&gt;
We have to &amp;quot;turn projects into programs&amp;quot; in order to create a solid and sustainable digital preservation initiative...but what the heck does that even mean? What does that look like?&lt;br /&gt;
&lt;br /&gt;
In this talk, members of the inaugural Boston cohort of the National Digital Stewardship Residency will discuss one piece of our digital preservation test kitchen: our stabs at creating digital workflows that will (hopefully) help our institutions turn digital preservation projects into programs. Specifically, we will talk about how difficult it is to create a general and overarching workflow for digital preservation tasks (e.g. ingest into repositories, format migrations, etc.) that incorporates various technical tools while also taking into account the myriad and unending list of possible exceptions or special scenarios. Turning these complicated, specific processes into a simplified and generalized workflow is an art. We haven't necessarily perfected that art yet, but in this talk, we'll share what has worked for us -- and what hasn't. We’ll also touch on the importance of documentation, and achieving that delicate balance of adequately thorough documentation that doesn’t pose the risk of information avalanche. These processes often create more questions than answers, but we'll share the answers that we (and our mentors) have found along the way!&lt;br /&gt;
&lt;br /&gt;
== Annotations as Linked Data with Fedora4 and Triannon (a Real Use Case for RDF!) ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, azaroth@stanford.edu,  Stanford University Libraries&lt;br /&gt;
* Naomi Dushay, ndushay@stanford.edu,  Stanford University Libraries&lt;br /&gt;
&lt;br /&gt;
Annotations on content resources allow users to contribute knowledge within the digital repository space.  W3C Open Annotation provides a comprehensive model for web annotation on all types of content, using Linked Data as a fundamental framework.  Annotation clients generate instances of this model, typically using a JSON serialization, but need to store that data somewhere using a standard interaction pattern so that best of breed clients, servers, and data can be mixed and matched.&lt;br /&gt;
&lt;br /&gt;
Stanford is using Fedora4 for managing Open Annotations, via a middleware component called Triannon.  Triannon receives the JSON data from the annotation client, and uses the Linked Data Platform API implementation in Fedora4 to create, retrieve, update and delete the constituent resources.  Triannon could be easily modified to use other LDP implementations, or could be modified to work with linked data other than annotations.&lt;br /&gt;
&lt;br /&gt;
== Hydras in the Wild: A survey of current projects ==&lt;br /&gt;
 &lt;br /&gt;
* Mark Bussey,  mark@curationexperts.com, Data Curation Experts&lt;br /&gt;
&lt;br /&gt;
You've seen the tutorials, but [https://github.com/projecthydra/hydra/wiki/Dive-into-Hydra Dive Into Hydra] seems to leave something wanting.  What can you really do using the Hydra Framework?  This talks looks at a number of current Hydra projects and highlights the design and functional features unique to each. Compare and contrast UX, design and functional capabilites from a range of hydra-based repositories including:&lt;br /&gt;
* Avalon for media discovery and distribution (Indiana and Northwestern Univiersities) &lt;br /&gt;
* HydraDam for media archive management(WGBH)&lt;br /&gt;
* HyHull for general Institutional Repository needs (University of Hull)&lt;br /&gt;
* T-DIL for slide library functions (Tufts University)&lt;br /&gt;
* Sufia &amp;amp; Scholarsphere as a bundled self-deposit IR solution (Pennsylvania State University)&lt;br /&gt;
* Curate &amp;amp; Worthwhile as general purpose repository platforms (Multiple Insititutions)&lt;br /&gt;
This will be a whirlwind tour aimed at providing ideas and inspiration for your own repository development project.&lt;br /&gt;
&lt;br /&gt;
== Hydra Makeovers! ==&lt;br /&gt;
&lt;br /&gt;
* Alicia Cozine, alicia@curationexperts.com, Data Curation Experts&lt;br /&gt;
* Patrick Feeley, pgf8@case.edu, Case Western Reserve University&lt;br /&gt;
&lt;br /&gt;
Compare two Hydra-based applications with the systems they replaced. Marvel at the Before and After snapshots of functionality, speed, and look &amp;amp; feel. &lt;br /&gt;
* '''Digital Case 2.0''' is an institutional repository, complete with administrative tools, derivatives transcoding, flexible XML metadata storage, embargo and lease capability, faceted searching, and content viewers for texts/TEI, images, audio recordings, and videos. Digital Case 2.0 is based on worthwhile, an open-source IR starter gem. &lt;br /&gt;
* '''The Tufts Digital Image Library''' is a specialized tool for art and art history resources, offering image collections with user access controls, image ordering, collection nesting, drag-and-drop organization, slideshows, and export capability. &lt;br /&gt;
Both new systems are built on hydra, the open-source Ruby-on-Rails repository solution that incorporates Fedora for storage, Solr for indexing, and Blacklight for search optimization. Their beauty is not just skin-deep!&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Helping Google (and scholars, researchers, educators, &amp;amp; the public) find archival audio ==&lt;br /&gt;
&lt;br /&gt;
* Anne Wootton, anne@popuparchive.org, Pop Up Archive (www.popuparchive.org)&lt;br /&gt;
&lt;br /&gt;
Culturally significant digital audio collections are hard to discover on the web. There are major barriers keeping this valuable media from scholars, researchers, and the general public:&lt;br /&gt;
&lt;br /&gt;
Audio is opaque: you can’t picture sound, or skim the words in a recording. &lt;br /&gt;
Audio is hard to share: there’s no text to interact with. &lt;br /&gt;
Audio is not text: but since text is the medium of the web, there’s no path for audiences to find content-rich audio.&lt;br /&gt;
Audio metadata is inconsistent and incomplete.&lt;br /&gt;
&lt;br /&gt;
At Pop Up Archive, we're helping solve this problem making the spoken word searchable. We began as a UC-Berkeley School of Information Master's thesis to provide better access to recorded sound for audio producers, journalists, and historians. Today, Pop Up Archive processes thousands of hours of sound from all over the web to create automatic, timestamped transcripts and keywords, working with media companies and institutions like NPR, KQED, HuffPost Live, Princeton, and Stanford. We're building collections of sound from journalists, media organizations, and oral history archives from around the world. Pop Up Archive is supported by the John S. and James L. Knight Foundation, the National Endowment for the Humanities, and 500 Startups.&lt;br /&gt;
&lt;br /&gt;
== Digital Content Integrated with ILS Data for User Discovery:  Lessons Learned ==&lt;br /&gt;
&lt;br /&gt;
* Naomi Dushay, ndushay@stanford.edu,  Stanford University Libraries&lt;br /&gt;
* Laney McGlohon, laneymcg@stanford.edu,  Stanford University Libraries (first-time presenter)&lt;br /&gt;
&lt;br /&gt;
So you want to expose your digital content in your discovery interface, integrated with the data from your ILS?  How do you make the best information user searchable?  How do you present complete, up to date search results with a minimum of duplicate entries?&lt;br /&gt;
&lt;br /&gt;
At Stanford, we have these cases and more:&lt;br /&gt;
* digital content with no metadata in ILS&lt;br /&gt;
* digital content for metadata in ILS&lt;br /&gt;
* digital content with its own metadata derived from ILS metadata.&lt;br /&gt;
&lt;br /&gt;
We will describe our efforts to accommodate multiple updatable metadata sources for materials in the ILS and our Digital Object Repository while presenting users with reduced duplication in SearchWorks.  Included will be some failures, some successes, and an honest assessment of where we are now.&lt;br /&gt;
&lt;br /&gt;
== Show All the Things: Kanban for Libraries == &lt;br /&gt;
&lt;br /&gt;
* Mike Hagedon, mhagedon@email.arizona.edu, University of Arizona Libraries (first-time presenter)&lt;br /&gt;
&lt;br /&gt;
The web developers at the University of Arizona Libraries had a problem: we were working on a major website rebuild project with no clear way to prioritize it against our other work. We knew we wanted to follow Agile principles and initially chose Scrum to organize and communicate about our work. But we found that certain core pieces of Scrum did not work for our team. Then we discovered Kanban, an Agile meta-process for organizing work (team or individual) that treats the work more as a flow than as a series of fixed time boxes. I’ll be talking about our journey toward finding a process that works for our team and how we’ve applied the principles of Kanban to better get our work done. Specifically, I'll discuss principles like how to visualize all your work, how to limit how much you’re doing (to get more done!), and how to optimize the flow of your work.&lt;br /&gt;
&lt;br /&gt;
== DIY Catalog ==&lt;br /&gt;
 &lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
With so many discovery layers being made available that didn’t quite fit user needs but still came with substantial costs, Hennepin County Library determined it would develop one that could handle its customizations.  So they built their own ILS API, Web services, and an interface that provides all the features (and existing custom features) of the old catalog, with new features planned, all within a responsive layout.  Get a look at the system architecture, how the front end communicates through the layers back to the databases, and how to manage the infrastructure.&lt;br /&gt;
&lt;br /&gt;
== OverDrive: Full Integration ==&lt;br /&gt;
 &lt;br /&gt;
* Amy Drayer, adrayer at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
How do you increase usage of your ebooks?  Seamless integration in the catalog certainly helps. Users can search for digital titles, limit to just those that are available, place a request, borrow, and manage all their digital titles from the library website alongside all the other physical titles the library has.  Get a quick demo and learn how this was all made possible.&lt;br /&gt;
&lt;br /&gt;
The [//www.hclib.org Hennepin County Library] is ranked one of [//www.thedigitalshift.com/2014/10/ebooks/overdrive-api-usage-indicates-growth-ebook-checkouts-via-opac/ the top users of OverDrive and OverDrive’s API] and is looking to more fully integrate 3M Cloud, OneClickDigital, and other digital resources.&lt;br /&gt;
&lt;br /&gt;
== Dynamic Indexing: a Tragic Solr Story ==&lt;br /&gt;
 &lt;br /&gt;
* Wayne Schneider, wschneider at [//www.hclib.org hclib.org], [//www.hclib.org Hennepin County Library]&lt;br /&gt;
&lt;br /&gt;
Loading data from an ILS into Solr isn’t so hard, unless it needs to be dynamic, fast, and hold more data than what can be found in 1.5 million MARC records. Some additional information we’ve incorporated are from Syndetics, ILS circulation, and OverDrive.  We’ll share the nitty gritty details and what we learned about dynamic Solr indexing, including how to get good performance, how to deal with indexing failures, how to schedule it all to keep the data up-to-date, and some things you can do with that data such as popularity ratings.&lt;br /&gt;
&lt;br /&gt;
== Fedora 4, Survey of Core Capabilities and External Modules: == &lt;br /&gt;
* Andrew Woods, awoods@duraspace.org, DuraSpace&lt;br /&gt;
&lt;br /&gt;
Now that the production release of Fedora 4.0 is out, the time is right to take a step back and review the core capabilities offered by Fedora:&lt;br /&gt;
* Basic CRUD&lt;br /&gt;
* Versioning&lt;br /&gt;
*- Transactions&lt;br /&gt;
* etc&lt;br /&gt;
&lt;br /&gt;
...and take a closer look at the supported external modules:&lt;br /&gt;
* Authorization&lt;br /&gt;
* Solr integration&lt;br /&gt;
* Triplestore integration&lt;br /&gt;
* Camel integration&lt;br /&gt;
* etc&lt;br /&gt;
&lt;br /&gt;
In addition to clarifying specific implementation details, this is an opportunity to surface community requirements which may have not been addressed in the initial release.&lt;br /&gt;
&lt;br /&gt;
== Bleeding edge beacons: redesign of a library tour with new technology ==&lt;br /&gt;
&lt;br /&gt;
* Neal Henshaw, Virginia Tech &lt;br /&gt;
* Somiah Lattimore, Straight Up Creative&lt;br /&gt;
* Keith Gilbertson, keith.gilbertson@vt.edu, Virginia Tech&lt;br /&gt;
&lt;br /&gt;
This group has not yet presented at Code4Lib&lt;br /&gt;
&lt;br /&gt;
A technology liaison, a designer, and a developer are working together to update a mobile audio tour of an academic library.&lt;br /&gt;
&lt;br /&gt;
The walking tour was originally conceived of and created by the technology liaison to provide an orientation experience for incoming students. In its current version, patrons visit several stations in the library and activate an audio description of each station by scanning a QR barcode with a mobile device.&lt;br /&gt;
&lt;br /&gt;
We are building a prototype to test with focus groups so that we can launch a new version of the tour in early 2015. The tour has been reimagined with a professional and communicative user interface that presents students with learning goals for each location in the library.  The new version of the tour is a mobile application with integrated support for micro-location technology provided through Bluetooth low energy beacon devices, known informally as iBeacons, installed throughout the building. As a touring student walks through the library, her location is noted by the app so that an appropriate video segment is automatically played according to the current location.&lt;br /&gt;
 &lt;br /&gt;
We will discuss perspectives on the design of the project, including&lt;br /&gt;
 &lt;br /&gt;
* Designing with technology to achieve learning goals&lt;br /&gt;
* Designing graphical elements, interaction, and user experience&lt;br /&gt;
* Designing to maintain compatibility with older technologies&lt;br /&gt;
&lt;br /&gt;
We’ll explain our efforts to make the application usable by those with hearing impairments and mobility impairments, reactions from our first users, and challenges in working with the relatively new beacon technology.&lt;br /&gt;
&lt;br /&gt;
== Distributed Remediation: Small tools for big problems: ==&lt;br /&gt;
 &lt;br /&gt;
* Matt Miller,  matthewmiller@nypl.org, New York Public Library, NYPL Labs&lt;br /&gt;
&lt;br /&gt;
Remediation of legacy data can be automated only so much. Certain essential cleanup tasks, such as aligning a name with the correct authorized version, is very difficult for a computer yet trivial for a person. While it is these types remediations that will allow an institution to take advantage and participate in the web of Linked Open Data, a wholly manual approach is unrealistic. However, by augmenting automated remediation with a light human touch we can quickly and efficiently reach our goals. This talk will look at tools and methods being developed at NYPL Labs to empower library staff and the public to help clean up our legacy metadata.&lt;/div&gt;</summary>
		<author><name>MatthewMiller</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39865</id>
		<title>2014 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39865"/>
				<updated>2013-11-08T21:31:17Z</updated>
		
		<summary type="html">&lt;p&gt;MatthewMiller: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
 &lt;br /&gt;
* ''Projects'' you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* ''Tools and technologies'' – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* ''Technical issues'' - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* ''Relevant non-technical issues'' – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
'''To Propose a Talk'''&lt;br /&gt;
* Log in to the wiki in order to submit a proposal. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
* Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
* If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist us in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
As in past years, the Code4Lib community will vote on proposals that they would like to see included in the program. This year, however, only the top 10 proposals will be guaranteed a slot at the conference. Additional presentations will be selected by the Program Committee in an effort to ensure diversity in program content. Community votes will, of course, still weigh heavily in these decisions.&lt;br /&gt;
&lt;br /&gt;
Presenters whose proposals are selected for inclusion in the program will be guaranteed an opportunity to register for the conference. The standard conference registration fee will still apply.&lt;br /&gt;
&lt;br /&gt;
''Proposals can be submitted through '''Friday, November 8, 2013, at 5pm PST'''''. Voting will commence on November 18, 2013 and continue through December 6, 2013. The final line-up of presentations will be announced in early January, 2014.&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
&lt;br /&gt;
==Creating a new Greek-Dutch dictionary==&lt;br /&gt;
* Caspar Treijtel, University of Amsterdam, c.treijtel@uva.nl&lt;br /&gt;
&lt;br /&gt;
At present, no complete dictionary of (ancient) Greek-Dutch is available online. A new dictionary is currently under construction at Leiden University, with software being developed at the University of Amsterdam. The team in Leiden has already begun preparation of the data, with at this moment about 6,000 approved lemmas. The ultimate goal is to produce both a print version and online open access version from the same source documents. The software needed for this has been made in a project that was funded by CLARIN-NL.&lt;br /&gt;
&lt;br /&gt;
Migrator&lt;br /&gt;
&lt;br /&gt;
For the production of lemmas we have implemented an advanced workflow. The (generally non-technical) users create lemmas using MS Word, which is both familiar and easy to use. We have developed a custom software module that carefully migrates the Word documents into deeply structured XML by analyzing the structure and semantics of the lemmas, and falling back on heuristics in ambiguous cases. While having initially envisioned the oXygen XML Author component as the main tool for creating new lemmas, we obtained excellent results with the migrator module, and decided therefore to continue using MS Word as the primary composition tool. The main advantage of this is that the editors are much more familiar with Word than with any other WYSIWYG editor. Lemmas that have been migrated to XML are stored in an XML database and can be further edited using oXygen XML Author.&lt;br /&gt;
&lt;br /&gt;
Lemmatizer&lt;br /&gt;
&lt;br /&gt;
Greek morphology is complicated. In order to use a dictionary effectively, a rather high level of initial language competence is necessary for the user to be able to relate the word form s/he finds in a text to the correct basic lemma form, where the definition of the word can be found. Using a Greek morphological database we have been able to facilitate the search for lemmas. A ‘lemmatizer’ module gives the possible parsings of the word forms and the lemmas they can be derived from. This enables the user to type in the word as found in the text and be redirected to the correct lemma.&lt;br /&gt;
&lt;br /&gt;
Visualization&lt;br /&gt;
&lt;br /&gt;
For the online dictionary we have implemented a visualization module that allows the user to view multiple lemmas at once. The implementation of this module has been done using the Javascript framework MooTools. The result is a viewer that performs really well and is run by maintainable Javascript code.&lt;br /&gt;
&lt;br /&gt;
The online dictionary is still being worked on, have a look at http://www.woordenboekgrieks.nl/ for the beta version. A newer test version with additional features can be found here: http://angel.ic.uva.nl:8600/.&lt;br /&gt;
&lt;br /&gt;
Credits&lt;br /&gt;
&lt;br /&gt;
* construction of the dictionary: Prof. Ineke Sluiter, Classics department of Leiden University; Prof. Albert Rijksbaron, University of Amsterdam&lt;br /&gt;
* publisher of the dictionary: Amsterdam University Press&lt;br /&gt;
* design/typesetting dictionary: TaT Zetwerk (http://www.tatzetwerk.nl/)&lt;br /&gt;
* software development: Digital Production Center, University Library, University of Amsterdam&lt;br /&gt;
* project funding: CLARIN-NL (http://www.clarin.nl/)&lt;br /&gt;
* morphological database for use by the lemmatizer: courtesy of Prof. Helma Dik, University of Chicago (based on data of the Perseus Project)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Using Drupal to drive alternative presentation systems ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Recently, we have been building systems that use angular.js, Rails, or other systems for presentation, while leveraging Drupal's sophisticated content management capabilities on the back end.&lt;br /&gt;
&lt;br /&gt;
So far, these have been one-way systems, but as we move to Drupal 8 we are beginning to explore ways to further decouple the presentation and CMS functions.&lt;br /&gt;
&lt;br /&gt;
== A Book, a Web Browser and a Tablet: How Bibliotheca Alexandrina's Book Viewer Framework Makes It Possible ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mohammed.abuouda|Mohammed Abu ouda]], Bibliotheca Alexandrina (The new Library of Alexandria)&lt;br /&gt;
&lt;br /&gt;
A lot of institutions around the world are engaged in multiple digitization projects aiming at preserving the human knowledge present in books and availing them through multiple channels to people around the whole globe. These efforts will sure help close the digital gap particularly with the arrival of affordable e-readers, mobile phones and network coverage. However, the digital reading experience has not yet arrived to its maximum potential. Many readers miss features they like in their good old books and wish to find them in their digital counterpart. In an attempt to create a unique digital reading experience, Bibliotheca Alexandria (BA) created a flexible book viewing framework that is currently used to access its current collection of more than 300,000 digital books in five different languages which includes the largest collection of digitized Arabic books.&lt;br /&gt;
&lt;br /&gt;
Using open source tools, BA used the framework to develop a modular book viewer that can be deployed in different environments and is currently at the heart of various BA projects. The Book viewer provides several features creating a more natural reading experience. As with physical books, the reader can now personalize the books he reads by adding annotations like highlights, underlines and sticky notes to capture his thoughts and ideas in addition to being able to share the book with friends on social networks. The reader can perform a search across the content of the book receiving highlighted search results within the pages of the book. More features can be further added to the book viewer through its plugin architecture.&lt;br /&gt;
&lt;br /&gt;
== Structured data NOW: seeding schema.org in library systems ==&lt;br /&gt;
 &lt;br /&gt;
* [http://coffeecode.net Dan Scott], Laurentian University&lt;br /&gt;
** Previous code4lib presentations: [https://archive.org/details/code4lib.conf.2008.pres.CouchDBsacrilege CouchDB is sacrilege... mmm, delicious sacrilege] at Code4Lib 2008&lt;br /&gt;
&lt;br /&gt;
The semantic web, linked data, and structured data are all fantastic ideas with a barrier imposed by implementation constraints. If their system does not allow customizations, or the institution lacks skilled human resources, it does not matter how enthused a given library might be about publishing structured data... it will not happen. However, if the software in use simply publishes structured data by default, then the web will be populated for free. Really! No extra resources necessary.&lt;br /&gt;
&lt;br /&gt;
This presentation highlights Dan's work with systems such as Evergreen, Koha, and VuFind to enable the publication of schema.org structured data out-of-the-box. Along the way, we reflect the current state of the W3C Schema.org Bibliographic Extension community group efforts to shape the evolution of the schema.org vocabulary. Finally, hold on tight as we contemplate next steps and the possibilities of a world where structured data is the norm on the web.&lt;br /&gt;
&lt;br /&gt;
== Towards Pasta Code Nirvana: Using JavaScript MVC to Fill Your Programming Ravioli ==&lt;br /&gt;
&lt;br /&gt;
* Bret Davidson, North Carolina State University Libraries, bret_davidson@ncsu.edu&lt;br /&gt;
** Previous Code4Lib Presentations: [http://wiki.code4lib.org/index.php/2013_talks_proposals#Data-Driven_Documents:_Visualizing_library_data_with_D3.js Visualizing library data with D3.js] at Code4Lib 2013&lt;br /&gt;
&lt;br /&gt;
JavaScript MVC frameworks are ushering in a golden age of robust and responsive web applications that take advantage of evergreen browsers, performant JS engines, and the unprecedented reach provided by billions of personal computing devices. The web browser has emerged as the world’s most popular application runtime and the complexity[1] and scope of JavaScript applications has exploded accordingly. Server-side web frameworks like Rails and Django have helped developers adhere to best practices like modularity, dependency injection, and unit testing for years, practices that are now being applied to JavaScript development through projects like Backbone[2], Ember[3], and Angular[4].&lt;br /&gt;
&lt;br /&gt;
This talk will discuss the issues JavaScript MVC frameworks are trying to solve, common features like data binding, implications for the future of web development[5], and the appropriateness of JavaScript MVC for library applications.&lt;br /&gt;
&lt;br /&gt;
*[1]http://en.wikipedia.org/wiki/Spaghetti_code&lt;br /&gt;
*[2]http://backbonejs.org&lt;br /&gt;
*[3]http://emberjs.com&lt;br /&gt;
*[4]http://angularjs.org&lt;br /&gt;
*[5]http://tomdale.net/2013/09/progressive-enhancement-is-dead/&lt;br /&gt;
&lt;br /&gt;
== WebSockets for Real-Time and Interactive Interfaces ==&lt;br /&gt;
&lt;br /&gt;
* [http://ronallo.com Jason Ronallo], NCSU Libraries, jason_ronallo@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/ronallo HTML5 Microdata and Schema.org] 2012&lt;br /&gt;
* [http://code4lib.org/conference/2013/ronallo HTML5 Video Now!] 2013&lt;br /&gt;
&lt;br /&gt;
Watching the Google Analytics Real-Time dashboard for the first time was mesmerizing. As soon as someone visited a site, I could see what page they were on. For a digital collections site with a lot of images, it was fun to see what visitors were looking at. But getting from Google Analytics to the image or other content of what was currently being viewed was cumbersome. The real-time experience was something I wanted share with others. I'll show you how I used a WebSocket service to create a real-time interface to digital collections.&lt;br /&gt;
&lt;br /&gt;
In the Hunt Library at NCSU we have some large video walls. I wanted to make HTML-based exhibits that featured viewer interactions. I'll show you how I converted Listen to Wikipedia [1] into an bring-your-own-device interactive exhibit. With WebSockets any HTML page can be remote controlled by any internet connected device.&lt;br /&gt;
&lt;br /&gt;
I will attempt to include real-time audience participation.&lt;br /&gt;
&lt;br /&gt;
[1] http://listen.hatnote.com/&lt;br /&gt;
&lt;br /&gt;
== Rapid Development of Automated Tasks with the File Analyzer ==&lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Libraries, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Libraries have customized the File Analyzer and Metadata Harvester application (https://github.com/Georgetown-University-Libraries/File-Analyzer) to solve a number of library automation challenges:&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census.  After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform.  Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules.  As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
Learn about the functionality that is available for download, how you can use this tool to automate workflows from digital collections to ILS ingests to electronic resources statistics and also discuss the opportunities to collaborate on enhancements to this application!&lt;br /&gt;
&lt;br /&gt;
== GeoHydra: How to Build a Geospatial Digital Library with Fedora ==&lt;br /&gt;
 &lt;br /&gt;
* [http://stanford.edu/~drh Darren Hardy], Stanford University, drh@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Geographically-rich data are exploding and putting fear in those trying to&lt;br /&gt;
tackle integrating them into existing digital library infrastructures.&lt;br /&gt;
Building a spatial data infrastructure that integrates with your digital&lt;br /&gt;
library infrastructure need not be a daunting task. We have successfully&lt;br /&gt;
deployed a geospatial digital library infrastructure using Fedora and&lt;br /&gt;
open-source geospatial software [1]. We'll discuss the primary design&lt;br /&gt;
decisions and technologies that led to a production deployment within a few&lt;br /&gt;
months. Briefly, our architecture revolves around discovery, delivery, and&lt;br /&gt;
metadata pipelines using open-source OpenGeoPortal [2], Solr [3], GeoServer&lt;br /&gt;
[4], PostGIS [5], and GeoNetwork [6] technologies, plus the proprietary ESRI&lt;br /&gt;
ArcMap [7] -- the GIS industry's workhorse. Finally, we'll discuss the key&lt;br /&gt;
skillsets needed to build and maintain a spatial data infrastructure.&lt;br /&gt;
&lt;br /&gt;
[1] http://foss4g.org&lt;br /&gt;
[2] http://opengeoportal.org&lt;br /&gt;
[3] http://lucene.apache.org/solr&lt;br /&gt;
[4] http://geoserver.org&lt;br /&gt;
[5] http://postgis.net&lt;br /&gt;
[6] http://geonetwork-opensource.org&lt;br /&gt;
[7] http://esri.com&lt;br /&gt;
&lt;br /&gt;
==Under the Hood of Hadoop Processing at OCLC Research ==&lt;br /&gt;
&lt;br /&gt;
[http://roytennant.com/ Roy Tennant]&lt;br /&gt;
&lt;br /&gt;
* Previous Code4Lib presentations: 2006: &amp;quot;The Case for Code4Lib 501c(3)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[http://hadoop.apache.org/ Apache Hadoop] is widely used by Yahoo!, Google, and many others to process massive amounts of data quickly. OCLC Research uses a 40-node compute cluster with Hadoop and HBase to process the 300 million MARC records of WorldCat in various ways. This presentation will explain how Hadoop MapReduce works and illustrate it with specific examples and code. The role of the jobtracker in both monitoring and reporting on processes will be explained. String searching WorldCat will also be demonstrated live.&lt;br /&gt;
&lt;br /&gt;
== Quick and Easy Data Visualization with Google Visualization API and Google Chart Libraries ==&lt;br /&gt;
 &lt;br /&gt;
[http://bohyunkim.net/blog Bohyun Kim], Florida International University, bohyun.kim@fiu.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Do most of the data that your library collects stay in spreadsheets or are published as a static table with a series of boring numbers? Do your library stakeholders spend more time collecting the data than using it as a decision-making tool because the data is presented in a way that makes it hard for them [http://developers.google.com/chart/interactive/docs/gallery to quickly grasp its significance? ]&lt;br /&gt;
&lt;br /&gt;
This talk will provide an overview of [http://developers.google.com/chart/interactive/docs/reference Google Visualization API] [2] and [http://developers.google.com/chart/ Google Chart Libraries] [3] to get you started on the way to quickly query and visualize your library data from remote data sources (e.g. a Google Spreadsheet or your own database) with (or without) cool-looking user-controls, animation effects, and even a dashboard.&lt;br /&gt;
&lt;br /&gt;
== Leap Motion + Rare Books: A hands-free way to view and interact with rare books in 3D ==&lt;br /&gt;
 &lt;br /&gt;
[http://http://www.youtube.com/user/jpdenzer Juan Denzer], Binghamton University, jdenzer@binghamton.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
As rare books become more delicate over time, making them available to the public becomes harder.  We at Binghamton University Library have developed an application that makes it easier to view rare books without ever having to touch them.  We have combined the Leap Motion hands-free device and 3D rendered models to create a new virtual experience for the viewer.&lt;br /&gt;
&lt;br /&gt;
The application allows the user to rotate and zoom in on a 3D representation of a rare book.  The user is also able to ‘open’ the virtual book and flip through it using a natural user interface.  Such as swiping the hand left or right to turn the page.&lt;br /&gt;
&lt;br /&gt;
The application is built on the .Net framework and is written in C#.  3D models are created using simple 3D software such as sketchup or Blender.  Scans of the book cover and spine are created using simple flatbed scanners.  The inside pages are scanned using overhead scanners. &lt;br /&gt;
&lt;br /&gt;
This talk with discuss the technologies used in developing the application and virtually any library could implement the application with virtually no coding at all. This presentation will have a demonstration of the software and also a chance for audience members to experience the Rare Book Leap Motion App themselves.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Course Reserves Unleashed! ==&lt;br /&gt;
 &lt;br /&gt;
* Bobbi Fox, Library Technology Services, Harvard University, bobbi_fox@harvard.edu&lt;br /&gt;
* Gloria Korsman, Andover-Harvard Theological Library&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Hey kids!  Remember when SOAP was used for something other than washing?  Our sophisticated (and highly functional) Course Reserves Request system does!&lt;br /&gt;
&lt;br /&gt;
However, while the system is great for submitting and  processing course reserve requests, the student-facing presentation through Havard’s home-grown -- and soon to be replaced -- LMS leaves a lot to be desired.  &lt;br /&gt;
&lt;br /&gt;
Follow along as we leverage Solr 4 as a No-SQL database, along with more progressive RESTful API techniques, to release Reserves data into the wild without interfering with reserves request processing -- and, in the process, open up the opportunity for other schools to feed their data in as well.&lt;br /&gt;
&lt;br /&gt;
== We Are All Disabled! Universal Web Design Making Web Services Accessible for Everyone ==&lt;br /&gt;
 &lt;br /&gt;
Cynthia Ng, Accessibility Librarian, CILS at Langara College&lt;br /&gt;
* No previous Code4Lib presentations (not counting lightning talks)&lt;br /&gt;
&lt;br /&gt;
We’re building and improving tools and services all the time, but do you only develop for the “average” user or add things for “disabled” users? We all use “assistive” technology accessing information in a multitude of ways with different platforms, devices, etc. Let’s focus on providing web services that are accessible to everyone without it being onerous or ugly. The aim is to get you thinking about what you can do to make web-based services and content more accessible for all from the beginning or with small amounts of effort whether you're a developer or not.&lt;br /&gt;
&lt;br /&gt;
The goal of the presentation is to provide both developers and content creators with information on simple, practical ways to make web content and web services more accessible. However, rather than thinking about putting in extra effort or making adjustment for those with disabilities, I want to help people think about how to make their websites more accessible for all users through universal web design.&lt;br /&gt;
&lt;br /&gt;
== Personalize your Google Analytics Data with Custom Events and Variables ==&lt;br /&gt;
&lt;br /&gt;
[http://joshwilson.net Josh Wilson], Systems Integration Librarian, State Library of North Carolina - joshwilsonnc@gmail.com&lt;br /&gt;
&lt;br /&gt;
At the State Library of North Carolina, we had more specific questions about the use of our digital collections than standard GA could provide. A few implementations of custom events and custom variables later, we have our answers.&lt;br /&gt;
&lt;br /&gt;
I'll demonstrate how these analytics add-ons work, and why implementation can sometimes be more complicated than just adding a few lines of JavaScript to your ga.js. I'll discuss some specific examples in use at the SLNC:&lt;br /&gt;
&lt;br /&gt;
* Capturing the content of specific metadata fields in CONTENTdm as Custom Events &lt;br /&gt;
* Recording Drupal taxonomy terms as Custom Variables&lt;br /&gt;
&lt;br /&gt;
In both instances, this data deepened our understanding of how our sites and collections were being used, and in turn, we were able to report usage more accurately to content contributors and other stakeholders.&lt;br /&gt;
&lt;br /&gt;
More on: [https://developers.google.com/analytics/devguides/collection/gajs/eventTrackerGuide GA Custom Events] | [https://developers.google.com/analytics/devguides/collection/gajs/gaTrackingCustomVariables GA Custom Variables]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Behold Fedora 4: The Incredible Shrinking Repository! ==&lt;br /&gt;
&lt;br /&gt;
Esmé Cowles, UC San Diego Library.  Previous talk: [http://code4lib.org/conference/2013/cowles-critchlow-westbrook All Teh Metadatas Re-Revisited] (2013)&lt;br /&gt;
&lt;br /&gt;
* One repository contains untold numbers of digital objects and powers many Hydra and Islandora apps&lt;br /&gt;
* It speaks RDF, but contains no triplestore! (triplestores sold separately, SPARQL Update may be involved, some restrictions apply)&lt;br /&gt;
* Flexible enough to tie itself in knots implementing storage and access control policies&lt;br /&gt;
* Witness feats of strength and scalability, with dramatically increased performance and clustering&lt;br /&gt;
* Plumb the depths of bottomless hierarchies, and marvel at the metadata woven into the very fabric of the repository&lt;br /&gt;
* Ponder the paradox of ingesting large files by not ingesting them&lt;br /&gt;
* Be amazed as Fedora 4 swallows other systems whole (including Fedora 3 repositories)&lt;br /&gt;
* Watch novice developers setup Fedora 4 from scratch, with just a handful of incantations to Git and Maven&lt;br /&gt;
&lt;br /&gt;
The Fedora Commons Repository is the foundation of many digital collections, e-research, digital library, archives, digital preservation, institutional repository and open access publishing systems.  This talk will focus on how Fedora 4 improves core repository functionality, adds new features, maintains backwards compatibility, and addresses the shortcomings of Fedora 3.&lt;br /&gt;
&lt;br /&gt;
== Organic Free-Range API Development - Making Web Services That You Will Actually Want to Consume ==&lt;br /&gt;
&lt;br /&gt;
Steve Meyer and Karen Coombs, OCLC&lt;br /&gt;
&lt;br /&gt;
Building web services can have great benefits by providing reusability of data and functionality. Underpinning your applications with a web service will allow you to write code once and support multiple environments: your library's web app, mobile applications, the embedded widget in your campus portal. However, building a web service is its own kind of artful programming. Doing it well requires attention to many of the same techniques and requirements as building web applications, though with different outcomes. &lt;br /&gt;
&lt;br /&gt;
So what are the usability principles for web services? How do you build a web service that you (and others) will actually want to use? In this talk, we’ll share some of the lessons learned - the good, the bad, and the ugly - through OCLC's work on the WorldCat Metadata API. This web service is a sophisticated API that provides external clients with read and write access to WorldCat data. It provides a model to help aspiring API creators navigate the potential complications of crafting a web service. We'll cover:&lt;br /&gt;
&lt;br /&gt;
* Loose coupling of data assets and resource-oriented data modeling at the core&lt;br /&gt;
* Coding to standards vs. exposure of an internal data model&lt;br /&gt;
* Authentication and security for web services: API Keys, Digital Signing, OAuth Flows&lt;br /&gt;
* Building web services that behave as a suite so it looks like the left hand knows what the right hand is doing&lt;br /&gt;
&lt;br /&gt;
So at the end of the day, your team will know your API is a very good egg after all. &lt;br /&gt;
&lt;br /&gt;
If accepted, the presenters intend to produce and share a Quick Guide for building a web service that will reflect content presented in the talk.&lt;br /&gt;
&lt;br /&gt;
== Lucene's Latest (for Libraries) ==&lt;br /&gt;
&lt;br /&gt;
erik.hatcher@lucidworks.com&lt;br /&gt;
&lt;br /&gt;
Lucene powers the search capabilities of practically all library discovery platforms, by way of Solr, etc.  The Lucene project evolves rapidly, and it's a full-time job to keep up with the ever improving features and scalability.   This talk will distill and showcase the most relevant(!) advancements to date.&lt;br /&gt;
&lt;br /&gt;
== The Why and How of Very Large Displays in Libraries. ==&lt;br /&gt;
&lt;br /&gt;
* Cory Lown, NCSU Libraries, cwlown@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib Presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/lown How People Search the Library from a Single Search Box]  2012&lt;br /&gt;
* [http://code4lib.org/conference/2010/orphanides_lown_lynema Enhancing Discoverability with Virtual Shelf Browse] 2010&lt;br /&gt;
&lt;br /&gt;
Built into the walls of NC State's new Hunt Library are several [http://www.christiedigital.com/en-us/digital-signage/products/microtiles/pages/microtiles-digital-signage-video-wall.aspx Christie MicroTile Display Wall Systems]. What does a library do with a display that's seven feet tall and over twenty feet wide? I'll talk about why libraries might want large displays like this, what we're doing with them right now, and what we might do with them in the future. I'll talk about how these displays factor into planning for new and existing web projects. And I'll get into the fun details of how you build web applications that scale from the very small browser window on a phone all the way up to a browser window with about 14 million pixels (about 10 million more than a dual 24&amp;quot; monitor desktop setup).&lt;br /&gt;
&lt;br /&gt;
== Discovering your Discovery System in Real Time. ==&lt;br /&gt;
&lt;br /&gt;
* Godmar Back, Virginia Tech, gback@vt.edu&lt;br /&gt;
* Annette Bailey, Virginia Tech, afbailey@vt.edu&lt;br /&gt;
&lt;br /&gt;
Practically all libraries today provide web-based discovery systems to their users;&lt;br /&gt;
users discover items and peruse or check them out by clicking on links.  Unlike&lt;br /&gt;
the traditional transaction of checking out a book at the circulation desk, this&lt;br /&gt;
interaction is largely invisible.  We have built a system that records user's&lt;br /&gt;
interaction with Summon in real-time, processes the resulting data with minimal delay,&lt;br /&gt;
and visualizes it in various ways using Google Charts and using various d3.js modules,&lt;br /&gt;
such as word clouds, tree maps, and others.&lt;br /&gt;
&lt;br /&gt;
These visualizations can be embedded in web sites, but are also suitable for&lt;br /&gt;
projection via large-scale displays or projectors right into the 'Learning Spaces'&lt;br /&gt;
many libraries are converted into.  The goal of this talk is to share the technology&lt;br /&gt;
and advocate the building of a cloud-based infrastructure that would make this&lt;br /&gt;
technology available to any library that uses a discovery system, rather than just&lt;br /&gt;
those who have the technological prowess for developing such systems and&lt;br /&gt;
visualizations in-house.  &lt;br /&gt;
&lt;br /&gt;
Previous presentations at Code4Lib:&lt;br /&gt;
* Talk: Code4Lib 2009 [http://code4lib.org/files/LibX2.0-Code4Lib-2009AsPresented.ppt LibX 2.0]&lt;br /&gt;
* Preconference: [http://wiki.code4lib.org/index.php/LibX_Preconference LibX 2.0, 2009]&lt;br /&gt;
* Preconference: Code4Lib 2010, On Widgets and Web Services&lt;br /&gt;
&lt;br /&gt;
== Your Library, Anywhere: A Modern, Responsive Library Catalogue at University of Toronto Libraries ==&lt;br /&gt;
&lt;br /&gt;
* Bilal Khalid, Gordon Belray, Lisa Gayhart (lisa.gayhart@utoronto.ca)&lt;br /&gt;
&lt;br /&gt;
* No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. &lt;br /&gt;
&lt;br /&gt;
[http://search.library.utoronto.ca The result]: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles. This talk will discuss: application development; service improvements; interface design; and user outreach, testing, and project communications. Feedback and questions from the audience are very welcome. If time runs short, we will be available for questions and conversation after the presentation.&lt;br /&gt;
&lt;br /&gt;
Note: A version of this content has been provisionally accepted as an article for Code4Lib Journal, January 2014 publication.)&lt;br /&gt;
&lt;br /&gt;
== All Tiled Up ==&lt;br /&gt;
&lt;br /&gt;
* Mike Graves, MIT Libraries (mgraves@mit.edu)&lt;br /&gt;
&lt;br /&gt;
You've got maps. You even scanned and georeferenced them. Now what? Running a full GIS stack can be expensive, and overkill in some cases. The good news is that you have a lot more options now than you did just a few years ago. I'd like to present some lighter weight solutions to making georeferenced images available on the Web.&lt;br /&gt;
&lt;br /&gt;
This talk will provide an introduction to MBTiles. I'll go over what they are, how you create them, how you use them and why you would use them.&lt;br /&gt;
&lt;br /&gt;
== The Great War: Image Interoperability to Facebook ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, Los Alamos National Laboratory (azaroth42@gmail.com)&lt;br /&gt;
** (Code4Lib 2006: [http://www.code4lib.org/2006/sanderson | Library Text Mining])&lt;br /&gt;
* Rob Warren, Carleton University&lt;br /&gt;
** No previous presentations&lt;br /&gt;
&lt;br /&gt;
Using a pipeline constructed from Linked Open Data and other interoperability specifications, it is possible to merge and re-use image and textual data from distributed library collections to build new, useful tools and applications.  Starting with the OAI-PMH interface to ContentDM, we will take you on a tour through the International Image Interoperability Framework and Shared Canvas, to a cross-institutional viewer, and image analysis for the purposes of building a historical Facebook from finding and tagging people in photographs.  The World War One collections are drawn from multiple institutions and merged by the machine learning code.&lt;br /&gt;
&lt;br /&gt;
The presentation will focus on the (open source) toolchain and the benefits of the use of standards throughout:  OAI-PMH to get the metadata, IIIF for interaction with the images, the Shared Canvas ontology for describing collections of digitized objects, Open Annotation for tagging things in the images and specialized ontologies that are specific to the contents.  The tools include standard RDF / OWL technologies, JSON-LD, imagemagick and OpenCV for image analysis.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Solr Search Results with D3.js for User-Friendly Navigation of Large Results Sets ==&lt;br /&gt;
&lt;br /&gt;
*Julia Bauder, Grinnell College Libraries (bauderj-at-grinnell-dot-edu)&lt;br /&gt;
*No previous presentations at national Code4Lib conferences&lt;br /&gt;
&lt;br /&gt;
As the corpus of articles, books, and other resources searched by discovery systems continues to get bigger, searchers are more and more frequently confronted with unmanageably large numbers of results. How can we help users make sense of 10,000 hits and find the ones they actually want? Facets help, but making sense of a gigantic sidebar of facets is not an easy task for users, either.&lt;br /&gt;
During this talk, I will explain how we will soon be using Solr 4’s pivot queries and hierarchical visualizations (e.g., treemaps) from D3.js to let patrons view and manipulate search results. We will be doing this with our VuFind 2.0 catalog, but this technique will work with any system running Solr 4. I will also talk about early student reaction to our tests of these visualization features.&lt;br /&gt;
&lt;br /&gt;
== PeerLibrary – open source cloud based collaborative library ==&lt;br /&gt;
&lt;br /&gt;
* [http://mitar.tnode.com/ Mitar Milutinovic], UC Berkeley, mitar.code4lib at tnode.com&lt;br /&gt;
* Not presented or attended code4lib before &lt;br /&gt;
&lt;br /&gt;
[https://github.com/peerlibrary/peerlibrary PeerLibrary is a new open source project] and a cloud service providing collaborative reading, sharing and storing. Users can upload publications they want to read (currently in PDF format), read them in the browser in real-time with others, highlight, annotate and organize their own or collaborative library. PeerLibrary provides a search engine to search over all uploaded open access publications. Additionally, it aims to collaboratively aggregate the open layer of knowledge on top of this publications through public annotations and references user will add to publications. In this way publications would not just be available to read, but accessible to the general public as well. Currently, it is aiming at scientific community and scientific publications.&lt;br /&gt;
&lt;br /&gt;
See [http://blog.peerlibrary.org/post/63458789185/screencast-previewing-the-peerlibrary-project screencast here]. [http://peerlibrary.org/ Subscribe to newsletter] to be a beta tester when we open.&lt;br /&gt;
&lt;br /&gt;
It is still in development and beta launch is planned at the end of November.&lt;br /&gt;
&lt;br /&gt;
== Who was where when, or finding biographical articles on Wikipedia by place and time ==&lt;br /&gt;
&lt;br /&gt;
* [http://morton-owens.info Emily Morton-Owens], The Seattle Public Library (presenting on work from NYU)&lt;br /&gt;
* No previous c4l presentations&lt;br /&gt;
&lt;br /&gt;
It's easy to answer the question &amp;quot;What important people were in Paris in 1939?&amp;quot; But what about Virginia in the 1750s or Scandinavia in the 14th century? I created a tool that allows you to search for biographies in a generally applicable way, using a map interface. I would like to present updates to my thesis project, which combines a crawler written in Java that extracts information from Wikipedia articles, with a MongoDB data store and a frontend in Python.&lt;br /&gt;
&lt;br /&gt;
The input to the project is freetext of entire articles in Wikipedia; this is important to allow us to pick up Benjamin Franklin not just in the single most obvious place of Philadelphia but also in London, Paris, Boston, etc. I can talk about my experiments disambiguating place names (approaches pioneered on newspaper articles were actually unhelpful on this type of text) and setting up a processing queue that does not become mired in the biographies of every human who ever played soccer. I also want to mitigate some of the implementation choices I made due to my academic deadline and improve the accuracy/usability.&lt;br /&gt;
&lt;br /&gt;
What I hope to show is that I was able to develop a novel and useful reference tool automatically, using fairly simple heuristics that are a far cry from hand-cataloging familiar to many librarians.&lt;br /&gt;
&lt;br /&gt;
You can try out [http://linserv1.cims.nyu.edu:48866/ the original version] (this server is inconveniently set to be updated/rebooted on 11/8--may be temporarily unavailable)&lt;br /&gt;
&lt;br /&gt;
== Good!, DRY, and Dynamic: Content Strategy for Libraries (Especially the Big Ones) ==&lt;br /&gt;
&lt;br /&gt;
*Michael Schofield, Nova Southeastern University Libraries, mschofield@nova.edu&lt;br /&gt;
*No previous code4lib presentations.&lt;br /&gt;
&lt;br /&gt;
The responsibilities of the #libweb are exploding [it’s a good thing] and it is no longer uncommon for libraries to manage or even home-grow multiple applications and sites. Often it is at this point where the web people begin to suffer the absence of a content strategy when, say, business hours need to be updated sitewide a half-dozen times.&lt;br /&gt;
&lt;br /&gt;
We were already feeling this crunch when we decided to further complicate the Nova Southeastern University Libraries by splitting the main library website into two. The Alvin Sherman Library, Research, and Information Technology Center is a unique joint-use facility that serves not only the academic community but the public of Broward County - and marketing a hyperblend of content through one portal just wasn't cutting it. With a web team of two, we knew that managing all this rehashed, disparate content was totally unsustainable.&lt;br /&gt;
&lt;br /&gt;
I want to share in this talk how I went about making our library content DRY (“don’t repeat yourself”): input content in one place--blurbs, policies, featured events, featured databases, book reviews, business hours, and so on.--and syndicate it everywhere - even, sometimes, dynamically target that content for specific audiences or context. It is a presentation that is a little about workflow, a little more about browser and context detection, a tangent about content-modeling the CMS, and a lot about APIs, syndication, and performance.&lt;br /&gt;
&lt;br /&gt;
== No code, no root, no problem? Adventures in SaaS and library discovery ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:erwhite@vcu.edu Erin White, VCU]&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
In 2012 VCU was an eager early adopter of Ex Libris' cloud service Alma as an ILS, ERM, link resolver, and single-stop, de-silo'd public-facing discovery tool. This has been a disruptive change that has shifted our systems staff's day-to-day work, relationships with others in the library, and relationships with vendors.&lt;br /&gt;
&lt;br /&gt;
I'll share some of our experiences and takeaways from implementing and maintaining a cloud service:&lt;br /&gt;
* Seeking disruption and finding it&lt;br /&gt;
* Changing expectations of service and the reality of unplanned downtime&lt;br /&gt;
* Communication and problem resolution with non-IT library staff&lt;br /&gt;
* Working with a vendor that uses agile development methodology&lt;br /&gt;
* Benefits and pitfalls of creating customizations and code workarounds&lt;br /&gt;
* Changes in library IT/coders' roles with SaaS&lt;br /&gt;
&lt;br /&gt;
...as well as thoughts on the philosophy of library discovery vs real-life experiences in moving to a single-search model.&lt;br /&gt;
&lt;br /&gt;
== Building for others (and ourselves):  the Avalon Media System ==&lt;br /&gt;
* [mailto:michael.klein@northwestern.edu Michael B Klein], Senior Software Developer, Northwestern University &lt;br /&gt;
** [http://code4lib.org/conference/2010/metz_klein Public Datasets in the Cloud] (code4lib 2010)&lt;br /&gt;
** [http://code4lib.org/conference/2013/klein-rogers The Avalon Media System: A Next Generation Hydra Head For Audio and Video Delivery] (code4lib 2013)&lt;br /&gt;
* [mailto:j-rudder@northwestern.edu Julie Rudder], Digital Initiatives Project Manager, Northwestern University&lt;br /&gt;
** no previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
[http://www.avalonmediasystem.org/ Avalon Media System] is a collaborative effort between development teams at Northwestern and Indiana Universities. Our goal is to produce an open source media management platform that works well for us, but is also widely adopted and contributed to by other institutions. We believe that building a strong user and contributor community is vital to the success and longevity of the project, and have developed the system with this goal in mind. We will share lessons learned, pains and successes we’ve had releasing two versions of the application since last year.  &lt;br /&gt;
&lt;br /&gt;
Our presentation will cover our experiences:&lt;br /&gt;
* providing flexible, admin-friendly distribution and installation options&lt;br /&gt;
* building with abstraction, customization and local integrations in mind&lt;br /&gt;
* prioritizing features (user stories)&lt;br /&gt;
* attracting code contributions from other institutions&lt;br /&gt;
* gathering community feedback &lt;br /&gt;
* creating a product rather than a bag of parts&lt;br /&gt;
&lt;br /&gt;
== How to check your data to provide a great data product? Data quality as a key product feature at Europeana ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:Peter.Kiraly@kb.nl Péter Király] portal backend developer, Europeana&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
[http://Europeana.eu/ Europeana.eu] - Europe's digital library, archive and museum - aggregates more than 30 million metadata records from more than 2200 institutions.  The records come from libraries, archives, museums and every other kind of cultural institution, from very different systems and metadata schemas, and are typically transformed several times until they are ingested into the Europeana data repository.  Europeana builds a consolidated database from these records, creating reliable and consistent services for end-users (a search portal, search widget, mobile apps, thematic sites etc.) and an API, which supports our strategic goeal of data for reuse in education, creative industries, and the cultural sector.  A reliable &amp;quot;data product&amp;quot; is thus at the core of our own software products, as well as those of our API partners.&lt;br /&gt;
&lt;br /&gt;
Much effort is needed to smooth out local differences in the metadata curation practice of our data providers. We need a solid framework to measure the consistency of our data and provide feedback to decision-makers inside and outside the organisation. We can also use this metrics framework to ask content providers to improve their own metadata. Of course, a data-quality-driven approach requires that we also improve the data transformation steps of the Europeana ingestion process itself. Data quality issues heavily define what new features we are able to create in our user interfaces and API, and might actually affect the design and implementation of our underlying data structure, the Europeana Data Model.&lt;br /&gt;
&lt;br /&gt;
In the presentation I briefly describe the Europeana metadata ingestion process, show the data quality metrics, the measuring techniques (using the Europeana API, Solr and MongoDB queries), some typical problems (both trivial and difficult ones), and finally the feedback mechanism we propose to deploy.&lt;br /&gt;
&lt;br /&gt;
Keywords: Europeana, data quality, EDM, API, Apache Solr, MongoDB, #opendata, #openglam&lt;br /&gt;
&lt;br /&gt;
== Teach your Fedora to Fly: scaling out a digital repository ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:acoburn@amherst.edu Aaron Coburn], Software Developer, Amherst College&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Fedora is a great repository system for managing large collections of digital objects, but what happens when a popular food magazine begins directing a large number of readers to a manuscript showing Emily Dickinson’s own recipe for doughnuts? While Fedora excels in its support of XML-based metadata, it doesn’t always perform well under a high volume of traffic. Nor is it especially tolerant of network or hardware failures.&lt;br /&gt;
&lt;br /&gt;
This presentation will show how we are making heavy use of a fedora repository while at the same time insulating it almost entirely from any web traffic. Starting with a distributed web front-end built with Node.js, and caching most of the user-accessible content from Fedora in an elastic, fault-tolerant Riak (NoSQL) cluster, we have eliminated nearly all single points of failure in the system. It also means that our production system is spread across twelve separate servers, where asynchrony and Map-Reduce are king. And aside from being blazing fast, it is also entirely Hydra-compliant.&lt;br /&gt;
&lt;br /&gt;
Furthermore, we will attempt to answer the question: if fedora crashes and the visitors to your site don’t notice, did it really fail?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using Open Source Software and Freeware to Preserve and Deliver Digital Videos ==&lt;br /&gt;
* [mailto:wfang@kinoy.rutgers.edu Wei Fang], Head of Digital Services, Rutgers University Law Library&lt;br /&gt;
* Jiebei Luo, Digital Projects Initiative Intern, Rutgers University&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The Rutgers University Law Library is the official digital repository of the New Jersey Supreme Court oral arguments since 2002. This large video collection contains approximately 3,000 videos with a total of 400 GB or 6,000 viewing hours. With the expansion of this collection, the existing database and the static website could not efficiently support the library’s daily operations and meet its patrons’ search needs. &lt;br /&gt;
By utilizing open source software and freeware such as Ubuntu, FFmpeg, Solr and Drupal, the library is able to develop a complete solution to re-encoding videos, embedding subtitles, incorporating  Solr search engine and content management system to support full-text subtitle search, automatically updating video metadata records in the library catalog system and eventually providing a plug-in free HTML 5-based Web interface for patrons to view the videos online.&lt;br /&gt;
The aspects below will be presented in detail at the conference:&lt;br /&gt;
*	Video codecs comparison &lt;br /&gt;
*	Server-end batch video encoding/re-encoding&lt;br /&gt;
*	HTML 5 video tag and embedding subtitles&lt;br /&gt;
*	Incorporating search engine Solr and content management tool 	Drupal with the database to retrieve videos by full-text search especially in subtitle files&lt;br /&gt;
*	Incorporating video metadata with the library catalog system&lt;br /&gt;
&lt;br /&gt;
== Shared Vision, Shared Resources: the Curate Institutional Repository ==&lt;br /&gt;
* Dan Brubaker Horst, University of Notre Dame &lt;br /&gt;
** [http://code4lib.org/conference/2011/JohnsonHorst A Community-Based Approach to Developing a Digital Exhibit at Notre Dame Using the Hydra Framework] &lt;br /&gt;
* Julie Rudder, Northwestern University&lt;br /&gt;
** no previous presentations&lt;br /&gt;
&lt;br /&gt;
Curate is being collaboratively developed by several institutions in the Hydra community who share the need and vision for a Fedora-backed Institutional Repository. The first release of Curate was a collaboration between Notre Dame and Northwestern University, along with Digital Curation Experts (DCE) - a vendor hired jointly by our two institutions. Powered by the Hydra engine Sufia, the team worked quickly to release the first version of Curate in October 2013 which provides a basic self-deposit system that has support for various content types, collection building, DOI minting, and user profile creation. From the very beginning we have built Curate to be easy to theme and extend in order to ease the process of installation and use by other institutions.&lt;br /&gt;
&lt;br /&gt;
In December 2013, additional partners will join the project including: Indiana University, the University of Cincinnati and the University of Virginia. Each institution contributes resources to the project in order to further our common goal to create a product that fits our needs and has a sustainable future.Together we will tackle additional content types (like complex data, software, media), administrative collections and more. &lt;br /&gt;
&lt;br /&gt;
Our presentation will include:&lt;br /&gt;
* a brief demonstration of Curate and technical overview&lt;br /&gt;
* why and how we work together&lt;br /&gt;
* why build Curate&lt;br /&gt;
* the future of the project&lt;br /&gt;
&lt;br /&gt;
== Solr, Cloud and Blacklight ==&lt;br /&gt;
* David Jiao, Library Information Systems, Indiana University at Bloomington, djiao@indiana.edu&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
SolrCloud refers to the distributed capabilities in Solr4. It is designed to offer a highly available, fault tolerant environment by organizing data into multiple pieces that can be hosted on multiple machines with replicas, and providing a centralized cluster configuration and management. &lt;br /&gt;
&lt;br /&gt;
At Indiana University, we are upgrading our Solr backend for our recently released Blacklight-based OPAC system from Solr 1.4 to Solr4, and we also put up efforts to build a private cloud of Solr4 servers. In this talk, I will persent certain features of SolrCloud, including distributed requests, fault tolerance, near real time indexing/searching, and configuration management with Zookeeper, and our experiences of utilizing these features to provide better performance and architecture for our OPAC system, which serves over 7 million bibliographic records to over 100 thousand students and faculty members. I will also discuss some practical lessons learned from our SolrCloud setup/upgrade and the integration of the new SolrCloud to our customized Blacklight system.  &lt;br /&gt;
&lt;br /&gt;
== Leveraging XSD's for Reflective, Live Dataset Support in Institutional Repositories ==&lt;br /&gt;
* [mailto:msulliva@ufl.edu Mark Sullivan], Library Information Technology, University of Florida&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
The University of Florida Libraries are currently adding support for active datasets into our METS-based institutional repository software.  This ongoing project enables the library to be a partner in current, or long-running, data-driven projects around the university by providing tangible short-term and long-term benefits to the projects.  The system assists project teams by storing and providing access to their data, while supporting online filtering and sorting of the data, custom queries, and adding and editing of the data by authorized users.  We are also exploring simple data visualizations to allow users to perform basic graphical and geographic queries.  Several different schemas were explored including DDI and EML, but ultimately the streamlined approach of using XSD's with some custom attributes was chosen, with all other data residing in the METS file portions.  Currently the system is being developed using XSD's describing XML datasets, but this model should easily scale to support SQL datasets or large datasets supported by Hadoop or iRODS.&lt;br /&gt;
&lt;br /&gt;
This work is being integrated in the open source [http://sobek.ufl.edu SobekCM Digital Content Management System] which is built on a pair-tree structure of METS resources with [http://ufdc.ufl.edu/design/webcontent/sobekcm/SobekCM_Resource_Object.pdf rich metadata support] including DC, MODS, MARC, VRACore, DarwinCore, IEE-LOM, GML/KML, schema.org microdata, and many other standard schemas.  The system has emphasized online, distributed creation and maintenance of resources including geo-placement and geographic searching of resources, building structure maps (table of contents) visually online, and a broad suite of curator tools.  &lt;br /&gt;
&lt;br /&gt;
This work is presented as a model which could be implemented in other systems as well.  We will demonstrate current support and discuss our upcoming roadmap to provide complete support.&lt;br /&gt;
&lt;br /&gt;
== Dead-simple Video Content Management: Let Your Filesystem Do The Work ==&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, NCSU Libraries (akorphan (at) ncsu.edu)&lt;br /&gt;
** (never led or soloed a C4L presentation)&lt;br /&gt;
&lt;br /&gt;
Content management is hard. To keep all the moving parts in order, and to maintain a layer of separation between the system and content creators (who are frequently not technical experts), we typically turn to content management systems like Drupal. But even Drupal and its kin require significant overhead and present a not inconsiderable learning curve for nontechnical users.&lt;br /&gt;
&lt;br /&gt;
In some contexts it's possible -- and desirable -- to manage content in a more streamlined, lightweight way, with a minimum of fuss and technical infrastructure. In this presentation I'll share a simple MVC-like architecture for managing video content for playback on the web, which uses a combination of Apache's mod_rewrite module and your server's filesystem structure to provide an automated approach to video content management that's easy to implement and provides a low barrier to content updates: friendly to content creators and technology implementors alike. Even better, the basic method is HTML5-friendly, and can be integrated into your favorite content management system if you've got permissions for creating templates.&lt;br /&gt;
&lt;br /&gt;
In the presentation I'll go into detail about the system structure and logic required to implement this approach. I'll detail the benefits and limitations of the system, as well as the challenges I encountered in developing its implementation. Audience members should come away with sufficient background to implement a similar system on their own servers. Implementation documentation and genericized code will also be shared, as available.&lt;br /&gt;
&lt;br /&gt;
== Managing Discovery ==&lt;br /&gt;
&lt;br /&gt;
* Andrew Pasterfield, Senior Programmer/Systems Analyst, University of Calgary Library, ampaster@ucalgary.ca&lt;br /&gt;
**No previous code4lib presentations &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
In fall 2012 the University of Calgary Library launched a new home page that incorporated a Summon powered&lt;br /&gt;
Single Search Box with customized “bento box” results display. Search at the U of C now combines a range of&lt;br /&gt;
metadata sources for discovery and customized mapping of a database recommender and LibGuide into a unified&lt;br /&gt;
display.  Further customizations include a non Google Analytics/non proxy method to log clicks.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss the technical details of bringing the various systems together into one display interface to increase discovery at the U of C Library.&lt;br /&gt;
&lt;br /&gt;
http://library.ucalgary.ca&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sorting it out: a piece of the User Centered Design Process ==&lt;br /&gt;
&lt;br /&gt;
* Cindy Beggs, [http://www.akendi.com/aboutus/management/ Akendi], cindy@akendi.com&lt;br /&gt;
&lt;br /&gt;
This talk is about how to apply a user centered design methodology to the process of creating an information architecture.  Participants learn the fundamentals of UCD and how card sorting and reverse card sorting enable us to isolate the content we present on screen from the layouts and visuals of those screens.  We talk about ways to identify who will be using the information architecture you are creating and why we need to know how it will be used.&lt;br /&gt;
 &lt;br /&gt;
What will attendees takes away from your talk?&lt;br /&gt;
The criticality of involving “real” end users in the process of creating an information architecture.  The basics of following a user-centered-design process in the creation of best in class, content-rich, digital products.&lt;br /&gt;
&lt;br /&gt;
Cindy Beggs has been working in the “information industry” for over 25 years.  A librarian by profession, she has spent decades helping users figure out how to find their way through large bodies of content.  Her insights into how people seek information, her empathy for those who find it a challenge and her practical experience helping organizations figure out how to best structure their content contribute to her success as an information architect with both clients and trainees.  (http://www.akendi.com/aboutus/management/)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Implementation of ArchivesSpace in University of Richmond==&lt;br /&gt;
&lt;br /&gt;
*Birong Ho, bho@richmond.edu&lt;br /&gt;
&lt;br /&gt;
University of Richmond implemented its archive collection management ArchivsSpace in the fall, 2013. As a charter member and the Head of Special Collection as the Board member, implementation of such an Open Source Software became a priority. &lt;br /&gt;
&lt;br /&gt;
Several aspects of implementation will be addressed in the talk. Among them, they are Collections and Repository, storage layer including data format, System resources requirements, Technical architecture, Customization, scaling and integrated with other systems in the library.&lt;br /&gt;
&lt;br /&gt;
The customization, scale, and integration with other systems such as Archeon and Exist on campus became a concern will be focused and elaborated in the talk.&lt;br /&gt;
&lt;br /&gt;
==Easy Wins for Modern Web Technologies in Libraries==&lt;br /&gt;
&lt;br /&gt;
*[mailto:trey.terrell@oregonstate.edu Trey Terrell], Analyst Programmer, Oregon State University&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Oregon State University is currently implementing an updated version of its room reservation system. In its development we've come across and implemented a variety of &amp;quot;easy wins&amp;quot; to make it more responsive, easier to maintain, less expensive to run, and just cooler to experience. While our particular system was in Ruby on Rails, this talk will address general methods and example utilities which can be used no matter your stack.&lt;br /&gt;
&lt;br /&gt;
I'll be talking about things like cache management, reverse proxies, publish/subscribe servers, WebSockets, responsive design, asynchronous processing, and keeping complicated stacks up and running with minimal effort.&lt;br /&gt;
&lt;br /&gt;
==Implementing Islandora at a Small Institution==&lt;br /&gt;
&lt;br /&gt;
*Megan Kudzia, Albion College Library&lt;br /&gt;
*Eddie Bachle, Albion College IT&lt;br /&gt;
**No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
Albion College (and particularly the Library/Archives and Special Collections) has a variety of needs which could be met by an open-source Institutional Repository system. Several months and lots of conversations later, we’re continuing to troubleshoot our way through Islandora. We’d like to talk about what has worked for us, where our frustrations have been, whether it’s even possible to install and develop a system like this at a small institution, and where the process has stalled. &lt;br /&gt;
&lt;br /&gt;
As of right now, we do have a semi-working installation. We’re not sure when it will be ready for our end users, but we'll talk about our development process and evaluate our progress.&lt;br /&gt;
''Contributions also by Nicole Smeltekop, Albion College Archives &amp;amp; Special Collections''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== PhantomJS+Selenium: Easy Automated Testing of AJAX-y UIs ==&lt;br /&gt;
&lt;br /&gt;
* Martin Haye, California Digital Library, martin.haye@ucop.edu&lt;br /&gt;
** Previous Code4Lib Presentation: [http://code4lib.org/conference/2012/collett Beyond code: Versioning data with Git and Mercurial] at Code4Lib 2012 (Martin co-presenting with Stephanie Collett)&lt;br /&gt;
* Mark Redar, California Digital Library, mark.redar@ucop.edu&lt;br /&gt;
&lt;br /&gt;
Web user interfaces are demanding ever-more dynamism and polish, combining HTML5, AJAX, lots of CSS and jQuery (or ilk) to create autocomplete drop-downs, intelligent buttons, stylish alert dialogs, etc. How can you make automated tests for these highly complex and interactive UIs?&lt;br /&gt;
&lt;br /&gt;
Part of the answer is PhantomJS. It’s a modern WebKit browser that’s “headless” (meaning it has no display) that can be driven from command-line Selenium unit tests. PhantomJS is dead simple to install, and its blazing speed and server-friendliness make continuous integration testing easy. You can write UI unit tests in {language-of-your-choice} and run them not just in PhantomJS but in Firefox and Chrome, plus a zillion browser/OS combinations at places like SauceLabs, TestingBot and BrowserStack.&lt;br /&gt;
&lt;br /&gt;
In this double-team live code talk, we’ll explain all that while we demonstrate the following in real time:&lt;br /&gt;
&lt;br /&gt;
* Start with nothing.&lt;br /&gt;
* Install Selenium bindings for Ruby and Python.&lt;br /&gt;
* In each language write a small test of an AJAX-y UI.&lt;br /&gt;
* Run the tests in Firefox, and fix bugs (in the test or UI) as needed.&lt;br /&gt;
* Install PhantomJS.&lt;br /&gt;
* Show the same tests running headless as part of a server-friendly test suite. &lt;br /&gt;
* (Wifi permitting) Show the same tests running on a couple different browser/OS combinations on the server cloud at SauceLabs – talking through a tunnel to the local firewalled application.&lt;br /&gt;
&lt;br /&gt;
==New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library==&lt;br /&gt;
&lt;br /&gt;
* Stephanie Walker – swalker@brooklyn.cuny.edu&lt;br /&gt;
* Howard Spivak – howards@brooklyn.cuny.edu&lt;br /&gt;
* Alex - Alex@brooklyn.cuny.edu&lt;br /&gt;
&lt;br /&gt;
Academic libraries are caught in budget squeezes and often struggle to find ways to communicate value to senior administration and others.  At Brooklyn College Library, we have taken an unusual, possibly unique, approach to these issues.  Our technology staff have long worked directly with librarians to develop products that meet library, faculty, and student needs, and we have shared many of our products with colleagues, including an award-winning website, e-resource, and content management system we call 4MyLibrary, which we shared for free with 8 CUNY colleges, and also an easy-to-use book scanner, which has proven overwhelming popular with students, faculty, other librarians, and numerous campus offices.  Recently, motivated by budget cuts, we decided that what worked for us might interest other libraries, and working with our Office of Technology Commercialization, we started selling 2 products:  our book scanners (at half the price of commercial alternatives), and a hosting service, whereby we could host and support 4MyLibrary for libraries with minimal technology staff.  Both succeeded, and yielded major benefits:  a steady revenue stream and the admiration and serious goodwill of our senior administration and others.   However, this presentation is neither a basic how-to, nor an advertisement.  With this presentation, we hope to spur a conversation for broader collaboration, especially regarding new technologies, among libraries.  We all have some level of technical expertise, most of us are struggling with rising prices and tight budgets, and many of us are unhappy with various technology products we use, from scanners to our ILS.  We believe – and can demonstrate – that with collaboration, we can solve many of our problems, and provide better services to boot. &lt;br /&gt;
&lt;br /&gt;
== Identifiers, Data, and Norse Gods ==&lt;br /&gt;
&lt;br /&gt;
* Ryan Scherle, [http://datadryad.org Dryad Digital Repository], ryan@datadryad.org&lt;br /&gt;
** previous Code4Lib talk [http://ryan.scherle.org/papers/2010-2-code4lib-HIVE.ppt  HIVE: A New Tool for Working With Vocabularies], at Code4Lib 2011.&lt;br /&gt;
&lt;br /&gt;
ORCID and DataCite provide stable identifiers for researchers and and data, respectively. Each system does a fine job of providing value to its users. But wouldn't it be great if they could link their systems to create something much more powerful? Perhaps even as powerful as a god?&lt;br /&gt;
&lt;br /&gt;
Enter [http://odin-project.eu/ ODIN], The ORCID and DataCite Interoperability Network. ODIN is a two-year project to unleash the power of persistent identifiers for researchers and the research they create. This talk will present recent work from the ODIN project, including several tools that can unleash the godlike power of identifiers at your institution. Current tools include:&lt;br /&gt;
* Metadata generator tool: allows repository staff to create DataCite metadata with embedded ORCIDs.&lt;br /&gt;
* Claiming tool: assists researchers in claiming their work within the ORCID system.  &lt;br /&gt;
* ORCID-feed: includes a list of ORCID works on any web page.&lt;br /&gt;
* ODIN's HAMR: assists in populating a DSpace repository with ORCIDs. Based on work from a Code4Lib hackathon!&lt;br /&gt;
&lt;br /&gt;
== Armed Bandits in the Digital Library ==&lt;br /&gt;
&lt;br /&gt;
* Roman Chyla, [http://labs.adsabs.harvard.edu/adsabs/ Astrophysics Data System], rchyla@cfa.harvard.edu&lt;br /&gt;
** Previous Code4Lib: [http://code4lib.org/conference/2013/chyla Citation search in SOLR and second-order operators]&lt;br /&gt;
&lt;br /&gt;
Many of us are using the excellent Lucene library (or SOLR appliance) to provide search functionality. These systems contain number of features to adjust relevancy ranking of hits, but we may not know how to use them. In this presentation, I'll present the available options - eg. what is the default ranking 'Vector space model, what are the alternatives (eg. BM25) and what are the other options we have to tweak and adjust the ranking of the hits (eg. boost factors, functions). But even if we know how to deploy these adjustments and tweaks, we are still left in dark. We do not know whether the change we've just rolled out had a significant (statistically significant) effect or maybe it was just a waste of time and resources? A/B testing is one option, but there may be a much better one - so called &amp;quot;Multi-Armed Bandits Approach&amp;quot;. And in this talk I'd like to show how we are experimenting with this strategy to adjust [http://labs.adsabs.harvard.edu/adsabs/ ADS search engine].&lt;br /&gt;
&lt;br /&gt;
== Building Worker Queues with AWS and Resque ==&lt;br /&gt;
&lt;br /&gt;
* Eric Rochester [http://scholarslab.org Scholars' Lab], erochest@virginia.edu&lt;br /&gt;
* Scott Turnbull [http://aptrust.org/ Academic Preservation Trust], scott.turnbull@aptrust.org &lt;br /&gt;
&lt;br /&gt;
A common task in larger systems is to be able to process large input files automatically. Often users can drop those files into a shared directory on AWS or on NFS or another shared drive. Those files need to be processed and potentially integrated into a system. This task has come up recently in the University of Virginia libraries in allowing users to add GIS data to the system and in setting up a system for the Academic Preservation Trust (http://aptrust.org/) that ingests files and resources into the preservation system.&lt;br /&gt;
&lt;br /&gt;
This system is built by loosely coupling a number of different technologies. This allows us to easily interoperate and communicate between different system and programming environments. Because the interfaces are well defined, it’s also fairly simple to switch out technologies as the requirements of the system change.&lt;br /&gt;
&lt;br /&gt;
The process is fairly simple:&lt;br /&gt;
&lt;br /&gt;
First, a Ruby daemon monitors an AWS S3 bucket that others can upload new files into. This daemon creates a Resque status task, adds a marker for the task in a database, and continues monitoring.&lt;br /&gt;
&lt;br /&gt;
Second, Resque mediates incoming job requests and routes them to the appropriate workers which may be in Java, Go, or Ruby.  The diversity of technologies that Resque can manage allows great latitude to leverage the appropriate tool for a specific job.  While processing, it updates the status for that job and coordinates processing with other jobs.&lt;br /&gt;
&lt;br /&gt;
Finally, a page that is integrated into a larger Rails app provides a novice-user-friendly view of the status of the workers and allows basic tasks such as restarting the job.&lt;br /&gt;
&lt;br /&gt;
This architecture allows us to swap in the technology that best fits each part of the process, and it makes it easier to maintain the system. We use this to integrate and coordinate between tasks handled in Java, Ruby, and Go, and it provides an effective way to interoperate with these programming languages and the respective strengths that they bring to this system.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sustaining your Open Source project through training ==&lt;br /&gt;
&lt;br /&gt;
* Bess Sadler (Stanford University Libraries) and Mark Bussey (Data Curation Experts) will discuss their experiences developing and delivering training for Project Hydra.&lt;br /&gt;
&lt;br /&gt;
Topics covered:&lt;br /&gt;
* Working practices for developing training materials&lt;br /&gt;
* Sharing the work when there are no dedicated resources&lt;br /&gt;
* Inviting community (and student) input to create higher quality content&lt;br /&gt;
* Strategies to keep training docs up-to date&lt;br /&gt;
* Strategies to make training materials available to the widest-possible audience&lt;br /&gt;
* Using surveys (Survey Monkey) to assess the effectiveness of your training program&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Piwik: Open source web analytics==&lt;br /&gt;
* Kirk Hess, University of Illinois at Urbana-Champaign (kirkhess@illinois.edu)&lt;br /&gt;
** (Code4Lib 2012: [http://code4lib.org/conference/2012/hess| Discovering Digital Library User Behavior with Google Analytics])&lt;br /&gt;
&lt;br /&gt;
While Google Analytics is synonymous with Web Analytics, fortunately today we have many other good options, and one option is Piwik, [http://piwik.org| piwik.org] a simple to install, open-source PHP/MySQL application with a tracking script that will sit alongside Google Analytics tracking the usual clicks, events and variables. In this presentation, I'd like to cover the usual analytics topics and also cover what makes Piwik powerful, such as importing and visualizing web logs from any system to incorporate both past and future data, easily tracking downloads, and the ability to write your own reports or dashboard. The visitor log data is stored securely on your own server so you have control over who looks at the data and how much or how little to keep. With an active and helpful developer community, Piwik has the potential for analytics which makes sense for libraries, not e-commerce.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Next Generation Catalogue - RDF as a Basis for New Services == &lt;br /&gt;
* Anne-Lena Westrum – digitalutvikling@gmail.com&lt;br /&gt;
* Benjamin Rokseth &lt;br /&gt;
* Asgeir Rekkavik &lt;br /&gt;
* Petter Goksøyr Åsen&lt;br /&gt;
&lt;br /&gt;
Oslo Public Library has converted the entire MARC-catalogue to RDF via the self-made conversion tool MARC2RDF.&lt;br /&gt;
 &lt;br /&gt;
[http://digital.deichman.no/data.deichman.no/| data.deichman.no], the enriched RDF version of the library catalogue including its authority files, forms the basis for two different mashups; The Active shelf and the Book recommendation database. The RDF catalogue is linked with various content and the dataset is updated daily to account for additions, deletions and changes made in the MARC catalogue. &lt;br /&gt;
&lt;br /&gt;
[http://vimeo.com/68687814| The Active shelf] is a physical touchscreen device that makes use of open source software, RFID technology, RDF data and external web service APIs to provide information about any library book a patron is curious to know more about.&lt;br /&gt;
&lt;br /&gt;
The Book recommendations database stores book recommendations written by library staff from all over Norway and links them to the RDF-representation of the MARC-catalogue.&lt;br /&gt;
&lt;br /&gt;
==Economics of Scale: Thinking about Metadata Quality and Completeness for Fun and Profit==&lt;br /&gt;
* William Hicks, University of North Texas (William.hicks@unt.edu)&lt;br /&gt;
&lt;br /&gt;
The UNT Libraries Digital Collections constitute three internet gateways, The Portal to Texas History, UNT Digital Library, and the Gateway to Oklahoma History, making available to the public a wide range of materials, from photographs and newspapers, to dissertations and recordings of music ensemble performances. The collections disseminate over 500,000 unique items, that were used over 9 millions times last year  and with growth trends in both areas shows no signs of slowing. &lt;br /&gt;
 &lt;br /&gt;
As the size and scope of our collections has grown, so to has a pressing need to think clearly about the quality of our metadata, the completeness of our records, and the most efficient way of doing metadata entry. Not surprisingly there have been a few things written on the subject and so over the last few months we’ve started writing new code and getting the infrastructure of our metadata  editing system to a place where we can begin to test these ideas on our ever expanding dataset.  What kinds of questions are we looking to answer, and what types of tools are we building?  That’s what this talk will be all about, but here are a few ideas to ponder:&lt;br /&gt;
 &lt;br /&gt;
* What kinds of tools have we built, or can we employ to standardize data entry and aid the user in their input needs?&lt;br /&gt;
* How close does a metadata record come to a “completeness” standard?  What does that even look like? What are the implications when we look at such a standard at scale?&lt;br /&gt;
* If we can identify what we think a “quality” metadata record “is”, historically speaking, how close do we get to that ideal? &lt;br /&gt;
* Does an item’s history matter? Can we quantify it and locate value in change through time?&lt;br /&gt;
* What are the economic costs of metadata entry? If we have enough quantifiable measures about the types of objects in our systems, and we can profile our data entry personnel, what can this say about optimizing staff time and return on investment?&lt;br /&gt;
* What sort of priorities are we setting for ourselves when we treat all items as equal, when clearly some types of materials get vastly more use by the public.&lt;br /&gt;
* Finally what kinds of analysis tools might we develop to gauge our overall metadata “health,” to steer projects, or to ultimately improve our systems for our end user’s needs?&lt;br /&gt;
&lt;br /&gt;
Most of our questions are still quite open ended, and honestly we are just getting started down this road. But as digital collections grow, and library budgets realign or shrink, it becomes increasingly important to back up our assertions and opinions with numbers, and find more efficient ways to work with the resources we have.&lt;br /&gt;
&lt;br /&gt;
==More Like This: Approaches to Recommending Related Items using Subject Headings==&lt;br /&gt;
* Kevin Beswick, NCSU Libraries (kdbeswic@ncsu.edu)&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
 &lt;br /&gt;
With a significant portion of the collection at our new Hunt Library being housed in an automated storage and retrieval system, several of us at NCSU Libraries have begun looking at ways to replace and improve upon the classic shelf browsing experience in order to make it easier for patrons to browse related materials. Our goal is to mimic popular services like Amazon and Netflix, which utilize recommendation engines to make it easy for users to find items similar to a particular item of interest. While there have been previous efforts in libraries to recreate this experience using circulation or call number data, we are currently investigating algorithms that focus on use of subject headings. Use of subject headings as an alternative can be particularly helpful in the case of electronic materials that do not always have call numbers or circulation data. In this talk, I will share:&lt;br /&gt;
 &lt;br /&gt;
* Details of the proposed algorithms &lt;br /&gt;
* How these algorithms were quickly and easily implemented using Solr. &lt;br /&gt;
* Our evaluation process and its outcomes in terms of the effectiveness of the algorithms.&lt;br /&gt;
* How this has (or could) impact presentation of recommended items in our discovery layer.&lt;br /&gt;
&lt;br /&gt;
== Questioning Authority: building a ruby gem to facilitate UI interactions with varied controlled vocabularies ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mhbussey|Mark Bussey]], Data Curation Experts, mark@curationexperts.com&lt;br /&gt;
&lt;br /&gt;
At a recent Hydra meeting, developers from five different institutions all realized that they had similar needs to support various types of UI fields based on a multiple of internal and external authorities and controlled vocabularies.  Their goals was to develop a tool that let them meet these needs in ways that minimized the need for custom coding for each vocabulary.  During an intense three-day working session, they minted the initial release of the [https://github.com/projecthydra/questioning_authority/blob/master/README.md questioning authority] gem.  &lt;br /&gt;
&lt;br /&gt;
The talk will cover both how cross-institutional development helped speed development and how the gem can be used for accessing both external vocabularies like LCSH and LCNA and for presenting internal vocabulary lists.  Although the developing institutions are all Hydra implementers, the gem itself doesn't have any Hydra dependencies and can be used in any Rails or Blacklight based application.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Building Hydra, a framework; a community ==&lt;br /&gt;
&lt;br /&gt;
Justin Coyne - Project Hydra contributor / Data Curation Experts - justin at curationexperts.com&lt;br /&gt;
no previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
More than just a repository, the [http://projecthydra.org Hydra Project] is a community of cultural heritage institutions dedicated to pooling knowledge and resources.  It is a completely open source project that has grown continuously for over 5 years.  Within this vibrant community, a number of conventions and practices have emerged that we believe will benefit others attempting to cultivate support for their community oriented projects. The Hydra Project is now a mature initiative which is producing shareable, reusable and customizable components as well as complete repository solutions.  In a time of tight budgets and growing demand for improved systems, we believe that &amp;quot;the Hydra way&amp;quot; is the exemplar case in the library community for how to work across institutions to deliver high quality services to our patrons.  This talk will cover both the technical and human processes that have sustained Hydra's continued development and growth.&lt;br /&gt;
&lt;br /&gt;
From [http://www.ohloh.net/p/projecthydra Ohloh.net]&lt;br /&gt;
In a Nutshell, Project Hydra...&lt;br /&gt;
* has had 8,364 commits made by 64 contributors representing 60,733 lines of code&lt;br /&gt;
* has a codebase with a long source history maintained by a very large development team with stable Y-O-Y commits&lt;br /&gt;
* took an estimated 15 years of effort (COCOMO model) starting with its first commit in October, 2009 ending with its most recent commit 7 days ago&lt;br /&gt;
&lt;br /&gt;
== JQuery XML Editor ==&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Presenter: Ben Pennell, UNC Chapel Hill Libraries (bbpennel@email.unc.edu)&lt;br /&gt;
no previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The jquery.xmleditor is a portable jquery widget developed by the University of North Carolina at Chapel Hill Libraries for the purpose of simplifying the description workflow for existing objects in our digital repository.  It does so by adding context and structure informed by an underlying XML schema.  Even more generally, it creates and modifies XML documents in your web browser.  &lt;br /&gt;
&lt;br /&gt;
It can be found here, including a live demo:&lt;br /&gt;
[https://github.com/UNC-Libraries/jquery.xmleditor]&lt;br /&gt;
&lt;br /&gt;
Features include:&lt;br /&gt;
* Graphical editor mode for displaying and modifying XML elements&lt;br /&gt;
* Text editor mode for directly modifying the underlying document (using the Cloud9 editor)&lt;br /&gt;
* Contextual, schema driven menus for adding new elements, subelements and attributes in both the graphical and text editing modes&lt;br /&gt;
* Fully javascript and CSS based, jquery widget&lt;br /&gt;
* AJAX submission of document modifications&lt;br /&gt;
* Ability to export XML document to a file in web browsers that support it&lt;br /&gt;
* Keyboard shortcuts for navigation and other operations&lt;br /&gt;
* Standalone tool for building JSON representations of XML schemas &lt;br /&gt;
&lt;br /&gt;
In our own implementation, the tool communicates with a Fedora based SWORD 2 enabled repository to receive the starting MODS document and to submit changes.  But it's all XML in the end, and includes options for exporting to file or submitting to any endpoint that accepts XML.&lt;br /&gt;
&lt;br /&gt;
This presentation will include an overview of the development process, technologies and issues involved, as well as a brief demonstration of the editor in use.  It will also touch on the tool backing the editor which constructs JSON objects from schemas.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Library Resources as Networks ==&lt;br /&gt;
* [mailto:matthewmiller@nypl.org Matt Miller] New York Public Library, NYPL Labs.&lt;br /&gt;
**No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Library resources are typically presented linearly in the form of a catalog search results page or an iterative list of subjects, books, special collections, etc. This talk explores the possibilities created when thinking of library resources as interconnected networks. We will look at the progress of a project to visualize NYPL resources such as catalog subject headings[1][2] as a network. We will also look at moving beyond visualizations into building network interfaces, such as our archival access term explorer[3] prototype. &lt;br /&gt;
&lt;br /&gt;
[1] [https://dl.dropboxusercontent.com/u/4070829/catalog-viz-subjects/seadragon.html Catalog Subject Headings Visualization]&lt;br /&gt;
&lt;br /&gt;
[2] [https://dl.dropboxusercontent.com/u/16562899/timelapse6.mp4 Time lapsed catalog network]&lt;br /&gt;
&lt;br /&gt;
[3] [http://archives.nypl.org/terms Archival access term explorer prototype.]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>MatthewMiller</name></author>	</entry>

	</feed>