<?xml version="1.0"?>
<feed xmlns="http://www.w3.org/2005/Atom" xml:lang="en">
		<id>https://wiki.code4lib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ryscher</id>
		<title>Code4Lib - User contributions [en]</title>
		<link rel="self" type="application/atom+xml" href="https://wiki.code4lib.org/api.php?action=feedcontributions&amp;feedformat=atom&amp;user=Ryscher"/>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/Special:Contributions/Ryscher"/>
		<updated>2026-04-09T20:23:16Z</updated>
		<subtitle>User contributions</subtitle>
		<generator>MediaWiki 1.26.2</generator>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_preconference_proposals&amp;diff=40117</id>
		<title>2014 preconference proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_preconference_proposals&amp;diff=40117"/>
				<updated>2013-12-09T19:51:11Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Managing Projects: Or I'm in charge, now what? (aka PM4Lib) */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;= PROPOSALS ARE CLOSED : PLEASE DO NOT ADD NEW PRECONFERENCES TO THIS PAGE =&lt;br /&gt;
&lt;br /&gt;
Proposals were accepted through December 6th, 2013.&lt;br /&gt;
&lt;br /&gt;
It would be really, super duper helpful if folks who think they might want to attend a pre-conference could indicate interest by adding your name to a session below. &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Note===&lt;br /&gt;
Attendance at a pre-conference will require a small fee ''due at the time of conference registration&amp;quot;.&lt;br /&gt;
 &lt;br /&gt;
Although this was specified in the email announcements relating to pre-conferences, it was not added to this page until December 2nd.  I (Adam C.) apologize for the omission and I hope this will not cause any &amp;quot;sticker shock.&amp;quot;  Putting your name on this list does not incur any obligation on your part, but we'll be using it to gauge interest and work out room assignments.&lt;br /&gt;
&lt;br /&gt;
Please put your pre-conference on the list in the following format:&lt;br /&gt;
&lt;br /&gt;
=Code4Lib 2014 Pre-Conference Proposals=&lt;br /&gt;
&lt;br /&gt;
===Drupal4lib Sub-con Barcamp===&lt;br /&gt;
=====Full Day=====&lt;br /&gt;
&lt;br /&gt;
* Contact [[User:highermath|Cary Gordon]], cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
This will be a full day of self-selected barcamp style sessions. Anyone who wants to present can write down the topic on an index card and, after the keynote, we will vote to choose what we want to see. Attendees can also pick a topic and attempt to talk someone else into presenting on it.&lt;br /&gt;
&lt;br /&gt;
This event is open to the library community. There will be a nominal fee (t/b/d) for non-Code4LibCon attendees (subject to organizer approval).&lt;br /&gt;
&lt;br /&gt;
[[resources to help you learn drupal]]&lt;br /&gt;
&lt;br /&gt;
====Interested in Attending:====&lt;br /&gt;
&lt;br /&gt;
=====All Day=====&lt;br /&gt;
&lt;br /&gt;
* Renna Tuten &lt;br /&gt;
&lt;br /&gt;
=====Morning=====&lt;br /&gt;
&lt;br /&gt;
* Kevin Reiss&lt;br /&gt;
* Charlie Morris (NCSU) - glad to see this again this year!&lt;br /&gt;
&lt;br /&gt;
=====Afternoon=====&lt;br /&gt;
&lt;br /&gt;
&amp;amp;nbsp;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Open Refine Hackfest===&lt;br /&gt;
'''&amp;quot;Half-Day&amp;quot;'''&lt;br /&gt;
* Contact [[User:bibliotechy|Chad Nelson]], chadbnelson@gmail.com&lt;br /&gt;
&lt;br /&gt;
[http://openrefine.org/ Open Refine] is a powerful open source tool for wrangling messy data that can also be used to help in the creation of Linked Data via the [https://github.com/OpenRefine/OpenRefine/wiki/Reconciliation-Service-API Reconciliation API]. It is possible to write reconciliation services against API's, like the [http://iphylo.blogspot.com/2013/04/reconciling-author-names-using-open.html VIAF service] or, even just against local authority files for helping maintain authority control&lt;br /&gt;
&lt;br /&gt;
The session would first introduce Open Refine, then walk through building a reconciliation service, and the rest of the session would be a hackfest where we build new reconciliation services for public consumption or local use. &lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&amp;lt;ol&amp;gt;&lt;br /&gt;
&amp;lt;li&amp;gt;Adam Constabaris&lt;br /&gt;
&amp;lt;li&amp;gt;Ray Schwartz&lt;br /&gt;
&amp;lt;li&amp;gt;Jason Stirnaman&lt;br /&gt;
&amp;lt;li&amp;gt;Joshua Gomez&lt;br /&gt;
&amp;lt;li&amp;gt;Sam Kome&lt;br /&gt;
&amp;lt;/ol&amp;gt;&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Responsive Design Hackfest===&lt;br /&gt;
'''&amp;quot;Half-Day [Afternoon]&amp;quot;''' &lt;br /&gt;
* Contact Jim Hahn, University of Illinois, jimfhahn@gmail.com&lt;br /&gt;
* Contact David Ward, University of Illinois, dh-ward@illinois.edu&lt;br /&gt;
&lt;br /&gt;
This structured hackfest will give attendees an opportunity to explore methods to create responsive mobile apps using the Bootstrap framework [http://getbootstrap.com/]and a set of APIs for accessing library data. We will start with an API template for creating space-based mobile tools that draw from work coming out of the IMLS funded Student/Library Collaborative grant [http://www.library.illinois.edu/nlg_student_apps]. Available APIs will include a room reservation template and codebase for implementing at any campus and the set of Minrva catalog APIs generating JSONP [http://minrvaproject.org/services.php]. &lt;br /&gt;
&lt;br /&gt;
Hosts will give a brief report of a study on student hacking projects and interests in mobile library apps that are the basis for the templates utilized in this Hackathon. By the end of the pre-conference attendees will have a sample responsive mobile web app in Bootstrap 3 to bring back to their campus which can plug into their site-based content.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Intro to Blacklight ===&lt;br /&gt;
'''&amp;quot;Half-Day [Morning]&amp;quot;''' &lt;br /&gt;
* Contact: Chris Beer, Stanford University, cabeer@stanford.edu&lt;br /&gt;
* TA: Bess Sadler, Stanford University, bess@stanford.edu&lt;br /&gt;
&lt;br /&gt;
This session will be walk-through of the architecture of Blacklight, the community, and an introduction to building a Blacklight-based application. Each participant will have the opportunity to build a simple Blacklight application, and make basic customizations, while using a test-driven approach.&lt;br /&gt;
&lt;br /&gt;
For more information about Blacklight see our wiki ( http://projectblacklight.org/ ) and our GitHub repo ( https://github.com/projectblacklight/blacklight ). We will also send out some brief instructions beforehand for those that would like to setup their environments to follow along and get Blacklight up and running on their local machines.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Megan Kudzia&lt;br /&gt;
# Bret Davidson&lt;br /&gt;
# Coral Sheldon-Hess&lt;br /&gt;
# Cory Lown&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Blacklight Hackfest===&lt;br /&gt;
'''&amp;quot;Half-Day [Afternoon]&amp;quot;''' &lt;br /&gt;
* Contact Chris Beer, Stanford University, cabeer@stanford.edu&lt;br /&gt;
&lt;br /&gt;
This afternoon hackfest is both a follow-on to the Intro to Blacklight morning session to continue building Blacklight-based applications, and also an opportunity for existing Blacklight contributors and members of the Blacklight community to exchange common patterns and approaches into reusable gems or incorporate customizations into Blacklight itself.&lt;br /&gt;
&lt;br /&gt;
For more information about Blacklight see our wiki ( http://projectblacklight.org/ ) and our GitHub repo ( https://github.com/projectblacklight/blacklight ).&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Shaun Ellis&lt;br /&gt;
# Kevin Reiss&lt;br /&gt;
# Megan Kudzia&lt;br /&gt;
# Erik Hatcher&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
===RailsBridge: Intro to programming in Ruby on Rails===&lt;br /&gt;
'''&amp;quot;Half-Day&amp;quot; [morning]'''&lt;br /&gt;
* Contact Justin Coyne, Data Curation Experts, justin@curationexperts.com&lt;br /&gt;
&lt;br /&gt;
Interested in learning how to program? Want to build your own web application? Never written a line of code before and are a little intimidated? There's no need to be! RailsBridge is a friendly place to get together and learn how to write some code.&lt;br /&gt;
&lt;br /&gt;
RailsBridge is a great workshop that opens the doors to projects like Blacklight and Hydra.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
1. Ayla Stein&lt;br /&gt;
&lt;br /&gt;
2. Heidi Dowding&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
===Managing Projects: Or I'm in charge, now what? (aka PM4Lib)===&lt;br /&gt;
'''Full-Day'''&lt;br /&gt;
&lt;br /&gt;
Contact: &lt;br /&gt;
* [[User:rosy1280|Rosalyn Metz]], rosalynmetz@gmail.com&lt;br /&gt;
* [[User:yoosebj|Becky Yoose]], yoosebec@grinnell.edu&lt;br /&gt;
&lt;br /&gt;
This will be a full day session on project management.  We'll cover&lt;br /&gt;
* '''Kicking off the Project''' -- project lifecycle, project constraints, scoping/goals, stakeholders, assessment&lt;br /&gt;
* '''Planning the Project''' -- project charters, work breakdown structures, responsibilities, estimating time, creating budgets&lt;br /&gt;
* '''Executing the Project''' -- status meeting, status reports, issue management&lt;br /&gt;
* '''Finishing the Project''' -- achieving the goal, post mortems, project v. product&lt;br /&gt;
This is a revival of rosy1280's LITA Forum Pre-Conference, but better (because iteration is good) and adapted to c4lib types.&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Robin Dean&lt;br /&gt;
# Erin White&lt;br /&gt;
# Andrew Darby&lt;br /&gt;
# Sam Kome&lt;br /&gt;
# Ryan Scherle&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Fail4Lib 2014===&lt;br /&gt;
'''Half Day [TBD, probably afternoon]'''&lt;br /&gt;
&lt;br /&gt;
Contacts: &lt;br /&gt;
* Andreas Orphanides, akorphan (at) ncsu.edu&lt;br /&gt;
* Jason Casden, jmcasden (at) ncsu.edu&lt;br /&gt;
&lt;br /&gt;
The task of design (and the work that we do as library coders) is intimately tied to failure. Failures, both big and small, motivate us to create and improve. Failures are also occasionally the result of our work. Understanding and embracing failure, encouraging enlightened risk-taking, and seeking out opportunities to fail and learn are essential to success in our field. At Fail4Lib, we'll talk about our own experiences with projects gone wrong, explore some famous design failures in the real world, and talk about how we can come to terms with the reality of failure, to make it part of our creative process -- rather than something to be feared.&lt;br /&gt;
&lt;br /&gt;
The schedule may include the following:&lt;br /&gt;
&lt;br /&gt;
* Case studies. We'll look at some classic failures from the literature: What can we learn from the mistakes of others?&lt;br /&gt;
* Confessionals, for those willing to share. Talk about your own experiences with rough starts, labor pains, and doomed projects in your own work: What can we learn from our own (and each others') failures?&lt;br /&gt;
* Group therapy. Let's talk about how to deal with risk management, failed projects, experimental endeavors, and more: How can we make ourselves, our colleagues, and our organizations more fault tolerant? How do we make sure we fail as productively as possible?&lt;br /&gt;
&lt;br /&gt;
''Interested in attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#Bret Davidson&lt;br /&gt;
#Mike Graves&lt;br /&gt;
#Ray Schwartz&lt;br /&gt;
#Jason Stirnaman&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
===CLLAM @ code4lib===&lt;br /&gt;
'''(Computational Linguistics for Libraries, Archives and Museums)'''&lt;br /&gt;
&lt;br /&gt;
'''Full Day'''&lt;br /&gt;
&lt;br /&gt;
Contacts: &lt;br /&gt;
* Douglas W. Oard (primary), oard (at) umd.edu &lt;br /&gt;
* Corey Harper, corey (dot) harper (at) nyu.edu&lt;br /&gt;
* Robert Sanderson, azaroth42 (at) gmail.com &lt;br /&gt;
* Robert Warren, rwarren (at) math.carleton.ca&lt;br /&gt;
&lt;br /&gt;
We will hack at the intersection of diverse content from Libraries, Archives and Museums and bleeding edge tools from computational linguistics for slicing and dicing that content. Did you just acquire the email archives of a startup company? Maybe you can automatically build an org chart. Have you got metadata in a slew of languages? Perhaps you can search it all using one query. Is name authority control for e-resources getting too costly? Let’s see if entity linking techniques can help. These are just a few teasers. &lt;br /&gt;
&lt;br /&gt;
There’ll be plenty of content and tools supplied, but please bring your own [data] too -- you’ll hack with it in new ways throughout the day. We’ll get started with some lightning talks on what we’ve brought,then we’ll break up into groups to experiment and work on the ideas that appeal. Three guaranteed outcomes: you’ll walk away with new ideas, new tools, and new people you’ll have met.&lt;br /&gt;
&lt;br /&gt;
''Interested in attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Devon Smith&lt;br /&gt;
# Kevin S. Clarke&lt;br /&gt;
# Jason Stirnaman&lt;br /&gt;
# Joshua Gomez&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== GeoHydra: Managing geospatial content ===&lt;br /&gt;
&lt;br /&gt;
'''Half-day [Afternoon]'''&lt;br /&gt;
&lt;br /&gt;
* Contact: Darren Hardy, Stanford University, drh@stanford.edu&lt;br /&gt;
* Moderator: Bess Sadler, Stanford University, bess@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Do you have digitized maps, GIS datasets like Shapefiles, aerial photography,&lt;br /&gt;
etc., all of which you want to integrate into your digital repository? In this&lt;br /&gt;
workshop, we will discuss how Hydra can provide discovery, delivery, and&lt;br /&gt;
management services for geospatial assets, as well as solicit questions about&lt;br /&gt;
your own GIS projects. We aim to help answer the following questions you might have about putting geospatial data into your Hydra-based digital library:&lt;br /&gt;
&lt;br /&gt;
* What are the types of geospatial data?&lt;br /&gt;
* How to dive into Hydra?&lt;br /&gt;
* How to model geospatial holdings with Hydra?&lt;br /&gt;
* How to discover and view geospatial data?&lt;br /&gt;
* How to build a geospatial data infrastructure?&lt;br /&gt;
* What are common approaches and problems?&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Esmé Cowles&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Technology, Librarianship, and Gender: Moving the conversation forward===&lt;br /&gt;
'''Full Day'''&lt;br /&gt;
&lt;br /&gt;
Contact: Lisa Rabey lisa @ biblyotheke dot net | [http://twitter.com/pnkrcklibrarian @pnkrcklibrarian]&lt;br /&gt;
&lt;br /&gt;
'''Description'''&lt;br /&gt;
&lt;br /&gt;
Librarianship is largely made up of women, yet women are significantly underrepresented in tech positions, on any level, within libraries themselves. Why? What are we doing to encourage women to become more involved in STEM within librarianship? What kind of message are we sending when library technology keynotes remain almost resolutely male? How are we changing the face of technology, not only within libraries, but with the field itself? How are we training our staff and colleagues in the areas of fairness and removal of bias? Our vendors?&lt;br /&gt;
&lt;br /&gt;
Lots of tough questions.&lt;br /&gt;
&lt;br /&gt;
While the conversation has been going on via various blogs and articles within the last few years, it was given a public face at [http://infotoday.com/il2013/day.asp?day=Monday#session_D105 Internet Librarian 2013] where a panel of 7 (four women, three men) gave personal experiences on the above and then opened up the conversation to the audience. As eye opening and enriching the conversation was, a 45 minute panel was not enough. One thing remains clear: We need to keep the conversation moving forward and start making some radical changes in the way we think, act, and how we need to harness this to start making real changes within librarianship itself.&lt;br /&gt;
&lt;br /&gt;
Topics to include:  Fairness, bias, impostor syndrome, code of conducts, sexual harassment, training opportunities, support systems,  mentoring, ally support, and more&lt;br /&gt;
&lt;br /&gt;
Those attending should expect: Begin with opening up the conversation of experiences and talking about what is most needed, spending remaining time putting together live, usable solutions to start implementing as well as pushing the conversation forward at local levels&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
=====All Day=====&lt;br /&gt;
1. Kate Kosturski&lt;br /&gt;
&lt;br /&gt;
2. Valerie Aurora&lt;br /&gt;
&lt;br /&gt;
3. Declan Fleming (I'd be good with a half day too)&lt;br /&gt;
&lt;br /&gt;
=====Morning=====&lt;br /&gt;
1. Shaun Ellis&lt;br /&gt;
&lt;br /&gt;
2. Jason Casden&lt;br /&gt;
&lt;br /&gt;
=====Afternoon=====&lt;br /&gt;
1. Ayla Stein&lt;br /&gt;
&lt;br /&gt;
2. Heidi Dowding&lt;br /&gt;
&lt;br /&gt;
3. Coral Sheldon-Hess&lt;br /&gt;
&lt;br /&gt;
4. Cory Lown&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===FileAnalyzer: Rapid Development of File Manipulation Tasks===&lt;br /&gt;
'''&amp;quot;Half-Day&amp;quot; [morning]'''&lt;br /&gt;
* Contact Terry Brady, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The FileAnalyzer (https://github.com/Georgetown-University-Libraries/File-Analyzer) is an application designed to solve a number of library automation challenges:&lt;br /&gt;
&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census. After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform. Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules. As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
The first half of this session will be targeted to potential users and developers.  The second half of the session will be targeted towards developers who are interested in developing custom rules for the application.&lt;br /&gt;
&lt;br /&gt;
''Session Overview''&lt;br /&gt;
* Overview of the application&lt;br /&gt;
* Running sample file tests/transformations through the application&lt;br /&gt;
* Compiling and building the application&lt;br /&gt;
* Coding a custom file processing task&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
#Ray Schwartz&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Collecting social media data with Social Feed Manager===&lt;br /&gt;
'''Half-Day [Morning]'''&lt;br /&gt;
&lt;br /&gt;
Contacts: &lt;br /&gt;
* Dan Chudnov, GW Libraries, dchud (at) gwu.edu&lt;br /&gt;
* Dan Kerchner, GW Libraries, kerchner (at) gwu.edu&lt;br /&gt;
* Laura Wrubel, GW Libraries, lwrubel (at) gwu.edu&lt;br /&gt;
&lt;br /&gt;
Social media data is a popular material for research and a new format for building collections.  What does it take to collect meaningfully from Twitter, Tumblr, YouTube, Weibo, Facebook, and other sites?  We will:&lt;br /&gt;
* Introduce options for collections, including both high- and low-end commercial offerings. Discuss what it means to collect these resources, covering boundaries, policies, and workflows required to develop a social media collection program in your institution.&lt;br /&gt;
* Explore the Twitter API in depth, with hands-on opportunities for those w/laptops and others who want to team up w/them&lt;br /&gt;
* Help you get started using the free [http://gwu-libraries.github.io/social-feed-manager Social Feed Manager] (SFM) app we're developing at GW to create your first collections. We’ll demo its use and demo a clean install (those w/environments can follow along)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Declan Fleming&lt;br /&gt;
# Esmé Cowles&lt;br /&gt;
# Jason Stirnaman&lt;br /&gt;
# Ray Schwartz&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Intro to Git ===&lt;br /&gt;
'''&amp;quot;Half-Day [tbd - probably afternoon]&amp;quot;''' &lt;br /&gt;
* Contact: Erin Fahy, Stanford University, efahy at stanford.edu&lt;br /&gt;
* TA: Michael Klein, Northwestern University, michael.klein at northwestern.edu&lt;br /&gt;
&lt;br /&gt;
This session will cover the fundamentals of git by discussing/going through (time allowing):&lt;br /&gt;
* what is a distributed version control system&lt;br /&gt;
* what is git and github&lt;br /&gt;
* initializing a repo on a remote server/github&lt;br /&gt;
* cloning an existing repo&lt;br /&gt;
* creating a branch&lt;br /&gt;
* contributing code to a repo&lt;br /&gt;
* how to handle merge conflicts&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
# Ray Schwartz&lt;br /&gt;
# Sam Kome&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
=== Archival discovery and use ===&lt;br /&gt;
'''Full Day''' &lt;br /&gt;
&lt;br /&gt;
Contacts: &lt;br /&gt;
* Tim Shearer, UNC Chapel Hill, tshearer at email.unc.edu, &lt;br /&gt;
* Will Sexton, Duke, will.sexton at duke.edu&lt;br /&gt;
&lt;br /&gt;
This is a full day pre-conference about archival collections and will cover the intersections of archives, workflows, technologies, discovery, and use.&lt;br /&gt;
&lt;br /&gt;
Morning agenda: focused talks around (but not limited to) issues such as:&lt;br /&gt;
* Crowd-sourcing description to enhance collecitons&lt;br /&gt;
* Linked data and authority&lt;br /&gt;
* Mass digitization and sustainable workflows&lt;br /&gt;
* Digitized objects in context (images and other objects in finding aids)&lt;br /&gt;
* Too many cooks in the kitchen: versioning&lt;br /&gt;
* Global-, intra-, and inter- discovery of archival materials via finding aids &lt;br /&gt;
* and more...&lt;br /&gt;
&lt;br /&gt;
Afternoon agenda:  Focused talks around specific tools followed by general discussion, connections, opportunities, aspirations, and planning.&lt;br /&gt;
&lt;br /&gt;
Tool examples:&lt;br /&gt;
* Archivespace&lt;br /&gt;
* STEADy&lt;br /&gt;
* &amp;quot;RAMP&amp;quot; (Remixing Archival Metadata Project)&lt;br /&gt;
* OpenRefine&lt;br /&gt;
* Aeon&lt;br /&gt;
&lt;br /&gt;
''Interested in Attending''&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here&lt;br /&gt;
&lt;br /&gt;
Morning:&lt;br /&gt;
* your name&lt;br /&gt;
&lt;br /&gt;
Afternoon:&lt;br /&gt;
* your name&lt;br /&gt;
&lt;br /&gt;
All day:&lt;br /&gt;
&lt;br /&gt;
# Josh Wilson&lt;br /&gt;
# Sam Kome&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===AV Content Slam===&lt;br /&gt;
'''Half-Day [morning]'''&lt;br /&gt;
Contacts:&lt;br /&gt;
* Kara Van Malssen, kara (at) avpreserve.com&lt;br /&gt;
* Lauren Sorenson, laurens (at) bavc.org&lt;br /&gt;
* Steven Villereal , villereal (at) gmail.com&lt;br /&gt;
A morning BarCamp/unconference for practitioners and coders who work with audiovisual content. The agenda will be attendee-driven, with a focus on sharing, synthesizing, and improving workflow strategies and documentation for software-based approaches to wrangling and providing access to audio and video content.&lt;br /&gt;
Possible topics of discussion might include:&lt;br /&gt;
* Use of format id and characterization/metadata extraction tools for AV&lt;br /&gt;
* Creating and using time-based metadata&lt;br /&gt;
* Managing (moving, fixity checking, etc) massive files (like uncompressed video)&lt;br /&gt;
For a better idea of the topics and concerns that have informed some past AV-themed events, check out the event wikis for [http://wiki.curatecamp.org/index.php/CURATEcamp_AVpres_2013 CURATEcamp AVpres 2013] as well as the [http://wiki.curatecamp.org/index.php/Association_of_Moving_Image_Archivists_%26_Digital_Library_Federation_Hack_Day_2013 AMIA/DLF 2013 Hack Day] .&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here:&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===OCLC Web Services Hackfest===&lt;br /&gt;
&lt;br /&gt;
&amp;quot;Half-Day&amp;quot; [afternoon]&lt;br /&gt;
&lt;br /&gt;
Contact: Shelley Hostetler, Community Manager, Developer Network hostetls[at]oclc.org&lt;br /&gt;
&lt;br /&gt;
This half-day hackfest will explore some of the OCLC Developer Network web services. We will provide an overview of some of the common topics such as the general REST-based architecture for most services and how to use some new authentication clients. The group can then decide to take a deep dive into a particular API and/or write a client library for the community.&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here:&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
===Obey the Testing Goat!: Test Driven Web Development From The Ground Up===&lt;br /&gt;
'''Half-Day [tbd - probably afternoon]'''&lt;br /&gt;
* Contact [[User:Mredar|Mark Redar]], mredar[at]gmail.com&lt;br /&gt;
&lt;br /&gt;
Test driven development is a proven method for producing better quality code. But I've found it hard to follow a strict TDD methodology when starting new web projects. How do you write that first test when there is no code or web pages created yet.&lt;br /&gt;
&lt;br /&gt;
In this session, we will follow the excellent book [http://shop.oreilly.com/product/0636920029533.do &amp;quot;Test-Driven Web Development with Python&amp;quot;] to create a simple web site in Django following TDD from the first character typed. Come ready to code and test. No prior knowledge of python or Django required.&lt;br /&gt;
&lt;br /&gt;
By the end of this session, you should be able to  [http://www.obeythetestinggoat.com/ &amp;quot;Obey the Testing Goat&amp;quot;] from the start to finish for your next project.&lt;br /&gt;
&lt;br /&gt;
If you would be interested in attending, please indicate by adding your name (but not email address, etc.) here:&lt;br /&gt;
&lt;br /&gt;
# Charlie Morris (NCSU)&lt;br /&gt;
# Jason Stirnaman&lt;br /&gt;
# Joshua Gomez&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
===Summon Camp===&lt;br /&gt;
Placeholder by Tim McGeary for Gillian Cain (Serials Solutions)&lt;br /&gt;
Description to be provided by Gillian after account issues resolved. &lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39856</id>
		<title>2014 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39856"/>
				<updated>2013-11-08T20:34:06Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Identifiers, Data, and Norse Gods */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
 &lt;br /&gt;
* ''Projects'' you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* ''Tools and technologies'' – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* ''Technical issues'' - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* ''Relevant non-technical issues'' – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
'''To Propose a Talk'''&lt;br /&gt;
* Log in to the wiki in order to submit a proposal. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
* Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
* If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist us in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
As in past years, the Code4Lib community will vote on proposals that they would like to see included in the program. This year, however, only the top 10 proposals will be guaranteed a slot at the conference. Additional presentations will be selected by the Program Committee in an effort to ensure diversity in program content. Community votes will, of course, still weigh heavily in these decisions.&lt;br /&gt;
&lt;br /&gt;
Presenters whose proposals are selected for inclusion in the program will be guaranteed an opportunity to register for the conference. The standard conference registration fee will still apply.&lt;br /&gt;
&lt;br /&gt;
''Proposals can be submitted through '''Friday, November 8, 2013, at 5pm PST'''''. Voting will commence on November 18, 2013 and continue through December 6, 2013. The final line-up of presentations will be announced in early January, 2014.&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
&lt;br /&gt;
==Creating a new Greek-Dutch dictionary==&lt;br /&gt;
* Caspar Treijtel, University of Amsterdam, c.treijtel@uva.nl&lt;br /&gt;
&lt;br /&gt;
At present, no complete dictionary of (ancient) Greek-Dutch is available online. A new dictionary is currently under construction at Leiden University, with software being developed at the University of Amsterdam. The team in Leiden has already begun preparation of the data, with at this moment about 6,000 approved lemmas. The ultimate goal is to produce both a print version and online open access version from the same source documents. The software needed for this has been made in a project that was funded by CLARIN-NL.&lt;br /&gt;
&lt;br /&gt;
Migrator&lt;br /&gt;
&lt;br /&gt;
For the production of lemmas we have implemented an advanced workflow. The (generally non-technical) users create lemmas using MS Word, which is both familiar and easy to use. We have developed a custom software module that carefully migrates the Word documents into deeply structured XML by analyzing the structure and semantics of the lemmas, and falling back on heuristics in ambiguous cases. While having initially envisioned the oXygen XML Author component as the main tool for creating new lemmas, we obtained excellent results with the migrator module, and decided therefore to continue using MS Word as the primary composition tool. The main advantage of this is that the editors are much more familiar with Word than with any other WYSIWYG editor. Lemmas that have been migrated to XML are stored in an XML database and can be further edited using oXygen XML Author.&lt;br /&gt;
&lt;br /&gt;
Lemmatizer&lt;br /&gt;
&lt;br /&gt;
Greek morphology is complicated. In order to use a dictionary effectively, a rather high level of initial language competence is necessary for the user to be able to relate the word form s/he finds in a text to the correct basic lemma form, where the definition of the word can be found. Using a Greek morphological database we have been able to facilitate the search for lemmas. A ‘lemmatizer’ module gives the possible parsings of the word forms and the lemmas they can be derived from. This enables the user to type in the word as found in the text and be redirected to the correct lemma.&lt;br /&gt;
&lt;br /&gt;
Visualization&lt;br /&gt;
&lt;br /&gt;
For the online dictionary we have implemented a visualization module that allows the user to view multiple lemmas at once. The implementation of this module has been done using the Javascript framework MooTools. The result is a viewer that performs really well and is run by maintainable Javascript code.&lt;br /&gt;
&lt;br /&gt;
The online dictionary is still being worked on, have a look at http://www.woordenboekgrieks.nl/ for the beta version. A newer test version with additional features can be found here: http://angel.ic.uva.nl:8600/.&lt;br /&gt;
&lt;br /&gt;
Credits&lt;br /&gt;
&lt;br /&gt;
* construction of the dictionary: Prof. Ineke Sluiter, Classics department of Leiden University; Prof. Albert Rijksbaron, University of Amsterdam&lt;br /&gt;
* publisher of the dictionary: Amsterdam University Press&lt;br /&gt;
* design/typesetting dictionary: TaT Zetwerk (http://www.tatzetwerk.nl/)&lt;br /&gt;
* software development: Digital Production Center, University Library, University of Amsterdam&lt;br /&gt;
* project funding: CLARIN-NL (http://www.clarin.nl/)&lt;br /&gt;
* morphological database for use by the lemmatizer: courtesy of Prof. Helma Dik, University of Chicago (based on data of the Perseus Project)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Using Drupal to drive alternative presentation systems ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Recently, we have been building systems that use angular.js, Rails, or other systems for presentation, while leveraging Drupal's sophisticated content management capabilities on the back end.&lt;br /&gt;
&lt;br /&gt;
So far, these have been one-way systems, but as we move to Drupal 8 we are beginning to explore ways to further decouple the presentation and CMS functions.&lt;br /&gt;
&lt;br /&gt;
== A Book, a Web Browser and a Tablet: How Bibliotheca Alexandrina's Book Viewer Framework Makes It Possible ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mohammed.abuouda|Mohammed Abu ouda]], Bibliotheca Alexandrina (The new Library of Alexandria)&lt;br /&gt;
&lt;br /&gt;
A lot of institutions around the world are engaged in multiple digitization projects aiming at preserving the human knowledge present in books and availing them through multiple channels to people around the whole globe. These efforts will sure help close the digital gap particularly with the arrival of affordable e-readers, mobile phones and network coverage. However, the digital reading experience has not yet arrived to its maximum potential. Many readers miss features they like in their good old books and wish to find them in their digital counterpart. In an attempt to create a unique digital reading experience, Bibliotheca Alexandria (BA) created a flexible book viewing framework that is currently used to access its current collection of more than 300,000 digital books in five different languages which includes the largest collection of digitized Arabic books.&lt;br /&gt;
&lt;br /&gt;
Using open source tools, BA used the framework to develop a modular book viewer that can be deployed in different environments and is currently at the heart of various BA projects. The Book viewer provides several features creating a more natural reading experience. As with physical books, the reader can now personalize the books he reads by adding annotations like highlights, underlines and sticky notes to capture his thoughts and ideas in addition to being able to share the book with friends on social networks. The reader can perform a search across the content of the book receiving highlighted search results within the pages of the book. More features can be further added to the book viewer through its plugin architecture.&lt;br /&gt;
&lt;br /&gt;
== Structured data NOW: seeding schema.org in library systems ==&lt;br /&gt;
 &lt;br /&gt;
* [http://coffeecode.net Dan Scott], Laurentian University&lt;br /&gt;
** Previous code4lib presentations: [https://archive.org/details/code4lib.conf.2008.pres.CouchDBsacrilege CouchDB is sacrilege... mmm, delicious sacrilege] at Code4Lib 2008&lt;br /&gt;
&lt;br /&gt;
The semantic web, linked data, and structured data are all fantastic ideas with a barrier imposed by implementation constraints. If their system does not allow customizations, or the institution lacks skilled human resources, it does not matter how enthused a given library might be about publishing structured data... it will not happen. However, if the software in use simply publishes structured data by default, then the web will be populated for free. Really! No extra resources necessary.&lt;br /&gt;
&lt;br /&gt;
This presentation highlights Dan's work with systems such as Evergreen, Koha, and VuFind to enable the publication of schema.org structured data out-of-the-box. Along the way, we reflect the current state of the W3C Schema.org Bibliographic Extension community group efforts to shape the evolution of the schema.org vocabulary. Finally, hold on tight as we contemplate next steps and the possibilities of a world where structured data is the norm on the web.&lt;br /&gt;
&lt;br /&gt;
== Towards Pasta Code Nirvana: Using JavaScript MVC to Fill Your Programming Ravioli ==&lt;br /&gt;
&lt;br /&gt;
* Bret Davidson, North Carolina State University Libraries, bret_davidson@ncsu.edu&lt;br /&gt;
** Previous Code4Lib Presentations: [http://wiki.code4lib.org/index.php/2013_talks_proposals#Data-Driven_Documents:_Visualizing_library_data_with_D3.js Visualizing library data with D3.js] at Code4Lib 2013&lt;br /&gt;
&lt;br /&gt;
JavaScript MVC frameworks are ushering in a golden age of robust and responsive web applications that take advantage of evergreen browsers, performant JS engines, and the unprecedented reach provided by billions of personal computing devices. The web browser has emerged as the world’s most popular application runtime and the complexity[1] and scope of JavaScript applications has exploded accordingly. Server-side web frameworks like Rails and Django have helped developers adhere to best practices like modularity, dependency injection, and unit testing for years, practices that are now being applied to JavaScript development through projects like Backbone[2], Ember[3], and Angular[4].&lt;br /&gt;
&lt;br /&gt;
This talk will discuss the issues JavaScript MVC frameworks are trying to solve, common features like data binding, implications for the future of web development[5], and the appropriateness of JavaScript MVC for library applications.&lt;br /&gt;
&lt;br /&gt;
*[1]http://en.wikipedia.org/wiki/Spaghetti_code&lt;br /&gt;
*[2]http://backbonejs.org&lt;br /&gt;
*[3]http://emberjs.com&lt;br /&gt;
*[4]http://angularjs.org&lt;br /&gt;
*[5]http://tomdale.net/2013/09/progressive-enhancement-is-dead/&lt;br /&gt;
&lt;br /&gt;
== WebSockets for Real-Time and Interactive Interfaces ==&lt;br /&gt;
&lt;br /&gt;
* [http://ronallo.com Jason Ronallo], NCSU Libraries, jason_ronallo@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/ronallo HTML5 Microdata and Schema.org] 2012&lt;br /&gt;
* [http://code4lib.org/conference/2013/ronallo HTML5 Video Now!] 2013&lt;br /&gt;
&lt;br /&gt;
Watching the Google Analytics Real-Time dashboard for the first time was mesmerizing. As soon as someone visited a site, I could see what page they were on. For a digital collections site with a lot of images, it was fun to see what visitors were looking at. But getting from Google Analytics to the image or other content of what was currently being viewed was cumbersome. The real-time experience was something I wanted share with others. I'll show you how I used a WebSocket service to create a real-time interface to digital collections.&lt;br /&gt;
&lt;br /&gt;
In the Hunt Library at NCSU we have some large video walls. I wanted to make HTML-based exhibits that featured viewer interactions. I'll show you how I converted Listen to Wikipedia [1] into an bring-your-own-device interactive exhibit. With WebSockets any HTML page can be remote controlled by any internet connected device.&lt;br /&gt;
&lt;br /&gt;
I will attempt to include real-time audience participation.&lt;br /&gt;
&lt;br /&gt;
[1] http://listen.hatnote.com/&lt;br /&gt;
&lt;br /&gt;
== Rapid Development of Automated Tasks with the File Analyzer ==&lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Libraries, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Libraries have customized the File Analyzer and Metadata Harvester application (https://github.com/Georgetown-University-Libraries/File-Analyzer) to solve a number of library automation challenges:&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census.  After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform.  Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules.  As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
Learn about the functionality that is available for download, how you can use this tool to automate workflows from digital collections to ILS ingests to electronic resources statistics and also discuss the opportunities to collaborate on enhancements to this application!&lt;br /&gt;
&lt;br /&gt;
== GeoHydra: How to Build a Geospatial Digital Library with Fedora ==&lt;br /&gt;
 &lt;br /&gt;
* [http://stanford.edu/~drh Darren Hardy], Stanford University, drh@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Geographically-rich data are exploding and putting fear in those trying to&lt;br /&gt;
tackle integrating them into existing digital library infrastructures.&lt;br /&gt;
Building a spatial data infrastructure that integrates with your digital&lt;br /&gt;
library infrastructure need not be a daunting task. We have successfully&lt;br /&gt;
deployed a geospatial digital library infrastructure using Fedora and&lt;br /&gt;
open-source geospatial software [1]. We'll discuss the primary design&lt;br /&gt;
decisions and technologies that led to a production deployment within a few&lt;br /&gt;
months. Briefly, our architecture revolves around discovery, delivery, and&lt;br /&gt;
metadata pipelines using open-source OpenGeoPortal [2], Solr [3], GeoServer&lt;br /&gt;
[4], PostGIS [5], and GeoNetwork [6] technologies, plus the proprietary ESRI&lt;br /&gt;
ArcMap [7] -- the GIS industry's workhorse. Finally, we'll discuss the key&lt;br /&gt;
skillsets needed to build and maintain a spatial data infrastructure.&lt;br /&gt;
&lt;br /&gt;
[1] http://foss4g.org&lt;br /&gt;
[2] http://opengeoportal.org&lt;br /&gt;
[3] http://lucene.apache.org/solr&lt;br /&gt;
[4] http://geoserver.org&lt;br /&gt;
[5] http://postgis.net&lt;br /&gt;
[6] http://geonetwork-opensource.org&lt;br /&gt;
[7] http://esri.com&lt;br /&gt;
&lt;br /&gt;
==Under the Hood of Hadoop Processing at OCLC Research ==&lt;br /&gt;
&lt;br /&gt;
[http://roytennant.com/ Roy Tennant]&lt;br /&gt;
&lt;br /&gt;
* Previous Code4Lib presentations: 2006: &amp;quot;The Case for Code4Lib 501c(3)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[http://hadoop.apache.org/ Apache Hadoop] is widely used by Yahoo!, Google, and many others to process massive amounts of data quickly. OCLC Research uses a 40-node compute cluster with Hadoop and HBase to process the 300 million MARC records of WorldCat in various ways. This presentation will explain how Hadoop MapReduce works and illustrate it with specific examples and code. The role of the jobtracker in both monitoring and reporting on processes will be explained. String searching WorldCat will also be demonstrated live.&lt;br /&gt;
&lt;br /&gt;
== Quick and Easy Data Visualization with Google Visualization API and Google Chart Libraries ==&lt;br /&gt;
 &lt;br /&gt;
[http://bohyunkim.net/blog Bohyun Kim], Florida International University, bohyun.kim@fiu.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Do most of the data that your library collects stay in spreadsheets or are published as a static table with a series of boring numbers? Do your library stakeholders spend more time collecting the data than using it as a decision-making tool because the data is presented in a way that makes it hard for them [http://developers.google.com/chart/interactive/docs/gallery to quickly grasp its significance? ]&lt;br /&gt;
&lt;br /&gt;
This talk will provide an overview of [http://developers.google.com/chart/interactive/docs/reference Google Visualization API] [2] and [http://developers.google.com/chart/ Google Chart Libraries] [3] to get you started on the way to quickly query and visualize your library data from remote data sources (e.g. a Google Spreadsheet or your own database) with (or without) cool-looking user-controls, animation effects, and even a dashboard.&lt;br /&gt;
&lt;br /&gt;
== Leap Motion + Rare Books: A hands-free way to view and interact with rare books in 3D ==&lt;br /&gt;
 &lt;br /&gt;
[http://http://www.youtube.com/user/jpdenzer Juan Denzer], Binghamton University, jdenzer@binghamton.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
As rare books become more delicate over time, making them available to the public becomes harder.  We at Binghamton University Library have developed an application that makes it easier to view rare books without ever having to touch them.  We have combined the Leap Motion hands-free device and 3D rendered models to create a new virtual experience for the viewer.&lt;br /&gt;
&lt;br /&gt;
The application allows the user to rotate and zoom in on a 3D representation of a rare book.  The user is also able to ‘open’ the virtual book and flip through it using a natural user interface.  Such as swiping the hand left or right to turn the page.&lt;br /&gt;
&lt;br /&gt;
The application is built on the .Net framework and is written in C#.  3D models are created using simple 3D software such as sketchup or Blender.  Scans of the book cover and spine are created using simple flatbed scanners.  The inside pages are scanned using overhead scanners. &lt;br /&gt;
&lt;br /&gt;
This talk with discuss the technologies used in developing the application and virtually any library could implement the application with virtually no coding at all. This presentation will have a demonstration of the software and also a chance for audience members to experience the Rare Book Leap Motion App themselves.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Course Reserves Unleashed! ==&lt;br /&gt;
 &lt;br /&gt;
* Bobbi Fox, Library Technology Services, Harvard University, bobbi_fox@harvard.edu&lt;br /&gt;
* Gloria Korsman, Andover-Harvard Theological Library&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Hey kids!  Remember when SOAP was used for something other than washing?  Our sophisticated (and highly functional) Course Reserves Request system does!&lt;br /&gt;
&lt;br /&gt;
However, while the system is great for submitting and  processing course reserve requests, the student-facing presentation through Havard’s home-grown -- and soon to be replaced -- LMS leaves a lot to be desired.  &lt;br /&gt;
&lt;br /&gt;
Follow along as we leverage Solr 4 as a No-SQL database, along with more progressive RESTful API techniques, to release Reserves data into the wild without interfering with reserves request processing -- and, in the process, open up the opportunity for other schools to feed their data in as well.&lt;br /&gt;
&lt;br /&gt;
== We Are All Disabled! Universal Web Design Making Web Services Accessible for Everyone ==&lt;br /&gt;
 &lt;br /&gt;
Cynthia Ng, Accessibility Librarian, CILS at Langara College&lt;br /&gt;
* No previous Code4Lib presentations (not counting lightning talks)&lt;br /&gt;
&lt;br /&gt;
We’re building and improving tools and services all the time, but do you only develop for the “average” user or add things for “disabled” users? We all use “assistive” technology accessing information in a multitude of ways with different platforms, devices, etc. Let’s focus on providing web services that are accessible to everyone without it being onerous or ugly. The aim is to get you thinking about what you can do to make web-based services and content more accessible for all from the beginning or with small amounts of effort whether you're a developer or not.&lt;br /&gt;
&lt;br /&gt;
The goal of the presentation is to provide both developers and content creators with information on simple, practical ways to make web content and web services more accessible. However, rather than thinking about putting in extra effort or making adjustment for those with disabilities, I want to help people think about how to make their websites more accessible for all users through universal web design.&lt;br /&gt;
&lt;br /&gt;
== Personalize your Google Analytics Data with Custom Events and Variables ==&lt;br /&gt;
&lt;br /&gt;
[http://joshwilson.net Josh Wilson], Systems Integration Librarian, State Library of North Carolina - joshwilsonnc@gmail.com&lt;br /&gt;
&lt;br /&gt;
At the State Library of North Carolina, we had more specific questions about the use of our digital collections than standard GA could provide. A few implementations of custom events and custom variables later, we have our answers.&lt;br /&gt;
&lt;br /&gt;
I'll demonstrate how these analytics add-ons work, and why implementation can sometimes be more complicated than just adding a few lines of JavaScript to your ga.js. I'll discuss some specific examples in use at the SLNC:&lt;br /&gt;
&lt;br /&gt;
* Capturing the content of specific metadata fields in CONTENTdm as Custom Events &lt;br /&gt;
* Recording Drupal taxonomy terms as Custom Variables&lt;br /&gt;
&lt;br /&gt;
In both instances, this data deepened our understanding of how our sites and collections were being used, and in turn, we were able to report usage more accurately to content contributors and other stakeholders.&lt;br /&gt;
&lt;br /&gt;
More on: [https://developers.google.com/analytics/devguides/collection/gajs/eventTrackerGuide GA Custom Events] | [https://developers.google.com/analytics/devguides/collection/gajs/gaTrackingCustomVariables GA Custom Variables]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Behold Fedora 4: The Incredible Shrinking Repository! ==&lt;br /&gt;
&lt;br /&gt;
Esmé Cowles, UC San Diego Library.  Previous talk: [http://code4lib.org/conference/2013/cowles-critchlow-westbrook All Teh Metadatas Re-Revisited] (2013)&lt;br /&gt;
&lt;br /&gt;
* One repository contains untold numbers of digital objects and powers many Hydra and Islandora apps&lt;br /&gt;
* It speaks RDF, but contains no triplestore! (triplestores sold separately, SPARQL Update may be involved, some restrictions apply)&lt;br /&gt;
* Flexible enough to tie itself in knots implementing storage and access control policies&lt;br /&gt;
* Witness feats of strength and scalability, with dramatically increased performance and clustering&lt;br /&gt;
* Plumb the depths of bottomless hierarchies, and marvel at the metadata woven into the very fabric of the repository&lt;br /&gt;
* Ponder the paradox of ingesting large files by not ingesting them&lt;br /&gt;
* Be amazed as Fedora 4 swallows other systems whole (including Fedora 3 repositories)&lt;br /&gt;
* Watch novice developers setup Fedora 4 from scratch, with just a handful of incantations to Git and Maven&lt;br /&gt;
&lt;br /&gt;
The Fedora Commons Repository is the foundation of many digital collections, e-research, digital library, archives, digital preservation, institutional repository and open access publishing systems.  This talk will focus on how Fedora 4 improves core repository functionality, adds new features, maintains backwards compatibility, and addresses the shortcomings of Fedora 3.&lt;br /&gt;
&lt;br /&gt;
== Organic Free-Range API Development - Making Web Services That You Will Actually Want to Consume ==&lt;br /&gt;
&lt;br /&gt;
Steve Meyer and Karen Coombs, OCLC&lt;br /&gt;
&lt;br /&gt;
Building web services can have great benefits by providing reusability of data and functionality. Underpinning your applications with a web service will allow you to write code once and support multiple environments: your library's web app, mobile applications, the embedded widget in your campus portal. However, building a web service is its own kind of artful programming. Doing it well requires attention to many of the same techniques and requirements as building web applications, though with different outcomes. &lt;br /&gt;
&lt;br /&gt;
So what are the usability principles for web services? How do you build a web service that you (and others) will actually want to use? In this talk, we’ll share some of the lessons learned - the good, the bad, and the ugly - through OCLC's work on the WorldCat Metadata API. This web service is a sophisticated API that provides external clients with read and write access to WorldCat data. It provides a model to help aspiring API creators navigate the potential complications of crafting a web service. We'll cover:&lt;br /&gt;
&lt;br /&gt;
* Loose coupling of data assets and resource-oriented data modeling at the core&lt;br /&gt;
* Coding to standards vs. exposure of an internal data model&lt;br /&gt;
* Authentication and security for web services: API Keys, Digital Signing, OAuth Flows&lt;br /&gt;
* Building web services that behave as a suite so it looks like the left hand knows what the right hand is doing&lt;br /&gt;
&lt;br /&gt;
So at the end of the day, your team will know your API is a very good egg after all. &lt;br /&gt;
&lt;br /&gt;
If accepted, the presenters intend to produce and share a Quick Guide for building a web service that will reflect content presented in the talk.&lt;br /&gt;
&lt;br /&gt;
== Lucene's Latest (for Libraries) ==&lt;br /&gt;
&lt;br /&gt;
erik.hatcher@lucidworks.com&lt;br /&gt;
&lt;br /&gt;
Lucene powers the search capabilities of practically all library discovery platforms, by way of Solr, etc.  The Lucene project evolves rapidly, and it's a full-time job to keep up with the ever improving features and scalability.   This talk will distill and showcase the most relevant(!) advancements to date.&lt;br /&gt;
&lt;br /&gt;
== The Why and How of Very Large Displays in Libraries. ==&lt;br /&gt;
&lt;br /&gt;
* Cory Lown, NCSU Libraries, cwlown@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib Presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/lown How People Search the Library from a Single Search Box]  2012&lt;br /&gt;
* [http://code4lib.org/conference/2010/orphanides_lown_lynema Enhancing Discoverability with Virtual Shelf Browse] 2010&lt;br /&gt;
&lt;br /&gt;
Built into the walls of NC State's new Hunt Library are several [http://www.christiedigital.com/en-us/digital-signage/products/microtiles/pages/microtiles-digital-signage-video-wall.aspx Christie MicroTile Display Wall Systems]. What does a library do with a display that's seven feet tall and over twenty feet wide? I'll talk about why libraries might want large displays like this, what we're doing with them right now, and what we might do with them in the future. I'll talk about how these displays factor into planning for new and existing web projects. And I'll get into the fun details of how you build web applications that scale from the very small browser window on a phone all the way up to a browser window with about 14 million pixels (about 10 million more than a dual 24&amp;quot; monitor desktop setup).&lt;br /&gt;
&lt;br /&gt;
== Discovering your Discovery System in Real Time. ==&lt;br /&gt;
&lt;br /&gt;
* Godmar Back, Virginia Tech, gback@vt.edu&lt;br /&gt;
* Annette Bailey, Virginia Tech, afbailey@vt.edu&lt;br /&gt;
&lt;br /&gt;
Practically all libraries today provide web-based discovery systems to their users;&lt;br /&gt;
users discover items and peruse or check them out by clicking on links.  Unlike&lt;br /&gt;
the traditional transaction of checking out a book at the circulation desk, this&lt;br /&gt;
interaction is largely invisible.  We have built a system that records user's&lt;br /&gt;
interaction with Summon in real-time, processes the resulting data with minimal delay,&lt;br /&gt;
and visualizes it in various ways using Google Charts and using various d3.js modules,&lt;br /&gt;
such as word clouds, tree maps, and others.&lt;br /&gt;
&lt;br /&gt;
These visualizations can be embedded in web sites, but are also suitable for&lt;br /&gt;
projection via large-scale displays or projectors right into the 'Learning Spaces'&lt;br /&gt;
many libraries are converted into.  The goal of this talk is to share the technology&lt;br /&gt;
and advocate the building of a cloud-based infrastructure that would make this&lt;br /&gt;
technology available to any library that uses a discovery system, rather than just&lt;br /&gt;
those who have the technological prowess for developing such systems and&lt;br /&gt;
visualizations in-house.  &lt;br /&gt;
&lt;br /&gt;
Previous presentations at Code4Lib:&lt;br /&gt;
* Talk: Code4Lib 2009 [http://code4lib.org/files/LibX2.0-Code4Lib-2009AsPresented.ppt LibX 2.0]&lt;br /&gt;
* Preconference: [http://wiki.code4lib.org/index.php/LibX_Preconference LibX 2.0, 2009]&lt;br /&gt;
* Preconference: Code4Lib 2010, On Widgets and Web Services&lt;br /&gt;
&lt;br /&gt;
== Your Library, Anywhere: A Modern, Responsive Library Catalogue at University of Toronto Libraries ==&lt;br /&gt;
&lt;br /&gt;
* Bilal Khalid, Gordon Belray, Lisa Gayhart (lisa.gayhart@utoronto.ca)&lt;br /&gt;
&lt;br /&gt;
* No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. &lt;br /&gt;
&lt;br /&gt;
[http://search.library.utoronto.ca The result]: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles. This talk will discuss: application development; service improvements; interface design; and user outreach, testing, and project communications. Feedback and questions from the audience are very welcome. If time runs short, we will be available for questions and conversation after the presentation.&lt;br /&gt;
&lt;br /&gt;
Note: A version of this content has been provisionally accepted as an article for Code4Lib Journal, January 2014 publication.)&lt;br /&gt;
&lt;br /&gt;
== All Tiled Up ==&lt;br /&gt;
&lt;br /&gt;
* Mike Graves, MIT Libraries (mgraves@mit.edu)&lt;br /&gt;
&lt;br /&gt;
You've got maps. You even scanned and georeferenced them. Now what? Running a full GIS stack can be expensive, and overkill in some cases. The good news is that you have a lot more options now than you did just a few years ago. I'd like to present some lighter weight solutions to making georeferenced images available on the Web.&lt;br /&gt;
&lt;br /&gt;
This talk will provide an introduction to MBTiles. I'll go over what they are, how you create them, how you use them and why you would use them.&lt;br /&gt;
&lt;br /&gt;
== The Great War: Image Interoperability to Facebook ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, Los Alamos National Laboratory (azaroth42@gmail.com)&lt;br /&gt;
** (Code4Lib 2006: [http://www.code4lib.org/2006/sanderson | Library Text Mining])&lt;br /&gt;
* Rob Warren, Carleton University&lt;br /&gt;
** No previous presentations&lt;br /&gt;
&lt;br /&gt;
Using a pipeline constructed from Linked Open Data and other interoperability specifications, it is possible to merge and re-use image and textual data from distributed library collections to build new, useful tools and applications.  Starting with the OAI-PMH interface to ContentDM, we will take you on a tour through the International Image Interoperability Framework and Shared Canvas, to a cross-institutional viewer, and image analysis for the purposes of building a historical Facebook from finding and tagging people in photographs.  The World War One collections are drawn from multiple institutions and merged by the machine learning code.&lt;br /&gt;
&lt;br /&gt;
The presentation will focus on the (open source) toolchain and the benefits of the use of standards throughout:  OAI-PMH to get the metadata, IIIF for interaction with the images, the Shared Canvas ontology for describing collections of digitized objects, Open Annotation for tagging things in the images and specialized ontologies that are specific to the contents.  The tools include standard RDF / OWL technologies, JSON-LD, imagemagick and OpenCV for image analysis.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Solr Search Results with D3.js for User-Friendly Navigation of Large Results Sets ==&lt;br /&gt;
&lt;br /&gt;
*Julia Bauder, Grinnell College Libraries (bauderj-at-grinnell-dot-edu)&lt;br /&gt;
*No previous presentations at national Code4Lib conferences&lt;br /&gt;
&lt;br /&gt;
As the corpus of articles, books, and other resources searched by discovery systems continues to get bigger, searchers are more and more frequently confronted with unmanageably large numbers of results. How can we help users make sense of 10,000 hits and find the ones they actually want? Facets help, but making sense of a gigantic sidebar of facets is not an easy task for users, either.&lt;br /&gt;
During this talk, I will explain how we will soon be using Solr 4’s pivot queries and hierarchical visualizations (e.g., treemaps) from D3.js to let patrons view and manipulate search results. We will be doing this with our VuFind 2.0 catalog, but this technique will work with any system running Solr 4. I will also talk about early student reaction to our tests of these visualization features.&lt;br /&gt;
&lt;br /&gt;
== PeerLibrary – open source cloud based collaborative library ==&lt;br /&gt;
&lt;br /&gt;
[https://github.com/peerlibrary/peerlibrary PeerLibrary is a new open source project] and a cloud service providing collaborative reading, sharing and storing. Users can upload publications they want to read (currently in PDF format), read them in the browser in real-time with others, highlight, annotate and organize their own or collaborative library. PeerLibrary provides a search engine to search over all uploaded open access publications. Additionally, it aims to collaboratively aggregate the open layer of knowledge on top of this publications through public annotations and references user will add to publications. In this way publications would not just be available to read, but accessible to the general public as well. Currently, it is aiming at scientific community and scientific publications.&lt;br /&gt;
&lt;br /&gt;
See [http://blog.peerlibrary.org/post/63458789185/screencast-previewing-the-peerlibrary-project screencast here].&lt;br /&gt;
&lt;br /&gt;
It is still in development and beta launch is planned at the end of November.&lt;br /&gt;
&lt;br /&gt;
== Who was where when, or finding biographical articles on Wikipedia by place and time ==&lt;br /&gt;
&lt;br /&gt;
* [http://morton-owens.info Emily Morton-Owens], The Seattle Public Library (presenting on work from NYU)&lt;br /&gt;
* No previous c4l presentations&lt;br /&gt;
&lt;br /&gt;
It's easy to answer the question &amp;quot;What important people were in Paris in 1939?&amp;quot; But what about Virginia in the 1750s or Scandinavia in the 14th century? I created a tool that allows you to search for biographies in a generally applicable way, using a map interface. I would like to present updates to my thesis project, which combines a crawler written in Java that extracts information from Wikipedia articles, with a MongoDB data store and a frontend in Python.&lt;br /&gt;
&lt;br /&gt;
The input to the project is freetext of entire articles in Wikipedia; this is important to allow us to pick up Benjamin Franklin not just in the single most obvious place of Philadelphia but also in London, Paris, Boston, etc. I can talk about my experiments disambiguating place names (approaches pioneered on newspaper articles were actually unhelpful on this type of text) and setting up a processing queue that does not become mired in the biographies of every human who ever played soccer. I also want to mitigate some of the implementation choices I made due to my academic deadline and improve the accuracy/usability.&lt;br /&gt;
&lt;br /&gt;
What I hope to show is that I was able to develop a novel and useful reference tool automatically, using fairly simple heuristics that are a far cry from hand-cataloging familiar to many librarians.&lt;br /&gt;
&lt;br /&gt;
You can try out [http://linserv1.cims.nyu.edu:48866/ the original version] (this server is inconveniently set to be updated/rebooted on 11/8--may be temporarily unavailable)&lt;br /&gt;
&lt;br /&gt;
== Good!, DRY, and Dynamic: Content Strategy for Libraries (Especially the Big Ones) ==&lt;br /&gt;
&lt;br /&gt;
*Michael Schofield, Nova Southeastern University Libraries, mschofield@nova.edu&lt;br /&gt;
*No previous code4lib presentations.&lt;br /&gt;
&lt;br /&gt;
The responsibilities of the #libweb are exploding [it’s a good thing] and it is no longer uncommon for libraries to manage or even home-grow multiple applications and sites. Often it is at this point where the web people begin to suffer the absence of a content strategy when, say, business hours need to be updated sitewide a half-dozen times.&lt;br /&gt;
&lt;br /&gt;
We were already feeling this crunch when we decided to further complicate the Nova Southeastern University Libraries by splitting the main library website into two. The Alvin Sherman Library, Research, and Information Technology Center is a unique joint-use facility that serves not only the academic community but the public of Broward County - and marketing a hyperblend of content through one portal just wasn't cutting it. With a web team of two, we knew that managing all this rehashed, disparate content was totally unsustainable.&lt;br /&gt;
&lt;br /&gt;
I want to share in this talk how I went about making our library content DRY (“don’t repeat yourself”): input content in one place--blurbs, policies, featured events, featured databases, book reviews, business hours, and so on.--and syndicate it everywhere - even, sometimes, dynamically target that content for specific audiences or context. It is a presentation that is a little about workflow, a little more about browser and context detection, a tangent about content-modeling the CMS, and a lot about APIs, syndication, and performance.&lt;br /&gt;
&lt;br /&gt;
== No code, no root, no problem? Adventures in SaaS and library discovery ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:erwhite@vcu.edu Erin White, VCU]&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
In 2012 VCU was an eager early adopter of Ex Libris' cloud service Alma as an ILS, ERM, link resolver, and single-stop, de-silo'd public-facing discovery tool. This has been a disruptive change that has shifted our systems staff's day-to-day work, relationships with others in the library, and relationships with vendors.&lt;br /&gt;
&lt;br /&gt;
I'll share some of our experiences and takeaways from implementing and maintaining a cloud service:&lt;br /&gt;
* Seeking disruption and finding it&lt;br /&gt;
* Changing expectations of service and the reality of unplanned downtime&lt;br /&gt;
* Communication and problem resolution with non-IT library staff&lt;br /&gt;
* Working with a vendor that uses agile development methodology&lt;br /&gt;
* Benefits and pitfalls of creating customizations and code workarounds&lt;br /&gt;
* Changes in library IT/coders' roles with SaaS&lt;br /&gt;
&lt;br /&gt;
...as well as thoughts on the philosophy of library discovery vs real-life experiences in moving to a single-search model.&lt;br /&gt;
&lt;br /&gt;
== Building for others (and ourselves):  the Avalon Media System ==&lt;br /&gt;
* [mailto:michael.klein@northwestern.edu Michael B Klein], Senior Software Developer, Northwestern University &lt;br /&gt;
** [http://code4lib.org/conference/2010/metz_klein Public Datasets in the Cloud] (code4lib 2010)&lt;br /&gt;
** [http://code4lib.org/conference/2013/klein-rogers The Avalon Media System: A Next Generation Hydra Head For Audio and Video Delivery] (code4lib 2013)&lt;br /&gt;
* [mailto:j-rudder@northwestern.edu Julie Rudder], Digital Initiatives Project Manager, Northwestern University&lt;br /&gt;
** no previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
[http://www.avalonmediasystem.org/ Avalon Media System] is a collaborative effort between development teams at Northwestern and Indiana Universities. Our goal is to produce an open source media management platform that works well for us, but is also widely adopted and contributed to by other institutions. We believe that building a strong user and contributor community is vital to the success and longevity of the project, and have developed the system with this goal in mind. We will share lessons learned, pains and successes we’ve had releasing two versions of the application since last year.  &lt;br /&gt;
&lt;br /&gt;
Our presentation will cover our experiences:&lt;br /&gt;
* providing flexible, admin-friendly distribution and installation options&lt;br /&gt;
* building with abstraction, customization and local integrations in mind&lt;br /&gt;
* prioritizing features (user stories)&lt;br /&gt;
* attracting code contributions from other institutions&lt;br /&gt;
* gathering community feedback &lt;br /&gt;
* creating a product rather than a bag of parts&lt;br /&gt;
&lt;br /&gt;
== How to check your data to provide a great data product? Data quality as a key product feature at Europeana ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:Peter.Kiraly@kb.nl Péter Király] portal backend developer, Europeana&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
[http://Europeana.eu/ Europeana.eu] - Europe's digital library, archive and museum - aggregates more than 30 million metadata records from more than 2200 institutions.  The records come from libraries, archives, museums and every other kind of cultural institution, from very different systems and metadata schemas, and are typically transformed several times until they are ingested into the Europeana data repository.  Europeana builds a consolidated database from these records, creating reliable and consistent services for end-users (a search portal, search widget, mobile apps, thematic sites etc.) and an API, which supports our strategic goeal of data for reuse in education, creative industries, and the cultural sector.  A reliable &amp;quot;data product&amp;quot; is thus at the core of our own software products, as well as those of our API partners.&lt;br /&gt;
&lt;br /&gt;
Much effort is needed to smooth out local differences in the metadata curation practice of our data providers. We need a solid framework to measure the consistency of our data and provide feedback to decision-makers inside and outside the organisation. We can also use this metrics framework to ask content providers to improve their own metadata. Of course, a data-quality-driven approach requires that we also improve the data transformation steps of the Europeana ingestion process itself. Data quality issues heavily define what new features we are able to create in our user interfaces and API, and might actually affect the design and implementation of our underlying data structure, the Europeana Data Model.&lt;br /&gt;
&lt;br /&gt;
In the presentation I briefly describe the Europeana metadata ingestion process, show the data quality metrics, the measuring techniques (using the Europeana API, Solr and MongoDB queries), some typical problems (both trivial and difficult ones), and finally the feedback mechanism we propose to deploy.&lt;br /&gt;
&lt;br /&gt;
Keywords: Europeana, data quality, EDM, API, Apache Solr, MongoDB, #opendata, #openglam&lt;br /&gt;
&lt;br /&gt;
== Teach your Fedora to Fly: scaling out a digital repository ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:acoburn@amherst.edu Aaron Coburn], Software Developer, Amherst College&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Fedora is a great repository system for managing large collections of digital objects, but what happens when a popular food magazine begins directing a large number of readers to a manuscript showing Emily Dickinson’s own recipe for doughnuts? While Fedora excels in its support of XML-based metadata, it doesn’t always perform well under a high volume of traffic. Nor is it especially tolerant of network or hardware failures.&lt;br /&gt;
&lt;br /&gt;
This presentation will show how we are making heavy use of a fedora repository while at the same time insulating it almost entirely from any web traffic. Starting with a distributed web front-end built with Node.js, and caching most of the user-accessible content from Fedora in an elastic, fault-tolerant Riak (NoSQL) cluster, we have eliminated nearly all single points of failure in the system. It also means that our production system is spread across twelve separate servers, where asynchrony and Map-Reduce are king. And aside from being blazing fast, it is also entirely Hydra-compliant.&lt;br /&gt;
&lt;br /&gt;
Furthermore, we will attempt to answer the question: if fedora crashes and the visitors to your site don’t notice, did it really fail?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using Open Source Software and Freeware to Preserve and Deliver Digital Videos ==&lt;br /&gt;
* [mailto:wfang@kinoy.rutgers.edu Wei Fang], Head of Digital Services, Rutgers University Law Library&lt;br /&gt;
* Jiebei Luo, Digital Projects Initiative Intern, Rutgers University&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The Rutgers University Law Library is the official digital repository of the New Jersey Supreme Court oral arguments since 2002. This large video collection contains approximately 3,000 videos with a total of 400 GB or 6,000 viewing hours. With the expansion of this collection, the existing database and the static website could not efficiently support the library’s daily operations and meet its patrons’ search needs. &lt;br /&gt;
By utilizing open source software and freeware such as Ubuntu, FFmpeg, Solr and Drupal, the library is able to develop a complete solution to re-encoding videos, embedding subtitles, incorporating  Solr search engine and content management system to support full-text subtitle search, automatically updating video metadata records in the library catalog system and eventually providing a plug-in free HTML 5-based Web interface for patrons to view the videos online.&lt;br /&gt;
The aspects below will be presented in detail at the conference:&lt;br /&gt;
*	Video codecs comparison &lt;br /&gt;
*	Server-end batch video encoding/re-encoding&lt;br /&gt;
*	HTML 5 video tag and embedding subtitles&lt;br /&gt;
*	Incorporating search engine Solr and content management tool 	Drupal with the database to retrieve videos by full-text search especially in subtitle files&lt;br /&gt;
*	Incorporating video metadata with the library catalog system&lt;br /&gt;
&lt;br /&gt;
== Shared Vision, Shared Resources: the Curate Institutional Repository ==&lt;br /&gt;
* Dan Brubaker Horst, University of Notre Dame &lt;br /&gt;
** [http://code4lib.org/conference/2011/JohnsonHorst A Community-Based Approach to Developing a Digital Exhibit at Notre Dame Using the Hydra Framework] &lt;br /&gt;
* Julie Rudder, Northwestern University&lt;br /&gt;
** no previous presentations&lt;br /&gt;
&lt;br /&gt;
Curate is being collaboratively developed by several institutions in the Hydra community who share the need and vision for a Fedora-backed Institutional Repository. The first release of Curate was a collaboration between Notre Dame and Northwestern University, along with Digital Curation Experts (DCE) - a vendor hired jointly by our two institutions. Powered by the Hydra engine Sufia, the team worked quickly to release the first version of Curate in October 2013 which provides a basic self-deposit system that has support for various content types, collection building, DOI minting, and user profile creation. From the very beginning we have built Curate to be easy to theme and extend in order to ease the process of installation and use by other institutions.&lt;br /&gt;
&lt;br /&gt;
In December 2013, additional partners will join the project including: Indiana University, the University of Cincinnati and the University of Virginia. Each institution contributes resources to the project in order to further our common goal to create a product that fits our needs and has a sustainable future.Together we will tackle additional content types (like complex data, software, media), administrative collections and more. &lt;br /&gt;
&lt;br /&gt;
Our presentation will include:&lt;br /&gt;
* a brief demonstration of Curate and technical overview&lt;br /&gt;
* why and how we work together&lt;br /&gt;
* why build Curate&lt;br /&gt;
* the future of the project&lt;br /&gt;
&lt;br /&gt;
== Solr, Cloud and Blacklight ==&lt;br /&gt;
* David Jiao, Library Information Systems, Indiana University at Bloomington, djiao@indiana.edu&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
SolrCloud refers to the distributed capabilities in Solr4. It is designed to offer a highly available, fault tolerant environment by organizing data into multiple pieces that can be hosted on multiple machines with replicas, and providing a centralized cluster configuration and management. &lt;br /&gt;
&lt;br /&gt;
At Indiana University, we are upgrading our Solr backend for our recently released Blacklight-based OPAC system from Solr 1.4 to Solr4, and we also put up efforts to build a private cloud of Solr4 servers. In this talk, I will persent certain features of SolrCloud, including distributed requests, fault tolerance, near real time indexing/searching, and configuration management with Zookeeper, and our experiences of utilizing these features to provide better performance and architecture for our OPAC system, which serves over 7 million bibliographic records to over 100 thousand students and faculty members. I will also discuss some practical lessons learned from our SolrCloud setup/upgrade and the integration of the new SolrCloud to our customized Blacklight system.  &lt;br /&gt;
&lt;br /&gt;
== Leveraging XSD's for Reflective, Live Dataset Support in Institutional Repositories ==&lt;br /&gt;
* [mailto:msulliva@ufl.edu Mark Sullivan], Library Information Technology, University of Florida&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
The University of Florida Libraries are currently adding support for active datasets into our METS-based institutional repository software.  This ongoing project enables the library to be a partner in current, or long-running, data-driven projects around the university by providing tangible short-term and long-term benefits to the projects.  The system assists project teams by storing and providing access to their data, while supporting online filtering and sorting of the data, custom queries, and adding and editing of the data by authorized users.  We are also exploring simple data visualizations to allow users to perform basic graphical and geographic queries.  Several different schemas were explored including DDI and EML, but ultimately the streamlined approach of using XSD's with some custom attributes was chosen, with all other data residing in the METS file portions.  Currently the system is being developed using XSD's describing XML datasets, but this model should easily scale to support SQL datasets or large datasets supported by Hadoop or iRODS.&lt;br /&gt;
&lt;br /&gt;
This work is being integrated in the open source [http://sobek.ufl.edu SobekCM Digital Content Management System] which is built on a pair-tree structure of METS resources with [http://ufdc.ufl.edu/design/webcontent/sobekcm/SobekCM_Resource_Object.pdf rich metadata support] including DC, MODS, MARC, VRACore, DarwinCore, IEE-LOM, GML/KML, schema.org microdata, and many other standard schemas.  The system has emphasized online, distributed creation and maintenance of resources including geo-placement and geographic searching of resources, building structure maps (table of contents) visually online, and a broad suite of curator tools.  &lt;br /&gt;
&lt;br /&gt;
This work is presented as a model which could be implemented in other systems as well.  We will demonstrate current support and discuss our upcoming roadmap to provide complete support.&lt;br /&gt;
&lt;br /&gt;
== Dead-simple Video Content Management: Let Your Filesystem Do The Work ==&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, NCSU Libraries (akorphan (at) ncsu.edu)&lt;br /&gt;
** (never led or soloed a C4L presentation)&lt;br /&gt;
&lt;br /&gt;
Content management is hard. To keep all the moving parts in order, and to maintain a layer of separation between the system and content creators (who are frequently not technical experts), we typically turn to content management systems like Drupal. But even Drupal and its kin require significant overhead and present a not inconsiderable learning curve for nontechnical users.&lt;br /&gt;
&lt;br /&gt;
In some contexts it's possible -- and desirable -- to manage content in a more streamlined, lightweight way, with a minimum of fuss and technical infrastructure. In this presentation I'll share a simple MVC-like architecture for managing video content for playback on the web, which uses a combination of Apache's mod_rewrite module and your server's filesystem structure to provide an automated approach to video content management that's easy to implement and provides a low barrier to content updates: friendly to content creators and technology implementors alike. Even better, the basic method is HTML5-friendly, and can be integrated into your favorite content management system if you've got permissions for creating templates.&lt;br /&gt;
&lt;br /&gt;
In the presentation I'll go into detail about the system structure and logic required to implement this approach. I'll detail the benefits and limitations of the system, as well as the challenges I encountered in developing its implementation. Audience members should come away with sufficient background to implement a similar system on their own servers. Implementation documentation and genericized code will also be shared, as available.&lt;br /&gt;
&lt;br /&gt;
== Managing Discovery ==&lt;br /&gt;
&lt;br /&gt;
* Andrew Pasterfield, Senior Programmer/Systems Analyst, University of Calgary Library, ampaster@ucalgary.ca&lt;br /&gt;
**No previous code4lib presentations &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
In fall 2012 the University of Calgary Library launched a new home page that incorporated a Summon powered&lt;br /&gt;
Single Search Box with customized “bento box” results display. Search at the U of C now combines a range of&lt;br /&gt;
metadata sources for discovery and customized mapping of a database recommender and LibGuide into a unified&lt;br /&gt;
display.  Further customizations include a non Google Analytics/non proxy method to log clicks.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss the technical details of bringing the various systems together into one display interface to increase discovery at the U of C Library.&lt;br /&gt;
&lt;br /&gt;
http://library.ucalgary.ca&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sorting it out: a piece of the User Centered Design Process ==&lt;br /&gt;
&lt;br /&gt;
* Cindy Beggs, [http://www.akendi.com/aboutus/management/ Akendi], cindy@akendi.com&lt;br /&gt;
&lt;br /&gt;
This talk is about how to apply a user centered design methodology to the process of creating an information architecture.  Participants learn the fundamentals of UCD and how card sorting and reverse card sorting enable us to isolate the content we present on screen from the layouts and visuals of those screens.  We talk about ways to identify who will be using the information architecture you are creating and why we need to know how it will be used.&lt;br /&gt;
 &lt;br /&gt;
What will attendees takes away from your talk?&lt;br /&gt;
The criticality of involving “real” end users in the process of creating an information architecture.  The basics of following a user-centered-design process in the creation of best in class, content-rich, digital products.&lt;br /&gt;
&lt;br /&gt;
Cindy Beggs has been working in the “information industry” for over 25 years.  A librarian by profession, she has spent decades helping users figure out how to find their way through large bodies of content.  Her insights into how people seek information, her empathy for those who find it a challenge and her practical experience helping organizations figure out how to best structure their content contribute to her success as an information architect with both clients and trainees.  (http://www.akendi.com/aboutus/management/)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Implementation of ArchivesSpace in University of Richmond==&lt;br /&gt;
&lt;br /&gt;
*Birong Ho, bho@richmond.edu&lt;br /&gt;
&lt;br /&gt;
University of Richmond implemented its archive collection management ArchivsSpace in the fall, 2013. As a charter member and the Head of Special Collection as the Board member, implementation of such an Open Source Software became a priority. &lt;br /&gt;
&lt;br /&gt;
Several aspects of implementation will be addressed in the talk. Among them, they are Collections and Repository, storage layer including data format, System resources requirements, Technical architecture, Customization, scaling and integrated with other systems in the library.&lt;br /&gt;
&lt;br /&gt;
The customization, scale, and integration with other systems such as Archeon and Exist on campus became a concern will be focused and elaborated in the talk.&lt;br /&gt;
&lt;br /&gt;
==Easy Wins for Modern Web Technologies in Libraries==&lt;br /&gt;
&lt;br /&gt;
*[mailto:trey.terrell@oregonstate.edu Trey Terrell], Analyst Programmer, Oregon State University&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Oregon State University is currently implementing an updated version of its room reservation system. In its development we've come across and implemented a variety of &amp;quot;easy wins&amp;quot; to make it more responsive, easier to maintain, less expensive to run, and just cooler to experience. While our particular system was in Ruby on Rails, this talk will address general methods and example utilities which can be used no matter your stack.&lt;br /&gt;
&lt;br /&gt;
I'll be talking about things like cache management, reverse proxies, publish/subscribe servers, WebSockets, responsive design, asynchronous processing, and keeping complicated stacks up and running with minimal effort.&lt;br /&gt;
&lt;br /&gt;
==Implementing Islandora at a Small Institution==&lt;br /&gt;
&lt;br /&gt;
*Megan Kudzia, Albion College Library&lt;br /&gt;
*Eddie Bachle, Albion College IT&lt;br /&gt;
**No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
Albion College (and particularly the Library/Archives and Special Collections) has a variety of needs which could be met by an open-source Institutional Repository system. Several months and lots of conversations later, we’re continuing to troubleshoot our way through Islandora. We’d like to talk about what has worked for us, where our frustrations have been, whether it’s even possible to install and develop a system like this at a small institution, and where the process has stalled. &lt;br /&gt;
&lt;br /&gt;
As of right now, we do have a semi-working installation. We’re not sure when it will be ready for our end users, but we'll talk about our development process and evaluate our progress.&lt;br /&gt;
''Contributions also by Nicole Smeltekop, Albion College Archives &amp;amp; Special Collections''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== PhantomJS+Selenium: Easy Automated Testing of AJAX-y UIs ==&lt;br /&gt;
&lt;br /&gt;
* Martin Haye, California Digital Library, martin.haye@ucop.edu&lt;br /&gt;
** Previous Code4Lib Presentation: [http://code4lib.org/conference/2012/collett Beyond code: Versioning data with Git and Mercurial] at Code4Lib 2012 (Martin co-presenting with Stephanie Collett)&lt;br /&gt;
* Mark Redar, California Digital Library, mark.redar@ucop.edu&lt;br /&gt;
&lt;br /&gt;
Web user interfaces are demanding ever-more dynamism and polish, combining HTML5, AJAX, lots of CSS and jQuery (or ilk) to create autocomplete drop-downs, intelligent buttons, stylish alert dialogs, etc. How can you make automated tests for these highly complex and interactive UIs?&lt;br /&gt;
&lt;br /&gt;
Part of the answer is PhantomJS. It’s a modern WebKit browser that’s “headless” (meaning it has no display) that can be driven from command-line Selenium unit tests. PhantomJS is dead simple to install, and its blazing speed and server-friendliness make continuous integration testing easy. You can write UI unit tests in {language-of-your-choice} and run them not just in PhantomJS but in Firefox and Chrome, plus a zillion browser/OS combinations at places like SauceLabs, TestingBot and BrowserStack.&lt;br /&gt;
&lt;br /&gt;
In this double-team live code talk, we’ll explain all that while we demonstrate the following in real time:&lt;br /&gt;
&lt;br /&gt;
* Start with nothing.&lt;br /&gt;
* Install Selenium bindings for Ruby and Python.&lt;br /&gt;
* In each language write a small test of an AJAX-y UI.&lt;br /&gt;
* Run the tests in Firefox, and fix bugs (in the test or UI) as needed.&lt;br /&gt;
* Install PhantomJS.&lt;br /&gt;
* Show the same tests running headless as part of a server-friendly test suite. &lt;br /&gt;
* (Wifi permitting) Show the same tests running on a couple different browser/OS combinations on the server cloud at SauceLabs – talking through a tunnel to the local firewalled application.&lt;br /&gt;
&lt;br /&gt;
==New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library==&lt;br /&gt;
&lt;br /&gt;
* Stephanie Walker – swalker@brooklyn.cuny.edu&lt;br /&gt;
* Howard Spivak – howards@brooklyn.cuny.edu&lt;br /&gt;
* Alex - Alex@brooklyn.cuny.edu&lt;br /&gt;
&lt;br /&gt;
Academic libraries are caught in budget squeezes and often struggle to find ways to communicate value to senior administration and others.  At Brooklyn College Library, we have taken an unusual, possibly unique, approach to these issues.  Our technology staff have long worked directly with librarians to develop products that meet library, faculty, and student needs, and we have shared many of our products with colleagues, including an award-winning website, e-resource, and content management system we call 4MyLibrary, which we shared for free with 8 CUNY colleges, and also an easy-to-use book scanner, which has proven overwhelming popular with students, faculty, other librarians, and numerous campus offices.  Recently, motivated by budget cuts, we decided that what worked for us might interest other libraries, and working with our Office of Technology Commercialization, we started selling 2 products:  our book scanners (at half the price of commercial alternatives), and a hosting service, whereby we could host and support 4MyLibrary for libraries with minimal technology staff.  Both succeeded, and yielded major benefits:  a steady revenue stream and the admiration and serious goodwill of our senior administration and others.   However, this presentation is neither a basic how-to, nor an advertisement.  With this presentation, we hope to spur a conversation for broader collaboration, especially regarding new technologies, among libraries.  We all have some level of technical expertise, most of us are struggling with rising prices and tight budgets, and many of us are unhappy with various technology products we use, from scanners to our ILS.  We believe – and can demonstrate – that with collaboration, we can solve many of our problems, and provide better services to boot. &lt;br /&gt;
&lt;br /&gt;
== Identifiers, Data, and Norse Gods ==&lt;br /&gt;
&lt;br /&gt;
* Ryan Scherle, [http://datadryad.org Dryad Digital Repository], ryan@datadryad.org&lt;br /&gt;
** previous Code4Lib talk [http://ryan.scherle.org/papers/2010-2-code4lib-HIVE.ppt  HIVE: A New Tool for Working With Vocabularies], at Code4Lib 2011.&lt;br /&gt;
&lt;br /&gt;
ORCID and DataCite provide stable identifiers for researchers and and data, respectively. Each system does a fine job of providing value to its users. But wouldn't it be great if they could link their systems to create something much more powerful? Perhaps even as powerful as a god?&lt;br /&gt;
&lt;br /&gt;
Enter [http://odin-project.eu/ ODIN], The ORCID and DataCite Interoperability Network. ODIN is a two-year project to unleash the power of persistent identifiers for researchers and the research they create. This talk will present recent work from the ODIN project, including several tools that can unleash the godlike power of identifiers at your institution. Current tools include:&lt;br /&gt;
* Metadata generator tool: allows repository staff to create DataCite metadata with embedded ORCIDs.&lt;br /&gt;
* Claiming tool: assists researchers in claiming their work within the ORCID system.  &lt;br /&gt;
* ORCID-feed: includes a list of ORCID works on any web page.&lt;br /&gt;
* ODIN's HAMR: assists in populating a DSpace repository with ORCIDs. Based on work from a Code4Lib hackathon!&lt;br /&gt;
&lt;br /&gt;
== Armed Bandits in the Digital Library ==&lt;br /&gt;
&lt;br /&gt;
* Roman Chyla, [http://labs.adsabs.harvard.edu/adsabs/ Astrophysics Data System], rchyla@cfa.harvard.edu&lt;br /&gt;
** Previous Code4Lib: [http://code4lib.org/conference/2013/chyla Citation search in SOLR and second-order operators]&lt;br /&gt;
&lt;br /&gt;
Many of us are using the excellent Lucene library (or SOLR appliance) to provide search functionality. These systems contain number of features to adjust relevancy ranking of hits, but we may not know how to use them. In this presentation, I'll present the available options - eg. what is the default ranking 'Vector space model, what are the alternatives (eg. BM25) and what are the other options we have to tweak and adjust the ranking of the hits (eg. boost factors, functions). But even if we know how to deploy these adjustments and tweaks, we are still left in dark. We do not know whether the change we've just rolled out had a significant (statistically significant) effect or maybe it was just a waste of time and resources? A/B testing is one option, but there may be a much better one - so called &amp;quot;Multi-Armed Bandits Approach&amp;quot;. And in this talk I'd like to show how we are experimenting with this strategy to adjust [http://labs.adsabs.harvard.edu/adsabs/ ADS search engine].&lt;br /&gt;
&lt;br /&gt;
== Building Worker Queues with AWS and Resque ==&lt;br /&gt;
&lt;br /&gt;
* Eric Rochester [http://scholarslab.org Scholars' Lab], erochest@virginia.edu&lt;br /&gt;
* Scott Turnbull [http://aptrust.org/ Academic Preservation Trust], scott.turnbull@aptrust.org &lt;br /&gt;
&lt;br /&gt;
A common task in larger systems is to be able to process large input files automatically. Often users can drop those files into a shared directory on AWS or on NFS or another shared drive. Those files need to be processed and potentially integrated into a system. This task has come up recently in the University of Virginia libraries in allowing users to add GIS data to the system and in setting up a system for the Academic Preservation Trust (http://aptrust.org/) that ingests files and resources into the preservation system.&lt;br /&gt;
&lt;br /&gt;
This system is built by loosely coupling a number of different technologies. This allows us to easily interoperate and communicate between different system and programming environments. Because the interfaces are well defined, it’s also fairly simple to switch out technologies as the requirements of the system change.&lt;br /&gt;
&lt;br /&gt;
The process is fairly simple:&lt;br /&gt;
&lt;br /&gt;
First, a Ruby daemon monitors an AWS S3 bucket that others can upload new files into. This daemon creates a Resque status task, adds a marker for the task in a database, and continues monitoring.&lt;br /&gt;
&lt;br /&gt;
Second, Resque mediates incoming job requests and routes them to the appropriate workers which may be in Java, Go, or Ruby.  The diversity of technologies that Resque can manage allows great latitude to leverage the appropriate tool for a specific job.  While processing, it updates the status for that job and coordinates processing with other jobs.&lt;br /&gt;
&lt;br /&gt;
Finally, a page that is integrated into a larger Rails app provides a novice-user-friendly view of the status of the workers and allows basic tasks such as restarting the job.&lt;br /&gt;
&lt;br /&gt;
This architecture allows us to swap in the technology that best fits each part of the process, and it makes it easier to maintain the system. We use this to integrate and coordinate between tasks handled in Java, Ruby, and Go, and it provides an effective way to interoperate with these programming languages and the respective strengths that they bring to this system.&lt;br /&gt;
&lt;br /&gt;
==Piwik: Open source web analytics==&lt;br /&gt;
* Kirk Hess, University of Illinois at Urbana-Champaign (kirkhess@illinois.edu)&lt;br /&gt;
** (Code4Lib 2012: [http://code4lib.org/conference/2012/hess| Discovering Digital Library User Behavior with Google Analytics])&lt;br /&gt;
&lt;br /&gt;
While Google Analytics is synonymous with Web Analytics, fortunately today we have many other good options, and one option is Piwik, [http://piwik.org| piwik.org] a simple to install, open-source PHP/MySQL application with a tracking script that will sit alongside Google Analytics tracking the usual clicks, events and variables. In this presentation, I'd like to cover the usual analytics topics and also cover what makes Piwik powerful, such as importing and visualizing web logs from any system to incorporate both past and future data, easily tracking downloads, and the ability to write your own reports or dashboard. The visitor log data is stored securely on your own server so you have control over who looks at the data and how much or how little to keep. With an active and helpful developer community, Piwik has the potential for analytics which makes sense for libraries, not e-commerce.&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39855</id>
		<title>2014 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39855"/>
				<updated>2013-11-08T20:31:50Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Identifiers, Data, and Norse Gods */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
 &lt;br /&gt;
* ''Projects'' you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* ''Tools and technologies'' – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* ''Technical issues'' - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* ''Relevant non-technical issues'' – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
'''To Propose a Talk'''&lt;br /&gt;
* Log in to the wiki in order to submit a proposal. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
* Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
* If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist us in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
As in past years, the Code4Lib community will vote on proposals that they would like to see included in the program. This year, however, only the top 10 proposals will be guaranteed a slot at the conference. Additional presentations will be selected by the Program Committee in an effort to ensure diversity in program content. Community votes will, of course, still weigh heavily in these decisions.&lt;br /&gt;
&lt;br /&gt;
Presenters whose proposals are selected for inclusion in the program will be guaranteed an opportunity to register for the conference. The standard conference registration fee will still apply.&lt;br /&gt;
&lt;br /&gt;
''Proposals can be submitted through '''Friday, November 8, 2013, at 5pm PST'''''. Voting will commence on November 18, 2013 and continue through December 6, 2013. The final line-up of presentations will be announced in early January, 2014.&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
&lt;br /&gt;
==Creating a new Greek-Dutch dictionary==&lt;br /&gt;
* Caspar Treijtel, University of Amsterdam, c.treijtel@uva.nl&lt;br /&gt;
&lt;br /&gt;
At present, no complete dictionary of (ancient) Greek-Dutch is available online. A new dictionary is currently under construction at Leiden University, with software being developed at the University of Amsterdam. The team in Leiden has already begun preparation of the data, with at this moment about 6,000 approved lemmas. The ultimate goal is to produce both a print version and online open access version from the same source documents. The software needed for this has been made in a project that was funded by CLARIN-NL.&lt;br /&gt;
&lt;br /&gt;
Migrator&lt;br /&gt;
&lt;br /&gt;
For the production of lemmas we have implemented an advanced workflow. The (generally non-technical) users create lemmas using MS Word, which is both familiar and easy to use. We have developed a custom software module that carefully migrates the Word documents into deeply structured XML by analyzing the structure and semantics of the lemmas, and falling back on heuristics in ambiguous cases. While having initially envisioned the oXygen XML Author component as the main tool for creating new lemmas, we obtained excellent results with the migrator module, and decided therefore to continue using MS Word as the primary composition tool. The main advantage of this is that the editors are much more familiar with Word than with any other WYSIWYG editor. Lemmas that have been migrated to XML are stored in an XML database and can be further edited using oXygen XML Author.&lt;br /&gt;
&lt;br /&gt;
Lemmatizer&lt;br /&gt;
&lt;br /&gt;
Greek morphology is complicated. In order to use a dictionary effectively, a rather high level of initial language competence is necessary for the user to be able to relate the word form s/he finds in a text to the correct basic lemma form, where the definition of the word can be found. Using a Greek morphological database we have been able to facilitate the search for lemmas. A ‘lemmatizer’ module gives the possible parsings of the word forms and the lemmas they can be derived from. This enables the user to type in the word as found in the text and be redirected to the correct lemma.&lt;br /&gt;
&lt;br /&gt;
Visualization&lt;br /&gt;
&lt;br /&gt;
For the online dictionary we have implemented a visualization module that allows the user to view multiple lemmas at once. The implementation of this module has been done using the Javascript framework MooTools. The result is a viewer that performs really well and is run by maintainable Javascript code.&lt;br /&gt;
&lt;br /&gt;
The online dictionary is still being worked on, have a look at http://www.woordenboekgrieks.nl/ for the beta version. A newer test version with additional features can be found here: http://angel.ic.uva.nl:8600/.&lt;br /&gt;
&lt;br /&gt;
Credits&lt;br /&gt;
&lt;br /&gt;
* construction of the dictionary: Prof. Ineke Sluiter, Classics department of Leiden University; Prof. Albert Rijksbaron, University of Amsterdam&lt;br /&gt;
* publisher of the dictionary: Amsterdam University Press&lt;br /&gt;
* design/typesetting dictionary: TaT Zetwerk (http://www.tatzetwerk.nl/)&lt;br /&gt;
* software development: Digital Production Center, University Library, University of Amsterdam&lt;br /&gt;
* project funding: CLARIN-NL (http://www.clarin.nl/)&lt;br /&gt;
* morphological database for use by the lemmatizer: courtesy of Prof. Helma Dik, University of Chicago (based on data of the Perseus Project)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Using Drupal to drive alternative presentation systems ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Recently, we have been building systems that use angular.js, Rails, or other systems for presentation, while leveraging Drupal's sophisticated content management capabilities on the back end.&lt;br /&gt;
&lt;br /&gt;
So far, these have been one-way systems, but as we move to Drupal 8 we are beginning to explore ways to further decouple the presentation and CMS functions.&lt;br /&gt;
&lt;br /&gt;
== A Book, a Web Browser and a Tablet: How Bibliotheca Alexandrina's Book Viewer Framework Makes It Possible ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mohammed.abuouda|Mohammed Abu ouda]], Bibliotheca Alexandrina (The new Library of Alexandria)&lt;br /&gt;
&lt;br /&gt;
A lot of institutions around the world are engaged in multiple digitization projects aiming at preserving the human knowledge present in books and availing them through multiple channels to people around the whole globe. These efforts will sure help close the digital gap particularly with the arrival of affordable e-readers, mobile phones and network coverage. However, the digital reading experience has not yet arrived to its maximum potential. Many readers miss features they like in their good old books and wish to find them in their digital counterpart. In an attempt to create a unique digital reading experience, Bibliotheca Alexandria (BA) created a flexible book viewing framework that is currently used to access its current collection of more than 300,000 digital books in five different languages which includes the largest collection of digitized Arabic books.&lt;br /&gt;
&lt;br /&gt;
Using open source tools, BA used the framework to develop a modular book viewer that can be deployed in different environments and is currently at the heart of various BA projects. The Book viewer provides several features creating a more natural reading experience. As with physical books, the reader can now personalize the books he reads by adding annotations like highlights, underlines and sticky notes to capture his thoughts and ideas in addition to being able to share the book with friends on social networks. The reader can perform a search across the content of the book receiving highlighted search results within the pages of the book. More features can be further added to the book viewer through its plugin architecture.&lt;br /&gt;
&lt;br /&gt;
== Structured data NOW: seeding schema.org in library systems ==&lt;br /&gt;
 &lt;br /&gt;
* [http://coffeecode.net Dan Scott], Laurentian University&lt;br /&gt;
** Previous code4lib presentations: [https://archive.org/details/code4lib.conf.2008.pres.CouchDBsacrilege CouchDB is sacrilege... mmm, delicious sacrilege] at Code4Lib 2008&lt;br /&gt;
&lt;br /&gt;
The semantic web, linked data, and structured data are all fantastic ideas with a barrier imposed by implementation constraints. If their system does not allow customizations, or the institution lacks skilled human resources, it does not matter how enthused a given library might be about publishing structured data... it will not happen. However, if the software in use simply publishes structured data by default, then the web will be populated for free. Really! No extra resources necessary.&lt;br /&gt;
&lt;br /&gt;
This presentation highlights Dan's work with systems such as Evergreen, Koha, and VuFind to enable the publication of schema.org structured data out-of-the-box. Along the way, we reflect the current state of the W3C Schema.org Bibliographic Extension community group efforts to shape the evolution of the schema.org vocabulary. Finally, hold on tight as we contemplate next steps and the possibilities of a world where structured data is the norm on the web.&lt;br /&gt;
&lt;br /&gt;
== Towards Pasta Code Nirvana: Using JavaScript MVC to Fill Your Programming Ravioli ==&lt;br /&gt;
&lt;br /&gt;
* Bret Davidson, North Carolina State University Libraries, bret_davidson@ncsu.edu&lt;br /&gt;
** Previous Code4Lib Presentations: [http://wiki.code4lib.org/index.php/2013_talks_proposals#Data-Driven_Documents:_Visualizing_library_data_with_D3.js Visualizing library data with D3.js] at Code4Lib 2013&lt;br /&gt;
&lt;br /&gt;
JavaScript MVC frameworks are ushering in a golden age of robust and responsive web applications that take advantage of evergreen browsers, performant JS engines, and the unprecedented reach provided by billions of personal computing devices. The web browser has emerged as the world’s most popular application runtime and the complexity[1] and scope of JavaScript applications has exploded accordingly. Server-side web frameworks like Rails and Django have helped developers adhere to best practices like modularity, dependency injection, and unit testing for years, practices that are now being applied to JavaScript development through projects like Backbone[2], Ember[3], and Angular[4].&lt;br /&gt;
&lt;br /&gt;
This talk will discuss the issues JavaScript MVC frameworks are trying to solve, common features like data binding, implications for the future of web development[5], and the appropriateness of JavaScript MVC for library applications.&lt;br /&gt;
&lt;br /&gt;
*[1]http://en.wikipedia.org/wiki/Spaghetti_code&lt;br /&gt;
*[2]http://backbonejs.org&lt;br /&gt;
*[3]http://emberjs.com&lt;br /&gt;
*[4]http://angularjs.org&lt;br /&gt;
*[5]http://tomdale.net/2013/09/progressive-enhancement-is-dead/&lt;br /&gt;
&lt;br /&gt;
== WebSockets for Real-Time and Interactive Interfaces ==&lt;br /&gt;
&lt;br /&gt;
* [http://ronallo.com Jason Ronallo], NCSU Libraries, jason_ronallo@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/ronallo HTML5 Microdata and Schema.org] 2012&lt;br /&gt;
* [http://code4lib.org/conference/2013/ronallo HTML5 Video Now!] 2013&lt;br /&gt;
&lt;br /&gt;
Watching the Google Analytics Real-Time dashboard for the first time was mesmerizing. As soon as someone visited a site, I could see what page they were on. For a digital collections site with a lot of images, it was fun to see what visitors were looking at. But getting from Google Analytics to the image or other content of what was currently being viewed was cumbersome. The real-time experience was something I wanted share with others. I'll show you how I used a WebSocket service to create a real-time interface to digital collections.&lt;br /&gt;
&lt;br /&gt;
In the Hunt Library at NCSU we have some large video walls. I wanted to make HTML-based exhibits that featured viewer interactions. I'll show you how I converted Listen to Wikipedia [1] into an bring-your-own-device interactive exhibit. With WebSockets any HTML page can be remote controlled by any internet connected device.&lt;br /&gt;
&lt;br /&gt;
I will attempt to include real-time audience participation.&lt;br /&gt;
&lt;br /&gt;
[1] http://listen.hatnote.com/&lt;br /&gt;
&lt;br /&gt;
== Rapid Development of Automated Tasks with the File Analyzer ==&lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Libraries, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Libraries have customized the File Analyzer and Metadata Harvester application (https://github.com/Georgetown-University-Libraries/File-Analyzer) to solve a number of library automation challenges:&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census.  After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform.  Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules.  As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
Learn about the functionality that is available for download, how you can use this tool to automate workflows from digital collections to ILS ingests to electronic resources statistics and also discuss the opportunities to collaborate on enhancements to this application!&lt;br /&gt;
&lt;br /&gt;
== GeoHydra: How to Build a Geospatial Digital Library with Fedora ==&lt;br /&gt;
 &lt;br /&gt;
* [http://stanford.edu/~drh Darren Hardy], Stanford University, drh@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Geographically-rich data are exploding and putting fear in those trying to&lt;br /&gt;
tackle integrating them into existing digital library infrastructures.&lt;br /&gt;
Building a spatial data infrastructure that integrates with your digital&lt;br /&gt;
library infrastructure need not be a daunting task. We have successfully&lt;br /&gt;
deployed a geospatial digital library infrastructure using Fedora and&lt;br /&gt;
open-source geospatial software [1]. We'll discuss the primary design&lt;br /&gt;
decisions and technologies that led to a production deployment within a few&lt;br /&gt;
months. Briefly, our architecture revolves around discovery, delivery, and&lt;br /&gt;
metadata pipelines using open-source OpenGeoPortal [2], Solr [3], GeoServer&lt;br /&gt;
[4], PostGIS [5], and GeoNetwork [6] technologies, plus the proprietary ESRI&lt;br /&gt;
ArcMap [7] -- the GIS industry's workhorse. Finally, we'll discuss the key&lt;br /&gt;
skillsets needed to build and maintain a spatial data infrastructure.&lt;br /&gt;
&lt;br /&gt;
[1] http://foss4g.org&lt;br /&gt;
[2] http://opengeoportal.org&lt;br /&gt;
[3] http://lucene.apache.org/solr&lt;br /&gt;
[4] http://geoserver.org&lt;br /&gt;
[5] http://postgis.net&lt;br /&gt;
[6] http://geonetwork-opensource.org&lt;br /&gt;
[7] http://esri.com&lt;br /&gt;
&lt;br /&gt;
==Under the Hood of Hadoop Processing at OCLC Research ==&lt;br /&gt;
&lt;br /&gt;
[http://roytennant.com/ Roy Tennant]&lt;br /&gt;
&lt;br /&gt;
* Previous Code4Lib presentations: 2006: &amp;quot;The Case for Code4Lib 501c(3)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[http://hadoop.apache.org/ Apache Hadoop] is widely used by Yahoo!, Google, and many others to process massive amounts of data quickly. OCLC Research uses a 40-node compute cluster with Hadoop and HBase to process the 300 million MARC records of WorldCat in various ways. This presentation will explain how Hadoop MapReduce works and illustrate it with specific examples and code. The role of the jobtracker in both monitoring and reporting on processes will be explained. String searching WorldCat will also be demonstrated live.&lt;br /&gt;
&lt;br /&gt;
== Quick and Easy Data Visualization with Google Visualization API and Google Chart Libraries ==&lt;br /&gt;
 &lt;br /&gt;
[http://bohyunkim.net/blog Bohyun Kim], Florida International University, bohyun.kim@fiu.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Do most of the data that your library collects stay in spreadsheets or are published as a static table with a series of boring numbers? Do your library stakeholders spend more time collecting the data than using it as a decision-making tool because the data is presented in a way that makes it hard for them [http://developers.google.com/chart/interactive/docs/gallery to quickly grasp its significance? ]&lt;br /&gt;
&lt;br /&gt;
This talk will provide an overview of [http://developers.google.com/chart/interactive/docs/reference Google Visualization API] [2] and [http://developers.google.com/chart/ Google Chart Libraries] [3] to get you started on the way to quickly query and visualize your library data from remote data sources (e.g. a Google Spreadsheet or your own database) with (or without) cool-looking user-controls, animation effects, and even a dashboard.&lt;br /&gt;
&lt;br /&gt;
== Leap Motion + Rare Books: A hands-free way to view and interact with rare books in 3D ==&lt;br /&gt;
 &lt;br /&gt;
[http://http://www.youtube.com/user/jpdenzer Juan Denzer], Binghamton University, jdenzer@binghamton.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
As rare books become more delicate over time, making them available to the public becomes harder.  We at Binghamton University Library have developed an application that makes it easier to view rare books without ever having to touch them.  We have combined the Leap Motion hands-free device and 3D rendered models to create a new virtual experience for the viewer.&lt;br /&gt;
&lt;br /&gt;
The application allows the user to rotate and zoom in on a 3D representation of a rare book.  The user is also able to ‘open’ the virtual book and flip through it using a natural user interface.  Such as swiping the hand left or right to turn the page.&lt;br /&gt;
&lt;br /&gt;
The application is built on the .Net framework and is written in C#.  3D models are created using simple 3D software such as sketchup or Blender.  Scans of the book cover and spine are created using simple flatbed scanners.  The inside pages are scanned using overhead scanners. &lt;br /&gt;
&lt;br /&gt;
This talk with discuss the technologies used in developing the application and virtually any library could implement the application with virtually no coding at all. This presentation will have a demonstration of the software and also a chance for audience members to experience the Rare Book Leap Motion App themselves.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Course Reserves Unleashed! ==&lt;br /&gt;
 &lt;br /&gt;
* Bobbi Fox, Library Technology Services, Harvard University, bobbi_fox@harvard.edu&lt;br /&gt;
* Gloria Korsman, Andover-Harvard Theological Library&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Hey kids!  Remember when SOAP was used for something other than washing?  Our sophisticated (and highly functional) Course Reserves Request system does!&lt;br /&gt;
&lt;br /&gt;
However, while the system is great for submitting and  processing course reserve requests, the student-facing presentation through Havard’s home-grown -- and soon to be replaced -- LMS leaves a lot to be desired.  &lt;br /&gt;
&lt;br /&gt;
Follow along as we leverage Solr 4 as a No-SQL database, along with more progressive RESTful API techniques, to release Reserves data into the wild without interfering with reserves request processing -- and, in the process, open up the opportunity for other schools to feed their data in as well.&lt;br /&gt;
&lt;br /&gt;
== We Are All Disabled! Universal Web Design Making Web Services Accessible for Everyone ==&lt;br /&gt;
 &lt;br /&gt;
Cynthia Ng, Accessibility Librarian, CILS at Langara College&lt;br /&gt;
* No previous Code4Lib presentations (not counting lightning talks)&lt;br /&gt;
&lt;br /&gt;
We’re building and improving tools and services all the time, but do you only develop for the “average” user or add things for “disabled” users? We all use “assistive” technology accessing information in a multitude of ways with different platforms, devices, etc. Let’s focus on providing web services that are accessible to everyone without it being onerous or ugly. The aim is to get you thinking about what you can do to make web-based services and content more accessible for all from the beginning or with small amounts of effort whether you're a developer or not.&lt;br /&gt;
&lt;br /&gt;
The goal of the presentation is to provide both developers and content creators with information on simple, practical ways to make web content and web services more accessible. However, rather than thinking about putting in extra effort or making adjustment for those with disabilities, I want to help people think about how to make their websites more accessible for all users through universal web design.&lt;br /&gt;
&lt;br /&gt;
== Personalize your Google Analytics Data with Custom Events and Variables ==&lt;br /&gt;
&lt;br /&gt;
[http://joshwilson.net Josh Wilson], Systems Integration Librarian, State Library of North Carolina - joshwilsonnc@gmail.com&lt;br /&gt;
&lt;br /&gt;
At the State Library of North Carolina, we had more specific questions about the use of our digital collections than standard GA could provide. A few implementations of custom events and custom variables later, we have our answers.&lt;br /&gt;
&lt;br /&gt;
I'll demonstrate how these analytics add-ons work, and why implementation can sometimes be more complicated than just adding a few lines of JavaScript to your ga.js. I'll discuss some specific examples in use at the SLNC:&lt;br /&gt;
&lt;br /&gt;
* Capturing the content of specific metadata fields in CONTENTdm as Custom Events &lt;br /&gt;
* Recording Drupal taxonomy terms as Custom Variables&lt;br /&gt;
&lt;br /&gt;
In both instances, this data deepened our understanding of how our sites and collections were being used, and in turn, we were able to report usage more accurately to content contributors and other stakeholders.&lt;br /&gt;
&lt;br /&gt;
More on: [https://developers.google.com/analytics/devguides/collection/gajs/eventTrackerGuide GA Custom Events] | [https://developers.google.com/analytics/devguides/collection/gajs/gaTrackingCustomVariables GA Custom Variables]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Behold Fedora 4: The Incredible Shrinking Repository! ==&lt;br /&gt;
&lt;br /&gt;
Esmé Cowles, UC San Diego Library.  Previous talk: [http://code4lib.org/conference/2013/cowles-critchlow-westbrook All Teh Metadatas Re-Revisited] (2013)&lt;br /&gt;
&lt;br /&gt;
* One repository contains untold numbers of digital objects and powers many Hydra and Islandora apps&lt;br /&gt;
* It speaks RDF, but contains no triplestore! (triplestores sold separately, SPARQL Update may be involved, some restrictions apply)&lt;br /&gt;
* Flexible enough to tie itself in knots implementing storage and access control policies&lt;br /&gt;
* Witness feats of strength and scalability, with dramatically increased performance and clustering&lt;br /&gt;
* Plumb the depths of bottomless hierarchies, and marvel at the metadata woven into the very fabric of the repository&lt;br /&gt;
* Ponder the paradox of ingesting large files by not ingesting them&lt;br /&gt;
* Be amazed as Fedora 4 swallows other systems whole (including Fedora 3 repositories)&lt;br /&gt;
* Watch novice developers setup Fedora 4 from scratch, with just a handful of incantations to Git and Maven&lt;br /&gt;
&lt;br /&gt;
The Fedora Commons Repository is the foundation of many digital collections, e-research, digital library, archives, digital preservation, institutional repository and open access publishing systems.  This talk will focus on how Fedora 4 improves core repository functionality, adds new features, maintains backwards compatibility, and addresses the shortcomings of Fedora 3.&lt;br /&gt;
&lt;br /&gt;
== Organic Free-Range API Development - Making Web Services That You Will Actually Want to Consume ==&lt;br /&gt;
&lt;br /&gt;
Steve Meyer and Karen Coombs, OCLC&lt;br /&gt;
&lt;br /&gt;
Building web services can have great benefits by providing reusability of data and functionality. Underpinning your applications with a web service will allow you to write code once and support multiple environments: your library's web app, mobile applications, the embedded widget in your campus portal. However, building a web service is its own kind of artful programming. Doing it well requires attention to many of the same techniques and requirements as building web applications, though with different outcomes. &lt;br /&gt;
&lt;br /&gt;
So what are the usability principles for web services? How do you build a web service that you (and others) will actually want to use? In this talk, we’ll share some of the lessons learned - the good, the bad, and the ugly - through OCLC's work on the WorldCat Metadata API. This web service is a sophisticated API that provides external clients with read and write access to WorldCat data. It provides a model to help aspiring API creators navigate the potential complications of crafting a web service. We'll cover:&lt;br /&gt;
&lt;br /&gt;
* Loose coupling of data assets and resource-oriented data modeling at the core&lt;br /&gt;
* Coding to standards vs. exposure of an internal data model&lt;br /&gt;
* Authentication and security for web services: API Keys, Digital Signing, OAuth Flows&lt;br /&gt;
* Building web services that behave as a suite so it looks like the left hand knows what the right hand is doing&lt;br /&gt;
&lt;br /&gt;
So at the end of the day, your team will know your API is a very good egg after all. &lt;br /&gt;
&lt;br /&gt;
If accepted, the presenters intend to produce and share a Quick Guide for building a web service that will reflect content presented in the talk.&lt;br /&gt;
&lt;br /&gt;
== Lucene's Latest (for Libraries) ==&lt;br /&gt;
&lt;br /&gt;
erik.hatcher@lucidworks.com&lt;br /&gt;
&lt;br /&gt;
Lucene powers the search capabilities of practically all library discovery platforms, by way of Solr, etc.  The Lucene project evolves rapidly, and it's a full-time job to keep up with the ever improving features and scalability.   This talk will distill and showcase the most relevant(!) advancements to date.&lt;br /&gt;
&lt;br /&gt;
== The Why and How of Very Large Displays in Libraries. ==&lt;br /&gt;
&lt;br /&gt;
* Cory Lown, NCSU Libraries, cwlown@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib Presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/lown How People Search the Library from a Single Search Box]  2012&lt;br /&gt;
* [http://code4lib.org/conference/2010/orphanides_lown_lynema Enhancing Discoverability with Virtual Shelf Browse] 2010&lt;br /&gt;
&lt;br /&gt;
Built into the walls of NC State's new Hunt Library are several [http://www.christiedigital.com/en-us/digital-signage/products/microtiles/pages/microtiles-digital-signage-video-wall.aspx Christie MicroTile Display Wall Systems]. What does a library do with a display that's seven feet tall and over twenty feet wide? I'll talk about why libraries might want large displays like this, what we're doing with them right now, and what we might do with them in the future. I'll talk about how these displays factor into planning for new and existing web projects. And I'll get into the fun details of how you build web applications that scale from the very small browser window on a phone all the way up to a browser window with about 14 million pixels (about 10 million more than a dual 24&amp;quot; monitor desktop setup).&lt;br /&gt;
&lt;br /&gt;
== Discovering your Discovery System in Real Time. ==&lt;br /&gt;
&lt;br /&gt;
* Godmar Back, Virginia Tech, gback@vt.edu&lt;br /&gt;
* Annette Bailey, Virginia Tech, afbailey@vt.edu&lt;br /&gt;
&lt;br /&gt;
Practically all libraries today provide web-based discovery systems to their users;&lt;br /&gt;
users discover items and peruse or check them out by clicking on links.  Unlike&lt;br /&gt;
the traditional transaction of checking out a book at the circulation desk, this&lt;br /&gt;
interaction is largely invisible.  We have built a system that records user's&lt;br /&gt;
interaction with Summon in real-time, processes the resulting data with minimal delay,&lt;br /&gt;
and visualizes it in various ways using Google Charts and using various d3.js modules,&lt;br /&gt;
such as word clouds, tree maps, and others.&lt;br /&gt;
&lt;br /&gt;
These visualizations can be embedded in web sites, but are also suitable for&lt;br /&gt;
projection via large-scale displays or projectors right into the 'Learning Spaces'&lt;br /&gt;
many libraries are converted into.  The goal of this talk is to share the technology&lt;br /&gt;
and advocate the building of a cloud-based infrastructure that would make this&lt;br /&gt;
technology available to any library that uses a discovery system, rather than just&lt;br /&gt;
those who have the technological prowess for developing such systems and&lt;br /&gt;
visualizations in-house.  &lt;br /&gt;
&lt;br /&gt;
Previous presentations at Code4Lib:&lt;br /&gt;
* Talk: Code4Lib 2009 [http://code4lib.org/files/LibX2.0-Code4Lib-2009AsPresented.ppt LibX 2.0]&lt;br /&gt;
* Preconference: [http://wiki.code4lib.org/index.php/LibX_Preconference LibX 2.0, 2009]&lt;br /&gt;
* Preconference: Code4Lib 2010, On Widgets and Web Services&lt;br /&gt;
&lt;br /&gt;
== Your Library, Anywhere: A Modern, Responsive Library Catalogue at University of Toronto Libraries ==&lt;br /&gt;
&lt;br /&gt;
* Bilal Khalid, Gordon Belray, Lisa Gayhart (lisa.gayhart@utoronto.ca)&lt;br /&gt;
&lt;br /&gt;
* No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. &lt;br /&gt;
&lt;br /&gt;
[http://search.library.utoronto.ca The result]: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles. This talk will discuss: application development; service improvements; interface design; and user outreach, testing, and project communications. Feedback and questions from the audience are very welcome. If time runs short, we will be available for questions and conversation after the presentation.&lt;br /&gt;
&lt;br /&gt;
Note: A version of this content has been provisionally accepted as an article for Code4Lib Journal, January 2014 publication.)&lt;br /&gt;
&lt;br /&gt;
== All Tiled Up ==&lt;br /&gt;
&lt;br /&gt;
* Mike Graves, MIT Libraries (mgraves@mit.edu)&lt;br /&gt;
&lt;br /&gt;
You've got maps. You even scanned and georeferenced them. Now what? Running a full GIS stack can be expensive, and overkill in some cases. The good news is that you have a lot more options now than you did just a few years ago. I'd like to present some lighter weight solutions to making georeferenced images available on the Web.&lt;br /&gt;
&lt;br /&gt;
This talk will provide an introduction to MBTiles. I'll go over what they are, how you create them, how you use them and why you would use them.&lt;br /&gt;
&lt;br /&gt;
== The Great War: Image Interoperability to Facebook ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, Los Alamos National Laboratory (azaroth42@gmail.com)&lt;br /&gt;
** (Code4Lib 2006: [http://www.code4lib.org/2006/sanderson | Library Text Mining])&lt;br /&gt;
* Rob Warren, Carleton University&lt;br /&gt;
** No previous presentations&lt;br /&gt;
&lt;br /&gt;
Using a pipeline constructed from Linked Open Data and other interoperability specifications, it is possible to merge and re-use image and textual data from distributed library collections to build new, useful tools and applications.  Starting with the OAI-PMH interface to ContentDM, we will take you on a tour through the International Image Interoperability Framework and Shared Canvas, to a cross-institutional viewer, and image analysis for the purposes of building a historical Facebook from finding and tagging people in photographs.  The World War One collections are drawn from multiple institutions and merged by the machine learning code.&lt;br /&gt;
&lt;br /&gt;
The presentation will focus on the (open source) toolchain and the benefits of the use of standards throughout:  OAI-PMH to get the metadata, IIIF for interaction with the images, the Shared Canvas ontology for describing collections of digitized objects, Open Annotation for tagging things in the images and specialized ontologies that are specific to the contents.  The tools include standard RDF / OWL technologies, JSON-LD, imagemagick and OpenCV for image analysis.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Solr Search Results with D3.js for User-Friendly Navigation of Large Results Sets ==&lt;br /&gt;
&lt;br /&gt;
*Julia Bauder, Grinnell College Libraries (bauderj-at-grinnell-dot-edu)&lt;br /&gt;
*No previous presentations at national Code4Lib conferences&lt;br /&gt;
&lt;br /&gt;
As the corpus of articles, books, and other resources searched by discovery systems continues to get bigger, searchers are more and more frequently confronted with unmanageably large numbers of results. How can we help users make sense of 10,000 hits and find the ones they actually want? Facets help, but making sense of a gigantic sidebar of facets is not an easy task for users, either.&lt;br /&gt;
During this talk, I will explain how we will soon be using Solr 4’s pivot queries and hierarchical visualizations (e.g., treemaps) from D3.js to let patrons view and manipulate search results. We will be doing this with our VuFind 2.0 catalog, but this technique will work with any system running Solr 4. I will also talk about early student reaction to our tests of these visualization features.&lt;br /&gt;
&lt;br /&gt;
== PeerLibrary – open source cloud based collaborative library ==&lt;br /&gt;
&lt;br /&gt;
[https://github.com/peerlibrary/peerlibrary PeerLibrary is a new open source project] and a cloud service providing collaborative reading, sharing and storing. Users can upload publications they want to read (currently in PDF format), read them in the browser in real-time with others, highlight, annotate and organize their own or collaborative library. PeerLibrary provides a search engine to search over all uploaded open access publications. Additionally, it aims to collaboratively aggregate the open layer of knowledge on top of this publications through public annotations and references user will add to publications. In this way publications would not just be available to read, but accessible to the general public as well. Currently, it is aiming at scientific community and scientific publications.&lt;br /&gt;
&lt;br /&gt;
See [http://blog.peerlibrary.org/post/63458789185/screencast-previewing-the-peerlibrary-project screencast here].&lt;br /&gt;
&lt;br /&gt;
It is still in development and beta launch is planned at the end of November.&lt;br /&gt;
&lt;br /&gt;
== Who was where when, or finding biographical articles on Wikipedia by place and time ==&lt;br /&gt;
&lt;br /&gt;
* [http://morton-owens.info Emily Morton-Owens], The Seattle Public Library (presenting on work from NYU)&lt;br /&gt;
* No previous c4l presentations&lt;br /&gt;
&lt;br /&gt;
It's easy to answer the question &amp;quot;What important people were in Paris in 1939?&amp;quot; But what about Virginia in the 1750s or Scandinavia in the 14th century? I created a tool that allows you to search for biographies in a generally applicable way, using a map interface. I would like to present updates to my thesis project, which combines a crawler written in Java that extracts information from Wikipedia articles, with a MongoDB data store and a frontend in Python.&lt;br /&gt;
&lt;br /&gt;
The input to the project is freetext of entire articles in Wikipedia; this is important to allow us to pick up Benjamin Franklin not just in the single most obvious place of Philadelphia but also in London, Paris, Boston, etc. I can talk about my experiments disambiguating place names (approaches pioneered on newspaper articles were actually unhelpful on this type of text) and setting up a processing queue that does not become mired in the biographies of every human who ever played soccer. I also want to mitigate some of the implementation choices I made due to my academic deadline and improve the accuracy/usability.&lt;br /&gt;
&lt;br /&gt;
What I hope to show is that I was able to develop a novel and useful reference tool automatically, using fairly simple heuristics that are a far cry from hand-cataloging familiar to many librarians.&lt;br /&gt;
&lt;br /&gt;
You can try out [http://linserv1.cims.nyu.edu:48866/ the original version] (this server is inconveniently set to be updated/rebooted on 11/8--may be temporarily unavailable)&lt;br /&gt;
&lt;br /&gt;
== Good!, DRY, and Dynamic: Content Strategy for Libraries (Especially the Big Ones) ==&lt;br /&gt;
&lt;br /&gt;
*Michael Schofield, Nova Southeastern University Libraries, mschofield@nova.edu&lt;br /&gt;
*No previous code4lib presentations.&lt;br /&gt;
&lt;br /&gt;
The responsibilities of the #libweb are exploding [it’s a good thing] and it is no longer uncommon for libraries to manage or even home-grow multiple applications and sites. Often it is at this point where the web people begin to suffer the absence of a content strategy when, say, business hours need to be updated sitewide a half-dozen times.&lt;br /&gt;
&lt;br /&gt;
We were already feeling this crunch when we decided to further complicate the Nova Southeastern University Libraries by splitting the main library website into two. The Alvin Sherman Library, Research, and Information Technology Center is a unique joint-use facility that serves not only the academic community but the public of Broward County - and marketing a hyperblend of content through one portal just wasn't cutting it. With a web team of two, we knew that managing all this rehashed, disparate content was totally unsustainable.&lt;br /&gt;
&lt;br /&gt;
I want to share in this talk how I went about making our library content DRY (“don’t repeat yourself”): input content in one place--blurbs, policies, featured events, featured databases, book reviews, business hours, and so on.--and syndicate it everywhere - even, sometimes, dynamically target that content for specific audiences or context. It is a presentation that is a little about workflow, a little more about browser and context detection, a tangent about content-modeling the CMS, and a lot about APIs, syndication, and performance.&lt;br /&gt;
&lt;br /&gt;
== No code, no root, no problem? Adventures in SaaS and library discovery ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:erwhite@vcu.edu Erin White, VCU]&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
In 2012 VCU was an eager early adopter of Ex Libris' cloud service Alma as an ILS, ERM, link resolver, and single-stop, de-silo'd public-facing discovery tool. This has been a disruptive change that has shifted our systems staff's day-to-day work, relationships with others in the library, and relationships with vendors.&lt;br /&gt;
&lt;br /&gt;
I'll share some of our experiences and takeaways from implementing and maintaining a cloud service:&lt;br /&gt;
* Seeking disruption and finding it&lt;br /&gt;
* Changing expectations of service and the reality of unplanned downtime&lt;br /&gt;
* Communication and problem resolution with non-IT library staff&lt;br /&gt;
* Working with a vendor that uses agile development methodology&lt;br /&gt;
* Benefits and pitfalls of creating customizations and code workarounds&lt;br /&gt;
* Changes in library IT/coders' roles with SaaS&lt;br /&gt;
&lt;br /&gt;
...as well as thoughts on the philosophy of library discovery vs real-life experiences in moving to a single-search model.&lt;br /&gt;
&lt;br /&gt;
== Building for others (and ourselves):  the Avalon Media System ==&lt;br /&gt;
* [mailto:michael.klein@northwestern.edu Michael B Klein], Senior Software Developer, Northwestern University &lt;br /&gt;
** [http://code4lib.org/conference/2010/metz_klein Public Datasets in the Cloud] (code4lib 2010)&lt;br /&gt;
** [http://code4lib.org/conference/2013/klein-rogers The Avalon Media System: A Next Generation Hydra Head For Audio and Video Delivery] (code4lib 2013)&lt;br /&gt;
* [mailto:j-rudder@northwestern.edu Julie Rudder], Digital Initiatives Project Manager, Northwestern University&lt;br /&gt;
** no previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
[http://www.avalonmediasystem.org/ Avalon Media System] is a collaborative effort between development teams at Northwestern and Indiana Universities. Our goal is to produce an open source media management platform that works well for us, but is also widely adopted and contributed to by other institutions. We believe that building a strong user and contributor community is vital to the success and longevity of the project, and have developed the system with this goal in mind. We will share lessons learned, pains and successes we’ve had releasing two versions of the application since last year.  &lt;br /&gt;
&lt;br /&gt;
Our presentation will cover our experiences:&lt;br /&gt;
* providing flexible, admin-friendly distribution and installation options&lt;br /&gt;
* building with abstraction, customization and local integrations in mind&lt;br /&gt;
* prioritizing features (user stories)&lt;br /&gt;
* attracting code contributions from other institutions&lt;br /&gt;
* gathering community feedback &lt;br /&gt;
* creating a product rather than a bag of parts&lt;br /&gt;
&lt;br /&gt;
== How to check your data to provide a great data product? Data quality as a key product feature at Europeana ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:Peter.Kiraly@kb.nl Péter Király] portal backend developer, Europeana&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
[http://Europeana.eu/ Europeana.eu] - Europe's digital library, archive and museum - aggregates more than 30 million metadata records from more than 2200 institutions.  The records come from libraries, archives, museums and every other kind of cultural institution, from very different systems and metadata schemas, and are typically transformed several times until they are ingested into the Europeana data repository.  Europeana builds a consolidated database from these records, creating reliable and consistent services for end-users (a search portal, search widget, mobile apps, thematic sites etc.) and an API, which supports our strategic goeal of data for reuse in education, creative industries, and the cultural sector.  A reliable &amp;quot;data product&amp;quot; is thus at the core of our own software products, as well as those of our API partners.&lt;br /&gt;
&lt;br /&gt;
Much effort is needed to smooth out local differences in the metadata curation practice of our data providers. We need a solid framework to measure the consistency of our data and provide feedback to decision-makers inside and outside the organisation. We can also use this metrics framework to ask content providers to improve their own metadata. Of course, a data-quality-driven approach requires that we also improve the data transformation steps of the Europeana ingestion process itself. Data quality issues heavily define what new features we are able to create in our user interfaces and API, and might actually affect the design and implementation of our underlying data structure, the Europeana Data Model.&lt;br /&gt;
&lt;br /&gt;
In the presentation I briefly describe the Europeana metadata ingestion process, show the data quality metrics, the measuring techniques (using the Europeana API, Solr and MongoDB queries), some typical problems (both trivial and difficult ones), and finally the feedback mechanism we propose to deploy.&lt;br /&gt;
&lt;br /&gt;
Keywords: Europeana, data quality, EDM, API, Apache Solr, MongoDB, #opendata, #openglam&lt;br /&gt;
&lt;br /&gt;
== Teach your Fedora to Fly: scaling out a digital repository ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:acoburn@amherst.edu Aaron Coburn], Software Developer, Amherst College&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Fedora is a great repository system for managing large collections of digital objects, but what happens when a popular food magazine begins directing a large number of readers to a manuscript showing Emily Dickinson’s own recipe for doughnuts? While Fedora excels in its support of XML-based metadata, it doesn’t always perform well under a high volume of traffic. Nor is it especially tolerant of network or hardware failures.&lt;br /&gt;
&lt;br /&gt;
This presentation will show how we are making heavy use of a fedora repository while at the same time insulating it almost entirely from any web traffic. Starting with a distributed web front-end built with Node.js, and caching most of the user-accessible content from Fedora in an elastic, fault-tolerant Riak (NoSQL) cluster, we have eliminated nearly all single points of failure in the system. It also means that our production system is spread across twelve separate servers, where asynchrony and Map-Reduce are king. And aside from being blazing fast, it is also entirely Hydra-compliant.&lt;br /&gt;
&lt;br /&gt;
Furthermore, we will attempt to answer the question: if fedora crashes and the visitors to your site don’t notice, did it really fail?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using Open Source Software and Freeware to Preserve and Deliver Digital Videos ==&lt;br /&gt;
* [mailto:wfang@kinoy.rutgers.edu Wei Fang], Head of Digital Services, Rutgers University Law Library&lt;br /&gt;
* Jiebei Luo, Digital Projects Initiative Intern, Rutgers University&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The Rutgers University Law Library is the official digital repository of the New Jersey Supreme Court oral arguments since 2002. This large video collection contains approximately 3,000 videos with a total of 400 GB or 6,000 viewing hours. With the expansion of this collection, the existing database and the static website could not efficiently support the library’s daily operations and meet its patrons’ search needs. &lt;br /&gt;
By utilizing open source software and freeware such as Ubuntu, FFmpeg, Solr and Drupal, the library is able to develop a complete solution to re-encoding videos, embedding subtitles, incorporating  Solr search engine and content management system to support full-text subtitle search, automatically updating video metadata records in the library catalog system and eventually providing a plug-in free HTML 5-based Web interface for patrons to view the videos online.&lt;br /&gt;
The aspects below will be presented in detail at the conference:&lt;br /&gt;
*	Video codecs comparison &lt;br /&gt;
*	Server-end batch video encoding/re-encoding&lt;br /&gt;
*	HTML 5 video tag and embedding subtitles&lt;br /&gt;
*	Incorporating search engine Solr and content management tool 	Drupal with the database to retrieve videos by full-text search especially in subtitle files&lt;br /&gt;
*	Incorporating video metadata with the library catalog system&lt;br /&gt;
&lt;br /&gt;
== Shared Vision, Shared Resources: the Curate Institutional Repository ==&lt;br /&gt;
* Dan Brubaker Horst, University of Notre Dame &lt;br /&gt;
** [http://code4lib.org/conference/2011/JohnsonHorst A Community-Based Approach to Developing a Digital Exhibit at Notre Dame Using the Hydra Framework] &lt;br /&gt;
* Julie Rudder, Northwestern University&lt;br /&gt;
** no previous presentations&lt;br /&gt;
&lt;br /&gt;
Curate is being collaboratively developed by several institutions in the Hydra community who share the need and vision for a Fedora-backed Institutional Repository. The first release of Curate was a collaboration between Notre Dame and Northwestern University, along with Digital Curation Experts (DCE) - a vendor hired jointly by our two institutions. Powered by the Hydra engine Sufia, the team worked quickly to release the first version of Curate in October 2013 which provides a basic self-deposit system that has support for various content types, collection building, DOI minting, and user profile creation. From the very beginning we have built Curate to be easy to theme and extend in order to ease the process of installation and use by other institutions.&lt;br /&gt;
&lt;br /&gt;
In December 2013, additional partners will join the project including: Indiana University, the University of Cincinnati and the University of Virginia. Each institution contributes resources to the project in order to further our common goal to create a product that fits our needs and has a sustainable future.Together we will tackle additional content types (like complex data, software, media), administrative collections and more. &lt;br /&gt;
&lt;br /&gt;
Our presentation will include:&lt;br /&gt;
* a brief demonstration of Curate and technical overview&lt;br /&gt;
* why and how we work together&lt;br /&gt;
* why build Curate&lt;br /&gt;
* the future of the project&lt;br /&gt;
&lt;br /&gt;
== Solr, Cloud and Blacklight ==&lt;br /&gt;
* David Jiao, Library Information Systems, Indiana University at Bloomington, djiao@indiana.edu&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
SolrCloud refers to the distributed capabilities in Solr4. It is designed to offer a highly available, fault tolerant environment by organizing data into multiple pieces that can be hosted on multiple machines with replicas, and providing a centralized cluster configuration and management. &lt;br /&gt;
&lt;br /&gt;
At Indiana University, we are upgrading our Solr backend for our recently released Blacklight-based OPAC system from Solr 1.4 to Solr4, and we also put up efforts to build a private cloud of Solr4 servers. In this talk, I will persent certain features of SolrCloud, including distributed requests, fault tolerance, near real time indexing/searching, and configuration management with Zookeeper, and our experiences of utilizing these features to provide better performance and architecture for our OPAC system, which serves over 7 million bibliographic records to over 100 thousand students and faculty members. I will also discuss some practical lessons learned from our SolrCloud setup/upgrade and the integration of the new SolrCloud to our customized Blacklight system.  &lt;br /&gt;
&lt;br /&gt;
== Leveraging XSD's for Reflective, Live Dataset Support in Institutional Repositories ==&lt;br /&gt;
* [mailto:msulliva@ufl.edu Mark Sullivan], Library Information Technology, University of Florida&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
The University of Florida Libraries are currently adding support for active datasets into our METS-based institutional repository software.  This ongoing project enables the library to be a partner in current, or long-running, data-driven projects around the university by providing tangible short-term and long-term benefits to the projects.  The system assists project teams by storing and providing access to their data, while supporting online filtering and sorting of the data, custom queries, and adding and editing of the data by authorized users.  We are also exploring simple data visualizations to allow users to perform basic graphical and geographic queries.  Several different schemas were explored including DDI and EML, but ultimately the streamlined approach of using XSD's with some custom attributes was chosen, with all other data residing in the METS file portions.  Currently the system is being developed using XSD's describing XML datasets, but this model should easily scale to support SQL datasets or large datasets supported by Hadoop or iRODS.&lt;br /&gt;
&lt;br /&gt;
This work is being integrated in the open source [http://sobek.ufl.edu SobekCM Digital Content Management System] which is built on a pair-tree structure of METS resources with [http://ufdc.ufl.edu/design/webcontent/sobekcm/SobekCM_Resource_Object.pdf rich metadata support] including DC, MODS, MARC, VRACore, DarwinCore, IEE-LOM, GML/KML, schema.org microdata, and many other standard schemas.  The system has emphasized online, distributed creation and maintenance of resources including geo-placement and geographic searching of resources, building structure maps (table of contents) visually online, and a broad suite of curator tools.  &lt;br /&gt;
&lt;br /&gt;
This work is presented as a model which could be implemented in other systems as well.  We will demonstrate current support and discuss our upcoming roadmap to provide complete support.&lt;br /&gt;
&lt;br /&gt;
== Dead-simple Video Content Management: Let Your Filesystem Do The Work ==&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, NCSU Libraries (akorphan (at) ncsu.edu)&lt;br /&gt;
** (never led or soloed a C4L presentation)&lt;br /&gt;
&lt;br /&gt;
Content management is hard. To keep all the moving parts in order, and to maintain a layer of separation between the system and content creators (who are frequently not technical experts), we typically turn to content management systems like Drupal. But even Drupal and its kin require significant overhead and present a not inconsiderable learning curve for nontechnical users.&lt;br /&gt;
&lt;br /&gt;
In some contexts it's possible -- and desirable -- to manage content in a more streamlined, lightweight way, with a minimum of fuss and technical infrastructure. In this presentation I'll share a simple MVC-like architecture for managing video content for playback on the web, which uses a combination of Apache's mod_rewrite module and your server's filesystem structure to provide an automated approach to video content management that's easy to implement and provides a low barrier to content updates: friendly to content creators and technology implementors alike. Even better, the basic method is HTML5-friendly, and can be integrated into your favorite content management system if you've got permissions for creating templates.&lt;br /&gt;
&lt;br /&gt;
In the presentation I'll go into detail about the system structure and logic required to implement this approach. I'll detail the benefits and limitations of the system, as well as the challenges I encountered in developing its implementation. Audience members should come away with sufficient background to implement a similar system on their own servers. Implementation documentation and genericized code will also be shared, as available.&lt;br /&gt;
&lt;br /&gt;
== Managing Discovery ==&lt;br /&gt;
&lt;br /&gt;
* Andrew Pasterfield, Senior Programmer/Systems Analyst, University of Calgary Library, ampaster@ucalgary.ca&lt;br /&gt;
**No previous code4lib presentations &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
In fall 2012 the University of Calgary Library launched a new home page that incorporated a Summon powered&lt;br /&gt;
Single Search Box with customized “bento box” results display. Search at the U of C now combines a range of&lt;br /&gt;
metadata sources for discovery and customized mapping of a database recommender and LibGuide into a unified&lt;br /&gt;
display.  Further customizations include a non Google Analytics/non proxy method to log clicks.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss the technical details of bringing the various systems together into one display interface to increase discovery at the U of C Library.&lt;br /&gt;
&lt;br /&gt;
http://library.ucalgary.ca&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sorting it out: a piece of the User Centered Design Process ==&lt;br /&gt;
&lt;br /&gt;
* Cindy Beggs, [http://www.akendi.com/aboutus/management/ Akendi], cindy@akendi.com&lt;br /&gt;
&lt;br /&gt;
This talk is about how to apply a user centered design methodology to the process of creating an information architecture.  Participants learn the fundamentals of UCD and how card sorting and reverse card sorting enable us to isolate the content we present on screen from the layouts and visuals of those screens.  We talk about ways to identify who will be using the information architecture you are creating and why we need to know how it will be used.&lt;br /&gt;
 &lt;br /&gt;
What will attendees takes away from your talk?&lt;br /&gt;
The criticality of involving “real” end users in the process of creating an information architecture.  The basics of following a user-centered-design process in the creation of best in class, content-rich, digital products.&lt;br /&gt;
&lt;br /&gt;
Cindy Beggs has been working in the “information industry” for over 25 years.  A librarian by profession, she has spent decades helping users figure out how to find their way through large bodies of content.  Her insights into how people seek information, her empathy for those who find it a challenge and her practical experience helping organizations figure out how to best structure their content contribute to her success as an information architect with both clients and trainees.  (http://www.akendi.com/aboutus/management/)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Implementation of ArchivesSpace in University of Richmond==&lt;br /&gt;
&lt;br /&gt;
*Birong Ho, bho@richmond.edu&lt;br /&gt;
&lt;br /&gt;
University of Richmond implemented its archive collection management ArchivsSpace in the fall, 2013. As a charter member and the Head of Special Collection as the Board member, implementation of such an Open Source Software became a priority. &lt;br /&gt;
&lt;br /&gt;
Several aspects of implementation will be addressed in the talk. Among them, they are Collections and Repository, storage layer including data format, System resources requirements, Technical architecture, Customization, scaling and integrated with other systems in the library.&lt;br /&gt;
&lt;br /&gt;
The customization, scale, and integration with other systems such as Archeon and Exist on campus became a concern will be focused and elaborated in the talk.&lt;br /&gt;
&lt;br /&gt;
==Easy Wins for Modern Web Technologies in Libraries==&lt;br /&gt;
&lt;br /&gt;
*[mailto:trey.terrell@oregonstate.edu Trey Terrell], Analyst Programmer, Oregon State University&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Oregon State University is currently implementing an updated version of its room reservation system. In its development we've come across and implemented a variety of &amp;quot;easy wins&amp;quot; to make it more responsive, easier to maintain, less expensive to run, and just cooler to experience. While our particular system was in Ruby on Rails, this talk will address general methods and example utilities which can be used no matter your stack.&lt;br /&gt;
&lt;br /&gt;
I'll be talking about things like cache management, reverse proxies, publish/subscribe servers, WebSockets, responsive design, asynchronous processing, and keeping complicated stacks up and running with minimal effort.&lt;br /&gt;
&lt;br /&gt;
==Implementing Islandora at a Small Institution==&lt;br /&gt;
&lt;br /&gt;
*Megan Kudzia, Albion College Library&lt;br /&gt;
*Eddie Bachle, Albion College IT&lt;br /&gt;
**No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
Albion College (and particularly the Library/Archives and Special Collections) has a variety of needs which could be met by an open-source Institutional Repository system. Several months and lots of conversations later, we’re continuing to troubleshoot our way through Islandora. We’d like to talk about what has worked for us, where our frustrations have been, whether it’s even possible to install and develop a system like this at a small institution, and where the process has stalled. &lt;br /&gt;
&lt;br /&gt;
As of right now, we do have a semi-working installation. We’re not sure when it will be ready for our end users, but we'll talk about our development process and evaluate our progress.&lt;br /&gt;
''Contributions also by Nicole Smeltekop, Albion College Archives &amp;amp; Special Collections''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== PhantomJS+Selenium: Easy Automated Testing of AJAX-y UIs ==&lt;br /&gt;
&lt;br /&gt;
* Martin Haye, California Digital Library, martin.haye@ucop.edu&lt;br /&gt;
** Previous Code4Lib Presentation: [http://code4lib.org/conference/2012/collett Beyond code: Versioning data with Git and Mercurial] at Code4Lib 2012 (Martin co-presenting with Stephanie Collett)&lt;br /&gt;
* Mark Redar, California Digital Library, mark.redar@ucop.edu&lt;br /&gt;
&lt;br /&gt;
Web user interfaces are demanding ever-more dynamism and polish, combining HTML5, AJAX, lots of CSS and jQuery (or ilk) to create autocomplete drop-downs, intelligent buttons, stylish alert dialogs, etc. How can you make automated tests for these highly complex and interactive UIs?&lt;br /&gt;
&lt;br /&gt;
Part of the answer is PhantomJS. It’s a modern WebKit browser that’s “headless” (meaning it has no display) that can be driven from command-line Selenium unit tests. PhantomJS is dead simple to install, and its blazing speed and server-friendliness make continuous integration testing easy. You can write UI unit tests in {language-of-your-choice} and run them not just in PhantomJS but in Firefox and Chrome, plus a zillion browser/OS combinations at places like SauceLabs, TestingBot and BrowserStack.&lt;br /&gt;
&lt;br /&gt;
In this double-team live code talk, we’ll explain all that while we demonstrate the following in real time:&lt;br /&gt;
&lt;br /&gt;
* Start with nothing.&lt;br /&gt;
* Install Selenium bindings for Ruby and Python.&lt;br /&gt;
* In each language write a small test of an AJAX-y UI.&lt;br /&gt;
* Run the tests in Firefox, and fix bugs (in the test or UI) as needed.&lt;br /&gt;
* Install PhantomJS.&lt;br /&gt;
* Show the same tests running headless as part of a server-friendly test suite. &lt;br /&gt;
* (Wifi permitting) Show the same tests running on a couple different browser/OS combinations on the server cloud at SauceLabs – talking through a tunnel to the local firewalled application.&lt;br /&gt;
&lt;br /&gt;
==New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library==&lt;br /&gt;
&lt;br /&gt;
* Stephanie Walker – swalker@brooklyn.cuny.edu&lt;br /&gt;
* Howard Spivak – howards@brooklyn.cuny.edu&lt;br /&gt;
* Alex - Alex@brooklyn.cuny.edu&lt;br /&gt;
&lt;br /&gt;
Academic libraries are caught in budget squeezes and often struggle to find ways to communicate value to senior administration and others.  At Brooklyn College Library, we have taken an unusual, possibly unique, approach to these issues.  Our technology staff have long worked directly with librarians to develop products that meet library, faculty, and student needs, and we have shared many of our products with colleagues, including an award-winning website, e-resource, and content management system we call 4MyLibrary, which we shared for free with 8 CUNY colleges, and also an easy-to-use book scanner, which has proven overwhelming popular with students, faculty, other librarians, and numerous campus offices.  Recently, motivated by budget cuts, we decided that what worked for us might interest other libraries, and working with our Office of Technology Commercialization, we started selling 2 products:  our book scanners (at half the price of commercial alternatives), and a hosting service, whereby we could host and support 4MyLibrary for libraries with minimal technology staff.  Both succeeded, and yielded major benefits:  a steady revenue stream and the admiration and serious goodwill of our senior administration and others.   However, this presentation is neither a basic how-to, nor an advertisement.  With this presentation, we hope to spur a conversation for broader collaboration, especially regarding new technologies, among libraries.  We all have some level of technical expertise, most of us are struggling with rising prices and tight budgets, and many of us are unhappy with various technology products we use, from scanners to our ILS.  We believe – and can demonstrate – that with collaboration, we can solve many of our problems, and provide better services to boot. &lt;br /&gt;
&lt;br /&gt;
== Identifiers, Data, and Norse Gods ==&lt;br /&gt;
&lt;br /&gt;
* Ryan Scherle, Dryad Digital Repository, ryan@datadryad.org&lt;br /&gt;
** previous Code4Lib talk [http://ryan.scherle.org/papers/2010-2-code4lib-HIVE.ppt  HIVE: A New Tool for Working With Vocabularies], at Code4Lib 2011.&lt;br /&gt;
&lt;br /&gt;
ORCID and DataCite provide stable identifiers for researchers and and data, respectively. Each system does a fine job of providing value to its users. But wouldn't it be great if they could link their systems to create something much more powerful? Perhaps even as powerful as a god?&lt;br /&gt;
&lt;br /&gt;
Enter [http://odin-project.eu/ ODIN], The ORCID and DataCite Interoperability Network. ODIN is a two-year project to unleash the power of persistent identifiers for researchers and the research they create. This talk will present recent work from the ODIN project, including several tools that can be used to unleash the godlike power of identifiers at your institution. Current tools include:&lt;br /&gt;
* Metadata generator tool: allows repository staff to create DataCite metadata with embedded ORCIDs.&lt;br /&gt;
* Claiming tool: assists researchers in claiming their work within the ORCID system.  &lt;br /&gt;
* ORCID-feed: includes a list of ORCID works on any web page.&lt;br /&gt;
* ODIN's HAMR: assists in populating a DSpace repository with ORCIDs. Based on work from a Code4Lib hackathon!&lt;br /&gt;
&lt;br /&gt;
== Armed Bandits in the Digital Library ==&lt;br /&gt;
&lt;br /&gt;
* Roman Chyla, [http://labs.adsabs.harvard.edu/adsabs/ Astrophysics Data System], rchyla@cfa.harvard.edu&lt;br /&gt;
** Previous Code4Lib: [http://code4lib.org/conference/2013/chyla Citation search in SOLR and second-order operators]&lt;br /&gt;
&lt;br /&gt;
Many of us are using the excellent Lucene library (or SOLR appliance) to provide search functionality. These systems contain number of features to adjust relevancy ranking of hits, but we may not know how to use them. In this presentation, I'll present the available options - eg. what is the default ranking 'Vector space model, what are the alternatives (eg. BM25) and what are the other options we have to tweak and adjust the ranking of the hits (eg. boost factors, functions). But even if we know how to deploy these adjustments and tweaks, we are still left in dark. We do not know whether the change we've just rolled out had a significant (statistically significant) effect or maybe it was just a waste of time and resources? A/B testing is one option, but there may be a much better one - so called &amp;quot;Multi-Armed Bandits Approach&amp;quot;. And in this talk I'd like to show how we are experimenting with this strategy to adjust [http://labs.adsabs.harvard.edu/adsabs/ ADS search engine].&lt;br /&gt;
&lt;br /&gt;
== Building Worker Queues with AWS and Resque ==&lt;br /&gt;
&lt;br /&gt;
* Eric Rochester [http://scholarslab.org Scholars' Lab], erochest@virginia.edu&lt;br /&gt;
* Scott Turnbull [http://aptrust.org/ Academic Preservation Trust], scott.turnbull@aptrust.org &lt;br /&gt;
&lt;br /&gt;
A common task in larger systems is to be able to process large input files automatically. Often users can drop those files into a shared directory on AWS or on NFS or another shared drive. Those files need to be processed and potentially integrated into a system. This task has come up recently in the University of Virginia libraries in allowing users to add GIS data to the system and in setting up a system for the Academic Preservation Trust (http://aptrust.org/) that ingests files and resources into the preservation system.&lt;br /&gt;
&lt;br /&gt;
This system is built by loosely coupling a number of different technologies. This allows us to easily interoperate and communicate between different system and programming environments. Because the interfaces are well defined, it’s also fairly simple to switch out technologies as the requirements of the system change.&lt;br /&gt;
&lt;br /&gt;
The process is fairly simple:&lt;br /&gt;
&lt;br /&gt;
First, a Ruby daemon monitors an AWS S3 bucket that others can upload new files into. This daemon creates a Resque status task, adds a marker for the task in a database, and continues monitoring.&lt;br /&gt;
&lt;br /&gt;
Second, Resque mediates incoming job requests and routes them to the appropriate workers which may be in Java, Go, or Ruby.  The diversity of technologies that Resque can manage allows great latitude to leverage the appropriate tool for a specific job.  While processing, it updates the status for that job and coordinates processing with other jobs.&lt;br /&gt;
&lt;br /&gt;
Finally, a page that is integrated into a larger Rails app provides a novice-user-friendly view of the status of the workers and allows basic tasks such as restarting the job.&lt;br /&gt;
&lt;br /&gt;
This architecture allows us to swap in the technology that best fits each part of the process, and it makes it easier to maintain the system. We use this to integrate and coordinate between tasks handled in Java, Ruby, and Go, and it provides an effective way to interoperate with these programming languages and the respective strengths that they bring to this system.&lt;br /&gt;
&lt;br /&gt;
==Piwik: Open source web analytics==&lt;br /&gt;
* Kirk Hess, University of Illinois at Urbana-Champaign (kirkhess@illinois.edu)&lt;br /&gt;
** (Code4Lib 2012: [http://code4lib.org/conference/2012/hess| Discovering Digital Library User Behavior with Google Analytics])&lt;br /&gt;
&lt;br /&gt;
While Google Analytics is synonymous with Web Analytics, fortunately today we have many other good options, and one option is Piwik, [http://piwik.org| piwik.org] a simple to install, open-source PHP/MySQL application with a tracking script that will sit alongside Google Analytics tracking the usual clicks, events and variables. In this presentation, I'd like to cover the usual analytics topics and also cover what makes Piwik powerful, such as importing and visualizing web logs from any system to incorporate both past and future data, easily tracking downloads, and the ability to write your own reports or dashboard. The visitor log data is stored securely on your own server so you have control over who looks at the data and how much or how little to keep. With an active and helpful developer community, Piwik has the potential for analytics which makes sense for libraries, not e-commerce.&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39850</id>
		<title>2014 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39850"/>
				<updated>2013-11-08T20:06:27Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Identifiers, Data, and Norse Gods */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
 &lt;br /&gt;
* ''Projects'' you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* ''Tools and technologies'' – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* ''Technical issues'' - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* ''Relevant non-technical issues'' – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
'''To Propose a Talk'''&lt;br /&gt;
* Log in to the wiki in order to submit a proposal. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
* Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
* If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist us in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
As in past years, the Code4Lib community will vote on proposals that they would like to see included in the program. This year, however, only the top 10 proposals will be guaranteed a slot at the conference. Additional presentations will be selected by the Program Committee in an effort to ensure diversity in program content. Community votes will, of course, still weigh heavily in these decisions.&lt;br /&gt;
&lt;br /&gt;
Presenters whose proposals are selected for inclusion in the program will be guaranteed an opportunity to register for the conference. The standard conference registration fee will still apply.&lt;br /&gt;
&lt;br /&gt;
''Proposals can be submitted through '''Friday, November 8, 2013, at 5pm PST'''''. Voting will commence on November 18, 2013 and continue through December 6, 2013. The final line-up of presentations will be announced in early January, 2014.&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
&lt;br /&gt;
==Creating a new Greek-Dutch dictionary==&lt;br /&gt;
* Caspar Treijtel, University of Amsterdam, c.treijtel@uva.nl&lt;br /&gt;
&lt;br /&gt;
At present, no complete dictionary of (ancient) Greek-Dutch is available online. A new dictionary is currently under construction at Leiden University, with software being developed at the University of Amsterdam. The team in Leiden has already begun preparation of the data, with at this moment about 6,000 approved lemmas. The ultimate goal is to produce both a print version and online open access version from the same source documents. The software needed for this has been made in a project that was funded by CLARIN-NL.&lt;br /&gt;
&lt;br /&gt;
Migrator&lt;br /&gt;
&lt;br /&gt;
For the production of lemmas we have implemented an advanced workflow. The (generally non-technical) users create lemmas using MS Word, which is both familiar and easy to use. We have developed a custom software module that carefully migrates the Word documents into deeply structured XML by analyzing the structure and semantics of the lemmas, and falling back on heuristics in ambiguous cases. While having initially envisioned the oXygen XML Author component as the main tool for creating new lemmas, we obtained excellent results with the migrator module, and decided therefore to continue using MS Word as the primary composition tool. The main advantage of this is that the editors are much more familiar with Word than with any other WYSIWYG editor. Lemmas that have been migrated to XML are stored in an XML database and can be further edited using oXygen XML Author.&lt;br /&gt;
&lt;br /&gt;
Lemmatizer&lt;br /&gt;
&lt;br /&gt;
Greek morphology is complicated. In order to use a dictionary effectively, a rather high level of initial language competence is necessary for the user to be able to relate the word form s/he finds in a text to the correct basic lemma form, where the definition of the word can be found. Using a Greek morphological database we have been able to facilitate the search for lemmas. A ‘lemmatizer’ module gives the possible parsings of the word forms and the lemmas they can be derived from. This enables the user to type in the word as found in the text and be redirected to the correct lemma.&lt;br /&gt;
&lt;br /&gt;
Visualization&lt;br /&gt;
&lt;br /&gt;
For the online dictionary we have implemented a visualization module that allows the user to view multiple lemmas at once. The implementation of this module has been done using the Javascript framework MooTools. The result is a viewer that performs really well and is run by maintainable Javascript code.&lt;br /&gt;
&lt;br /&gt;
The online dictionary is still being worked on, have a look at http://www.woordenboekgrieks.nl/ for the beta version. A newer test version with additional features can be found here: http://angel.ic.uva.nl:8600/.&lt;br /&gt;
&lt;br /&gt;
Credits&lt;br /&gt;
&lt;br /&gt;
* construction of the dictionary: Prof. Ineke Sluiter, Classics department of Leiden University; Prof. Albert Rijksbaron, University of Amsterdam&lt;br /&gt;
* publisher of the dictionary: Amsterdam University Press&lt;br /&gt;
* design/typesetting dictionary: TaT Zetwerk (http://www.tatzetwerk.nl/)&lt;br /&gt;
* software development: Digital Production Center, University Library, University of Amsterdam&lt;br /&gt;
* project funding: CLARIN-NL (http://www.clarin.nl/)&lt;br /&gt;
* morphological database for use by the lemmatizer: courtesy of Prof. Helma Dik, University of Chicago (based on data of the Perseus Project)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Using Drupal to drive alternative presentation systems ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Recently, we have been building systems that use angular.js, Rails, or other systems for presentation, while leveraging Drupal's sophisticated content management capabilities on the back end.&lt;br /&gt;
&lt;br /&gt;
So far, these have been one-way systems, but as we move to Drupal 8 we are beginning to explore ways to further decouple the presentation and CMS functions.&lt;br /&gt;
&lt;br /&gt;
== A Book, a Web Browser and a Tablet: How Bibliotheca Alexandrina's Book Viewer Framework Makes It Possible ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mohammed.abuouda|Mohammed Abu ouda]], Bibliotheca Alexandrina (The new Library of Alexandria)&lt;br /&gt;
&lt;br /&gt;
A lot of institutions around the world are engaged in multiple digitization projects aiming at preserving the human knowledge present in books and availing them through multiple channels to people around the whole globe. These efforts will sure help close the digital gap particularly with the arrival of affordable e-readers, mobile phones and network coverage. However, the digital reading experience has not yet arrived to its maximum potential. Many readers miss features they like in their good old books and wish to find them in their digital counterpart. In an attempt to create a unique digital reading experience, Bibliotheca Alexandria (BA) created a flexible book viewing framework that is currently used to access its current collection of more than 300,000 digital books in five different languages which includes the largest collection of digitized Arabic books.&lt;br /&gt;
&lt;br /&gt;
Using open source tools, BA used the framework to develop a modular book viewer that can be deployed in different environments and is currently at the heart of various BA projects. The Book viewer provides several features creating a more natural reading experience. As with physical books, the reader can now personalize the books he reads by adding annotations like highlights, underlines and sticky notes to capture his thoughts and ideas in addition to being able to share the book with friends on social networks. The reader can perform a search across the content of the book receiving highlighted search results within the pages of the book. More features can be further added to the book viewer through its plugin architecture.&lt;br /&gt;
&lt;br /&gt;
== Structured data NOW: seeding schema.org in library systems ==&lt;br /&gt;
 &lt;br /&gt;
* [http://coffeecode.net Dan Scott], Laurentian University&lt;br /&gt;
** Previous code4lib presentations: [https://archive.org/details/code4lib.conf.2008.pres.CouchDBsacrilege CouchDB is sacrilege... mmm, delicious sacrilege] at Code4Lib 2008&lt;br /&gt;
&lt;br /&gt;
The semantic web, linked data, and structured data are all fantastic ideas with a barrier imposed by implementation constraints. If their system does not allow customizations, or the institution lacks skilled human resources, it does not matter how enthused a given library might be about publishing structured data... it will not happen. However, if the software in use simply publishes structured data by default, then the web will be populated for free. Really! No extra resources necessary.&lt;br /&gt;
&lt;br /&gt;
This presentation highlights Dan's work with systems such as Evergreen, Koha, and VuFind to enable the publication of schema.org structured data out-of-the-box. Along the way, we reflect the current state of the W3C Schema.org Bibliographic Extension community group efforts to shape the evolution of the schema.org vocabulary. Finally, hold on tight as we contemplate next steps and the possibilities of a world where structured data is the norm on the web.&lt;br /&gt;
&lt;br /&gt;
== Towards Pasta Code Nirvana: Using JavaScript MVC to Fill Your Programming Ravioli ==&lt;br /&gt;
&lt;br /&gt;
* Bret Davidson, North Carolina State University Libraries, bret_davidson@ncsu.edu&lt;br /&gt;
** Previous Code4Lib Presentations: [http://wiki.code4lib.org/index.php/2013_talks_proposals#Data-Driven_Documents:_Visualizing_library_data_with_D3.js Visualizing library data with D3.js] at Code4Lib 2013&lt;br /&gt;
&lt;br /&gt;
JavaScript MVC frameworks are ushering in a golden age of robust and responsive web applications that take advantage of evergreen browsers, performant JS engines, and the unprecedented reach provided by billions of personal computing devices. The web browser has emerged as the world’s most popular application runtime and the complexity[1] and scope of JavaScript applications has exploded accordingly. Server-side web frameworks like Rails and Django have helped developers adhere to best practices like modularity, dependency injection, and unit testing for years, practices that are now being applied to JavaScript development through projects like Backbone[2], Ember[3], and Angular[4].&lt;br /&gt;
&lt;br /&gt;
This talk will discuss the issues JavaScript MVC frameworks are trying to solve, common features like data binding, implications for the future of web development[5], and the appropriateness of JavaScript MVC for library applications.&lt;br /&gt;
&lt;br /&gt;
*[1]http://en.wikipedia.org/wiki/Spaghetti_code&lt;br /&gt;
*[2]http://backbonejs.org&lt;br /&gt;
*[3]http://emberjs.com&lt;br /&gt;
*[4]http://angularjs.org&lt;br /&gt;
*[5]http://tomdale.net/2013/09/progressive-enhancement-is-dead/&lt;br /&gt;
&lt;br /&gt;
== WebSockets for Real-Time and Interactive Interfaces ==&lt;br /&gt;
&lt;br /&gt;
* [http://ronallo.com Jason Ronallo], NCSU Libraries, jason_ronallo@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/ronallo HTML5 Microdata and Schema.org] 2012&lt;br /&gt;
* [http://code4lib.org/conference/2013/ronallo HTML5 Video Now!] 2013&lt;br /&gt;
&lt;br /&gt;
Watching the Google Analytics Real-Time dashboard for the first time was mesmerizing. As soon as someone visited a site, I could see what page they were on. For a digital collections site with a lot of images, it was fun to see what visitors were looking at. But getting from Google Analytics to the image or other content of what was currently being viewed was cumbersome. The real-time experience was something I wanted share with others. I'll show you how I used a WebSocket service to create a real-time interface to digital collections.&lt;br /&gt;
&lt;br /&gt;
In the Hunt Library at NCSU we have some large video walls. I wanted to make HTML-based exhibits that featured viewer interactions. I'll show you how I converted Listen to Wikipedia [1] into an bring-your-own-device interactive exhibit. With WebSockets any HTML page can be remote controlled by any internet connected device.&lt;br /&gt;
&lt;br /&gt;
I will attempt to include real-time audience participation.&lt;br /&gt;
&lt;br /&gt;
[1] http://listen.hatnote.com/&lt;br /&gt;
&lt;br /&gt;
== Rapid Development of Automated Tasks with the File Analyzer ==&lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Libraries, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Libraries have customized the File Analyzer and Metadata Harvester application (https://github.com/Georgetown-University-Libraries/File-Analyzer) to solve a number of library automation challenges:&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census.  After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform.  Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules.  As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
Learn about the functionality that is available for download, how you can use this tool to automate workflows from digital collections to ILS ingests to electronic resources statistics and also discuss the opportunities to collaborate on enhancements to this application!&lt;br /&gt;
&lt;br /&gt;
== GeoHydra: How to Build a Geospatial Digital Library with Fedora ==&lt;br /&gt;
 &lt;br /&gt;
* [http://stanford.edu/~drh Darren Hardy], Stanford University, drh@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Geographically-rich data are exploding and putting fear in those trying to&lt;br /&gt;
tackle integrating them into existing digital library infrastructures.&lt;br /&gt;
Building a spatial data infrastructure that integrates with your digital&lt;br /&gt;
library infrastructure need not be a daunting task. We have successfully&lt;br /&gt;
deployed a geospatial digital library infrastructure using Fedora and&lt;br /&gt;
open-source geospatial software [1]. We'll discuss the primary design&lt;br /&gt;
decisions and technologies that led to a production deployment within a few&lt;br /&gt;
months. Briefly, our architecture revolves around discovery, delivery, and&lt;br /&gt;
metadata pipelines using open-source OpenGeoPortal [2], Solr [3], GeoServer&lt;br /&gt;
[4], PostGIS [5], and GeoNetwork [6] technologies, plus the proprietary ESRI&lt;br /&gt;
ArcMap [7] -- the GIS industry's workhorse. Finally, we'll discuss the key&lt;br /&gt;
skillsets needed to build and maintain a spatial data infrastructure.&lt;br /&gt;
&lt;br /&gt;
[1] http://foss4g.org&lt;br /&gt;
[2] http://opengeoportal.org&lt;br /&gt;
[3] http://lucene.apache.org/solr&lt;br /&gt;
[4] http://geoserver.org&lt;br /&gt;
[5] http://postgis.net&lt;br /&gt;
[6] http://geonetwork-opensource.org&lt;br /&gt;
[7] http://esri.com&lt;br /&gt;
&lt;br /&gt;
==Under the Hood of Hadoop Processing at OCLC Research ==&lt;br /&gt;
&lt;br /&gt;
[http://roytennant.com/ Roy Tennant]&lt;br /&gt;
&lt;br /&gt;
* Previous Code4Lib presentations: 2006: &amp;quot;The Case for Code4Lib 501c(3)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[http://hadoop.apache.org/ Apache Hadoop] is widely used by Yahoo!, Google, and many others to process massive amounts of data quickly. OCLC Research uses a 40-node compute cluster with Hadoop and HBase to process the 300 million MARC records of WorldCat in various ways. This presentation will explain how Hadoop MapReduce works and illustrate it with specific examples and code. The role of the jobtracker in both monitoring and reporting on processes will be explained. String searching WorldCat will also be demonstrated live.&lt;br /&gt;
&lt;br /&gt;
== Quick and Easy Data Visualization with Google Visualization API and Google Chart Libraries ==&lt;br /&gt;
 &lt;br /&gt;
[http://bohyunkim.net/blog Bohyun Kim], Florida International University, bohyun.kim@fiu.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Do most of the data that your library collects stay in spreadsheets or are published as a static table with a series of boring numbers? Do your library stakeholders spend more time collecting the data than using it as a decision-making tool because the data is presented in a way that makes it hard for them [http://developers.google.com/chart/interactive/docs/gallery to quickly grasp its significance? ]&lt;br /&gt;
&lt;br /&gt;
This talk will provide an overview of [http://developers.google.com/chart/interactive/docs/reference Google Visualization API] [2] and [http://developers.google.com/chart/ Google Chart Libraries] [3] to get you started on the way to quickly query and visualize your library data from remote data sources (e.g. a Google Spreadsheet or your own database) with (or without) cool-looking user-controls, animation effects, and even a dashboard.&lt;br /&gt;
&lt;br /&gt;
== Leap Motion + Rare Books: A hands-free way to view and interact with rare books in 3D ==&lt;br /&gt;
 &lt;br /&gt;
[http://http://www.youtube.com/user/jpdenzer Juan Denzer], Binghamton University, jdenzer@binghamton.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
As rare books become more delicate over time, making them available to the public becomes harder.  We at Binghamton University Library have developed an application that makes it easier to view rare books without ever having to touch them.  We have combined the Leap Motion hands-free device and 3D rendered models to create a new virtual experience for the viewer.&lt;br /&gt;
&lt;br /&gt;
The application allows the user to rotate and zoom in on a 3D representation of a rare book.  The user is also able to ‘open’ the virtual book and flip through it using a natural user interface.  Such as swiping the hand left or right to turn the page.&lt;br /&gt;
&lt;br /&gt;
The application is built on the .Net framework and is written in C#.  3D models are created using simple 3D software such as sketchup or Blender.  Scans of the book cover and spine are created using simple flatbed scanners.  The inside pages are scanned using overhead scanners. &lt;br /&gt;
&lt;br /&gt;
This talk with discuss the technologies used in developing the application and virtually any library could implement the application with virtually no coding at all. This presentation will have a demonstration of the software and also a chance for audience members to experience the Rare Book Leap Motion App themselves.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Course Reserves Unleashed! ==&lt;br /&gt;
 &lt;br /&gt;
* Bobbi Fox, Library Technology Services, Harvard University, bobbi_fox@harvard.edu&lt;br /&gt;
* Gloria Korsman, Andover-Harvard Theological Library&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Hey kids!  Remember when SOAP was used for something other than washing?  Our sophisticated (and highly functional) Course Reserves Request system does!&lt;br /&gt;
&lt;br /&gt;
However, while the system is great for submitting and  processing course reserve requests, the student-facing presentation through Havard’s home-grown -- and soon to be replaced -- LMS leaves a lot to be desired.  &lt;br /&gt;
&lt;br /&gt;
Follow along as we leverage Solr 4 as a No-SQL database, along with more progressive RESTful API techniques, to release Reserves data into the wild without interfering with reserves request processing -- and, in the process, open up the opportunity for other schools to feed their data in as well.&lt;br /&gt;
&lt;br /&gt;
== We Are All Disabled! Universal Web Design Making Web Services Accessible for Everyone ==&lt;br /&gt;
 &lt;br /&gt;
Cynthia Ng, Accessibility Librarian, CILS at Langara College&lt;br /&gt;
* No previous Code4Lib presentations (not counting lightning talks)&lt;br /&gt;
&lt;br /&gt;
We’re building and improving tools and services all the time, but do you only develop for the “average” user or add things for “disabled” users? We all use “assistive” technology accessing information in a multitude of ways with different platforms, devices, etc. Let’s focus on providing web services that are accessible to everyone without it being onerous or ugly. The aim is to get you thinking about what you can do to make web-based services and content more accessible for all from the beginning or with small amounts of effort whether you're a developer or not.&lt;br /&gt;
&lt;br /&gt;
The goal of the presentation is to provide both developers and content creators with information on simple, practical ways to make web content and web services more accessible. However, rather than thinking about putting in extra effort or making adjustment for those with disabilities, I want to help people think about how to make their websites more accessible for all users through universal web design.&lt;br /&gt;
&lt;br /&gt;
== Personalize your Google Analytics Data with Custom Events and Variables ==&lt;br /&gt;
&lt;br /&gt;
[http://joshwilson.net Josh Wilson], Systems Integration Librarian, State Library of North Carolina - joshwilsonnc@gmail.com&lt;br /&gt;
&lt;br /&gt;
At the State Library of North Carolina, we had more specific questions about the use of our digital collections than standard GA could provide. A few implementations of custom events and custom variables later, we have our answers.&lt;br /&gt;
&lt;br /&gt;
I'll demonstrate how these analytics add-ons work, and why implementation can sometimes be more complicated than just adding a few lines of JavaScript to your ga.js. I'll discuss some specific examples in use at the SLNC:&lt;br /&gt;
&lt;br /&gt;
* Capturing the content of specific metadata fields in CONTENTdm as Custom Events &lt;br /&gt;
* Recording Drupal taxonomy terms as Custom Variables&lt;br /&gt;
&lt;br /&gt;
In both instances, this data deepened our understanding of how our sites and collections were being used, and in turn, we were able to report usage more accurately to content contributors and other stakeholders.&lt;br /&gt;
&lt;br /&gt;
More on: [https://developers.google.com/analytics/devguides/collection/gajs/eventTrackerGuide GA Custom Events] | [https://developers.google.com/analytics/devguides/collection/gajs/gaTrackingCustomVariables GA Custom Variables]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Behold Fedora 4: The Incredible Shrinking Repository! ==&lt;br /&gt;
&lt;br /&gt;
Esmé Cowles, UC San Diego Library.  Previous talk: [http://code4lib.org/conference/2013/cowles-critchlow-westbrook All Teh Metadatas Re-Revisited] (2013)&lt;br /&gt;
&lt;br /&gt;
* One repository contains untold numbers of digital objects and powers many Hydra and Islandora apps&lt;br /&gt;
* It speaks RDF, but contains no triplestore! (triplestores sold separately, SPARQL Update may be involved, some restrictions apply)&lt;br /&gt;
* Flexible enough to tie itself in knots implementing storage and access control policies&lt;br /&gt;
* Witness feats of strength and scalability, with dramatically increased performance and clustering&lt;br /&gt;
* Plumb the depths of bottomless hierarchies, and marvel at the metadata woven into the very fabric of the repository&lt;br /&gt;
* Ponder the paradox of ingesting large files by not ingesting them&lt;br /&gt;
* Be amazed as Fedora 4 swallows other systems whole (including Fedora 3 repositories)&lt;br /&gt;
* Watch novice developers setup Fedora 4 from scratch, with just a handful of incantations to Git and Maven&lt;br /&gt;
&lt;br /&gt;
The Fedora Commons Repository is the foundation of many digital collections, e-research, digital library, archives, digital preservation, institutional repository and open access publishing systems.  This talk will focus on how Fedora 4 improves core repository functionality, adds new features, maintains backwards compatibility, and addresses the shortcomings of Fedora 3.&lt;br /&gt;
&lt;br /&gt;
== Organic Free-Range API Development - Making Web Services That You Will Actually Want to Consume ==&lt;br /&gt;
&lt;br /&gt;
Steve Meyer and Karen Coombs, OCLC&lt;br /&gt;
&lt;br /&gt;
Building web services can have great benefits by providing reusability of data and functionality. Underpinning your applications with a web service will allow you to write code once and support multiple environments: your library's web app, mobile applications, the embedded widget in your campus portal. However, building a web service is its own kind of artful programming. Doing it well requires attention to many of the same techniques and requirements as building web applications, though with different outcomes. &lt;br /&gt;
&lt;br /&gt;
So what are the usability principles for web services? How do you build a web service that you (and others) will actually want to use? In this talk, we’ll share some of the lessons learned - the good, the bad, and the ugly - through OCLC's work on the WorldCat Metadata API. This web service is a sophisticated API that provides external clients with read and write access to WorldCat data. It provides a model to help aspiring API creators navigate the potential complications of crafting a web service. We'll cover:&lt;br /&gt;
&lt;br /&gt;
* Loose coupling of data assets and resource-oriented data modeling at the core&lt;br /&gt;
* Coding to standards vs. exposure of an internal data model&lt;br /&gt;
* Authentication and security for web services: API Keys, Digital Signing, OAuth Flows&lt;br /&gt;
* Building web services that behave as a suite so it looks like the left hand knows what the right hand is doing&lt;br /&gt;
&lt;br /&gt;
So at the end of the day, your team will know your API is a very good egg after all. &lt;br /&gt;
&lt;br /&gt;
If accepted, the presenters intend to produce and share a Quick Guide for building a web service that will reflect content presented in the talk.&lt;br /&gt;
&lt;br /&gt;
== Lucene's Latest (for Libraries) ==&lt;br /&gt;
&lt;br /&gt;
erik.hatcher@lucidworks.com&lt;br /&gt;
&lt;br /&gt;
Lucene powers the search capabilities of practically all library discovery platforms, by way of Solr, etc.  The Lucene project evolves rapidly, and it's a full-time job to keep up with the ever improving features and scalability.   This talk will distill and showcase the most relevant(!) advancements to date.&lt;br /&gt;
&lt;br /&gt;
== The Why and How of Very Large Displays in Libraries. ==&lt;br /&gt;
&lt;br /&gt;
* Cory Lown, NCSU Libraries, cwlown@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib Presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/lown How People Search the Library from a Single Search Box]  2012&lt;br /&gt;
* [http://code4lib.org/conference/2010/orphanides_lown_lynema Enhancing Discoverability with Virtual Shelf Browse] 2010&lt;br /&gt;
&lt;br /&gt;
Built into the walls of NC State's new Hunt Library are several [http://www.christiedigital.com/en-us/digital-signage/products/microtiles/pages/microtiles-digital-signage-video-wall.aspx Christie MicroTile Display Wall Systems]. What does a library do with a display that's seven feet tall and over twenty feet wide? I'll talk about why libraries might want large displays like this, what we're doing with them right now, and what we might do with them in the future. I'll talk about how these displays factor into planning for new and existing web projects. And I'll get into the fun details of how you build web applications that scale from the very small browser window on a phone all the way up to a browser window with about 14 million pixels (about 10 million more than a dual 24&amp;quot; monitor desktop setup).&lt;br /&gt;
&lt;br /&gt;
== Discovering your Discovery System in Real Time. ==&lt;br /&gt;
&lt;br /&gt;
* Godmar Back, Virginia Tech, gback@vt.edu&lt;br /&gt;
* Annette Bailey, Virginia Tech, afbailey@vt.edu&lt;br /&gt;
&lt;br /&gt;
Practically all libraries today provide web-based discovery systems to their users;&lt;br /&gt;
users discover items and peruse or check them out by clicking on links.  Unlike&lt;br /&gt;
the traditional transaction of checking out a book at the circulation desk, this&lt;br /&gt;
interaction is largely invisible.  We have built a system that records user's&lt;br /&gt;
interaction with Summon in real-time, processes the resulting data with minimal delay,&lt;br /&gt;
and visualizes it in various ways using Google Charts and using various d3.js modules,&lt;br /&gt;
such as word clouds, tree maps, and others.&lt;br /&gt;
&lt;br /&gt;
These visualizations can be embedded in web sites, but are also suitable for&lt;br /&gt;
projection via large-scale displays or projectors right into the 'Learning Spaces'&lt;br /&gt;
many libraries are converted into.  The goal of this talk is to share the technology&lt;br /&gt;
and advocate the building of a cloud-based infrastructure that would make this&lt;br /&gt;
technology available to any library that uses a discovery system, rather than just&lt;br /&gt;
those who have the technological prowess for developing such systems and&lt;br /&gt;
visualizations in-house.  &lt;br /&gt;
&lt;br /&gt;
Previous presentations at Code4Lib:&lt;br /&gt;
* Talk: Code4Lib 2009 [http://code4lib.org/files/LibX2.0-Code4Lib-2009AsPresented.ppt LibX 2.0]&lt;br /&gt;
* Preconference: [http://wiki.code4lib.org/index.php/LibX_Preconference LibX 2.0, 2009]&lt;br /&gt;
* Preconference: Code4Lib 2010, On Widgets and Web Services&lt;br /&gt;
&lt;br /&gt;
== Your Library, Anywhere: A Modern, Responsive Library Catalogue at University of Toronto Libraries ==&lt;br /&gt;
&lt;br /&gt;
* Bilal Khalid, Gordon Belray, Lisa Gayhart (lisa.gayhart@utoronto.ca)&lt;br /&gt;
&lt;br /&gt;
* No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. &lt;br /&gt;
&lt;br /&gt;
[http://search.library.utoronto.ca The result]: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles. This talk will discuss: application development; service improvements; interface design; and user outreach, testing, and project communications. Feedback and questions from the audience are very welcome. If time runs short, we will be available for questions and conversation after the presentation.&lt;br /&gt;
&lt;br /&gt;
Note: A version of this content has been provisionally accepted as an article for Code4Lib Journal, January 2014 publication.)&lt;br /&gt;
&lt;br /&gt;
== All Tiled Up ==&lt;br /&gt;
&lt;br /&gt;
* Mike Graves, MIT Libraries (mgraves@mit.edu)&lt;br /&gt;
&lt;br /&gt;
You've got maps. You even scanned and georeferenced them. Now what? Running a full GIS stack can be expensive, and overkill in some cases. The good news is that you have a lot more options now than you did just a few years ago. I'd like to present some lighter weight solutions to making georeferenced images available on the Web.&lt;br /&gt;
&lt;br /&gt;
This talk will provide an introduction to MBTiles. I'll go over what they are, how you create them, how you use them and why you would use them.&lt;br /&gt;
&lt;br /&gt;
== The Great War: Image Interoperability to Facebook ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, Los Alamos National Laboratory (azaroth42@gmail.com)&lt;br /&gt;
** (Code4Lib 2006: [http://www.code4lib.org/2006/sanderson | Library Text Mining])&lt;br /&gt;
* Rob Warren, Carleton University&lt;br /&gt;
** No previous presentations&lt;br /&gt;
&lt;br /&gt;
Using a pipeline constructed from Linked Open Data and other interoperability specifications, it is possible to merge and re-use image and textual data from distributed library collections to build new, useful tools and applications.  Starting with the OAI-PMH interface to ContentDM, we will take you on a tour through the International Image Interoperability Framework and Shared Canvas, to a cross-institutional viewer, and image analysis for the purposes of building a historical Facebook from finding and tagging people in photographs.  The World War One collections are drawn from multiple institutions and merged by the machine learning code.&lt;br /&gt;
&lt;br /&gt;
The presentation will focus on the (open source) toolchain and the benefits of the use of standards throughout:  OAI-PMH to get the metadata, IIIF for interaction with the images, the Shared Canvas ontology for describing collections of digitized objects, Open Annotation for tagging things in the images and specialized ontologies that are specific to the contents.  The tools include standard RDF / OWL technologies, JSON-LD, imagemagick and OpenCV for image analysis.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Solr Search Results with D3.js for User-Friendly Navigation of Large Results Sets ==&lt;br /&gt;
&lt;br /&gt;
*Julia Bauder, Grinnell College Libraries (bauderj-at-grinnell-dot-edu)&lt;br /&gt;
*No previous presentations at national Code4Lib conferences&lt;br /&gt;
&lt;br /&gt;
As the corpus of articles, books, and other resources searched by discovery systems continues to get bigger, searchers are more and more frequently confronted with unmanageably large numbers of results. How can we help users make sense of 10,000 hits and find the ones they actually want? Facets help, but making sense of a gigantic sidebar of facets is not an easy task for users, either.&lt;br /&gt;
During this talk, I will explain how we will soon be using Solr 4’s pivot queries and hierarchical visualizations (e.g., treemaps) from D3.js to let patrons view and manipulate search results. We will be doing this with our VuFind 2.0 catalog, but this technique will work with any system running Solr 4. I will also talk about early student reaction to our tests of these visualization features.&lt;br /&gt;
&lt;br /&gt;
== PeerLibrary – open source cloud based collaborative library ==&lt;br /&gt;
&lt;br /&gt;
[https://github.com/peerlibrary/peerlibrary PeerLibrary is a new open source project] and a cloud service providing collaborative reading, sharing and storing. Users can upload publications they want to read (currently in PDF format), read them in the browser in real-time with others, highlight, annotate and organize their own or collaborative library. PeerLibrary provides a search engine to search over all uploaded open access publications. Additionally, it aims to collaboratively aggregate the open layer of knowledge on top of this publications through public annotations and references user will add to publications. In this way publications would not just be available to read, but accessible to the general public as well. Currently, it is aiming at scientific community and scientific publications.&lt;br /&gt;
&lt;br /&gt;
See [http://blog.peerlibrary.org/post/63458789185/screencast-previewing-the-peerlibrary-project screencast here].&lt;br /&gt;
&lt;br /&gt;
It is still in development and beta launch is planned at the end of November.&lt;br /&gt;
&lt;br /&gt;
== Who was where when, or finding biographical articles on Wikipedia by place and time ==&lt;br /&gt;
&lt;br /&gt;
* [http://morton-owens.info Emily Morton-Owens], The Seattle Public Library (presenting on work from NYU)&lt;br /&gt;
* No previous c4l presentations&lt;br /&gt;
&lt;br /&gt;
It's easy to answer the question &amp;quot;What important people were in Paris in 1939?&amp;quot; But what about Virginia in the 1750s or Scandinavia in the 14th century? I created a tool that allows you to search for biographies in a generally applicable way, using a map interface. I would like to present updates to my thesis project, which combines a crawler written in Java that extracts information from Wikipedia articles, with a MongoDB data store and a frontend in Python.&lt;br /&gt;
&lt;br /&gt;
The input to the project is freetext of entire articles in Wikipedia; this is important to allow us to pick up Benjamin Franklin not just in the single most obvious place of Philadelphia but also in London, Paris, Boston, etc. I can talk about my experiments disambiguating place names (approaches pioneered on newspaper articles were actually unhelpful on this type of text) and setting up a processing queue that does not become mired in the biographies of every human who ever played soccer. I also want to mitigate some of the implementation choices I made due to my academic deadline and improve the accuracy/usability.&lt;br /&gt;
&lt;br /&gt;
What I hope to show is that I was able to develop a novel and useful reference tool automatically, using fairly simple heuristics that are a far cry from hand-cataloging familiar to many librarians.&lt;br /&gt;
&lt;br /&gt;
You can try out [http://linserv1.cims.nyu.edu:48866/ the original version] (this server is inconveniently set to be updated/rebooted on 11/8--may be temporarily unavailable)&lt;br /&gt;
&lt;br /&gt;
== Good!, DRY, and Dynamic: Content Strategy for Libraries (Especially the Big Ones) ==&lt;br /&gt;
&lt;br /&gt;
*Michael Schofield, Nova Southeastern University Libraries, mschofield@nova.edu&lt;br /&gt;
*No previous code4lib presentations.&lt;br /&gt;
&lt;br /&gt;
The responsibilities of the #libweb are exploding [it’s a good thing] and it is no longer uncommon for libraries to manage or even home-grow multiple applications and sites. Often it is at this point where the web people begin to suffer the absence of a content strategy when, say, business hours need to be updated sitewide a half-dozen times.&lt;br /&gt;
&lt;br /&gt;
We were already feeling this crunch when we decided to further complicate the Nova Southeastern University Libraries by splitting the main library website into two. The Alvin Sherman Library, Research, and Information Technology Center is a unique joint-use facility that serves not only the academic community but the public of Broward County - and marketing a hyperblend of content through one portal just wasn't cutting it. With a web team of two, we knew that managing all this rehashed, disparate content was totally unsustainable.&lt;br /&gt;
&lt;br /&gt;
I want to share in this talk how I went about making our library content DRY (“don’t repeat yourself”): input content in one place--blurbs, policies, featured events, featured databases, book reviews, business hours, and so on.--and syndicate it everywhere - even, sometimes, dynamically target that content for specific audiences or context. It is a presentation that is a little about workflow, a little more about browser and context detection, a tangent about content-modeling the CMS, and a lot about APIs, syndication, and performance.&lt;br /&gt;
&lt;br /&gt;
== No code, no root, no problem? Adventures in SaaS and library discovery ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:erwhite@vcu.edu Erin White, VCU]&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
In 2012 VCU was an eager early adopter of Ex Libris' cloud service Alma as an ILS, ERM, link resolver, and single-stop, de-silo'd public-facing discovery tool. This has been a disruptive change that has shifted our systems staff's day-to-day work, relationships with others in the library, and relationships with vendors.&lt;br /&gt;
&lt;br /&gt;
I'll share some of our experiences and takeaways from implementing and maintaining a cloud service:&lt;br /&gt;
* Seeking disruption and finding it&lt;br /&gt;
* Changing expectations of service and the reality of unplanned downtime&lt;br /&gt;
* Communication and problem resolution with non-IT library staff&lt;br /&gt;
* Working with a vendor that uses agile development methodology&lt;br /&gt;
* Benefits and pitfalls of creating customizations and code workarounds&lt;br /&gt;
* Changes in library IT/coders' roles with SaaS&lt;br /&gt;
&lt;br /&gt;
...as well as thoughts on the philosophy of library discovery vs real-life experiences in moving to a single-search model.&lt;br /&gt;
&lt;br /&gt;
== Building for others (and ourselves):  the Avalon Media System ==&lt;br /&gt;
* [mailto:michael.klein@northwestern.edu Michael B Klein], Senior Software Developer, Northwestern University &lt;br /&gt;
** [http://code4lib.org/conference/2010/metz_klein Public Datasets in the Cloud] (code4lib 2010)&lt;br /&gt;
** [http://code4lib.org/conference/2013/klein-rogers The Avalon Media System: A Next Generation Hydra Head For Audio and Video Delivery] (code4lib 2013)&lt;br /&gt;
* [mailto:j-rudder@northwestern.edu Julie Rudder], Digital Initiatives Project Manager, Northwestern University&lt;br /&gt;
** no previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
[http://www.avalonmediasystem.org/ Avalon Media System] is a collaborative effort between development teams at Northwestern and Indiana Universities. Our goal is to produce an open source media management platform that works well for us, but is also widely adopted and contributed to by other institutions. We believe that building a strong user and contributor community is vital to the success and longevity of the project, and have developed the system with this goal in mind. We will share lessons learned, pains and successes we’ve had releasing two versions of the application since last year.  &lt;br /&gt;
&lt;br /&gt;
Our presentation will cover our experiences:&lt;br /&gt;
* providing flexible, admin-friendly distribution and installation options&lt;br /&gt;
* building with abstraction, customization and local integrations in mind&lt;br /&gt;
* prioritizing features (user stories)&lt;br /&gt;
* attracting code contributions from other institutions&lt;br /&gt;
* gathering community feedback &lt;br /&gt;
* creating a product rather than a bag of parts&lt;br /&gt;
&lt;br /&gt;
== How to check your data to provide a great data product? Data quality as a key product feature at Europeana ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:Peter.Kiraly@kb.nl Péter Király] portal backend developer, Europeana&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
[http://Europeana.eu/ Europeana.eu] - Europe's digital library, archive and museum - aggregates more than 30 million metadata records from more than 2200 institutions.  The records come from libraries, archives, museums and every other kind of cultural institution, from very different systems and metadata schemas, and are typically transformed several times until they are ingested into the Europeana data repository.  Europeana builds a consolidated database from these records, creating reliable and consistent services for end-users (a search portal, search widget, mobile apps, thematic sites etc.) and an API, which supports our strategic goeal of data for reuse in education, creative industries, and the cultural sector.  A reliable &amp;quot;data product&amp;quot; is thus at the core of our own software products, as well as those of our API partners.&lt;br /&gt;
&lt;br /&gt;
Much effort is needed to smooth out local differences in the metadata curation practice of our data providers. We need a solid framework to measure the consistency of our data and provide feedback to decision-makers inside and outside the organisation. We can also use this metrics framework to ask content providers to improve their own metadata. Of course, a data-quality-driven approach requires that we also improve the data transformation steps of the Europeana ingestion process itself. Data quality issues heavily define what new features we are able to create in our user interfaces and API, and might actually affect the design and implementation of our underlying data structure, the Europeana Data Model.&lt;br /&gt;
&lt;br /&gt;
In the presentation I briefly describe the Europeana metadata ingestion process, show the data quality metrics, the measuring techniques (using the Europeana API, Solr and MongoDB queries), some typical problems (both trivial and difficult ones), and finally the feedback mechanism we propose to deploy.&lt;br /&gt;
&lt;br /&gt;
Keywords: Europeana, data quality, EDM, API, Apache Solr, MongoDB, #opendata, #openglam&lt;br /&gt;
&lt;br /&gt;
== Teach your Fedora to Fly: scaling out your digital repository ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:acoburn@amherst.edu Aaron Coburn], Software Developer, Amherst College&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Fedora is a great repository system for managing large collections of digital objects, but what happens when a popular food magazine begins directing a large number of readers to a manuscript showing Emily Dickinson’s own recipe for doughnuts? While Fedora excels in its support of XML-based metadata, it doesn’t always perform well under a high volume of traffic. Nor is it especially tolerant of network or hardware failures.&lt;br /&gt;
&lt;br /&gt;
This presentation will show how we are making heavy use of a fedora repository while at the same time insulating it almost entirely from any web traffic. Starting with a distributed web front-end built with Node.js, and caching most of the user-accessible content from Fedora in an elastic, fault-tolerant Riak (NoSQL) cluster, we have eliminated nearly all single points of failure in the system. It also means that our production system is spread across twelve separate servers, where asynchrony and Map-Reduce are king. And aside from being blazing fast, it is also entirely Hydra-compliant.&lt;br /&gt;
&lt;br /&gt;
Furthermore, we will attempt to answer the question: if fedora crashes and the visitors to your site don’t notice, did it really fail?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using Open Source Software and Freeware to Preserve and Deliver Digital Videos ==&lt;br /&gt;
* [mailto:wfang@kinoy.rutgers.edu Wei Fang], Head of Digital Services, Rutgers University Law Library&lt;br /&gt;
* Jiebei Luo, Digital Projects Initiative Intern, Rutgers University&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The Rutgers University Law Library is the official digital repository of the New Jersey Supreme Court oral arguments since 2002. This large video collection contains approximately 3,000 videos with a total of 400 GB or 6,000 viewing hours. With the expansion of this collection, the existing database and the static website could not efficiently support the library’s daily operations and meet its patrons’ search needs. &lt;br /&gt;
By utilizing open source software and freeware such as Ubuntu, FFmpeg, Solr and Drupal, the library is able to develop a complete solution to re-encoding videos, embedding subtitles, incorporating  Solr search engine and content management system to support full-text subtitle search, automatically updating video metadata records in the library catalog system and eventually providing a plug-in free HTML 5-based Web interface for patrons to view the videos online.&lt;br /&gt;
The aspects below will be presented in detail at the conference:&lt;br /&gt;
*	Video codecs comparison &lt;br /&gt;
*	Server-end batch video encoding/re-encoding&lt;br /&gt;
*	HTML 5 video tag and embedding subtitles&lt;br /&gt;
*	Incorporating search engine Solr and content management tool 	Drupal with the database to retrieve videos by full-text search especially in subtitle files&lt;br /&gt;
*	Incorporating video metadata with the library catalog system&lt;br /&gt;
&lt;br /&gt;
== Shared Vision, Shared Resources: the Curate Institutional Repository ==&lt;br /&gt;
* Dan Brubaker Horst, University of Notre Dame &lt;br /&gt;
** [http://code4lib.org/conference/2011/JohnsonHorst A Community-Based Approach to Developing a Digital Exhibit at Notre Dame Using the Hydra Framework] &lt;br /&gt;
* Julie Rudder, Northwestern University&lt;br /&gt;
** no previous presentations&lt;br /&gt;
&lt;br /&gt;
Curate is being collaboratively developed by several institutions in the Hydra community who share the need and vision for a Fedora-backed Institutional Repository. The first release of Curate was a collaboration between Notre Dame and Northwestern University, along with Digital Curation Experts (DCE) - a vendor hired jointly by our two institutions. Powered by the Hydra engine Sufia, the team worked quickly to release the first version of Curate in October 2013 which provides a basic self-deposit system that has support for various content types, collection building, DOI minting, and user profile creation. From the very beginning we have built Curate to be easy to theme and extend in order to ease the process of installation and use by other institutions.&lt;br /&gt;
&lt;br /&gt;
In December 2013, additional partners will join the project including: Indiana University, the University of Cincinnati and the University of Virginia. Each institution contributes resources to the project in order to further our common goal to create a product that fits our needs and has a sustainable future.Together we will tackle additional content types (like complex data, software, media), administrative collections and more. &lt;br /&gt;
&lt;br /&gt;
Our presentation will include:&lt;br /&gt;
* a brief demonstration of Curate and technical overview&lt;br /&gt;
* why and how we work together&lt;br /&gt;
* why build Curate&lt;br /&gt;
* the future of the project&lt;br /&gt;
&lt;br /&gt;
== Solr, Cloud and Blacklight ==&lt;br /&gt;
* David Jiao, Library Information Systems, Indiana University at Bloomington, djiao@indiana.edu&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
SolrCloud refers to the distributed capabilities in Solr4. It is designed to offer a highly available, fault tolerant environment by organizing data into multiple pieces that can be hosted on multiple machines with replicas, and providing a centralized cluster configuration and management. &lt;br /&gt;
&lt;br /&gt;
At Indiana University, we are upgrading our Solr backend for our recently released Blacklight-based OPAC system from Solr 1.4 to Solr4, and we also put up efforts to build a private cloud of Solr4 servers. In this talk, I will persent certain features of SolrCloud, including distributed requests, fault tolerance, near real time indexing/searching, and configuration management with Zookeeper, and our experiences of utilizing these features to provide better performance and architecture for our OPAC system, which serves over 7 million bibliographic records to over 100 thousand students and faculty members. I will also discuss some practical lessons learned from our SolrCloud setup/upgrade and the integration of the new SolrCloud to our customized Blacklight system.  &lt;br /&gt;
&lt;br /&gt;
== Leveraging XSD's for Reflective, Live Dataset Support in Institutional Repositories ==&lt;br /&gt;
* [mailto:msulliva@ufl.edu Mark Sullivan], Library Information Technology, University of Florida&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
The University of Florida Libraries are currently adding support for active datasets into our METS-based institutional repository software.  This ongoing project enables the library to be a partner in current, or long-running, data-driven projects around the university by providing tangible short-term and long-term benefits to the projects.  The system assists project teams by storing and providing access to their data, while supporting online filtering and sorting of the data, custom queries, and adding and editing of the data by authorized users.  We are also exploring simple data visualizations to allow users to perform basic graphical and geographic queries.  Several different schemas were explored including DDI and EML, but ultimately the streamlined approach of using XSD's with some custom attributes was chosen, with all other data residing in the METS file portions.  Currently the system is being developed using XSD's describing XML datasets, but this model should easily scale to support SQL datasets or large datasets supported by Hadoop or iRODS.&lt;br /&gt;
&lt;br /&gt;
This work is being integrated in the open source [http://sobek.ufl.edu SobekCM Digital Content Management System] which is built on a pair-tree structure of METS resources with [http://ufdc.ufl.edu/design/webcontent/sobekcm/SobekCM_Resource_Object.pdf rich metadata support] including DC, MODS, MARC, VRACore, DarwinCore, IEE-LOM, GML/KML, schema.org microdata, and many other standard schemas.  The system has emphasized online, distributed creation and maintenance of resources including geo-placement and geographic searching of resources, building structure maps (table of contents) visually online, and a broad suite of curator tools.  &lt;br /&gt;
&lt;br /&gt;
This work is presented as a model which could be implemented in other systems as well.  We will demonstrate current support and discuss our upcoming roadmap to provide complete support.&lt;br /&gt;
&lt;br /&gt;
== Dead-simple Video Content Management: Let Your Filesystem Do The Work ==&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, NCSU Libraries (akorphan (at) ncsu.edu)&lt;br /&gt;
** (never led or soloed a C4L presentation)&lt;br /&gt;
&lt;br /&gt;
Content management is hard. To keep all the moving parts in order, and to maintain a layer of separation between the system and content creators (who are frequently not technical experts), we typically turn to content management systems like Drupal. But even Drupal and its kin require significant overhead and present a not inconsiderable learning curve for nontechnical users.&lt;br /&gt;
&lt;br /&gt;
In some contexts it's possible -- and desirable -- to manage content in a more streamlined, lightweight way, with a minimum of fuss and technical infrastructure. In this presentation I'll share a simple MVC-like architecture for managing video content for playback on the web, which uses a combination of Apache's mod_rewrite module and your server's filesystem structure to provide an automated approach to video content management that's easy to implement and provides a low barrier to content updates: friendly to content creators and technology implementors alike. Even better, the basic method is HTML5-friendly, and can be integrated into your favorite content management system if you've got permissions for creating templates.&lt;br /&gt;
&lt;br /&gt;
In the presentation I'll go into detail about the system structure and logic required to implement this approach. I'll detail the benefits and limitations of the system, as well as the challenges I encountered in developing its implementation. Audience members should come away with sufficient background to implement a similar system on their own servers. Implementation documentation and genericized code will also be shared, as available.&lt;br /&gt;
&lt;br /&gt;
== Managing Discovery ==&lt;br /&gt;
&lt;br /&gt;
* Andrew Pasterfield, Senior Programmer/Systems Analyst, University of Calgary Library, ampaster@ucalgary.ca&lt;br /&gt;
**No previous code4lib presentations &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
In fall 2012 the University of Calgary Library launched a new home page that incorporated a Summon powered&lt;br /&gt;
Single Search Box with customized “bento box” results display. Search at the U of C now combines a range of&lt;br /&gt;
metadata sources for discovery and customized mapping of a database recommender and LibGuide into a unified&lt;br /&gt;
display.  Further customizations include a non Google Analytics/non proxy method to log clicks.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss the technical details of bringing the various systems together into one display interface to increase discovery at the U of C Library.&lt;br /&gt;
&lt;br /&gt;
http://library.ucalgary.ca&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sorting it out: a piece of the User Centered Design Process ==&lt;br /&gt;
&lt;br /&gt;
* Cindy Beggs, [http://www.akendi.com/aboutus/management/ Akendi], cindy@akendi.com&lt;br /&gt;
&lt;br /&gt;
This talk is about how to apply a user centered design methodology to the process of creating an information architecture.  Participants learn the fundamentals of UCD and how card sorting and reverse card sorting enable us to isolate the content we present on screen from the layouts and visuals of those screens.  We talk about ways to identify who will be using the information architecture you are creating and why we need to know how it will be used.&lt;br /&gt;
 &lt;br /&gt;
What will attendees takes away from your talk?&lt;br /&gt;
The criticality of involving “real” end users in the process of creating an information architecture.  The basics of following a user-centered-design process in the creation of best in class, content-rich, digital products.&lt;br /&gt;
&lt;br /&gt;
Cindy Beggs has been working in the “information industry” for over 25 years.  A librarian by profession, she has spent decades helping users figure out how to find their way through large bodies of content.  Her insights into how people seek information, her empathy for those who find it a challenge and her practical experience helping organizations figure out how to best structure their content contribute to her success as an information architect with both clients and trainees.  (http://www.akendi.com/aboutus/management/)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Implementation of ArchivesSpace in University of Richmond==&lt;br /&gt;
&lt;br /&gt;
*Birong Ho, bho@richmond.edu&lt;br /&gt;
&lt;br /&gt;
University of Richmond implemented its archive collection management ArchivsSpace in the fall, 2013. As a charter member and the Head of Special Collection as the Board member, implementation of such an Open Source Software became a priority. &lt;br /&gt;
&lt;br /&gt;
Several aspects of implementation will be addressed in the talk. Among them, they are Collections and Repository, storage layer including data format, System resources requirements, Technical architecture, Customization, scaling and integrated with other systems in the library.&lt;br /&gt;
&lt;br /&gt;
The customization, scale, and integration with other systems such as Archeon and Exist on campus became a concern will be focused and elaborated in the talk.&lt;br /&gt;
&lt;br /&gt;
==Easy Wins for Modern Web Technologies in Libraries==&lt;br /&gt;
&lt;br /&gt;
*[mailto:trey.terrell@oregonstate.edu Trey Terrell], Analyst Programmer, Oregon State University&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Oregon State University is currently implementing an updated version of its room reservation system. In its development we've come across and implemented a variety of &amp;quot;easy wins&amp;quot; to make it more responsive, easier to maintain, less expensive to run, and just cooler to experience. While our particular system was in Ruby on Rails, this talk will address general methods and example utilities which can be used no matter your stack.&lt;br /&gt;
&lt;br /&gt;
I'll be talking about things like cache management, reverse proxies, publish/subscribe servers, WebSockets, responsive design, asynchronous processing, and keeping complicated stacks up and running with minimal effort.&lt;br /&gt;
&lt;br /&gt;
==Implementing Islandora at a Small Institution==&lt;br /&gt;
&lt;br /&gt;
*Megan Kudzia, Albion College Library&lt;br /&gt;
*Eddie Bachle, Albion College IT&lt;br /&gt;
**No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
Albion College (and particularly the Library/Archives and Special Collections) has a variety of needs which could be met by an open-source Institutional Repository system. Several months and lots of conversations later, we’re continuing to troubleshoot our way through Islandora. We’d like to talk about what has worked for us, where our frustrations have been, whether it’s even possible to install and develop a system like this at a small institution, and where the process has stalled. &lt;br /&gt;
&lt;br /&gt;
As of right now, we do have a semi-working installation. We’re not sure when it will be ready for our end users, but we'll talk about our development process and evaluate our progress.&lt;br /&gt;
''Contributions also by Nicole Smeltekop, Albion College Archives &amp;amp; Special Collections''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== PhantomJS+Selenium: Easy Automated Testing of AJAX-y UIs ==&lt;br /&gt;
&lt;br /&gt;
* Martin Haye, California Digital Library, martin.haye@ucop.edu&lt;br /&gt;
** Previous Code4Lib Presentation: [http://code4lib.org/conference/2012/collett Beyond code: Versioning data with Git and Mercurial] at Code4Lib 2012 (Martin co-presenting with Stephanie Collett)&lt;br /&gt;
* Mark Redar, California Digital Library, mark.redar@ucop.edu&lt;br /&gt;
&lt;br /&gt;
Web user interfaces are demanding ever-more dynamism and polish, combining HTML5, AJAX, lots of CSS and jQuery (or ilk) to create autocomplete drop-downs, intelligent buttons, stylish alert dialogs, etc. How can you make automated tests for these highly complex and interactive UIs?&lt;br /&gt;
&lt;br /&gt;
Part of the answer is PhantomJS. It’s a modern WebKit browser that’s “headless” (meaning it has no display) that can be driven from command-line Selenium unit tests. PhantomJS is dead simple to install, and its blazing speed and server-friendliness make continuous integration testing easy. You can write UI unit tests in {language-of-your-choice} and run them not just in PhantomJS but in Firefox and Chrome, plus a zillion browser/OS combinations at places like SauceLabs, TestingBot and BrowserStack.&lt;br /&gt;
&lt;br /&gt;
In this double-team live code talk, we’ll explain all that while we demonstrate the following in real time:&lt;br /&gt;
&lt;br /&gt;
* Start with nothing.&lt;br /&gt;
* Install Selenium bindings for Ruby and Python.&lt;br /&gt;
* In each language write a small test of an AJAX-y UI.&lt;br /&gt;
* Run the tests in Firefox, and fix bugs (in the test or UI) as needed.&lt;br /&gt;
* Install PhantomJS.&lt;br /&gt;
* Show the same tests running headless as part of a server-friendly test suite. &lt;br /&gt;
* (Wifi permitting) Show the same tests running on a couple different browser/OS combinations on the server cloud at SauceLabs – talking through a tunnel to the local firewalled application.&lt;br /&gt;
&lt;br /&gt;
==New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library==&lt;br /&gt;
&lt;br /&gt;
* Stephanie Walker – swalker@brooklyn.cuny.edu&lt;br /&gt;
* Howard Spivak – howards@brooklyn.cuny.edu&lt;br /&gt;
* Alex - Alex@brooklyn.cuny.edu&lt;br /&gt;
&lt;br /&gt;
Academic libraries are caught in budget squeezes and often struggle to find ways to communicate value to senior administration and others.  At Brooklyn College Library, we have taken an unusual, possibly unique, approach to these issues.  Our technology staff have long worked directly with librarians to develop products that meet library, faculty, and student needs, and we have shared many of our products with colleagues, including an award-winning website, e-resource, and content management system we call 4MyLibrary, which we shared for free with 8 CUNY colleges, and also an easy-to-use book scanner, which has proven overwhelming popular with students, faculty, other librarians, and numerous campus offices.  Recently, motivated by budget cuts, we decided that what worked for us might interest other libraries, and working with our Office of Technology Commercialization, we started selling 2 products:  our book scanners (at half the price of commercial alternatives), and a hosting service, whereby we could host and support 4MyLibrary for libraries with minimal technology staff.  Both succeeded, and yielded major benefits:  a steady revenue stream and the admiration and serious goodwill of our senior administration and others.   However, this presentation is neither a basic how-to, nor an advertisement.  With this presentation, we hope to spur a conversation for broader collaboration, especially regarding new technologies, among libraries.  We all have some level of technical expertise, most of us are struggling with rising prices and tight budgets, and many of us are unhappy with various technology products we use, from scanners to our ILS.  We believe – and can demonstrate – that with collaboration, we can solve many of our problems, and provide better services to boot. &lt;br /&gt;
&lt;br /&gt;
== Identifiers, Data, and Norse Gods ==&lt;br /&gt;
&lt;br /&gt;
* Ryan Scherle, Dryad Digital Repository, ryan@datadryad.org&lt;br /&gt;
&lt;br /&gt;
ORCID and DataCite are provide stable identifiers for researchers and and data, respectively. Each system does a fine job of providing value to its users. But wouldn't it be great if they could link their systems to create something much more powerful? Perhaps even as powerful as a god?&lt;br /&gt;
&lt;br /&gt;
Enter [http://odin-project.eu/ ODIN], The ORCID and DataCite Interoperability Network. ODIN is a two-year project to unleash the power of persistent identifiers for researchers and the research they create. This talk will present recent work from the ODIN project, including several tools that can be used to unleash the godlike power of identifiers at your institution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39849</id>
		<title>2014 Prepared Talk Proposals</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2014_Prepared_Talk_Proposals&amp;diff=39849"/>
				<updated>2013-11-08T20:03:02Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''Proposals for Prepared Talks:'''&lt;br /&gt;
&lt;br /&gt;
Prepared talks are 20 minutes (including setup and questions), and should focus on one or more of the following areas:&lt;br /&gt;
 &lt;br /&gt;
* ''Projects'' you've worked on which incorporate innovative implementation of existing technologies and/or development of new software&lt;br /&gt;
* ''Tools and technologies'' – How to get the most out of existing tools, standards and protocols (and ideas on how to make them better)&lt;br /&gt;
* ''Technical issues'' - Big issues in library technology that should be addressed or better understood&lt;br /&gt;
* ''Relevant non-technical issues'' – Concerns of interest to the Code4Lib community which are not strictly technical in nature, e.g. collaboration, diversity, organizational challenges, etc.&lt;br /&gt;
&lt;br /&gt;
'''To Propose a Talk'''&lt;br /&gt;
* Log in to the wiki in order to submit a proposal. If you are not already registered, follow the instructions to do so.&lt;br /&gt;
* Provide a title and brief (500 words or fewer) description of your proposed talk.&lt;br /&gt;
* If you so choose, you may also indicate when, if ever, you have presented at a prior Code4Lib conference. This information is completely optional, but it may assist us in opening the conference to new presenters.&lt;br /&gt;
&lt;br /&gt;
As in past years, the Code4Lib community will vote on proposals that they would like to see included in the program. This year, however, only the top 10 proposals will be guaranteed a slot at the conference. Additional presentations will be selected by the Program Committee in an effort to ensure diversity in program content. Community votes will, of course, still weigh heavily in these decisions.&lt;br /&gt;
&lt;br /&gt;
Presenters whose proposals are selected for inclusion in the program will be guaranteed an opportunity to register for the conference. The standard conference registration fee will still apply.&lt;br /&gt;
&lt;br /&gt;
''Proposals can be submitted through '''Friday, November 8, 2013, at 5pm PST'''''. Voting will commence on November 18, 2013 and continue through December 6, 2013. The final line-up of presentations will be announced in early January, 2014.&lt;br /&gt;
&lt;br /&gt;
'''Talk Proposals'''&lt;br /&gt;
&lt;br /&gt;
==Creating a new Greek-Dutch dictionary==&lt;br /&gt;
* Caspar Treijtel, University of Amsterdam, c.treijtel@uva.nl&lt;br /&gt;
&lt;br /&gt;
At present, no complete dictionary of (ancient) Greek-Dutch is available online. A new dictionary is currently under construction at Leiden University, with software being developed at the University of Amsterdam. The team in Leiden has already begun preparation of the data, with at this moment about 6,000 approved lemmas. The ultimate goal is to produce both a print version and online open access version from the same source documents. The software needed for this has been made in a project that was funded by CLARIN-NL.&lt;br /&gt;
&lt;br /&gt;
Migrator&lt;br /&gt;
&lt;br /&gt;
For the production of lemmas we have implemented an advanced workflow. The (generally non-technical) users create lemmas using MS Word, which is both familiar and easy to use. We have developed a custom software module that carefully migrates the Word documents into deeply structured XML by analyzing the structure and semantics of the lemmas, and falling back on heuristics in ambiguous cases. While having initially envisioned the oXygen XML Author component as the main tool for creating new lemmas, we obtained excellent results with the migrator module, and decided therefore to continue using MS Word as the primary composition tool. The main advantage of this is that the editors are much more familiar with Word than with any other WYSIWYG editor. Lemmas that have been migrated to XML are stored in an XML database and can be further edited using oXygen XML Author.&lt;br /&gt;
&lt;br /&gt;
Lemmatizer&lt;br /&gt;
&lt;br /&gt;
Greek morphology is complicated. In order to use a dictionary effectively, a rather high level of initial language competence is necessary for the user to be able to relate the word form s/he finds in a text to the correct basic lemma form, where the definition of the word can be found. Using a Greek morphological database we have been able to facilitate the search for lemmas. A ‘lemmatizer’ module gives the possible parsings of the word forms and the lemmas they can be derived from. This enables the user to type in the word as found in the text and be redirected to the correct lemma.&lt;br /&gt;
&lt;br /&gt;
Visualization&lt;br /&gt;
&lt;br /&gt;
For the online dictionary we have implemented a visualization module that allows the user to view multiple lemmas at once. The implementation of this module has been done using the Javascript framework MooTools. The result is a viewer that performs really well and is run by maintainable Javascript code.&lt;br /&gt;
&lt;br /&gt;
The online dictionary is still being worked on, have a look at http://www.woordenboekgrieks.nl/ for the beta version. A newer test version with additional features can be found here: http://angel.ic.uva.nl:8600/.&lt;br /&gt;
&lt;br /&gt;
Credits&lt;br /&gt;
&lt;br /&gt;
* construction of the dictionary: Prof. Ineke Sluiter, Classics department of Leiden University; Prof. Albert Rijksbaron, University of Amsterdam&lt;br /&gt;
* publisher of the dictionary: Amsterdam University Press&lt;br /&gt;
* design/typesetting dictionary: TaT Zetwerk (http://www.tatzetwerk.nl/)&lt;br /&gt;
* software development: Digital Production Center, University Library, University of Amsterdam&lt;br /&gt;
* project funding: CLARIN-NL (http://www.clarin.nl/)&lt;br /&gt;
* morphological database for use by the lemmatizer: courtesy of Prof. Helma Dik, University of Chicago (based on data of the Perseus Project)&lt;br /&gt;
&lt;br /&gt;
----&lt;br /&gt;
&lt;br /&gt;
== Using Drupal to drive alternative presentation systems ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Highermath|Cary Gordon]], The Cherry Hill Company, cgordon@chillco.com&lt;br /&gt;
&lt;br /&gt;
Recently, we have been building systems that use angular.js, Rails, or other systems for presentation, while leveraging Drupal's sophisticated content management capabilities on the back end.&lt;br /&gt;
&lt;br /&gt;
So far, these have been one-way systems, but as we move to Drupal 8 we are beginning to explore ways to further decouple the presentation and CMS functions.&lt;br /&gt;
&lt;br /&gt;
== A Book, a Web Browser and a Tablet: How Bibliotheca Alexandrina's Book Viewer Framework Makes It Possible ==&lt;br /&gt;
 &lt;br /&gt;
* [[User:Mohammed.abuouda|Mohammed Abu ouda]], Bibliotheca Alexandrina (The new Library of Alexandria)&lt;br /&gt;
&lt;br /&gt;
A lot of institutions around the world are engaged in multiple digitization projects aiming at preserving the human knowledge present in books and availing them through multiple channels to people around the whole globe. These efforts will sure help close the digital gap particularly with the arrival of affordable e-readers, mobile phones and network coverage. However, the digital reading experience has not yet arrived to its maximum potential. Many readers miss features they like in their good old books and wish to find them in their digital counterpart. In an attempt to create a unique digital reading experience, Bibliotheca Alexandria (BA) created a flexible book viewing framework that is currently used to access its current collection of more than 300,000 digital books in five different languages which includes the largest collection of digitized Arabic books.&lt;br /&gt;
&lt;br /&gt;
Using open source tools, BA used the framework to develop a modular book viewer that can be deployed in different environments and is currently at the heart of various BA projects. The Book viewer provides several features creating a more natural reading experience. As with physical books, the reader can now personalize the books he reads by adding annotations like highlights, underlines and sticky notes to capture his thoughts and ideas in addition to being able to share the book with friends on social networks. The reader can perform a search across the content of the book receiving highlighted search results within the pages of the book. More features can be further added to the book viewer through its plugin architecture.&lt;br /&gt;
&lt;br /&gt;
== Structured data NOW: seeding schema.org in library systems ==&lt;br /&gt;
 &lt;br /&gt;
* [http://coffeecode.net Dan Scott], Laurentian University&lt;br /&gt;
** Previous code4lib presentations: [https://archive.org/details/code4lib.conf.2008.pres.CouchDBsacrilege CouchDB is sacrilege... mmm, delicious sacrilege] at Code4Lib 2008&lt;br /&gt;
&lt;br /&gt;
The semantic web, linked data, and structured data are all fantastic ideas with a barrier imposed by implementation constraints. If their system does not allow customizations, or the institution lacks skilled human resources, it does not matter how enthused a given library might be about publishing structured data... it will not happen. However, if the software in use simply publishes structured data by default, then the web will be populated for free. Really! No extra resources necessary.&lt;br /&gt;
&lt;br /&gt;
This presentation highlights Dan's work with systems such as Evergreen, Koha, and VuFind to enable the publication of schema.org structured data out-of-the-box. Along the way, we reflect the current state of the W3C Schema.org Bibliographic Extension community group efforts to shape the evolution of the schema.org vocabulary. Finally, hold on tight as we contemplate next steps and the possibilities of a world where structured data is the norm on the web.&lt;br /&gt;
&lt;br /&gt;
== Towards Pasta Code Nirvana: Using JavaScript MVC to Fill Your Programming Ravioli ==&lt;br /&gt;
&lt;br /&gt;
* Bret Davidson, North Carolina State University Libraries, bret_davidson@ncsu.edu&lt;br /&gt;
** Previous Code4Lib Presentations: [http://wiki.code4lib.org/index.php/2013_talks_proposals#Data-Driven_Documents:_Visualizing_library_data_with_D3.js Visualizing library data with D3.js] at Code4Lib 2013&lt;br /&gt;
&lt;br /&gt;
JavaScript MVC frameworks are ushering in a golden age of robust and responsive web applications that take advantage of evergreen browsers, performant JS engines, and the unprecedented reach provided by billions of personal computing devices. The web browser has emerged as the world’s most popular application runtime and the complexity[1] and scope of JavaScript applications has exploded accordingly. Server-side web frameworks like Rails and Django have helped developers adhere to best practices like modularity, dependency injection, and unit testing for years, practices that are now being applied to JavaScript development through projects like Backbone[2], Ember[3], and Angular[4].&lt;br /&gt;
&lt;br /&gt;
This talk will discuss the issues JavaScript MVC frameworks are trying to solve, common features like data binding, implications for the future of web development[5], and the appropriateness of JavaScript MVC for library applications.&lt;br /&gt;
&lt;br /&gt;
*[1]http://en.wikipedia.org/wiki/Spaghetti_code&lt;br /&gt;
*[2]http://backbonejs.org&lt;br /&gt;
*[3]http://emberjs.com&lt;br /&gt;
*[4]http://angularjs.org&lt;br /&gt;
*[5]http://tomdale.net/2013/09/progressive-enhancement-is-dead/&lt;br /&gt;
&lt;br /&gt;
== WebSockets for Real-Time and Interactive Interfaces ==&lt;br /&gt;
&lt;br /&gt;
* [http://ronallo.com Jason Ronallo], NCSU Libraries, jason_ronallo@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/ronallo HTML5 Microdata and Schema.org] 2012&lt;br /&gt;
* [http://code4lib.org/conference/2013/ronallo HTML5 Video Now!] 2013&lt;br /&gt;
&lt;br /&gt;
Watching the Google Analytics Real-Time dashboard for the first time was mesmerizing. As soon as someone visited a site, I could see what page they were on. For a digital collections site with a lot of images, it was fun to see what visitors were looking at. But getting from Google Analytics to the image or other content of what was currently being viewed was cumbersome. The real-time experience was something I wanted share with others. I'll show you how I used a WebSocket service to create a real-time interface to digital collections.&lt;br /&gt;
&lt;br /&gt;
In the Hunt Library at NCSU we have some large video walls. I wanted to make HTML-based exhibits that featured viewer interactions. I'll show you how I converted Listen to Wikipedia [1] into an bring-your-own-device interactive exhibit. With WebSockets any HTML page can be remote controlled by any internet connected device.&lt;br /&gt;
&lt;br /&gt;
I will attempt to include real-time audience participation.&lt;br /&gt;
&lt;br /&gt;
[1] http://listen.hatnote.com/&lt;br /&gt;
&lt;br /&gt;
== Rapid Development of Automated Tasks with the File Analyzer ==&lt;br /&gt;
&lt;br /&gt;
* Terry Brady, Georgetown University Libraries, twb27@georgetown.edu&lt;br /&gt;
&lt;br /&gt;
The Georgetown University Libraries have customized the File Analyzer and Metadata Harvester application (https://github.com/Georgetown-University-Libraries/File-Analyzer) to solve a number of library automation challenges:&lt;br /&gt;
* validating digitized and reformatted files&lt;br /&gt;
* validating vendor statistics for counter compliance&lt;br /&gt;
* preparing collections of digital files for archiving and ingest&lt;br /&gt;
* manipulating ILS import and export files&lt;br /&gt;
&lt;br /&gt;
The File Analyzer application was used by the US National Archives to validate 3.5 million digitized images from the 1940 Census.  After implementing a customized ingest workflow within the File Analyzer, the Georgetown University Libraries was able to process an ingest backlog of over a thousand files of digital resources into DigitalGeorgetown, the Libraries’ Digital Collections and Institutional Repository platform.  Georgetown is currently developing customized workflows that integrate Apache Tika, BagIt, and Marc conversion utilities.&lt;br /&gt;
&lt;br /&gt;
The File Analyzer is a desktop application with a powerful framework for implementing customized file validation and transformation rules.  As new rules are deployed, they are presented to users within a user interface that is easy (and powerful) to use.&lt;br /&gt;
&lt;br /&gt;
Learn about the functionality that is available for download, how you can use this tool to automate workflows from digital collections to ILS ingests to electronic resources statistics and also discuss the opportunities to collaborate on enhancements to this application!&lt;br /&gt;
&lt;br /&gt;
== GeoHydra: How to Build a Geospatial Digital Library with Fedora ==&lt;br /&gt;
 &lt;br /&gt;
* [http://stanford.edu/~drh Darren Hardy], Stanford University, drh@stanford.edu&lt;br /&gt;
&lt;br /&gt;
Geographically-rich data are exploding and putting fear in those trying to&lt;br /&gt;
tackle integrating them into existing digital library infrastructures.&lt;br /&gt;
Building a spatial data infrastructure that integrates with your digital&lt;br /&gt;
library infrastructure need not be a daunting task. We have successfully&lt;br /&gt;
deployed a geospatial digital library infrastructure using Fedora and&lt;br /&gt;
open-source geospatial software [1]. We'll discuss the primary design&lt;br /&gt;
decisions and technologies that led to a production deployment within a few&lt;br /&gt;
months. Briefly, our architecture revolves around discovery, delivery, and&lt;br /&gt;
metadata pipelines using open-source OpenGeoPortal [2], Solr [3], GeoServer&lt;br /&gt;
[4], PostGIS [5], and GeoNetwork [6] technologies, plus the proprietary ESRI&lt;br /&gt;
ArcMap [7] -- the GIS industry's workhorse. Finally, we'll discuss the key&lt;br /&gt;
skillsets needed to build and maintain a spatial data infrastructure.&lt;br /&gt;
&lt;br /&gt;
[1] http://foss4g.org&lt;br /&gt;
[2] http://opengeoportal.org&lt;br /&gt;
[3] http://lucene.apache.org/solr&lt;br /&gt;
[4] http://geoserver.org&lt;br /&gt;
[5] http://postgis.net&lt;br /&gt;
[6] http://geonetwork-opensource.org&lt;br /&gt;
[7] http://esri.com&lt;br /&gt;
&lt;br /&gt;
==Under the Hood of Hadoop Processing at OCLC Research ==&lt;br /&gt;
&lt;br /&gt;
[http://roytennant.com/ Roy Tennant]&lt;br /&gt;
&lt;br /&gt;
* Previous Code4Lib presentations: 2006: &amp;quot;The Case for Code4Lib 501c(3)&amp;quot;&lt;br /&gt;
&lt;br /&gt;
[http://hadoop.apache.org/ Apache Hadoop] is widely used by Yahoo!, Google, and many others to process massive amounts of data quickly. OCLC Research uses a 40-node compute cluster with Hadoop and HBase to process the 300 million MARC records of WorldCat in various ways. This presentation will explain how Hadoop MapReduce works and illustrate it with specific examples and code. The role of the jobtracker in both monitoring and reporting on processes will be explained. String searching WorldCat will also be demonstrated live.&lt;br /&gt;
&lt;br /&gt;
== Quick and Easy Data Visualization with Google Visualization API and Google Chart Libraries ==&lt;br /&gt;
 &lt;br /&gt;
[http://bohyunkim.net/blog Bohyun Kim], Florida International University, bohyun.kim@fiu.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Do most of the data that your library collects stay in spreadsheets or are published as a static table with a series of boring numbers? Do your library stakeholders spend more time collecting the data than using it as a decision-making tool because the data is presented in a way that makes it hard for them [http://developers.google.com/chart/interactive/docs/gallery to quickly grasp its significance? ]&lt;br /&gt;
&lt;br /&gt;
This talk will provide an overview of [http://developers.google.com/chart/interactive/docs/reference Google Visualization API] [2] and [http://developers.google.com/chart/ Google Chart Libraries] [3] to get you started on the way to quickly query and visualize your library data from remote data sources (e.g. a Google Spreadsheet or your own database) with (or without) cool-looking user-controls, animation effects, and even a dashboard.&lt;br /&gt;
&lt;br /&gt;
== Leap Motion + Rare Books: A hands-free way to view and interact with rare books in 3D ==&lt;br /&gt;
 &lt;br /&gt;
[http://http://www.youtube.com/user/jpdenzer Juan Denzer], Binghamton University, jdenzer@binghamton.edu&lt;br /&gt;
* 'No' previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
As rare books become more delicate over time, making them available to the public becomes harder.  We at Binghamton University Library have developed an application that makes it easier to view rare books without ever having to touch them.  We have combined the Leap Motion hands-free device and 3D rendered models to create a new virtual experience for the viewer.&lt;br /&gt;
&lt;br /&gt;
The application allows the user to rotate and zoom in on a 3D representation of a rare book.  The user is also able to ‘open’ the virtual book and flip through it using a natural user interface.  Such as swiping the hand left or right to turn the page.&lt;br /&gt;
&lt;br /&gt;
The application is built on the .Net framework and is written in C#.  3D models are created using simple 3D software such as sketchup or Blender.  Scans of the book cover and spine are created using simple flatbed scanners.  The inside pages are scanned using overhead scanners. &lt;br /&gt;
&lt;br /&gt;
This talk with discuss the technologies used in developing the application and virtually any library could implement the application with virtually no coding at all. This presentation will have a demonstration of the software and also a chance for audience members to experience the Rare Book Leap Motion App themselves.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Course Reserves Unleashed! ==&lt;br /&gt;
 &lt;br /&gt;
* Bobbi Fox, Library Technology Services, Harvard University, bobbi_fox@harvard.edu&lt;br /&gt;
* Gloria Korsman, Andover-Harvard Theological Library&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Hey kids!  Remember when SOAP was used for something other than washing?  Our sophisticated (and highly functional) Course Reserves Request system does!&lt;br /&gt;
&lt;br /&gt;
However, while the system is great for submitting and  processing course reserve requests, the student-facing presentation through Havard’s home-grown -- and soon to be replaced -- LMS leaves a lot to be desired.  &lt;br /&gt;
&lt;br /&gt;
Follow along as we leverage Solr 4 as a No-SQL database, along with more progressive RESTful API techniques, to release Reserves data into the wild without interfering with reserves request processing -- and, in the process, open up the opportunity for other schools to feed their data in as well.&lt;br /&gt;
&lt;br /&gt;
== We Are All Disabled! Universal Web Design Making Web Services Accessible for Everyone ==&lt;br /&gt;
 &lt;br /&gt;
Cynthia Ng, Accessibility Librarian, CILS at Langara College&lt;br /&gt;
* No previous Code4Lib presentations (not counting lightning talks)&lt;br /&gt;
&lt;br /&gt;
We’re building and improving tools and services all the time, but do you only develop for the “average” user or add things for “disabled” users? We all use “assistive” technology accessing information in a multitude of ways with different platforms, devices, etc. Let’s focus on providing web services that are accessible to everyone without it being onerous or ugly. The aim is to get you thinking about what you can do to make web-based services and content more accessible for all from the beginning or with small amounts of effort whether you're a developer or not.&lt;br /&gt;
&lt;br /&gt;
The goal of the presentation is to provide both developers and content creators with information on simple, practical ways to make web content and web services more accessible. However, rather than thinking about putting in extra effort or making adjustment for those with disabilities, I want to help people think about how to make their websites more accessible for all users through universal web design.&lt;br /&gt;
&lt;br /&gt;
== Personalize your Google Analytics Data with Custom Events and Variables ==&lt;br /&gt;
&lt;br /&gt;
[http://joshwilson.net Josh Wilson], Systems Integration Librarian, State Library of North Carolina - joshwilsonnc@gmail.com&lt;br /&gt;
&lt;br /&gt;
At the State Library of North Carolina, we had more specific questions about the use of our digital collections than standard GA could provide. A few implementations of custom events and custom variables later, we have our answers.&lt;br /&gt;
&lt;br /&gt;
I'll demonstrate how these analytics add-ons work, and why implementation can sometimes be more complicated than just adding a few lines of JavaScript to your ga.js. I'll discuss some specific examples in use at the SLNC:&lt;br /&gt;
&lt;br /&gt;
* Capturing the content of specific metadata fields in CONTENTdm as Custom Events &lt;br /&gt;
* Recording Drupal taxonomy terms as Custom Variables&lt;br /&gt;
&lt;br /&gt;
In both instances, this data deepened our understanding of how our sites and collections were being used, and in turn, we were able to report usage more accurately to content contributors and other stakeholders.&lt;br /&gt;
&lt;br /&gt;
More on: [https://developers.google.com/analytics/devguides/collection/gajs/eventTrackerGuide GA Custom Events] | [https://developers.google.com/analytics/devguides/collection/gajs/gaTrackingCustomVariables GA Custom Variables]&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Behold Fedora 4: The Incredible Shrinking Repository! ==&lt;br /&gt;
&lt;br /&gt;
Esmé Cowles, UC San Diego Library.  Previous talk: [http://code4lib.org/conference/2013/cowles-critchlow-westbrook All Teh Metadatas Re-Revisited] (2013)&lt;br /&gt;
&lt;br /&gt;
* One repository contains untold numbers of digital objects and powers many Hydra and Islandora apps&lt;br /&gt;
* It speaks RDF, but contains no triplestore! (triplestores sold separately, SPARQL Update may be involved, some restrictions apply)&lt;br /&gt;
* Flexible enough to tie itself in knots implementing storage and access control policies&lt;br /&gt;
* Witness feats of strength and scalability, with dramatically increased performance and clustering&lt;br /&gt;
* Plumb the depths of bottomless hierarchies, and marvel at the metadata woven into the very fabric of the repository&lt;br /&gt;
* Ponder the paradox of ingesting large files by not ingesting them&lt;br /&gt;
* Be amazed as Fedora 4 swallows other systems whole (including Fedora 3 repositories)&lt;br /&gt;
* Watch novice developers setup Fedora 4 from scratch, with just a handful of incantations to Git and Maven&lt;br /&gt;
&lt;br /&gt;
The Fedora Commons Repository is the foundation of many digital collections, e-research, digital library, archives, digital preservation, institutional repository and open access publishing systems.  This talk will focus on how Fedora 4 improves core repository functionality, adds new features, maintains backwards compatibility, and addresses the shortcomings of Fedora 3.&lt;br /&gt;
&lt;br /&gt;
== Organic Free-Range API Development - Making Web Services That You Will Actually Want to Consume ==&lt;br /&gt;
&lt;br /&gt;
Steve Meyer and Karen Coombs, OCLC&lt;br /&gt;
&lt;br /&gt;
Building web services can have great benefits by providing reusability of data and functionality. Underpinning your applications with a web service will allow you to write code once and support multiple environments: your library's web app, mobile applications, the embedded widget in your campus portal. However, building a web service is its own kind of artful programming. Doing it well requires attention to many of the same techniques and requirements as building web applications, though with different outcomes. &lt;br /&gt;
&lt;br /&gt;
So what are the usability principles for web services? How do you build a web service that you (and others) will actually want to use? In this talk, we’ll share some of the lessons learned - the good, the bad, and the ugly - through OCLC's work on the WorldCat Metadata API. This web service is a sophisticated API that provides external clients with read and write access to WorldCat data. It provides a model to help aspiring API creators navigate the potential complications of crafting a web service. We'll cover:&lt;br /&gt;
&lt;br /&gt;
* Loose coupling of data assets and resource-oriented data modeling at the core&lt;br /&gt;
* Coding to standards vs. exposure of an internal data model&lt;br /&gt;
* Authentication and security for web services: API Keys, Digital Signing, OAuth Flows&lt;br /&gt;
* Building web services that behave as a suite so it looks like the left hand knows what the right hand is doing&lt;br /&gt;
&lt;br /&gt;
So at the end of the day, your team will know your API is a very good egg after all. &lt;br /&gt;
&lt;br /&gt;
If accepted, the presenters intend to produce and share a Quick Guide for building a web service that will reflect content presented in the talk.&lt;br /&gt;
&lt;br /&gt;
== Lucene's Latest (for Libraries) ==&lt;br /&gt;
&lt;br /&gt;
erik.hatcher@lucidworks.com&lt;br /&gt;
&lt;br /&gt;
Lucene powers the search capabilities of practically all library discovery platforms, by way of Solr, etc.  The Lucene project evolves rapidly, and it's a full-time job to keep up with the ever improving features and scalability.   This talk will distill and showcase the most relevant(!) advancements to date.&lt;br /&gt;
&lt;br /&gt;
== The Why and How of Very Large Displays in Libraries. ==&lt;br /&gt;
&lt;br /&gt;
* Cory Lown, NCSU Libraries, cwlown@ncsu.edu&lt;br /&gt;
&lt;br /&gt;
Previous Code4Lib Presentations:&lt;br /&gt;
* [http://code4lib.org/conference/2012/lown How People Search the Library from a Single Search Box]  2012&lt;br /&gt;
* [http://code4lib.org/conference/2010/orphanides_lown_lynema Enhancing Discoverability with Virtual Shelf Browse] 2010&lt;br /&gt;
&lt;br /&gt;
Built into the walls of NC State's new Hunt Library are several [http://www.christiedigital.com/en-us/digital-signage/products/microtiles/pages/microtiles-digital-signage-video-wall.aspx Christie MicroTile Display Wall Systems]. What does a library do with a display that's seven feet tall and over twenty feet wide? I'll talk about why libraries might want large displays like this, what we're doing with them right now, and what we might do with them in the future. I'll talk about how these displays factor into planning for new and existing web projects. And I'll get into the fun details of how you build web applications that scale from the very small browser window on a phone all the way up to a browser window with about 14 million pixels (about 10 million more than a dual 24&amp;quot; monitor desktop setup).&lt;br /&gt;
&lt;br /&gt;
== Discovering your Discovery System in Real Time. ==&lt;br /&gt;
&lt;br /&gt;
* Godmar Back, Virginia Tech, gback@vt.edu&lt;br /&gt;
* Annette Bailey, Virginia Tech, afbailey@vt.edu&lt;br /&gt;
&lt;br /&gt;
Practically all libraries today provide web-based discovery systems to their users;&lt;br /&gt;
users discover items and peruse or check them out by clicking on links.  Unlike&lt;br /&gt;
the traditional transaction of checking out a book at the circulation desk, this&lt;br /&gt;
interaction is largely invisible.  We have built a system that records user's&lt;br /&gt;
interaction with Summon in real-time, processes the resulting data with minimal delay,&lt;br /&gt;
and visualizes it in various ways using Google Charts and using various d3.js modules,&lt;br /&gt;
such as word clouds, tree maps, and others.&lt;br /&gt;
&lt;br /&gt;
These visualizations can be embedded in web sites, but are also suitable for&lt;br /&gt;
projection via large-scale displays or projectors right into the 'Learning Spaces'&lt;br /&gt;
many libraries are converted into.  The goal of this talk is to share the technology&lt;br /&gt;
and advocate the building of a cloud-based infrastructure that would make this&lt;br /&gt;
technology available to any library that uses a discovery system, rather than just&lt;br /&gt;
those who have the technological prowess for developing such systems and&lt;br /&gt;
visualizations in-house.  &lt;br /&gt;
&lt;br /&gt;
Previous presentations at Code4Lib:&lt;br /&gt;
* Talk: Code4Lib 2009 [http://code4lib.org/files/LibX2.0-Code4Lib-2009AsPresented.ppt LibX 2.0]&lt;br /&gt;
* Preconference: [http://wiki.code4lib.org/index.php/LibX_Preconference LibX 2.0, 2009]&lt;br /&gt;
* Preconference: Code4Lib 2010, On Widgets and Web Services&lt;br /&gt;
&lt;br /&gt;
== Your Library, Anywhere: A Modern, Responsive Library Catalogue at University of Toronto Libraries ==&lt;br /&gt;
&lt;br /&gt;
* Bilal Khalid, Gordon Belray, Lisa Gayhart (lisa.gayhart@utoronto.ca)&lt;br /&gt;
&lt;br /&gt;
* No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
With the recent surge in the mobile device market and an ever expanding patron base with increasingly divergent levels of technical ability, the University of Toronto Libraries embarked on the development of a new catalogue discovery layer to fit the needs of its diverse users. &lt;br /&gt;
&lt;br /&gt;
[http://search.library.utoronto.ca The result]: a mobile-friendly, flexible and intuitive web application that brings the full power of a faceted library catalogue to users without compromising quality or performance, employing Responsive Web Design principles. This talk will discuss: application development; service improvements; interface design; and user outreach, testing, and project communications. Feedback and questions from the audience are very welcome. If time runs short, we will be available for questions and conversation after the presentation.&lt;br /&gt;
&lt;br /&gt;
Note: A version of this content has been provisionally accepted as an article for Code4Lib Journal, January 2014 publication.)&lt;br /&gt;
&lt;br /&gt;
== All Tiled Up ==&lt;br /&gt;
&lt;br /&gt;
* Mike Graves, MIT Libraries (mgraves@mit.edu)&lt;br /&gt;
&lt;br /&gt;
You've got maps. You even scanned and georeferenced them. Now what? Running a full GIS stack can be expensive, and overkill in some cases. The good news is that you have a lot more options now than you did just a few years ago. I'd like to present some lighter weight solutions to making georeferenced images available on the Web.&lt;br /&gt;
&lt;br /&gt;
This talk will provide an introduction to MBTiles. I'll go over what they are, how you create them, how you use them and why you would use them.&lt;br /&gt;
&lt;br /&gt;
== The Great War: Image Interoperability to Facebook ==&lt;br /&gt;
&lt;br /&gt;
* Rob Sanderson, Los Alamos National Laboratory (azaroth42@gmail.com)&lt;br /&gt;
** (Code4Lib 2006: [http://www.code4lib.org/2006/sanderson | Library Text Mining])&lt;br /&gt;
* Rob Warren, Carleton University&lt;br /&gt;
** No previous presentations&lt;br /&gt;
&lt;br /&gt;
Using a pipeline constructed from Linked Open Data and other interoperability specifications, it is possible to merge and re-use image and textual data from distributed library collections to build new, useful tools and applications.  Starting with the OAI-PMH interface to ContentDM, we will take you on a tour through the International Image Interoperability Framework and Shared Canvas, to a cross-institutional viewer, and image analysis for the purposes of building a historical Facebook from finding and tagging people in photographs.  The World War One collections are drawn from multiple institutions and merged by the machine learning code.&lt;br /&gt;
&lt;br /&gt;
The presentation will focus on the (open source) toolchain and the benefits of the use of standards throughout:  OAI-PMH to get the metadata, IIIF for interaction with the images, the Shared Canvas ontology for describing collections of digitized objects, Open Annotation for tagging things in the images and specialized ontologies that are specific to the contents.  The tools include standard RDF / OWL technologies, JSON-LD, imagemagick and OpenCV for image analysis.&lt;br /&gt;
&lt;br /&gt;
== Visualizing Solr Search Results with D3.js for User-Friendly Navigation of Large Results Sets ==&lt;br /&gt;
&lt;br /&gt;
*Julia Bauder, Grinnell College Libraries (bauderj-at-grinnell-dot-edu)&lt;br /&gt;
*No previous presentations at national Code4Lib conferences&lt;br /&gt;
&lt;br /&gt;
As the corpus of articles, books, and other resources searched by discovery systems continues to get bigger, searchers are more and more frequently confronted with unmanageably large numbers of results. How can we help users make sense of 10,000 hits and find the ones they actually want? Facets help, but making sense of a gigantic sidebar of facets is not an easy task for users, either.&lt;br /&gt;
During this talk, I will explain how we will soon be using Solr 4’s pivot queries and hierarchical visualizations (e.g., treemaps) from D3.js to let patrons view and manipulate search results. We will be doing this with our VuFind 2.0 catalog, but this technique will work with any system running Solr 4. I will also talk about early student reaction to our tests of these visualization features.&lt;br /&gt;
&lt;br /&gt;
== PeerLibrary – open source cloud based collaborative library ==&lt;br /&gt;
&lt;br /&gt;
[https://github.com/peerlibrary/peerlibrary PeerLibrary is a new open source project] and a cloud service providing collaborative reading, sharing and storing. Users can upload publications they want to read (currently in PDF format), read them in the browser in real-time with others, highlight, annotate and organize their own or collaborative library. PeerLibrary provides a search engine to search over all uploaded open access publications. Additionally, it aims to collaboratively aggregate the open layer of knowledge on top of this publications through public annotations and references user will add to publications. In this way publications would not just be available to read, but accessible to the general public as well. Currently, it is aiming at scientific community and scientific publications.&lt;br /&gt;
&lt;br /&gt;
See [http://blog.peerlibrary.org/post/63458789185/screencast-previewing-the-peerlibrary-project screencast here].&lt;br /&gt;
&lt;br /&gt;
It is still in development and beta launch is planned at the end of November.&lt;br /&gt;
&lt;br /&gt;
== Who was where when, or finding biographical articles on Wikipedia by place and time ==&lt;br /&gt;
&lt;br /&gt;
* [http://morton-owens.info Emily Morton-Owens], The Seattle Public Library (presenting on work from NYU)&lt;br /&gt;
* No previous c4l presentations&lt;br /&gt;
&lt;br /&gt;
It's easy to answer the question &amp;quot;What important people were in Paris in 1939?&amp;quot; But what about Virginia in the 1750s or Scandinavia in the 14th century? I created a tool that allows you to search for biographies in a generally applicable way, using a map interface. I would like to present updates to my thesis project, which combines a crawler written in Java that extracts information from Wikipedia articles, with a MongoDB data store and a frontend in Python.&lt;br /&gt;
&lt;br /&gt;
The input to the project is freetext of entire articles in Wikipedia; this is important to allow us to pick up Benjamin Franklin not just in the single most obvious place of Philadelphia but also in London, Paris, Boston, etc. I can talk about my experiments disambiguating place names (approaches pioneered on newspaper articles were actually unhelpful on this type of text) and setting up a processing queue that does not become mired in the biographies of every human who ever played soccer. I also want to mitigate some of the implementation choices I made due to my academic deadline and improve the accuracy/usability.&lt;br /&gt;
&lt;br /&gt;
What I hope to show is that I was able to develop a novel and useful reference tool automatically, using fairly simple heuristics that are a far cry from hand-cataloging familiar to many librarians.&lt;br /&gt;
&lt;br /&gt;
You can try out [http://linserv1.cims.nyu.edu:48866/ the original version] (this server is inconveniently set to be updated/rebooted on 11/8--may be temporarily unavailable)&lt;br /&gt;
&lt;br /&gt;
== Good!, DRY, and Dynamic: Content Strategy for Libraries (Especially the Big Ones) ==&lt;br /&gt;
&lt;br /&gt;
*Michael Schofield, Nova Southeastern University Libraries, mschofield@nova.edu&lt;br /&gt;
*No previous code4lib presentations.&lt;br /&gt;
&lt;br /&gt;
The responsibilities of the #libweb are exploding [it’s a good thing] and it is no longer uncommon for libraries to manage or even home-grow multiple applications and sites. Often it is at this point where the web people begin to suffer the absence of a content strategy when, say, business hours need to be updated sitewide a half-dozen times.&lt;br /&gt;
&lt;br /&gt;
We were already feeling this crunch when we decided to further complicate the Nova Southeastern University Libraries by splitting the main library website into two. The Alvin Sherman Library, Research, and Information Technology Center is a unique joint-use facility that serves not only the academic community but the public of Broward County - and marketing a hyperblend of content through one portal just wasn't cutting it. With a web team of two, we knew that managing all this rehashed, disparate content was totally unsustainable.&lt;br /&gt;
&lt;br /&gt;
I want to share in this talk how I went about making our library content DRY (“don’t repeat yourself”): input content in one place--blurbs, policies, featured events, featured databases, book reviews, business hours, and so on.--and syndicate it everywhere - even, sometimes, dynamically target that content for specific audiences or context. It is a presentation that is a little about workflow, a little more about browser and context detection, a tangent about content-modeling the CMS, and a lot about APIs, syndication, and performance.&lt;br /&gt;
&lt;br /&gt;
== No code, no root, no problem? Adventures in SaaS and library discovery ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:erwhite@vcu.edu Erin White, VCU]&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
In 2012 VCU was an eager early adopter of Ex Libris' cloud service Alma as an ILS, ERM, link resolver, and single-stop, de-silo'd public-facing discovery tool. This has been a disruptive change that has shifted our systems staff's day-to-day work, relationships with others in the library, and relationships with vendors.&lt;br /&gt;
&lt;br /&gt;
I'll share some of our experiences and takeaways from implementing and maintaining a cloud service:&lt;br /&gt;
* Seeking disruption and finding it&lt;br /&gt;
* Changing expectations of service and the reality of unplanned downtime&lt;br /&gt;
* Communication and problem resolution with non-IT library staff&lt;br /&gt;
* Working with a vendor that uses agile development methodology&lt;br /&gt;
* Benefits and pitfalls of creating customizations and code workarounds&lt;br /&gt;
* Changes in library IT/coders' roles with SaaS&lt;br /&gt;
&lt;br /&gt;
...as well as thoughts on the philosophy of library discovery vs real-life experiences in moving to a single-search model.&lt;br /&gt;
&lt;br /&gt;
== Building for others (and ourselves):  the Avalon Media System ==&lt;br /&gt;
* [mailto:michael.klein@northwestern.edu Michael B Klein], Senior Software Developer, Northwestern University &lt;br /&gt;
** [http://code4lib.org/conference/2010/metz_klein Public Datasets in the Cloud] (code4lib 2010)&lt;br /&gt;
** [http://code4lib.org/conference/2013/klein-rogers The Avalon Media System: A Next Generation Hydra Head For Audio and Video Delivery] (code4lib 2013)&lt;br /&gt;
* [mailto:j-rudder@northwestern.edu Julie Rudder], Digital Initiatives Project Manager, Northwestern University&lt;br /&gt;
** no previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
[http://www.avalonmediasystem.org/ Avalon Media System] is a collaborative effort between development teams at Northwestern and Indiana Universities. Our goal is to produce an open source media management platform that works well for us, but is also widely adopted and contributed to by other institutions. We believe that building a strong user and contributor community is vital to the success and longevity of the project, and have developed the system with this goal in mind. We will share lessons learned, pains and successes we’ve had releasing two versions of the application since last year.  &lt;br /&gt;
&lt;br /&gt;
Our presentation will cover our experiences:&lt;br /&gt;
* providing flexible, admin-friendly distribution and installation options&lt;br /&gt;
* building with abstraction, customization and local integrations in mind&lt;br /&gt;
* prioritizing features (user stories)&lt;br /&gt;
* attracting code contributions from other institutions&lt;br /&gt;
* gathering community feedback &lt;br /&gt;
* creating a product rather than a bag of parts&lt;br /&gt;
&lt;br /&gt;
== How to check your data to provide a great data product? Data quality as a key product feature at Europeana ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:Peter.Kiraly@kb.nl Péter Király] portal backend developer, Europeana&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
[http://Europeana.eu/ Europeana.eu] - Europe's digital library, archive and museum - aggregates more than 30 million metadata records from more than 2200 institutions.  The records come from libraries, archives, museums and every other kind of cultural institution, from very different systems and metadata schemas, and are typically transformed several times until they are ingested into the Europeana data repository.  Europeana builds a consolidated database from these records, creating reliable and consistent services for end-users (a search portal, search widget, mobile apps, thematic sites etc.) and an API, which supports our strategic goeal of data for reuse in education, creative industries, and the cultural sector.  A reliable &amp;quot;data product&amp;quot; is thus at the core of our own software products, as well as those of our API partners.&lt;br /&gt;
&lt;br /&gt;
Much effort is needed to smooth out local differences in the metadata curation practice of our data providers. We need a solid framework to measure the consistency of our data and provide feedback to decision-makers inside and outside the organisation. We can also use this metrics framework to ask content providers to improve their own metadata. Of course, a data-quality-driven approach requires that we also improve the data transformation steps of the Europeana ingestion process itself. Data quality issues heavily define what new features we are able to create in our user interfaces and API, and might actually affect the design and implementation of our underlying data structure, the Europeana Data Model.&lt;br /&gt;
&lt;br /&gt;
In the presentation I briefly describe the Europeana metadata ingestion process, show the data quality metrics, the measuring techniques (using the Europeana API, Solr and MongoDB queries), some typical problems (both trivial and difficult ones), and finally the feedback mechanism we propose to deploy.&lt;br /&gt;
&lt;br /&gt;
Keywords: Europeana, data quality, EDM, API, Apache Solr, MongoDB, #opendata, #openglam&lt;br /&gt;
&lt;br /&gt;
== Teach your Fedora to Fly: scaling out your digital repository ==&lt;br /&gt;
&lt;br /&gt;
*[mailto:acoburn@amherst.edu Aaron Coburn], Software Developer, Amherst College&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
Fedora is a great repository system for managing large collections of digital objects, but what happens when a popular food magazine begins directing a large number of readers to a manuscript showing Emily Dickinson’s own recipe for doughnuts? While Fedora excels in its support of XML-based metadata, it doesn’t always perform well under a high volume of traffic. Nor is it especially tolerant of network or hardware failures.&lt;br /&gt;
&lt;br /&gt;
This presentation will show how we are making heavy use of a fedora repository while at the same time insulating it almost entirely from any web traffic. Starting with a distributed web front-end built with Node.js, and caching most of the user-accessible content from Fedora in an elastic, fault-tolerant Riak (NoSQL) cluster, we have eliminated nearly all single points of failure in the system. It also means that our production system is spread across twelve separate servers, where asynchrony and Map-Reduce are king. And aside from being blazing fast, it is also entirely Hydra-compliant.&lt;br /&gt;
&lt;br /&gt;
Furthermore, we will attempt to answer the question: if fedora crashes and the visitors to your site don’t notice, did it really fail?&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Using Open Source Software and Freeware to Preserve and Deliver Digital Videos ==&lt;br /&gt;
* [mailto:wfang@kinoy.rutgers.edu Wei Fang], Head of Digital Services, Rutgers University Law Library&lt;br /&gt;
* Jiebei Luo, Digital Projects Initiative Intern, Rutgers University&lt;br /&gt;
*No previous C4L presentations&lt;br /&gt;
&lt;br /&gt;
The Rutgers University Law Library is the official digital repository of the New Jersey Supreme Court oral arguments since 2002. This large video collection contains approximately 3,000 videos with a total of 400 GB or 6,000 viewing hours. With the expansion of this collection, the existing database and the static website could not efficiently support the library’s daily operations and meet its patrons’ search needs. &lt;br /&gt;
By utilizing open source software and freeware such as Ubuntu, FFmpeg, Solr and Drupal, the library is able to develop a complete solution to re-encoding videos, embedding subtitles, incorporating  Solr search engine and content management system to support full-text subtitle search, automatically updating video metadata records in the library catalog system and eventually providing a plug-in free HTML 5-based Web interface for patrons to view the videos online.&lt;br /&gt;
The aspects below will be presented in detail at the conference:&lt;br /&gt;
*	Video codecs comparison &lt;br /&gt;
*	Server-end batch video encoding/re-encoding&lt;br /&gt;
*	HTML 5 video tag and embedding subtitles&lt;br /&gt;
*	Incorporating search engine Solr and content management tool 	Drupal with the database to retrieve videos by full-text search especially in subtitle files&lt;br /&gt;
*	Incorporating video metadata with the library catalog system&lt;br /&gt;
&lt;br /&gt;
== Shared Vision, Shared Resources: the Curate Institutional Repository ==&lt;br /&gt;
* Dan Brubaker Horst, University of Notre Dame &lt;br /&gt;
** [http://code4lib.org/conference/2011/JohnsonHorst A Community-Based Approach to Developing a Digital Exhibit at Notre Dame Using the Hydra Framework] &lt;br /&gt;
* Julie Rudder, Northwestern University&lt;br /&gt;
** no previous presentations&lt;br /&gt;
&lt;br /&gt;
Curate is being collaboratively developed by several institutions in the Hydra community who share the need and vision for a Fedora-backed Institutional Repository. The first release of Curate was a collaboration between Notre Dame and Northwestern University, along with Digital Curation Experts (DCE) - a vendor hired jointly by our two institutions. Powered by the Hydra engine Sufia, the team worked quickly to release the first version of Curate in October 2013 which provides a basic self-deposit system that has support for various content types, collection building, DOI minting, and user profile creation. From the very beginning we have built Curate to be easy to theme and extend in order to ease the process of installation and use by other institutions.&lt;br /&gt;
&lt;br /&gt;
In December 2013, additional partners will join the project including: Indiana University, the University of Cincinnati and the University of Virginia. Each institution contributes resources to the project in order to further our common goal to create a product that fits our needs and has a sustainable future.Together we will tackle additional content types (like complex data, software, media), administrative collections and more. &lt;br /&gt;
&lt;br /&gt;
Our presentation will include:&lt;br /&gt;
* a brief demonstration of Curate and technical overview&lt;br /&gt;
* why and how we work together&lt;br /&gt;
* why build Curate&lt;br /&gt;
* the future of the project&lt;br /&gt;
&lt;br /&gt;
== Solr, Cloud and Blacklight ==&lt;br /&gt;
* David Jiao, Library Information Systems, Indiana University at Bloomington, djiao@indiana.edu&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
SolrCloud refers to the distributed capabilities in Solr4. It is designed to offer a highly available, fault tolerant environment by organizing data into multiple pieces that can be hosted on multiple machines with replicas, and providing a centralized cluster configuration and management. &lt;br /&gt;
&lt;br /&gt;
At Indiana University, we are upgrading our Solr backend for our recently released Blacklight-based OPAC system from Solr 1.4 to Solr4, and we also put up efforts to build a private cloud of Solr4 servers. In this talk, I will persent certain features of SolrCloud, including distributed requests, fault tolerance, near real time indexing/searching, and configuration management with Zookeeper, and our experiences of utilizing these features to provide better performance and architecture for our OPAC system, which serves over 7 million bibliographic records to over 100 thousand students and faculty members. I will also discuss some practical lessons learned from our SolrCloud setup/upgrade and the integration of the new SolrCloud to our customized Blacklight system.  &lt;br /&gt;
&lt;br /&gt;
== Leveraging XSD's for Reflective, Live Dataset Support in Institutional Repositories ==&lt;br /&gt;
* [mailto:msulliva@ufl.edu Mark Sullivan], Library Information Technology, University of Florida&lt;br /&gt;
** No previous code4lib presentations&lt;br /&gt;
&lt;br /&gt;
The University of Florida Libraries are currently adding support for active datasets into our METS-based institutional repository software.  This ongoing project enables the library to be a partner in current, or long-running, data-driven projects around the university by providing tangible short-term and long-term benefits to the projects.  The system assists project teams by storing and providing access to their data, while supporting online filtering and sorting of the data, custom queries, and adding and editing of the data by authorized users.  We are also exploring simple data visualizations to allow users to perform basic graphical and geographic queries.  Several different schemas were explored including DDI and EML, but ultimately the streamlined approach of using XSD's with some custom attributes was chosen, with all other data residing in the METS file portions.  Currently the system is being developed using XSD's describing XML datasets, but this model should easily scale to support SQL datasets or large datasets supported by Hadoop or iRODS.&lt;br /&gt;
&lt;br /&gt;
This work is being integrated in the open source [http://sobek.ufl.edu SobekCM Digital Content Management System] which is built on a pair-tree structure of METS resources with [http://ufdc.ufl.edu/design/webcontent/sobekcm/SobekCM_Resource_Object.pdf rich metadata support] including DC, MODS, MARC, VRACore, DarwinCore, IEE-LOM, GML/KML, schema.org microdata, and many other standard schemas.  The system has emphasized online, distributed creation and maintenance of resources including geo-placement and geographic searching of resources, building structure maps (table of contents) visually online, and a broad suite of curator tools.  &lt;br /&gt;
&lt;br /&gt;
This work is presented as a model which could be implemented in other systems as well.  We will demonstrate current support and discuss our upcoming roadmap to provide complete support.&lt;br /&gt;
&lt;br /&gt;
== Dead-simple Video Content Management: Let Your Filesystem Do The Work ==&lt;br /&gt;
&lt;br /&gt;
* Andreas Orphanides, NCSU Libraries (akorphan (at) ncsu.edu)&lt;br /&gt;
** (never led or soloed a C4L presentation)&lt;br /&gt;
&lt;br /&gt;
Content management is hard. To keep all the moving parts in order, and to maintain a layer of separation between the system and content creators (who are frequently not technical experts), we typically turn to content management systems like Drupal. But even Drupal and its kin require significant overhead and present a not inconsiderable learning curve for nontechnical users.&lt;br /&gt;
&lt;br /&gt;
In some contexts it's possible -- and desirable -- to manage content in a more streamlined, lightweight way, with a minimum of fuss and technical infrastructure. In this presentation I'll share a simple MVC-like architecture for managing video content for playback on the web, which uses a combination of Apache's mod_rewrite module and your server's filesystem structure to provide an automated approach to video content management that's easy to implement and provides a low barrier to content updates: friendly to content creators and technology implementors alike. Even better, the basic method is HTML5-friendly, and can be integrated into your favorite content management system if you've got permissions for creating templates.&lt;br /&gt;
&lt;br /&gt;
In the presentation I'll go into detail about the system structure and logic required to implement this approach. I'll detail the benefits and limitations of the system, as well as the challenges I encountered in developing its implementation. Audience members should come away with sufficient background to implement a similar system on their own servers. Implementation documentation and genericized code will also be shared, as available.&lt;br /&gt;
&lt;br /&gt;
== Managing Discovery ==&lt;br /&gt;
&lt;br /&gt;
* Andrew Pasterfield, Senior Programmer/Systems Analyst, University of Calgary Library, ampaster@ucalgary.ca&lt;br /&gt;
**No previous code4lib presentations &amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
In fall 2012 the University of Calgary Library launched a new home page that incorporated a Summon powered&lt;br /&gt;
Single Search Box with customized “bento box” results display. Search at the U of C now combines a range of&lt;br /&gt;
metadata sources for discovery and customized mapping of a database recommender and LibGuide into a unified&lt;br /&gt;
display.  Further customizations include a non Google Analytics/non proxy method to log clicks.&amp;lt;br&amp;gt;&amp;lt;br&amp;gt;&lt;br /&gt;
&lt;br /&gt;
This presentation will discuss the technical details of bringing the various systems together into one display interface to increase discovery at the U of C Library.&lt;br /&gt;
&lt;br /&gt;
http://library.ucalgary.ca&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== Sorting it out: a piece of the User Centered Design Process ==&lt;br /&gt;
&lt;br /&gt;
* Cindy Beggs, [http://www.akendi.com/aboutus/management/ Akendi], cindy@akendi.com&lt;br /&gt;
&lt;br /&gt;
This talk is about how to apply a user centered design methodology to the process of creating an information architecture.  Participants learn the fundamentals of UCD and how card sorting and reverse card sorting enable us to isolate the content we present on screen from the layouts and visuals of those screens.  We talk about ways to identify who will be using the information architecture you are creating and why we need to know how it will be used.&lt;br /&gt;
 &lt;br /&gt;
What will attendees takes away from your talk?&lt;br /&gt;
The criticality of involving “real” end users in the process of creating an information architecture.  The basics of following a user-centered-design process in the creation of best in class, content-rich, digital products.&lt;br /&gt;
&lt;br /&gt;
Cindy Beggs has been working in the “information industry” for over 25 years.  A librarian by profession, she has spent decades helping users figure out how to find their way through large bodies of content.  Her insights into how people seek information, her empathy for those who find it a challenge and her practical experience helping organizations figure out how to best structure their content contribute to her success as an information architect with both clients and trainees.  (http://www.akendi.com/aboutus/management/)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
==Implementation of ArchivesSpace in University of Richmond==&lt;br /&gt;
&lt;br /&gt;
*Birong Ho, bho@richmond.edu&lt;br /&gt;
&lt;br /&gt;
University of Richmond implemented its archive collection management ArchivsSpace in the fall, 2013. As a charter member and the Head of Special Collection as the Board member, implementation of such an Open Source Software became a priority. &lt;br /&gt;
&lt;br /&gt;
Several aspects of implementation will be addressed in the talk. Among them, they are Collections and Repository, storage layer including data format, System resources requirements, Technical architecture, Customization, scaling and integrated with other systems in the library.&lt;br /&gt;
&lt;br /&gt;
The customization, scale, and integration with other systems such as Archeon and Exist on campus became a concern will be focused and elaborated in the talk.&lt;br /&gt;
&lt;br /&gt;
==Easy Wins for Modern Web Technologies in Libraries==&lt;br /&gt;
&lt;br /&gt;
*[mailto:trey.terrell@oregonstate.edu Trey Terrell], Analyst Programmer, Oregon State University&lt;br /&gt;
** No previous Code4Lib presentations &lt;br /&gt;
&lt;br /&gt;
Oregon State University is currently implementing an updated version of its room reservation system. In its development we've come across and implemented a variety of &amp;quot;easy wins&amp;quot; to make it more responsive, easier to maintain, less expensive to run, and just cooler to experience. While our particular system was in Ruby on Rails, this talk will address general methods and example utilities which can be used no matter your stack.&lt;br /&gt;
&lt;br /&gt;
I'll be talking about things like cache management, reverse proxies, publish/subscribe servers, WebSockets, responsive design, asynchronous processing, and keeping complicated stacks up and running with minimal effort.&lt;br /&gt;
&lt;br /&gt;
==Implementing Islandora at a Small Institution==&lt;br /&gt;
&lt;br /&gt;
*Megan Kudzia, Albion College Library&lt;br /&gt;
*Eddie Bachle, Albion College IT&lt;br /&gt;
**No previous Code4Lib presentations&lt;br /&gt;
&lt;br /&gt;
Albion College (and particularly the Library/Archives and Special Collections) has a variety of needs which could be met by an open-source Institutional Repository system. Several months and lots of conversations later, we’re continuing to troubleshoot our way through Islandora. We’d like to talk about what has worked for us, where our frustrations have been, whether it’s even possible to install and develop a system like this at a small institution, and where the process has stalled. &lt;br /&gt;
&lt;br /&gt;
As of right now, we do have a semi-working installation. We’re not sure when it will be ready for our end users, but we'll talk about our development process and evaluate our progress.&lt;br /&gt;
''Contributions also by Nicole Smeltekop, Albion College Archives &amp;amp; Special Collections''&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
== PhantomJS+Selenium: Easy Automated Testing of AJAX-y UIs ==&lt;br /&gt;
&lt;br /&gt;
* Martin Haye, California Digital Library, martin.haye@ucop.edu&lt;br /&gt;
** Previous Code4Lib Presentation: [http://code4lib.org/conference/2012/collett Beyond code: Versioning data with Git and Mercurial] at Code4Lib 2012 (Martin co-presenting with Stephanie Collett)&lt;br /&gt;
* Mark Redar, California Digital Library, mark.redar@ucop.edu&lt;br /&gt;
&lt;br /&gt;
Web user interfaces are demanding ever-more dynamism and polish, combining HTML5, AJAX, lots of CSS and jQuery (or ilk) to create autocomplete drop-downs, intelligent buttons, stylish alert dialogs, etc. How can you make automated tests for these highly complex and interactive UIs?&lt;br /&gt;
&lt;br /&gt;
Part of the answer is PhantomJS. It’s a modern WebKit browser that’s “headless” (meaning it has no display) that can be driven from command-line Selenium unit tests. PhantomJS is dead simple to install, and its blazing speed and server-friendliness make continuous integration testing easy. You can write UI unit tests in {language-of-your-choice} and run them not just in PhantomJS but in Firefox and Chrome, plus a zillion browser/OS combinations at places like SauceLabs, TestingBot and BrowserStack.&lt;br /&gt;
&lt;br /&gt;
In this double-team live code talk, we’ll explain all that while we demonstrate the following in real time:&lt;br /&gt;
&lt;br /&gt;
* Start with nothing.&lt;br /&gt;
* Install Selenium bindings for Ruby and Python.&lt;br /&gt;
* In each language write a small test of an AJAX-y UI.&lt;br /&gt;
* Run the tests in Firefox, and fix bugs (in the test or UI) as needed.&lt;br /&gt;
* Install PhantomJS.&lt;br /&gt;
* Show the same tests running headless as part of a server-friendly test suite. &lt;br /&gt;
* (Wifi permitting) Show the same tests running on a couple different browser/OS combinations on the server cloud at SauceLabs – talking through a tunnel to the local firewalled application.&lt;br /&gt;
&lt;br /&gt;
==New Technologies, Collaboration, &amp;amp; Entrepreneurship in Libraries:  Harnessing Their Power to Help Your Library==&lt;br /&gt;
&lt;br /&gt;
* Stephanie Walker – swalker@brooklyn.cuny.edu&lt;br /&gt;
* Howard Spivak – howards@brooklyn.cuny.edu&lt;br /&gt;
* Alex - Alex@brooklyn.cuny.edu&lt;br /&gt;
&lt;br /&gt;
Academic libraries are caught in budget squeezes and often struggle to find ways to communicate value to senior administration and others.  At Brooklyn College Library, we have taken an unusual, possibly unique, approach to these issues.  Our technology staff have long worked directly with librarians to develop products that meet library, faculty, and student needs, and we have shared many of our products with colleagues, including an award-winning website, e-resource, and content management system we call 4MyLibrary, which we shared for free with 8 CUNY colleges, and also an easy-to-use book scanner, which has proven overwhelming popular with students, faculty, other librarians, and numerous campus offices.  Recently, motivated by budget cuts, we decided that what worked for us might interest other libraries, and working with our Office of Technology Commercialization, we started selling 2 products:  our book scanners (at half the price of commercial alternatives), and a hosting service, whereby we could host and support 4MyLibrary for libraries with minimal technology staff.  Both succeeded, and yielded major benefits:  a steady revenue stream and the admiration and serious goodwill of our senior administration and others.   However, this presentation is neither a basic how-to, nor an advertisement.  With this presentation, we hope to spur a conversation for broader collaboration, especially regarding new technologies, among libraries.  We all have some level of technical expertise, most of us are struggling with rising prices and tight budgets, and many of us are unhappy with various technology products we use, from scanners to our ILS.  We believe – and can demonstrate – that with collaboration, we can solve many of our problems, and provide better services to boot. &lt;br /&gt;
&lt;br /&gt;
== Identifiers, Data, and Norse Gods ==&lt;br /&gt;
&lt;br /&gt;
* Ryan Scherle, Dryad Digital Repository, ryan@datadryad.org&lt;br /&gt;
&lt;br /&gt;
ORCID and DataCite are provide stable identifiers for researchers and and data, respectively. Each system does a fine job of providing value to its users. But wouldn't it be great if they could link their systems to create something much more powerful? Perhaps even as powerful as a god?&lt;br /&gt;
&lt;br /&gt;
Enter ODIN, The ORCID and DataCite Interoperability Network. ODIN is a two-year project to unleash the power of persistent identifiers for researchers and the research they create. This talk will present recent work from the ODIN project, including several tools that can be used to unleash the godlike power of identifiers at your institution.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[:Category:Code4Lib2014]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=11551</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=11551"/>
				<updated>2012-03-09T20:29:10Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Initial design/prototype by: Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
== UI Prototype (uses static data) ==&lt;br /&gt;
http://dl.dropbox.com/u/9074989/code4lib/unverified.html&lt;br /&gt;
&lt;br /&gt;
== Basic design ==&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for an initial usable version:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
** Useful tool for finding PubMed IDs: http://www.ncbi.nlm.nih.gov/entrez/getids.cgi&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
** Unfortunately, it seems this wouldn't work out.  In order to perform querying of external authorities, they'd all need to support [http://en.wikipedia.org/wiki/JSON#JSONP JSONP] or similar (and they don't)&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
=== Draft Matching Algorithm ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
function compareRecords(localDubCore, authDubCore)&lt;br /&gt;
    recordMatches = []&lt;br /&gt;
    for each element-type:&lt;br /&gt;
        loc = array of local values&lt;br /&gt;
        auth = array of authority values&lt;br /&gt;
        // arrays are actually lists of dictionaries&lt;br /&gt;
        // a1&lt;br /&gt;
        // 0    value=&amp;quot;Benson, Arnold&amp;quot;, match=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
        // 1    value=&amp;quot;Terrence, D.&amp;quot;, match=&amp;quot;a2[3]&amp;quot;, strength=&amp;quot;100%&amp;quot;&lt;br /&gt;
        elementMatches = compareElements(loc, auth)&lt;br /&gt;
        recordMatches.add(elementMatches)&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
function compareElements(loc, auth)&lt;br /&gt;
    output = []&lt;br /&gt;
    //nested loops run through values and assigns strongest matches to each element&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        for each element in auth&lt;br /&gt;
            strength = string distance between the two elements&lt;br /&gt;
            if strength = 100&lt;br /&gt;
                //if match is perfect go ahead pop each element and add their values to output array&lt;br /&gt;
                //output array is also list of dictionaries&lt;br /&gt;
                //0    loc=&amp;quot;Hector&amp;quot;, auth=&amp;quot;Hector&amp;quot;, strength=&amp;quot;100&amp;quot;&lt;br /&gt;
                //1    loc=&amp;quot;Albert&amp;quot;, auth=&amp;quot;Alberto&amp;quot;, strength=&amp;quot;90&amp;quot;&lt;br /&gt;
           if strength &amp;gt; auth element's current strength value&lt;br /&gt;
               overwrite auth element's strength and match values&lt;br /&gt;
           if strength &amp;gt; loc element's current strength value&lt;br /&gt;
               overwrite loc element's strength and match values&lt;br /&gt;
    //this second set of non-nested loops pull out the strongest matches&lt;br /&gt;
    for each item in auth&lt;br /&gt;
        //x = some arbitrary barrier for a decent enough match&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the loc list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    for each item in loc&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the auth list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    //now do cleanup and look for values that have no decent matches&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;Heyward&amp;quot;, auth=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    for each element in auth&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;&amp;quot;, auth=&amp;quot;Perry&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    return output           &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
## crosswalk from authority format to simple dc&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=8581</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=8581"/>
				<updated>2011-06-03T20:21:35Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
== UI Prototype (uses static data) ==&lt;br /&gt;
http://dl.dropbox.com/u/9074989/code4lib/unverified.html&lt;br /&gt;
&lt;br /&gt;
== Basic design ==&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for an initial usable version:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
** Useful tool for finding PubMed IDs: http://www.ncbi.nlm.nih.gov/entrez/getids.cgi&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
** Unfortunately, it seems this wouldn't work out.  In order to perform querying of external authorities, they'd all need to support [http://en.wikipedia.org/wiki/JSON#JSONP JSONP] or similar (and they don't)&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
=== Draft Matching Algorithm ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
function compareRecords(localDubCore, authDubCore)&lt;br /&gt;
    recordMatches = []&lt;br /&gt;
    for each element-type:&lt;br /&gt;
        loc = array of local values&lt;br /&gt;
        auth = array of authority values&lt;br /&gt;
        // arrays are actually lists of dictionaries&lt;br /&gt;
        // a1&lt;br /&gt;
        // 0    value=&amp;quot;Benson, Arnold&amp;quot;, match=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
        // 1    value=&amp;quot;Terrence, D.&amp;quot;, match=&amp;quot;a2[3]&amp;quot;, strength=&amp;quot;100%&amp;quot;&lt;br /&gt;
        elementMatches = compareElements(loc, auth)&lt;br /&gt;
        recordMatches.add(elementMatches)&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
function compareElements(loc, auth)&lt;br /&gt;
    output = []&lt;br /&gt;
    //nested loops run through values and assigns strongest matches to each element&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        for each element in auth&lt;br /&gt;
            strength = string distance between the two elements&lt;br /&gt;
            if strength = 100&lt;br /&gt;
                //if match is perfect go ahead pop each element and add their values to output array&lt;br /&gt;
                //output array is also list of dictionaries&lt;br /&gt;
                //0    loc=&amp;quot;Hector&amp;quot;, auth=&amp;quot;Hector&amp;quot;, strength=&amp;quot;100&amp;quot;&lt;br /&gt;
                //1    loc=&amp;quot;Albert&amp;quot;, auth=&amp;quot;Alberto&amp;quot;, strength=&amp;quot;90&amp;quot;&lt;br /&gt;
           if strength &amp;gt; auth element's current strength value&lt;br /&gt;
               overwrite auth element's strength and match values&lt;br /&gt;
           if strength &amp;gt; loc element's current strength value&lt;br /&gt;
               overwrite loc element's strength and match values&lt;br /&gt;
    //this second set of non-nested loops pull out the strongest matches&lt;br /&gt;
    for each item in auth&lt;br /&gt;
        //x = some arbitrary barrier for a decent enough match&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the loc list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    for each item in loc&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the auth list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    //now do cleanup and look for values that have no decent matches&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;Heyward&amp;quot;, auth=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    for each element in auth&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;&amp;quot;, auth=&amp;quot;Perry&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    return output           &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
## crosswalk from authority format to simple dc&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7546</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7546"/>
				<updated>2011-02-17T04:24:42Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
** Useful tool for finding PubMed IDs: http://www.ncbi.nlm.nih.gov/entrez/getids.cgi&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
** Unfortunately, it seems this wouldn't work out.  In order to perform querying of external authorities, they'd all need to support [http://en.wikipedia.org/wiki/JSON#JSONP JSONP] or similar (and they don't)&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
=== Draft Matching Algorithm ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
function compareRecords(localDubCore, authDubCore)&lt;br /&gt;
    recordMatches = []&lt;br /&gt;
    for each element-type:&lt;br /&gt;
        loc = array of local values&lt;br /&gt;
        auth = array of authority values&lt;br /&gt;
        // arrays are actually lists of dictionaries&lt;br /&gt;
        // a1&lt;br /&gt;
        // 0    value=&amp;quot;Benson, Arnold&amp;quot;, match=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
        // 1    value=&amp;quot;Terrence, D.&amp;quot;, match=&amp;quot;a2[3]&amp;quot;, strength=&amp;quot;100%&amp;quot;&lt;br /&gt;
        elementMatches = compareElements(loc, auth)&lt;br /&gt;
        recordMatches.add(elementMatches)&lt;br /&gt;
        &lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
function compareElements(loc, auth)&lt;br /&gt;
    output = []&lt;br /&gt;
    //nested loops run through values and assigns strongest matches to each element&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        for each element in auth&lt;br /&gt;
            strength = string distance between the two elements&lt;br /&gt;
            if strength = 100&lt;br /&gt;
                //if match is perfect go ahead pop each element and add their values to output array&lt;br /&gt;
                //output array is also list of dictionaries&lt;br /&gt;
                //0    loc=&amp;quot;Hector&amp;quot;, auth=&amp;quot;Hector&amp;quot;, strength=&amp;quot;100&amp;quot;&lt;br /&gt;
                //1    loc=&amp;quot;Albert&amp;quot;, auth=&amp;quot;Alberto&amp;quot;, strength=&amp;quot;90&amp;quot;&lt;br /&gt;
           if strength &amp;gt; auth element's current strength value&lt;br /&gt;
               overwrite auth element's strength and match values&lt;br /&gt;
           if strength &amp;gt; loc element's current strength value&lt;br /&gt;
               overwrite loc element's strength and match values&lt;br /&gt;
    //this second set of non-nested loops pull out the strongest matches&lt;br /&gt;
    for each item in auth&lt;br /&gt;
        //x = some arbitrary barrier for a decent enough match&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the loc list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    for each item in loc&lt;br /&gt;
        if element strength &amp;gt; x AND if matching element is still in the auth list&lt;br /&gt;
            pop each element and add their values to output array&lt;br /&gt;
    //now do cleanup and look for values that have no decent matches&lt;br /&gt;
    for each element in loc&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;Heyward&amp;quot;, auth=&amp;quot;&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    for each element in auth&lt;br /&gt;
        pop element and add to output array without match //x   loc=&amp;quot;&amp;quot;, auth=&amp;quot;Perry&amp;quot;, strength=&amp;quot;&amp;quot;&lt;br /&gt;
    return output           &lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== UI Example (static) ==&lt;br /&gt;
http://dl.dropbox.com/u/9074989/code4lib/unverified.html&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
## crosswalk from authority format to simple dc&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=C4L2011_social_activities&amp;diff=7514</id>
		<title>C4L2011 social activities</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=C4L2011_social_activities&amp;diff=7514"/>
				<updated>2011-02-10T02:52:27Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Werewolf Signup */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;'''This page is under development'''&lt;br /&gt;
&lt;br /&gt;
==Code4Lib 2011 social activities ==&lt;br /&gt;
&lt;br /&gt;
[http://maps.google.com/maps/ms?hl=en&amp;amp;ie=UTF8&amp;amp;msa=0&amp;amp;msid=201419377696104618083.000499eeb466dfb962201&amp;amp;ll=39.169964,-86.53008&amp;amp;spn=0.025286,0.03077&amp;amp;z=15 Code4Lib 2011 Google Map] - It's the ultimate plot device!  See what others are recommending and recommend your own places to see and things to do!&lt;br /&gt;
&lt;br /&gt;
[http://www.imu.indiana.edu/pdfs/imu/pdfs/IMU%20Map%202011.pdf Map of the Indiana Memorial Union] - The conference is happening [http://www.imu.indiana.edu/event/AlumniHall.mov here], and [http://www.imu.indiana.edu/img/rooms/whittenberger.jpg here], [http://www.imu.indiana.edu/img/rooms/solarium.jpg here], and [http://www.imu.indiana.edu/event/TreeSuitesRoom.mov here].  Use this map and you can probably find it all.&lt;br /&gt;
&lt;br /&gt;
==Local events==&lt;br /&gt;
&lt;br /&gt;
* [https://wiki.dlib.indiana.edu/confluence/display/EVENTS/Code4Lib+2011+Proposal Original Proposal (suggests some nearby events)]&lt;br /&gt;
* February 4-5, 8-12: [http://www.indiana.edu/~thtr/productions/2010/angelsInAmerica.shtml Angels in America: Part One] at IU Wells-Metz Theatre&lt;br /&gt;
* February 5: [http://music.indiana.edu/events/?e=9221 New Music Ensemble] at Auer Hall, 8 p.m, with [http://www.parnasmusic.com/Index.html duo parnas] performing.&lt;br /&gt;
* February 6: [http://www.indiana.edu/~iucinema/calendar.shtml Who the #$&amp;amp;% is Jackson Pollock] and [http://www.indiana.edu/~iucinema/events2011.21.shtml And Everything is Going Fine] at the new IU Cinema&lt;br /&gt;
* February 7: [http://www.indiana.edu/~iucinema/events2011.33.shtml Reign of Terror] at the new IU Cinema&lt;br /&gt;
* February 7: [http://music.indiana.edu/events/?e=9224 Jazz Ensemble] at Musical Arts Center, 8 p.m.&lt;br /&gt;
* February 7: [https://onestart.iu.edu/ccl-prd/EventMaintenance.do?methodToCall=viewEvent&amp;amp;eventId=488273&amp;amp;pubCalId=GRP1445 Michael Chabon] at Fine Arts Auditorium, 5:30 p.m.&lt;br /&gt;
* February 8: [http://www.theroommovie.com/screeningspop.html Tommy Wiseau's Love Is Blind Tour] (showing of &amp;quot;[[wikipedia:The Room (film)|The Room]]&amp;quot; w/ Q&amp;amp;amp;A!)&lt;br /&gt;
* February 8: [http://music.indiana.edu/events/share.php?e=9225 IU Wind Ensemble] at Musical Arts Center - FREE!&lt;br /&gt;
* February 9: [http://music.indiana.edu/events/share.php?e=9226 IU University Orchestra] at Musical Arts Center - FREE!&lt;br /&gt;
* February 8 and 9: [http://www.iuauditorium.com/site/show-fiddler.html Fiddler on the Roof] at the IU Auditorium - $38-60&lt;br /&gt;
* February 9: [http://dylanettinger.bandcamp.com/ Dylan Ettinger]/[http://www.myspace.com/kamkama Kam Kama] at The Bishop &lt;br /&gt;
* February 10: Bob Marley's band [http://wailers.com/ Legendary Wailers] at [http://www.thebluebird.ws/ Bluebird Nightclub] - $20&lt;br /&gt;
* February 10: Amy Schumer at [http://www.comedyattic.com/index.php?option=com_k2&amp;amp;view=item&amp;amp;id=84:amy-schumer&amp;amp;Itemid=3 The Comedy Attic] - $13&lt;br /&gt;
&lt;br /&gt;
Things to see:&lt;br /&gt;
* Indiana University Art Museum: [http://www.iub.edu/~iuam/section.php?navSection=galleries New in the Galleries]&lt;br /&gt;
* Indiana University SoFA Gallery: [http://www.indiana.edu/~sofa/exhibitions/iu-school-of-fine-arts-student-shows-1/ MFA Painting, Metals, Graphic Design, Ceramics, and Textiles]&lt;br /&gt;
* [http://www.indiana.edu/~mathers/=Indiana University Mathers Museum]&lt;br /&gt;
&lt;br /&gt;
==Recommended Restaurants/Bars (no particular order)==&lt;br /&gt;
&lt;br /&gt;
* [http://www.the-uptown.com/ Uptown Café] - great for breakfast, lunch, or wine&lt;br /&gt;
* [http://www.villagedeli.biz/ Village Deli] - can accommodate large crowds for breakfast or lunch&lt;br /&gt;
* [http://grazieitalianeatery.com/Welcome.html Grazie] - Italian, great wine selection&lt;br /&gt;
* Japanee (320 N. Walnut St) - Bento Box lunch – yum!&lt;br /&gt;
* [http://www.maxsplace.info/ Max’s Place] - pizza and beer&lt;br /&gt;
* [http://www.samirasrestaurant.com/ Samira] - Afghanistan cuisine – lunch buffet w/ roasted chicken&lt;br /&gt;
* Shanti (221 E. Kirkwood Ave) - great for lunch&lt;br /&gt;
* [http://www.esanthairestaurant.com/ Esan Thai] - delicious, but slow service&lt;br /&gt;
* [http://www.cafedjango.us/ Café Django] - Indian and Thai, great noodle dishes&lt;br /&gt;
* [http://www.stefanoscafe.com/ Stefano’s Ice Café] - best chicken salad sandwich in town&lt;br /&gt;
* [http://www.thelaughingplanetcafe.com/ Laughing Planet] - yummy burritos/nachos&lt;br /&gt;
* [http://www.bbcbagel.com/ Bloomington Bagel Co.] - best bagels in town&lt;br /&gt;
* RockIt's Pizza (222 N. Walnut St) - open late for a slice after visiting nearby bars&lt;br /&gt;
* [http://www.farm-bloomington.com/ Farm] - fun little hipster whiskey bar in basement&lt;br /&gt;
* [http://www.nicksenglishhut.com/ Nick's] - local meat, in-house batter for deep fried goodness&lt;br /&gt;
* [http://www.bbcbloomington.com/ Lennie’s Bar &amp;amp; Grill] - local brew&lt;br /&gt;
* [http://squaredonuts.com Square Donuts] - donuts that are square and fresh and you eat them zomg&lt;br /&gt;
* [http://uplandbeer.com/ Upland Brewing Co.] - more local brew (different from Lennie's)&lt;br /&gt;
* [http://www.irishlion.com/ The Irish Lion] - Irish pub, Guinness on tap&lt;br /&gt;
* [http://www.bloomingpedia.org/wiki/Bloomington_Sandwich_Company Bloomington Sandwich Co.] - great for lunch, yummy reuben&lt;br /&gt;
* Bub’s Burgers and Ice Cream (480 N. Morton St) - eat the giganto burger in a certain time limit, get your picture on the wall (if that's what you're into)&lt;br /&gt;
* [http://www.yogis.com/ Yogi's Grill &amp;amp; Bar] - bazillion beer choices&lt;br /&gt;
* Restaurant Ami (1500 E. 3rd St) - Japanese/Korean, great for lunch&lt;br /&gt;
* [http://www.turkuazcafe.com/ Turkuaz Cafe] - Turkish, pides great for lunch&lt;br /&gt;
* [http://www.crazyhorseindiana.com/ Crazy Horse] - also bazillion beer choices&lt;br /&gt;
* [http://www.finchsbrasserie.com/ Finch's] - great wine selection&lt;br /&gt;
&lt;br /&gt;
==Planned events==&lt;br /&gt;
&lt;br /&gt;
Plan one if you like! Either on your own or you can [[2011 committees_sign-up_page|join the social activities committee]].&lt;br /&gt;
&lt;br /&gt;
=== [[Craft Brew Drinkup]] ===&lt;br /&gt;
''''Tuesday 2/8, 8:30PM, Hospitality suite'''' Like good beer? Bring some in your luggage! Some of us are planning on bringing on bottles of our favorite local brews to share. Interested? Sign up on the [[Craft Brew Drinkup]] page!&lt;br /&gt;
&lt;br /&gt;
=== Newcomer Dinner ===&lt;br /&gt;
First time at code4lib? Join fellow c4l newbies and veterans for an evening of food, socializing, and stimulating &amp;lt;strike&amp;gt;discussions about&amp;lt;/strike&amp;gt; demonstrations of the many uses of &amp;lt;strike&amp;gt;bacon&amp;lt;/strike&amp;gt; dongles.&lt;br /&gt;
&lt;br /&gt;
Code4Lib veterans, you're invited too. Join us in welcoming the newcomers!&lt;br /&gt;
&lt;br /&gt;
'''Plans'''&lt;br /&gt;
* When: Monday evening (2/7)&lt;br /&gt;
* Time: 6 PM (ish)&lt;br /&gt;
* Mastermind (if you have any questions): [mailto:yoosebj@muohio.edu Becky Yoose]&lt;br /&gt;
&lt;br /&gt;
''Guidelines:''&lt;br /&gt;
*Max of '''6''' per location&lt;br /&gt;
**Please, no waitlisting :(&lt;br /&gt;
*ID yourselves so we can jget a good mix of new people and veterans in each group&lt;br /&gt;
**New folks - n&lt;br /&gt;
**c4l vets - v&lt;br /&gt;
*One leader needed for each location (declare yourself! - '''Vets are highly encouraged to lead the group :)''')&lt;br /&gt;
**Leader duties&lt;br /&gt;
***Make reservations if required; otherwise make sure that the restaurant can handle a group of 6 rowdy library coders &lt;br /&gt;
***Herd folks from hotel to restaurant (know where you're going!)&lt;br /&gt;
&lt;br /&gt;
'''Restaurants'''&lt;br /&gt;
&lt;br /&gt;
'''West Side of Campus (towards downtown)'''&lt;br /&gt;
&lt;br /&gt;
'''Indiana Avenue''', across street from campus (5-10 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.buffalouies.com/home.html Buffa Louie's] (Wings/Subs/Sandwiches) - Gables location on Indiana Avenue is historic site&lt;br /&gt;
&lt;br /&gt;
'''4th St. between IMU and downtown square''' (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.anyetsangs.com/ Anyetsang's Little Tibet] (Tibetan/Thai/Indian)&lt;br /&gt;
&lt;br /&gt;
**MEET ON THE MEZZANINE LEVEL NEAR THE COMFY CHAIRS**&lt;br /&gt;
&lt;br /&gt;
*Dot Porter (leader) - n&lt;br /&gt;
*Jason Ronallo - n&lt;br /&gt;
*Ben Anderson - n&lt;br /&gt;
*Bill Dueber - v&lt;br /&gt;
*Jakub Skoczen - n&lt;br /&gt;
*[mailto:birkin_diana@brown.edu Birkin] - v&lt;br /&gt;
&lt;br /&gt;
[http://www.siamhousebloomington.com/ Siam House] (Thai) ''Meet in the Biddle hotel lobby at 6pm. Look for the short woman in a trench coat and wide brim hat ~Becky''&lt;br /&gt;
*Becky Yoose (leader) - v&lt;br /&gt;
*Margaret Heller - n&lt;br /&gt;
*Bohyun Kim - n&lt;br /&gt;
*Karen Hanson - v&lt;br /&gt;
*Daniel Lovins - v&lt;br /&gt;
*Gerald Snyder - v&lt;br /&gt;
*&amp;lt;strike&amp;gt;Wayne Schneider - v&amp;lt;/strike&amp;gt; off to hear [https://onestart.iu.edu/ccl-prd/EventMaintenance.do?methodToCall=viewEvent&amp;amp;eventId=488273&amp;amp;pubCalId=GRP1445 Michael Chabon] instead.&lt;br /&gt;
&lt;br /&gt;
Puccini's La Dolce Vita (Italian)&lt;br /&gt;
&lt;br /&gt;
'''Kirkwood Ave. (5th St) between IMU and downtown square''' (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.nicksenglishhut.com/ Nick's English Hut] (American Pub) - NOTE: web site plays IU fight song&lt;br /&gt;
&lt;br /&gt;
[http://www.cafepizzaria.com/ Cafe Pizzaria] (Pizza)&lt;br /&gt;
&lt;br /&gt;
[http://www.falafelsonline.com/Falafels/www.FalafelsOnline.com.html Falafels] (Middle Eastern)&lt;br /&gt;
&lt;br /&gt;
[http://www.finchsbrasserie.com/ Finch's] (Gastropub-ish) Meet in the mezzanine lounge at 6 PM, carefully avoiding contact with the man-eating comfy chairs. ''(8 min walk. Res. at 6:15 PM)''&lt;br /&gt;
* [mailto:cgordon@chillco.com Cary Gordon] (leader) - v (818-694-1626) ''(big guy wearing yellow jacket)''&lt;br /&gt;
* Michael Doran - v&lt;br /&gt;
* Tim Daniels - n&lt;br /&gt;
* Joshua Gomez - n&lt;br /&gt;
* Jenny Reiswig - n&lt;br /&gt;
* Kosuke Tanabe - n&lt;br /&gt;
&lt;br /&gt;
Note: FARMbloomingon is closed on Mondays&lt;br /&gt;
&lt;br /&gt;
[http://www.the-uptown.com/ Michael's Uptown Cafe] (Cajun/Creole/American) ''To walk to restaurant, meet in lounge upstairs from hotel lobby in IMU at 6pm.  Reservation for 6:30pm at Uptown (under Julie Hardesty).''&lt;br /&gt;
*[mailto:jlhardes@indiana.edu Julie Hardesty] (leader) - n&lt;br /&gt;
*Jean Rainwater - v&lt;br /&gt;
*D Ruth Bavousett - n&lt;br /&gt;
* Theodor T - n&lt;br /&gt;
*Sarah Weeks - n&lt;br /&gt;
*Mark Mounts - v&lt;br /&gt;
* Takanori Hayashi -n&lt;br /&gt;
&lt;br /&gt;
[http://www.thetrojanhorse.com/ Trojan Horse] (Greek)&lt;br /&gt;
&lt;br /&gt;
Shanti (Indian, Vegetarian options)&lt;br /&gt;
&lt;br /&gt;
'''6th St. between IMU and downtown square''' (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.runciblespoonrestaurant.com/ Runcible Spoon] (Variety, Vegetarian/Vegan options)&lt;br /&gt;
&lt;br /&gt;
'''Grant St. between 3rd St. and 6th St.''', about halfway between IMU and downtown square (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://cafedjango.us/ Cafe Django] (Indian, Thai, Vegetarian options) ''Meet in the Biddle hotel lobby at 6pm.  Look for the guy with the bright blue knitted Mets hat.''&lt;br /&gt;
* Maccabee Levine (leader) - v&lt;br /&gt;
* Andrew Darby - v&lt;br /&gt;
* Michael Slone - n&lt;br /&gt;
* Linda Ballinger - n&lt;br /&gt;
* Nell Taylor - n&lt;br /&gt;
* Richard Anderson - n&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[http://www.bloomingpedia.org/wiki/Snow_Lion Snow Lion] (Tibetan, Indian, Vegetarian options)&lt;br /&gt;
&lt;br /&gt;
'''Lincoln St. between 3rd St. and 6th St.''', about halfway between IMU and downtown square (15-20 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.esanthairestaurant.com/ Esan Thai] (Thai, Vegetarian options) ''Meet outside Whittenberger Auditorium at 6pm''&lt;br /&gt;
* Ryan Scherle (leader) - v&lt;br /&gt;
* Jason Stirnaman (drummer) - v&lt;br /&gt;
* Jon Dunn - n&lt;br /&gt;
* Richard Maiti - n&lt;br /&gt;
* Mike Stroming - n&lt;br /&gt;
* Trish Rose-Sandler (can drive if needed) - n&lt;br /&gt;
&lt;br /&gt;
'''Downtown Square''' (20-25 minute walk)&lt;br /&gt;
&lt;br /&gt;
&amp;lt;strike&amp;gt;[http://uplandbeer.com/ Upland Brewing]&amp;lt;/strike&amp;gt; (Microbrewery)&lt;br /&gt;
&lt;br /&gt;
'''CHANGE OF PLAN: Going to [http://yogis.com/ Yogi's] instead; meet in lobby at 5:30PM'''&lt;br /&gt;
&lt;br /&gt;
:Note: Lennie's (below) is also a Microbrewery/restaurant&lt;br /&gt;
&lt;br /&gt;
* THE ARCHITECT'S GROUP&lt;br /&gt;
** Mike Giarlo (leader) - v&lt;br /&gt;
** Jay Luker (wall dodger) - v&lt;br /&gt;
** Benoit Thiell (french tickler) - n&lt;br /&gt;
** Dan Suchy (designated non-driver) - v&lt;br /&gt;
** Matt Critchlow (lupulin addict) - n&lt;br /&gt;
** Esme Cowles (designated non-walker) -v&lt;br /&gt;
* THE ARCHIVIST'S GROUP&lt;br /&gt;
** Mark Matienzo (leader) - v&lt;br /&gt;
** Joe Atzberger (agitator) - v&lt;br /&gt;
** Hillel Arnold (intern) - n&lt;br /&gt;
** Mark Custer (title tk) - n&lt;br /&gt;
** Patrick Force (TBD) - n&lt;br /&gt;
** Adam Wead (zymurgist) - n&lt;br /&gt;
** Christopher Chagal (#7 apparently, but I had to squeeze into one of these groups.  Exhibit A: [http://awesome.good.is/transparency/web/1102/beer-map/flat.html The United States of GOOD beer]) - v&lt;br /&gt;
* THE DIRECTOR'S GROUP (where DIRECTOR = people cooler than Giarlo or Matienzo.  Which is pretty much anyone.  Except McDonald.)&lt;br /&gt;
** Declan Fleming - v&lt;br /&gt;
** Mark Phillips - v&lt;br /&gt;
** Tim Donohue (not a &amp;quot;director&amp;quot;, per se, but I enjoy microbrew beer) - v&lt;br /&gt;
** &amp;lt;strike&amp;gt;Andrea Schurr - n&amp;lt;/strike&amp;gt; - Last minute ship jumping so that I can go to Lennie's.  Have fun!&lt;br /&gt;
** Larry Baerveldt (also not a director, but I like beer, and this group isn't full yet) - n&lt;br /&gt;
** Shaun Ellis (all I know is that I'm following Declan when beer is involved) - n &lt;br /&gt;
** Jon Stroop - v&lt;br /&gt;
&lt;br /&gt;
[http://www.irishlion.com/ Irish Lion] (Irish Pub)&lt;br /&gt;
&lt;br /&gt;
[http://www.crazyhorseindiana.com/ Crazy Horse] (American) ''Meet near Walnut Meeting Room (M015) at 6pm''&lt;br /&gt;
* Joel Richard (leader, usurped. richardjm at si dot edu) - v &lt;br /&gt;
* Genevieve Francis - n&lt;br /&gt;
* Ben Shum (bshum AT biblio DOT org)- n&lt;br /&gt;
* Francis Kayiwa - (former leader francis dot kayiwa AT gmail) n&lt;br /&gt;
* Roberto Hoyle - n&lt;br /&gt;
* Hui Zhang - n (hz3 AT indiana DOT edu)&lt;br /&gt;
&lt;br /&gt;
[http://www.grazieitalianeatery.com/Welcome.html Grazie] (Italian)&lt;br /&gt;
&lt;br /&gt;
[http://www.samirasrestaurant.com/ Samira] (Afghanistan cuisine) - ''let's meet at 6pm in the mezzanine (above the hotel lobby)''&lt;br /&gt;
* Ranti Junus (ranti dot junus at gmail) (lead) - v&lt;br /&gt;
* Andreas Orphanides - v&lt;br /&gt;
* Matt Connolly - v ( back in the game.)&lt;br /&gt;
* Toke Eskildsen - intermediate&lt;br /&gt;
* Mads Villadsen - intermediate&lt;br /&gt;
* Will Kurt - n&lt;br /&gt;
&lt;br /&gt;
[http://www.opietaylors.com/welcome.html Opie Taylors] (American)&lt;br /&gt;
&lt;br /&gt;
[http://www.malibugrill.net/ Malibu Grill] (California)&lt;br /&gt;
&lt;br /&gt;
[http://www.littlezagreb.com/ Janko's Little Zagreb] (Steakhouse)&lt;br /&gt;
&lt;br /&gt;
[http://www.bloomingpedia.org/wiki/El_Norteño El Norteño] (Mexican)&lt;br /&gt;
&lt;br /&gt;
[http://www.maxsplace.info/ Max's Place] (Pizza)&lt;br /&gt;
&lt;br /&gt;
[http://www.restauranttallent.com/ Tallent] (Local/Seasonal)&lt;br /&gt;
''7:00 Reservation for 2 tables, 6 people each; will remix at dessert. Leave from hotel lobby at 6:45''&lt;br /&gt;
&lt;br /&gt;
Leaving Hotel Lobby at 6:25.  Otherwise, meet us there. [http://maps.google.com/maps?f=d&amp;amp;source=s_d&amp;amp;saddr=IU+memorial+union&amp;amp;daddr=208+North+Walnut+Street,+Bloomington,+IN+47404+(Restaurant+Tallent)&amp;amp;hl=en&amp;amp;geocode=FfamVQIdwsXX-iElSoPsShsQPA%3BFfynVQIdlJrX-iFsH61zxeNixA&amp;amp;mra=ltm&amp;amp;dirflg=w&amp;amp;sll=39.167352,-86.527956&amp;amp;sspn=0.009283,0.01811&amp;amp;ie=UTF8&amp;amp;ll=39.1682,-86.527956&amp;amp;spn=0.009283,0.01811&amp;amp;z=16| directions]&lt;br /&gt;
&lt;br /&gt;
ULCERATIVE LOONS GROUP&lt;br /&gt;
* Matt Zumwalt (leader) - v&lt;br /&gt;
* Michael Levy - n&lt;br /&gt;
* Devon Smith - v&lt;br /&gt;
* Sean Hannan - v&lt;br /&gt;
* Steven Miles - n&lt;br /&gt;
* Naomi Dushay - v&lt;br /&gt;
&lt;br /&gt;
TROUBLEMAKERS GROUP&lt;br /&gt;
* Eric Hellman (leader- camo ski jacket) - v&lt;br /&gt;
* Michael Klein (troublemaker) - v&lt;br /&gt;
* Matt Cordial - v&lt;br /&gt;
* Scot Colford (natalie wood to round out the james deans of troublemakers) - v &lt;br /&gt;
* Rachel Frick&lt;br /&gt;
* Ken Irwin - (semi-v) &lt;br /&gt;
&lt;br /&gt;
Roots on the Square (Vegan/Vegetarian)&lt;br /&gt;
&lt;br /&gt;
'''East Side of Campus (away from downtown)'''&lt;br /&gt;
&lt;br /&gt;
'''10th St., east of Wells Library''' (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.bbcbloomington.com/ Lennie's Bar &amp;amp; Grill] (Pizza/American) + (Microbrewery) - ''Meet in the Biddle hotel lobby at 6:15pm. Reservation made for 6 at 6:45pm''&lt;br /&gt;
&lt;br /&gt;
* [mailto:ryanwick@gmail.com Ryan Wick] (leader) - v&lt;br /&gt;
* June Rayner - n&lt;br /&gt;
* Roni Shwaish - n&lt;br /&gt;
* Kirk Hess (I'm a local &amp;amp; I think Lennie's has better beer &amp;amp; food than Upland.) - n&lt;br /&gt;
* Aaron Collie (Listens to locals; wonders who will be the last person to sign up) - n&lt;br /&gt;
* Jon Gorman - v - I know I make 7, but hope that's not a huge issue. If it turns out to be I'll grab a bar seat.&lt;br /&gt;
Tried to make a reservation for another group, but no go. Will be going to Yogi's.&lt;br /&gt;
* Chelsea Lobdell - n&lt;br /&gt;
* Andrea Schurr - n - jumped ship from Yogi's and made a reservation &lt;br /&gt;
* Mark Leggott - n    (would be 10, if we're doing 12?)&lt;br /&gt;
* Zac Howarth - n   11?&lt;br /&gt;
     (taking the bull by the horns, I just made a second reservation for 6 at 6:45 - pub-side)&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''3rd St., east of Music School''' (10-15 minute walk)&lt;br /&gt;
&lt;br /&gt;
[http://www.motherbearspizza.com/index2.html Mother Bear's Pizza] (Pizza)&lt;br /&gt;
&lt;br /&gt;
[http://www.macgyros.com/ The MAC] (Mediterranean)&lt;br /&gt;
&lt;br /&gt;
[http://www.yelp.com/biz/red-chopsticks-bloomington Red Chopsticks] (Sushi/Asian Fusion)&lt;br /&gt;
&lt;br /&gt;
Cafe Ami/Domo (Japanese/Korean)&lt;br /&gt;
&lt;br /&gt;
=== &amp;quot;Social Network Dine Arounds&amp;quot; ===&lt;br /&gt;
&lt;br /&gt;
Wednesday night.  Not sure what these are.  Make it up as you go.  Be social, network, dine around.&lt;br /&gt;
&lt;br /&gt;
'''Snow Lion Tibetan Restaurant (run by Dalai Rama's brother's neighbor!!!)'''&lt;br /&gt;
&lt;br /&gt;
Wednesday 6PM meeting at the mezzannine above the lobby &lt;br /&gt;
&lt;br /&gt;
The only agenda is relaxing and having a good time :))&lt;br /&gt;
[http://local.yahoo.com/info-16065573-snow-lion-restaurant-bloomington More info]&lt;br /&gt;
* Ranti Junus&lt;br /&gt;
* Becky Yoose&lt;br /&gt;
* Bohyun Kim&lt;br /&gt;
* Dileshni Jayasinghe&lt;br /&gt;
* Will Kurt&lt;br /&gt;
* Ian Mulvany&lt;br /&gt;
&lt;br /&gt;
'''Dine with Hydra'''&lt;br /&gt;
Wednesday Evening &lt;br /&gt;
&lt;br /&gt;
Location: [http://www.irishlion.com/ Irish Lion] upstairs&lt;br /&gt;
Time: 6:30-8:00&lt;br /&gt;
&lt;br /&gt;
Join us if you want to connect with the [http://wiki.duraspace.org/display/hydra/The+Hydra+Project Hydra] developer community.  Questions, ideas, suggestions, etc. all welcome.  Please add yourself to the list if you're coming.&lt;br /&gt;
&lt;br /&gt;
* Matt Zumwalt&lt;br /&gt;
* Bess Sadler&lt;br /&gt;
* Adam Wead &lt;br /&gt;
* Jeremy Nelson&lt;br /&gt;
* Rick Johnson&lt;br /&gt;
* Dan Brubaker Horst&lt;br /&gt;
* Scot Colford&lt;br /&gt;
* Mike Stroming&lt;br /&gt;
* Christopher Curry&lt;br /&gt;
* Banu Lakshminarayanan&lt;br /&gt;
* Richard Anderson&lt;br /&gt;
* Michael Levy&lt;br /&gt;
* Jason Stirnaman&lt;br /&gt;
* Daniel Lovins&lt;br /&gt;
&lt;br /&gt;
'''Anyetsang's Little Tibet''' Wednesday, leaving from the mezzanine above the lobby at 6:30 PM.  Menu available (in a really cheesy set of pictures of the actual menu [[http://www.anyetsangs.com/|right here]].  No agenda, no plan for discussion, just social time and Tibetan/Thai/Indian grub.  Add yourself if you want, or just show up.  If you sign up, we won't leave without you!&lt;br /&gt;
* D Ruth Bavousett (suggested it, so is going)&lt;br /&gt;
* Mark Mounts&lt;br /&gt;
* Roberto Hoyle&lt;br /&gt;
* Ian Walls&lt;br /&gt;
* Keith Nickum&lt;br /&gt;
* Ken Irwin&lt;br /&gt;
* Sarah Weeks&lt;br /&gt;
* Jason Fowler&lt;br /&gt;
* Ben Anderson&lt;br /&gt;
&lt;br /&gt;
'''Run to White Castle''' We need cars. &lt;br /&gt;
* Declan Fleming&lt;br /&gt;
* Matt Critchlow&lt;br /&gt;
&lt;br /&gt;
'''[http://uplandbeer.com/ Upland Brewing]''' - Leaving from lobby by hotel registration desk at 6:30 PM&lt;br /&gt;
&lt;br /&gt;
Just throwing this out there to gauge interest. No specific agenda.&lt;br /&gt;
&lt;br /&gt;
* Hillel Arnold&lt;br /&gt;
* Mark Matienzo&lt;br /&gt;
* Andrea Schurr - I was just going to do the same thing.  I'm in.&lt;br /&gt;
* Jon Gorman - i'd be in. Do they serve food? Ya betcha - just linked to website.&lt;br /&gt;
* Chelsea Lobdell - I might be a little later than 6:30 so I will meet everyone there. &lt;br /&gt;
* Joe Atzberger (atz)&lt;br /&gt;
* Wayne Schneider - will meet you there&lt;br /&gt;
* gluejar (Me)&lt;br /&gt;
* Spencer Lamm&lt;br /&gt;
* Benoit Thiell - I'll meet you there.&lt;br /&gt;
* Patrick Force&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''[http://www.finchsbrasserie.com/ Finch's]'''&lt;br /&gt;
Meeting in lobby at 7, leaving no later then 7:15 - reservation made for 8 at 7:30&lt;br /&gt;
* decasm&lt;br /&gt;
* ndushay&lt;br /&gt;
* cbeer&lt;br /&gt;
* MrDys&lt;br /&gt;
* jrochkind&lt;br /&gt;
* tburton-west&lt;br /&gt;
* Mads Villadsen&lt;br /&gt;
* Toke Eskildsen&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''UNC/Dook Basketball Game'''&lt;br /&gt;
Wednesday - TIPOFF @ 9pm - [http://www.nicksenglishhut.com/ Nick's English Hut] - just show up!&lt;br /&gt;
&lt;br /&gt;
'''Thursday breakfast at Village Deli''' Meet in hotel lobby 7:45 AM.&lt;br /&gt;
* Devon Smith/decasm&lt;br /&gt;
* Jason Stirnaman/jstirnaman&lt;br /&gt;
* Keith Nickum&lt;br /&gt;
&lt;br /&gt;
[[Category: Code4Lib2011]]&lt;br /&gt;
&lt;br /&gt;
=== Hanging out in the Hospitality Suite ===&lt;br /&gt;
* Located in Biddle Hotel room 116/117&lt;br /&gt;
* Keyholders: Gabriel Farrell (gsf, @g5f), Mike Giarlo (mjgiarlo, @mjgiarlo), Mark Matienzo (anarchivist, @anarchivist), Devon Smith (decasm)&lt;br /&gt;
&lt;br /&gt;
== Werewolf! ==&lt;br /&gt;
It wouldn't be a tech conference unless we got together one evening to turn into a gang of murdering beasts and hyper-suspicious victims. Facilitated by the one and only mbklein.&lt;br /&gt;
&lt;br /&gt;
* When: Wednesday evening&lt;br /&gt;
* Time: 10-10:30 PM (or whenever enough people wander in)&lt;br /&gt;
* Where: Planning to Meet in Hospitality Suite and move on from there is necessary. Check this space for updates.&lt;br /&gt;
&lt;br /&gt;
=== About Werewolf ===&lt;br /&gt;
&lt;br /&gt;
Werewolf (also known as Mafia) is a parlor game that has become [http://www.wired.co.uk/wired-magazine/archive/2010/03/features/werewolf.aspx the obsession of techie conferences everywhere]. At it's most basic, it's a game of information asymmetry -- a battle between an uninformed majority (the townspeople) and an informed minority (the werewolves who live and hunt among them). At its best, it's an off-the-wall paranoid screaming match. There are dozens of variations -- we'll start with the basics, and depending on everyone's stamina and desire to keep playing, save the tricky stuff for later. Hopefully by the end of the evening all the participants will be jibbering, jumpy, sleep-deprived lunatics incapable of trusting even their closest friends.&lt;br /&gt;
&lt;br /&gt;
In other words, ''good times.''&lt;br /&gt;
&lt;br /&gt;
=== Werewolf Signup ===&lt;br /&gt;
&lt;br /&gt;
If you're planning on/thinking about attending, put your name here.&lt;br /&gt;
&lt;br /&gt;
# Michael Klein&lt;br /&gt;
# anarchivist&lt;br /&gt;
# cbeer&lt;br /&gt;
# knickum&lt;br /&gt;
# vampire1&lt;br /&gt;
# vampire2&lt;br /&gt;
# mike D&lt;br /&gt;
# Sarah Weeks&lt;br /&gt;
# my middle name is werewolf&lt;br /&gt;
# do we really hafta sign up?&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=2011_Breakout_Sessions&amp;diff=7410</id>
		<title>2011 Breakout Sessions</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=2011_Breakout_Sessions&amp;diff=7410"/>
				<updated>2011-02-09T16:58:27Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;''''''NOTE: Breakout sessions are usually proposed at the conference or shortly befrore the conference begins&amp;quot;''&lt;br /&gt;
&lt;br /&gt;
Those interested in the same project/problem can hang out in a space together for 70 minute blocks. Generally the person who suggests the topic will take on the role as moderator to begin and moderate the discussion. Anyone can propose a breakout session - please think about whether you would want a session to be held on Tuesday or Wednesday, depending on the order of talks and who you hope will attend. There are lots of spaces in the IMU where small groups can congregate and we do have a couple of rooms including the large Alumni Hall space for this. We will route different proposed sessions to the different rooms depending on a quick show-of-hands survey just before each one begins.&lt;br /&gt;
&lt;br /&gt;
This page will list any sessions proposed, but there will also be flip charts outside the meeting room where more sessions can be proposed. '''Please include your name when proposing a session.&lt;br /&gt;
&lt;br /&gt;
'''Tuesday 16:00-17:00:''' Alumni Hall (1-2 groups); Solarium (2-3 groups); Whittenberger Auditorium&lt;br /&gt;
&lt;br /&gt;
* Drupal in Libraries - Cary Gordon and a cast of &amp;lt;strike&amp;gt;1,000s&amp;lt;/strike&amp;gt; &amp;lt;strike&amp;gt;100s&amp;lt;/strike&amp;gt; some - Solarium. &lt;br /&gt;
* Digital Video for research - William Cowan, Indiana University - Can we get beyond YouTube? - Solarium&lt;br /&gt;
* Plone / Zope in Libraries - Maccabee Levine, University of Wisconsin Oshkosh - Alumni Hall&lt;br /&gt;
* Usability research/designing for user experience - what are we doing about it? - Erin White, VCU - Solarium&lt;br /&gt;
* [[Can we hack on this: Open Extensible Proxy: going beyond EZProxy?]] - Terry Reese (Oregon State) and Jeremy Frumkin (University of Arizona) - Alumni Hall&lt;br /&gt;
&lt;br /&gt;
'''Wednesday 14:40-15:50:''' Alumni Hall (1-2 groups); Solarium (2-3 groups); Maple Room; Walnut Room&lt;br /&gt;
&lt;br /&gt;
*  Solr--Tom Burton-West University of Michigan (HathiTrust)&lt;br /&gt;
* Supporting Open Source in Libraries - Peter Murray and Tim Daniels, LYRASIS&lt;br /&gt;
* Digital Library Federation Community Update - Rachel L. Frick, CLIR/DLF&lt;br /&gt;
* Student thesis self-submission with Pylons, Fedora Commons, &amp;amp;amp; MARC - Jeremy Nelson (Colorado College)&lt;br /&gt;
* Collection, appraisal and tools like the [https://github.com/UNC-Libraries/Curators-Workbench Curator's Workbench] - Greg Jansen, UNC Chapel Hill - Solarium&lt;br /&gt;
* How OpenSocial apps and open APIs/data can personalize and enhance search and discovery on ScienceDirect,Scopus,Hub- Remko Caprio, Developer Platform Evangelist at [http://developer.sciverse.com SciVerse] &lt;br /&gt;
* Blacklight/Hydra - Bess Sadler &amp;amp; Matt Zumwalt&lt;br /&gt;
* [[Can we hack on this: Open Extensible Proxy: going beyond EZProxy?]] Part II - Reese, Frumkin, et al.&lt;br /&gt;
* VuFind Discussion - Demian Katz, Villanova University&lt;br /&gt;
* Data curation - Eric Lease Morgan, and I am hoping we can discuss and share best practices for curating research data across the enterprise. Copyright. Storage. Metadata. Privacy. Etc.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
'''Not yet scheduled:''' ''Organizers: Please move your sessions to Tuesday or Wednesday above.''&lt;br /&gt;
&lt;br /&gt;
* ColdFusion in Libraries - Daria Norris, Free Library of Philadelphia&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
[[Category:Code4Lib2011]]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7274</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7274"/>
				<updated>2011-02-07T20:23:59Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Need to do */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
## crosswalk from authority format to simple dc&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7273</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7273"/>
				<updated>2011-02-07T20:23:18Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Need to do */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7272</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7272"/>
				<updated>2011-02-07T20:22:59Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Need to do */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Create basic code framework&lt;br /&gt;
# Implement metadata retrieval from authority ''(done for crossref in ryan's code)''&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7270</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7270"/>
				<updated>2011-02-07T20:19:23Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Need to do */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Create basic code framework&lt;br /&gt;
# Implement metadata retrieval from authority (done for crossref in ryan's code)&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7269</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7269"/>
				<updated>2011-02-07T20:18:58Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Need to do */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
* 1 to 1 matching.  Even if you want to eventually match with multiple authorities, you'd only do one at a time&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Output Spec ==&lt;br /&gt;
&lt;br /&gt;
* We will use a simple XML output consisting of paired (and possibly unpaired) values.&lt;br /&gt;
* The root element will contain an attribute signifying the source of the authority metadata.&lt;br /&gt;
* The &amp;lt;match&amp;gt; element will be used to pair values, with a strength attribute to signify the string distance.&lt;br /&gt;
* Within each match element will be exactly 2 metadata elements with attributes signifying the source of each value: either the local input or the remote authority data.&lt;br /&gt;
* An &amp;lt;nonmatch&amp;gt; element will be used for unpaired values.&lt;br /&gt;
&lt;br /&gt;
=== Sample Output ===&lt;br /&gt;
&amp;lt;pre&amp;gt;&lt;br /&gt;
&amp;lt;hamr authority=&amp;quot;PubMed&amp;quot;&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;100%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;input&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
        &amp;lt;creator src=&amp;quot;authority&amp;quot;&amp;gt;Trojan, Tommy&amp;lt;/creator&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;match strength=&amp;quot;90%&amp;quot;&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;input&amp;quot;&amp;gt;Great American Article&amp;lt;/title&amp;gt;&lt;br /&gt;
        &amp;lt;title src=&amp;quot;authority&amp;quot;&amp;gt;Great American Article, The&amp;lt;/title&amp;gt;&lt;br /&gt;
    &amp;lt;/match&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;subject src=&amp;quot;input&amp;quot;&amp;gt;Medical Stuff&amp;lt;/subject&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
    &amp;lt;nonmatch&amp;gt;&lt;br /&gt;
        &amp;lt;type src=&amp;quot;authority&amp;quot;&amp;gt;text&amp;lt;/type&amp;gt;&lt;br /&gt;
    &amp;lt;/nonmatch&amp;gt;&lt;br /&gt;
&amp;lt;/hamr&amp;gt;&lt;br /&gt;
&amp;lt;/pre&amp;gt;&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Create basic code framework&lt;br /&gt;
# Implement metadata retrieval from authority&lt;br /&gt;
# Design structure of plugins&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7251</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7251"/>
				<updated>2011-02-07T19:11:35Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Code */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;br /&gt;
&lt;br /&gt;
== Need to do ==&lt;br /&gt;
&lt;br /&gt;
# Create basic code framework&lt;br /&gt;
# Implement metadata retrieval from authority&lt;br /&gt;
# Design matching algorithm&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7231</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7231"/>
				<updated>2011-02-07T18:41:11Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: /* Code */&lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [http://gitref.org/ quick reference for Git]&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7230</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7230"/>
				<updated>2011-02-07T18:40:01Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;br /&gt;
&lt;br /&gt;
== Code ==&lt;br /&gt;
&lt;br /&gt;
* [https://github.com/ryscher/hamr Ryan's really stupid scratch implementation]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7217</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7217"/>
				<updated>2011-02-07T16:12:41Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Narrowing the focus for today:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7216</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7216"/>
				<updated>2011-02-07T16:11:59Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Focus:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* PubMed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* CrossRef&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
** [http://labs.crossref.org/site/quick_and_dirty_api_guide.html OpenURL search]&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;br /&gt;
&lt;br /&gt;
Thoughts / Questions:&lt;br /&gt;
* Is there a way to do most/all of this via Javascript/AJAX/JQuery?  Could it be a simple Javascript framework you could &amp;quot;drop&amp;quot; into any metadata editing interface?&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	<entry>
		<id>https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7212</id>
		<title>HAMR: Human/Authority Metadata Reconciliation</title>
		<link rel="alternate" type="text/html" href="https://wiki.code4lib.org/index.php?title=HAMR:_Human/Authority_Metadata_Reconciliation&amp;diff=7212"/>
				<updated>2011-02-07T15:59:50Z</updated>
		
		<summary type="html">&lt;p&gt;Ryscher: &lt;/p&gt;
&lt;hr /&gt;
&lt;div&gt;[[HAMR: Human/Authority Metadata Reconciliation]]&lt;br /&gt;
&lt;br /&gt;
Sean Chen, Tim Donohue, Joshua Gomez, Ranti Junus, Ryan Scherle&lt;br /&gt;
&lt;br /&gt;
A tool for a curator to determine whether the various fields of a metadata record are correct. Takes a metadata record, locates any identifiers (e.g., DOI, PMID). Retrieves a copy of the metadata record from an authoritative source (e.g., CrossRef, PubMed). Displays a human-readable page that compares fields in the initial record with fields in the authoritative record. Each field is color-coded based on how well it matches, so the curator can quickly identify discrepancies.&lt;br /&gt;
&lt;br /&gt;
&lt;br /&gt;
Focus:&lt;br /&gt;
* Dublin core (maybe qualified)&lt;br /&gt;
* framework that allows multiple authority sources&lt;br /&gt;
* NOT focusing on author names ([http://www.orcid.org/ ORCID] is already working on this), except the fact that they are strings, and we'll do basic string matching&lt;br /&gt;
&lt;br /&gt;
Possible authority sources:&lt;br /&gt;
* pubmed&lt;br /&gt;
** Sample pubmed query (in Java): [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-PubmedPrefillStep.java DSpace PubMedPrefillStep.java] (From [https://wiki.duraspace.org/display/DSPACE/PopulateMetadataFromPubMed Populate Metadata from PubMed])&lt;br /&gt;
*** See 'retrievePubmedXML()' in above java code for actual call to PubMed&lt;br /&gt;
*** Mapping happens here: See [https://wiki.duraspace.org/display/DSPACE/PubMedPrefill-pmid+dim.xsl pmid-to-dim.xsl] for a sample XSLT crosswalk to translate PubMed format to a qualified dublin core (internal DSpace metadata format)&lt;br /&gt;
** More examples of querying PubMed: http://www.my-whiteboard.com/how-to-automate-pubmed-search-using-perl-php-or-java/&lt;br /&gt;
* crossref&lt;br /&gt;
** simply send the DOI to crossref, and get JSON/XML back&lt;br /&gt;
*** http://api.labs.crossref.org/10.1111/j.1558-5646.2009.00626.x.json&lt;br /&gt;
*** http://api.labs.crossref.org/10.2307/1935157.xml&lt;br /&gt;
*** [http://code.google.com/p/dryad/source/browse/trunk/dryad/dspace/modules/doi/dspace-doi-webapp/src/main/java/org/dspace/doi/DOIServlet.java java code that includes a lookup]&lt;br /&gt;
** [http://labs.crossref.org/site/crossref_metadata_search.html Metadata Search] -- send a text query, receive a list of matching records&lt;br /&gt;
* google scholar - does it have an API?&lt;br /&gt;
* [http://www.mendeley.com mendeley] - [http://dev.mendeley.com/ Mendeley API]&lt;br /&gt;
* [http://vivoweb.org/ vivo]&lt;br /&gt;
* [http://bibapp.org/ bibapp]&lt;/div&gt;</summary>
		<author><name>Ryscher</name></author>	</entry>

	</feed>