Changes

Jump to: navigation, search

2010talks Submissions

16,236 bytes added, 19:09, 17 November 2009
m
Adding 2010 category
== Submissions Deadline for 20-Minute Talk Slots ==talk submission was '''Friday, November 13'''. Edits to existing proposals are no longer allowed as these are being processed for the voting system.
'''Please follow the formatting guidelines:'''
Edit this page to submit your proposal for a 20-minute talk at the Code4Lib 2010 Conference. For more information, see the [[2010talkscall_Call_for_Submissions|Call for submissions]]. '''Please follow the formatting guidelines:'''<pre>'''== Talk Title:'''==
* Speaker'''Speaker s name(s), affiliation(s), and email address(es):'''  * Second speaker'''Abstract of no more than 500 words:''' Place your submission at the bottom of the page below this line:----s name, affiliation, email address, if second speaker
'''Talk Title:'''Abstract of no more than 500 words.</pre>
Mobile Web App Design: Getting Started
== Mobile Web App Design: Getting Started ==
'''Speaker name, affiliation, and email address:''' * Michael Doran, University of Texas at Arlington, doran@uta.edu, http://rocky.uta.edu/doran/  '''Abstract:'''
Creating or adapting library web applications for mobile devices such as the iPhone, Android, and Palm Pre is not hard, but it does require learning some new tools, new techniques, and new approaches. From the Tao of mobile web app design to using mobile device SDKs for their emulators, this presentation will give you a jump-start on mobile cross-platform design, development, and testing. And all illustrated with a real-world mobile library web application.
----
'''Talk Title== Drupal 7:'''A more powerful platform for building library applications ==
Drupal 7: A more powerful platform for building library applications* Cary Gordon, The Cherry Hill Company, cgordon@chillco.com
'''Speaker nameThe release of Drupal 7 brings with it a big increase in utility for this already very useful and well-accepted content management framework. Specifically, the addition of fields in core, the inclusion of RDFa, affiliationthe use of the PHP_db abstraction layer, and email address:'''the promotion of files to first class objects facilitate the development of richer applications directly in Drupal without the need to integrate external products.
Cary Gordon, The Cherry Hill Company, cgordon@chillco.com
'''Abstract== Fiwalk with Me:'''Using Automatic Forensics Tools and Python for Digital Curation Triage == * Mark Matienzo, The New York Public Library, mark@matienzo.org
The release of Drupal 7 brings with it a big increase Building on Simson Garfinkel's work in utility for this already very useful Automated Document and well-accepted content management framework. SpecificallyMedia Exploitation (ADOMEX), the addition this project investigates digital curation applications of fields open source tools used in coredigital forensics. Specifically, the inclusion of RDFawe will be using [http://afflib.org AFFLib]'s fiwalk ("file and inode walk") application and its corresponding Python library to develop a basic triage workflow for accessioned hard drives, the use of the PHP_db abstraction layerremovable media, or disk images. These tools will allow us to create a simple, Web-based "digital curation workbench" application to do preliminary analysis and the promotion processing of files to first class objects facilitate the development of richer applications directly in Drupal without the need to integrate external productsthis data.----
'''Talk Title:''' Fiwalk == Do it Yourself Cloud Computing with Me: Using Automatic Forensics Tools Apache and Python for Digital Curation TriageR ==
'''Speaker name* Harrison Dekker, affiliationUniversity of California, and email address:'''Berkeley, hdekker@library.berkeley.edu
Mark Matienzo[http://cran.r-project.org/ R] is a popular, The New York Public Librarypowerful, mark@matienzoand extensible open source statistical analysis application. [http://biostat.mc.vanderbilt.edu/rapache/ Rapache], software developed at Vanderbilt University, allows web developers to leverage the data analysis and visualization capabilities of R in real-time through simple Apache server requests. This presentation will provide an overview of both R and rapache and will explore how these tools might be used to develop applications for the library community.org
'''Abstract of no more than 500 words:'''== Metadata editing - a truly extensible solution ==
Building on Simson Garfinkel's work in Automated Document and Media Exploitation (ADOMEX)* David Kennedy, this project investigates digital curation applications of open source tools used in digital forensics. SpecificallyDuke University, we will be using [http://afflibdavid.org AFFLib]'s fiwalk ("file and inode walk") application and its corresponding Python library to develop a basic triage workflow for accessioned hard driveskennedy@duke.edu* David Chandek-Stark, removable mediaDuke University, or disk imagesdavid. These tools will allow us to create a simple, Web-based "digital curation workbench" application to do preliminary analysis and processing of this datachandek.----stark@duke.edu
http://library.duke.edu/trac/dc/wiki/Trident
'''Talk Title:'''We set out in the Trident project to create a metadata tool that scales. In doing so we have conceived of the metadata application profile, a profile which provides instructions for software on how to edit metadata. We have built a set of web services and some web-based tools for editing metadata. The metadata application profile allows these tools to extend across different metadata schemes, and allows for different rules to be established for editing items of different collections. Some features of the tools include integration with authority lists, auto-complete fields, validation and clean integration of batch editing with Excel. I know, I know, Excel, but in the right hands, this is a powerful tool for cleanup and batch editing.
Do it Yourself Cloud Computing with Apache and R '''Speaker nameIn this talk, affiliationwe want to introduce the concepts of the metadata application profile, and gather feedback on its merits, as well as demonstrate some of the tools we have developed and email address:'''how they work together to manage the metadata in our Fedora repository.
Harrison Dekker, University of California, Berkeley, hdekker@library.berkeley.edu
== Flickr'''Abstract of no more than 500 words:'''ing the Switch ==
R is a powerful and extensible open source statistical analysis application. Rapache* Dianne Dietrich, software developed at Vanderbilt Cornell UniversityLibrary, allows web developers to leverage the numeric processing and graphical capabilities of R in real-time through simple Apache server requests. This presentation will provide an overview of both R and rapache and will explore how these tools are relevant to the library communitydd388@cornell.----edu
We started out with a simple dream &mdash; to pilot a handful of images from our collection in Flickr. Since June 2009, we've grown that dream from its humble beginnings into something bigger: we now have a Flickr collection of over two thousand images. We added geocoding and tags, repurposed our awesome structured metadata, and screenscraped the rest. This talk will focus on the code, which made most of this possible.
'''Talk Title:'''This includes (and is certainly not limited to) using the Python Flickr API, various geocoding tools, crafting Flickr metadata by restructuring XML data from Luna Insight, screenscraping any descriptive text we could get our hands on, negotiating naming conventions for thousands of images, thinking cleverly in order to batch update images on Flickr at a later point (we had to do this more than once), using digital forensic tools to save malformed tifs (that were digitized in 1998!), and, finally, our efforts at scaling everything up so we can integrate our Flickr project into the regular workflow at technical services.
Metadata editing - a truly extensible solution
'''Speaker name, affiliation and email address== library/mobile:'''Developing a Mobile Catalog == * Kim Griggs, Oregon State University Libraries, kim.griggs@oregonstate.edu
David Kennedy, Duke University, david.kennedy@duke.edu<br>David Chandek-Stark, Duke University, david.chandek.stark@duke.edu<br>http://The increased use of mobile devices provides an untapped resource for delivering libraryresources to patrons.dukeThe mobile catalog is the next step for libraries in providing universal access to resources and information.edu/trac/dc/wiki/Trident
This talk will share Oregon State University (OSU) Libraries'''Abstract experience creating a custom mobile catalog. The discussion will first make the case for mobile catalogs, discuss the context of no more than 500 words:''mobile search, and give an overview of vendor and custom mobile catalogs. The second half of the talk will look under the hood of OSU Libraries'custom mobile catalog to provide implementation strategies and discuss tools, techniques, requirements, and guidelines for creating an optimal mobile catalog experience that offers services that support time critical and location sensitive activities.
We set out in the Trident project to create a metadata tool that scales. In doing so we have conceived of the metadata application profile, a profile which provides instructions for software on how to edit metadata. We have built a set of web services and some web-based tools for editing metadata. The metadata application profile allows these tools to extend across different metadata schemes, and allows for different rules to be established for editing items of different collections. Some features of the tools include integration with authority lists, auto-complete fields, validation and clean integration of batch editing with Excel. I know, I know, Excel, but in the right hands, this is a powerful tool for cleanup and batch editing.
In this talk== Enhancing discoverability with virtual shelf browse == * Andreas Orphanides, we want to introduce the concepts of the metadata application profileNCSU Libraries, and gather feedback on its meritsandreas_orphanides@ncsu.edu* Cory Lown, NCSU Libraries, cory_lown@ncsu.edu* Emily Lynema, NCSU Libraries, as well as demonstrate some of the tools we have developed and how they work together to manage the metadata in our Fedora repositoryemily_lynema@ncsu.edu
With collections turning digital, and libraries transforming into collaborative spaces, the physical shelf is disappearing. NCSU Libraries has implemented a virtual shelf browse tool, re----'''Talk Title:''' creating the benefits of physical browsing in an online environment and enabling users to explore digital and physical materials side by side. We hope that this is a first step towards enabling patrons familiar with Amazon and Netflix recommendations to "find more" in the library.
Flickr'ing We will provide an overview of the Switcharchitecture of the front-end application, which uses Syndetics cover images to provide a "cover flow" view and allows the entire "shelf" to be browsed dynamically. We will describe what we learned while wrangling multiple jQuery plugins, manipulating an ever-growing (and ever-slower) DOM, and dealing with unpredictable response times of third-party services. The front-end application is supported by a web service that provides access to a shelf-ordered index of our catalog. We will discuss our strategy for extracting data from the catalog, processing it, and storing it to create a queryable shelf order index.
'''Speaker name, affiliation and email address:'''
Dianne Dietrich== Where do mobile apps go when they die? or, Cornell University Library, dd388@cornellThe app with a thousand faces.edu==
'''Abstract of no more than 500 words:''' * Jason Casden, North Carolina State University Libraries, jason_casden@ncsu.edu
We started out with a simple dream – New capabilities in both native and web-based mobile platforms are rapidly expanding the possibilities for mobile library services. In addition to pilot a handful developing small-screen versions of images from our collection in Flickr. Since June 2009current services, at NCSU Libraries we've grown attempt to develop new services that dream from its humble beginnings into something bigger: we now have a Flickr collection take unique advantage of over two thousand images. We added geocoding and tags, repurposed our awesome structured metadata, and screenscraped the restmobile user context. This talk will focus on Some of these ideas may require capabilities that are not exposed to the codemobile browser. Smart technical planning can help to make sound development decisions when experimenting with mobile-enhanced development, which made most of this possiblewhile remaining agile when faced with constantly changing technical and non-technical restraints and opportunities.
This includes (and is certainly not limited to) using the Python Flickr API, various geocoding tools, crafting Flickr metadata by restructuring XML data from Luna Insight, screenscraping any descriptive text we could get our hands talk will be based on, negotiating naming conventions for thousands my experience as a developer of images, thinking cleverly in order to batch update images on Flickr both native iPhone and web-based mobile library apps at a later point (we had to do this more than once)NCSU Libraries, using digital forensic tools and with the effort to save malformed tifs (that were digitized in 1998!), and, finally, port our efforts at scaling everything up so we can integrate our Flickr project into geo-mobile WolfWalk iPhone app to the regular workflow at technical servicesweb. I will also discuss some opportunities being created by other platforms, particularly Android-based devices.
----
'''Talk Title:'''
library/mobile: Developing a Mobile Catalog '''Speaker name(s), affiliation(s), and email address(es):'''== Using Google Voice for Library SMS ==
Kim Griggs* Eric Sessoms, Oregon State University LibrariesNub Games, kimInc.griggs, nubgames@gmail.com* Pam Sessoms, UNC Chapel Hill, psessoms@oregonstategmail.educom
'''Abstract The LibraryH3lp Google Voice/SMS gateway (free, full AGPL source available at http://github.com/esessoms/gvgw, works with any XMPP server, LibraryH3lp subscription not required) enables libraries to easily integrate texting services into their normal IM workflow. This talk will review the challenges we faced, especially issues involved with interfacing to a Google service lacking a published API, and will outline the design of no the software with particular emphasis on features that help the gateway to be more than 500 words:''responsive to users. Because the gateway is written in the Clojure programming language, we'll close by highlighting which features of the language and available tools had the greatest positive and negative impacts on our development process.
The increased use of mobile devices provides an untapped resource for delivering library resources to patrons. The mobile catalog is the next step for libraries in providing universal access to resources and information.
This talk will share Oregon State University (OSU) Libraries’ experience creating == Building a custom mobile catalog. The discussion will first make the case for mobile catalogs, discuss the context of mobile search, and give an overview of vendor and custom mobile catalogs. The second half of the talk will look under the hood of OSU Libraries' custom mobile catalog to provide implementation strategies and discuss tools, techniques, requirements, and guidelines for creating an optimal mobile catalog experience that offers services that support time critical and location sensitive activities.discovery system with Meresco open source components ==
----* Karin Clavel, TU Delft Library, The Netherlands, c.l.clavel@tudelft.nl* Etienne Posthumus, TU Delft Library, The Netherlands, e.posthumus@tudelft.nl
'''Talk TitleTU Delft Library uses Meresco, an open source component library for metadata management, to implement a custom integrated search solution called [http:'''//discover.tudelft.nl/ Discover]). In Discover, different Meresco components are configured to work together in an efficient observer pattern, defined in what is called Meresco DNA (written in Python). The process is as follows: metadata is harvested from different sources using the Meresco harvester. It is then cross-walked into (any format you like, but we chose) MODS, then normalized, stored and indexed in three distinct but integrated indexes: a full-text Lucene index, a facet index and N-gram index for suggestions and fixing spelling mistakes. The facet index supports multiple algoritmes: drilldown, Jaccard, Mutual Information (or Information Gain) and Χ². One of the facets is used to cluster the search results by subject by using the Jaccard and Mutual Information algorithms.<br/>
Enhancing discoverability with virtual shelf browse '''Speaker name(s), affiliation(s)The query parser component automatically detects and supports Google-like, Boolean and field-specific queries. Different XML documents describing the same content item coalesce to provide the user interface with an easy way to access metadata from either the original or normalized metadata or from user generated metadata such as ratings or tags. Other Meresco components provide an SRU and email address(es):'''a RSS interface.<br/>
Andreas OrphanidesDiscover currently holds all catalogue records, NCSU Librariesthe institutional repository metadata, andreas_orphanides@ncsuan architecture bibliography and a test-set of Science Direct articles.edu <br/>Cory LownIn 2010, NCSU Librariesit is expected to grow to over 10 million records with content from Elsevier, cory_lown@ncsuIEEE and Springer (subject to negotiatons with these publishers) and various open access resources. We will also add the university's multimedia collection, ranging from digitized historical maps, drawing and photographs to recent (vod- and) podcasts.edu <br/>Emily Lynema, NCSU Libraries, emily_lynema@ncsu.edu
'''Abstract In the proposed session, we would like to show you some examples of no more than 500 words:'''above mentioned functionality and explain how Meresco components work together to create this flexible system.
With collections turning digital, and libraries transforming into collaborative spaces, the physical shelf is disappearing. NCSU Libraries has implemented a virtual shelf browse tool, re-creating the benefits of physical browsing in an online environment and enabling users to explore digital and physical materials side by side. We hope that this is a first step towards enabling patrons familiar with Amazon and Netflix recommendations to "find more" in the library.
We will provide an overview == Take control of the architecture of the front-end application, which uses Syndetics cover images to provide a "cover flow" view library metadata and allows websites using the entire "shelf" to be browsed dynamically. We will describe what we learned while wrangling multiple jQuery plugins, manipulating an ever-growing (and ever-slower) DOM, and dealing with unpredictable response times of third-party services. The front-end application is supported by a web service that provides access to a shelf-ordered index of our catalog. We will discuss our strategy for extracting data from the catalog, processing it, and storing it to create a queryable shelf order index.eXtensible Catalog ==
----* Jennifer Bowen, University of Rochester, jbowen@library.rochester.edu
'''Talk Title:'''The eXtensible Catalog Project has developed four open-source software toolkits that enable libraries to build and share their own web- and metadata-focused applications on top of a service-oriented architecture that incorporates Solr in Drupal, a robust metadata management platform, and OAI-PMH and NCIP-compatible tools that interact with legacy library systems in real-time.
Where do mobile apps go when they die? XC's robust metadata management platform allows libraries to orchestrate and sequence metadata processing services on large batches of metadata. Libraries can build their own services using the available "service-writers toolkit" orchoose from our initial set of metadata services that clean up and "FRBRize" MARC metadata. Another service will aggregate metadata from multiple repositories to prepare it for use in unified discovery applications. XC software provides an RDA metadata test bed and a Solr-based metadata "navigator" that can aggregate and browse metadata (or data) in any XML format. XC's user interface platform is the first suite of Drupal modules that treat both web content and library metadata as native Drupal nodes, The app allowing libraries to build web-applications that interact with metadata from library catalogs and institutional repositories as well as with library web pages. XC's Drupal modules enable Solr in a thousand facesFRBRized data environment, as a first step toward a full implementation of RDA. Other currently-available XC toolkits expose legacy ILS metadata, circulation, and patron functionality via web services for III, Voyager and Aleph (to date) using standard protocols (OAI-PMH and NCIP), allowing libraries to easily and regularly extract MARC data from an ILS in valid MARCXML and keep the metadata in their discovery applications "in sync" with source repositories.
This presentation will showcase XC'''Speaker name, affiliations metadata processing services, the metadata "navigator" and email address:'''the Drupal user interface platform. The presentation will also describe how libraries and their developers can get started using and contributing to the XC code.
Jason Casden, North Carolina State University Libraries, jason_casden@ncsu.edu
'''Abstract== I Am Not Your Mother:'''Write Your Test Code ==
New capabilities in both native and web-based mobile platforms are rapidly expanding the possibilities for mobile library services* Naomi Dushay, Stanford University, ndushay@stanford. In addition to developing small-screen versions of our current servicesedu* Willy Mene, at NCSU Libraries we attempt to develop new services that take unique advantage of the mobile user contextStanford University, wmene@stanford. Some of these ideas may require capabilities that are not exposed to the mobile browser. Smart technical planning can help to make sound development decisions when experimenting with mobile-enhanced developmentedu* Jessie Keck, Stanford University, while remaining agile when faced with constantly changing technical and non-technical restraints and opportunitiesjkeck@stanford.edu
This talk will be based on my experience as How is it worth it to slow down your code development to write tests? Won't it take you a developer of both native iPhone and web-based mobile library apps at NCSU Librarieslong time to learn how to write tests? Won't it take longer if you have to write tests AND develop new features, and with the effort fix bugs? Isn't it hard to port write test code? To maintain test code? We will address these questions as we talk about how test code is crucial for our geo-mobile WolfWalk iPhone app to the websoftware. I By way of illustration, we will also discuss some opportunities being created by other platformsshow how it has played a vital role in making Blacklight a true community collaboration, particularly Android-based devicesas well as how it has positively impacted coding projects in the Stanford Libraries.
----== How To Implement A Virtual Bookshelf With Solr ==
'''Talk Title:'''* Naomi Dushay, Stanford University, ndushay@stanford.edu* Jessie Keck, Stanford University, jkeck@stanford.edu
Using Google Voice Browsing bookshelves has long been a useful research technique as well as an activity many users enjoy. As larger and larger portions of our physical library materials migrate to offsite storage, having a browse-able virtual shelf organized by call number is a much-desired feature. I will talk about how we implemented nearby-on-shelf in Blacklight at Stanford, using Solr and SolrMarc: # the code to get shelfkeys out of call numbers# the code to lop volume data off the end of call numbers to avoid clutter in the browse # what I indexed in Solr given we have## multiple call numbers for Library SMSa single bib record ## multiple bib records for a single call number# Solr configuration, requests and responses to get call numbers before and after a given starting point as well as the desired information for display.# Other code needed to implement this feature in Blacklight (concepts easily ported to other UIs).
'''Speaker nameThis virtual shelf is not only browsable across locations, affiliation, and email address:'''but includes any item with a call number in our collection (digital or physical materials).
Eric SessomsAll code is available, Nub Games, Inc., nubgames@gmail.com<br />Pam Sessoms, UNC Chapel Hill, psessoms@gmailor will be by Code4Lib 2010.com
'''Abstract:'''
The LibraryH3lp Google Voice/SMS gateway (free, full AGPL source available at http://github.com/esessoms/gvgw, works with any XMPP server, LibraryH3lp subscription not required) enables libraries to easily integrate texting services into their normal IM workflow. This talk will review the challenges we faced, especially issues involved with interfacing to a Google service lacking a published API, and will outline the design of the software with particular emphasis on features that help the gateway to be more responsive to users. Because the gateway is written in the Clojure programming language, we'll close by highlighting which features of the language and available tools had the greatest positive and negative impacts on our development process.== A Better Advanced Search ==
----* Naomi Dushay, Stanford University, ndushay@stanford.edu* Jessie Keck, Stanford University, jkeck@stanford.edu
Even though we'd love to get basic searches working so well that advanced search wouldn't be necessary, there will always be a small set of users that want it, and there will always be some library searching needs that basic searching can'Talk Titlet serve. Our user interface designer was dissatisfied with many aspects of advanced search as currently available in most library discovery software; the form she designed was excellent but challenging to implement. See http://searchworks.stanford.edu/advancedWe'''ll share details of how we implemented Advanced Search in Blacklight:# non-techie designed html form for the user# boolean syntax while using Solr dismax magic (dismax does not speak Boolean)# checkbox facets (multiple facet value selection)# fielded searching while using Solr dismax magic (dismax allows complex weighting formulae across multiple author/title/subject/... fields, but does not allow "fielded" searching in the way lucene does)## easily configured in solrconfig.xml# manipulating user entered queries before sending them to Solr# making advanced search results look like other search results: breadcrumbs, selectable facets, and other fun.
Building a discovery system with Meresco open source components== Scholarly annotation services using AtomPub and Fedora ==
'''Speaker name* Andrew Ashton, affiliationBrown University, and email address:'''andrew_ashton@brown.edu
Karin ClavelWe are building a framework for doing granular annotations of objects housed in Brown's Digital Repository. Beginning with our TEI-encoded text collections, TU Delft Libraryand eventually expanding to other media, The Netherlandsthese scholarly annotations are themselves objects stored and preserved in the repository. They are linked to other resources via URI references, c.l.clavel@tudelft.nl<br and deployed using AtomPub services as part of Fedora's Service/>Etienne Posthumus, TU Delft Library, The Netherlands, e.posthumus@tudelftDissemination model.nl
'''Abstract:'''This effort stems from the recognition that standard web annotation techniques (e.g. tagging, Google Sidebar, page-level commenting, etc.) are not flexible or persistent enough to handle scholarly annotations as an organic part of natively digital research collections. We are developing solutions to several challenges that arise with this approach; particularly, how do we address highly granular portions of digital objects in a way that is applicable to different types of media (encoded texts, images, video, etc.). This presentation will provide an overview of the architecture, a discussion of the possibilities and problems we face in implementing this framework, and a demo of a live project using Atom annotations with a digital research collection.
TU Delft Library uses Meresco, an open source component library for metadata management, to implement a custom integrated search solution called [http://discover.tudelft.nl/ Discover]).
In Discover, different Meresco components are configured to work together in an efficient observer pattern, defined in what is called Meresco DNA (written in Python). The process is as follows: metadata is harvested from different sources using the Meresco harvester. It is then cross-walked into (any format you like, but we chose) MODS, then normalized, stored and indexed in three distinct but integrated indexes: a full-text Lucene index, a facet index and N-gram index for suggestions and fixing spelling mistakes. The facet index supports multiple algoritmes: drilldown, Jaccard, Mutual Information (or Information Gain) and Χ². One of the facets is used to cluster the search results by subject by using the Jaccard and Mutual Information algorithms.<br/>
The query parser component automatically detects and supports Google-like, Boolean and field-specific queries== With Great Power. Different XML documents describing the same content item coalesce to provide the user interface with an easy way to access metadata from either the original or normalized metadata or from user generated metadata such as ratings or tags. Other Meresco components provide . Managing an SRU and Open-Source ILS in a RSS interfacestate-wide consortium.<br== * Emily A. Almond, Software Development Manager, PINES/>Georgia Public Library Service, ealmond@georgialibraries.org
Discover currently holds all catalogue records, the institutional repository metadata, an architecture bibliography and Using agile software development methodology + project management to achieve a test-set balance of Science Direct articles. In 2010, it is expected to grow to over 10 million records with content from Elsevier, IEEE support and Springer (subject to negotiatons with these publishers) and various open access resourcesexpertise. We will also add Lessons learned after implementation that inform how the university’s multimedia collection, ranging from digitized historical maps, drawing consortium should evolve so that you can utilize your new ILS for the benefit of all stakeholders. Topics covered: -- troubleshooting and photographs to recent (vodhelp desk support- - development project plans-- roles and responsibility shifts-- re-branding the ILS and) podcastsrelated organizations.<br/>
In the proposed session, we would like to show you some examples of above mentioned functionality and explain how Meresco components work together to create this flexible system.
----== Data Modeling; Logical Versus Physical; Why Do I Care? ==
'''Talk Title:'''* Steve Dressler, Georgia Public Library Services, sdressler@georgialibraries.org
Take control I am sure we have all been in the situation of library metadata having mountains of data stored in our database, needing a piece of information and websites using yet being unable to determine how to get what we need. Computerized databases have been around for decades now and there are several architectures available; however, the ability of a database developer, regardless of the eXtensible Catalogarchitecture, to store data in a format that is comprehensible to a businessperson yet readily accessible through software applications remains an impossible challenge.
'''Speaker name(s)Topics to be discussed includeo Components comprising a logical model, affiliation(s)how it is developed and how is it used?o Components comprising a physical model, how it is developed and how is it used?o What does a logical model look like?o What does a physical model look like?o Who works with a logical model and why?o Who works with a physical model and why?o What is the relationship between the logical model and the physical model?o What kind of a time investment is required to develop and maintain logical and email address(es):'''physical models?o What are the challenges of keeping the two models in sync as the software application evolves?
Jennifer Bowen, University Although data modeling is a huge discipline and presents research topics for millions of Rochestertheses and dissertations, jbowen@library.rochesterthis twenty-minute snapshot view will allow anyone, technical or business, to sit through a development meeting and be able to grasp what is being discussed as well as gain a better understanding of logical and physical business flows.edu
'''Abstract of no more than 500 words:'''
The eXtensible Catalog Project has developed four open-source software toolkits that enable libraries to build and share their own web- and metadata-focused applications on top of a service-oriented architecture that incorporates Solr in Drupal== Media, a robust metadata management platformBlacklight, and OAI-PMH and NCIP-compatible tools that interact with legacy library systems in real-timeviewers like you.==
XC’s robust metadata management platform allows libraries to orchestrate and sequence metadata processing services on large batches of metadata. Libraries can build their own services using the available “service-writers toolkit” or choose from our initial set of metadata services that clean up and “FRBRize” MARC metadata. Another service will aggregate metadata from multiple repositories to prepare it for use in unified discovery applications. XC software provides an RDA metadata test bed and a Solr-based metadata “navigator” that can aggregate and browse metadata (or data) in any XML format. XC’s user interface platform is the first suite of Drupal modules that treat both web content and library metadata as native Drupal nodes, allowing libraries to build web-applications that interact with metadata from library catalogs and institutional repositories as well as with library web pages. XC’s Drupal modules enable Solr in a FRBRized data environment, as a first step toward a full implementation of RDA. Other currently-available XC toolkits expose legacy ILS metadata, circulation, and patron functionality via web services for III* Chris Beer, Voyager and Aleph (to date) using standard protocols (OAI-PMH and NCIP)WGBH, allowing libraries to easily and regularly extract MARC data from an ILS in valid MARCXML and keep the metadata in their discovery applications “in sync” with source repositorieschris_beer@wgbh. org
This presentation will showcase XC’s metadata processing services, the metadata “navigator” There are many shared problems (and solutions) for libraries and archives in the interest of helping the Drupal user interface platform. The presentation will There are also describe how libraries many "new" developments in the archives world that the library communities have been working on for ages, including item-level cataloging, metadata standards, and their developers can get started using and contributing asset management. Even with these similarities, media archives have additional issues that are less relevant to libraries: the XC codechoice of video players, large file sizes, proprietary file formats, challenges of time-based media, etc. In developing a web presence, many archives, including the WGBH Media Library and Archives, have created custom digital library applications to expose material online. In 2008, we began a prototyping phase for developing scholarly interfaces by creating a custom-written PHP front-end to our Fedora repository.
In late 2009, we finally saw the (black)light, and after some initial experimentation, decided to build a new, public website to support our IMLS-funded /Vietnam: A Television History/ archive (as well as existing legacy content). In this session, we will share our experience of and challenges with customizing Blacklight as an archival interface, including work in rights management, how we integrated existing Ruby on Rails user---generated content plugins, and the development of media components to support a rich user experience.
'''Talk Title:'''
I Am Not Your Mother: Write Your Test Code== DAMS PAS - Digital Asset Management System, Public Access System ==
'''Speaker name* Declan Fleming, affiliationUniversity of California, and email address:'''San Diego, dfleming@ucsd.edu* Esmé Cowles, University of California, San Diego, ecowles@ucsd.edu
Naomi DushayAfter years of describing our DAMS with Powerpoint, Stanford Universitywe finally have a public access system that we can show our mothers. And code4lib! The UCSD Libraries DAMS is an RDF based asset repository containing over 250, ndushay@stanford000 items and their derivatives. We describe the core system, the metadata and storage challenges involved in managing hundreds of thousands of items, and the interesting political aspects involved in releasing subsets to the public. We also describe the caching approach we used to ensure performance and access control.edu
'''Abstract:'''
How is it worth it to slow down your code development to write tests? == You Either Surf or You Fight: Integrating Library Services with Google Wave == Won’t it take you a long time to learn how to write tests? Won’t it take longer if you have to write tests AND develop new features* Sean Hannan, fix bugs? Isn’t it hard to write test code? To maintain test code? I will try to answer these questions as I talk about how test code is crucial for our software. By way of illustrationSheridan Libaries, I will show how it has played a vital role in making Blacklight a true community collaborationJohns Hopkins University, as well as how it has positively impacted coding projects in the Stanford Librariesshannan@jhu.edu
----So Google Wave is a new shiny web toy, but did you know that it's also a great platform for collaboration and research? (I bet you did.) ...And what platform for collaboration and research would not be complete without some library tools to aid and abet that process? I will talk about how to take your library web services and integrate them with Google Wave to create bots that users can interact with to get at your resources as part of their social and collaborative work.
'''Talk Title:'''
How To Implement A Virtual Bookshelf With Solr== The Linked Library Data Cloud: Stop talking and start doing ==
'''Speaker name* Ross Singer, affiliationTalis, and email address:'''ross.singer@talis.com
Naomi DushayA year later and how far has Linked Library Data come? With the emergence of large, Stanford University, ndushay@stanfordcentralized sources (id.edu<br loc.gov/>Jessie Keckauthorities/, viaf.org, among others) entry to the Linked Data cloud might be easier than you think. This presentation will describe various projects that are out in the wild that can bridge the gap between our legacy data and the semantic web, Stanford Universityincremental steps we can take modeling our data, jkeck@stanfordwhy linked data matters and a demonstration of how a small template changes can contribute to the Linked Data cloud.edu
'''Abstract:'''== A code4lib Manifesto ==
Browsing bookshelves has long been a useful research technique as well as an activity many users enjoy. As larger and larger portions of our physical library materials migrate to offsite storage* Dan Chudnov, having a browse-able virtual shelf organized by call number is a much-desired feature. I will talk about how we implemented nearby-on-shelf in Blacklight No Fixed Hairstyle, dchud at Stanford, using Solr and SolrMarc: # the code to get shelfkeys out of call numbers# the code to lop volume data off the end of call numbers to avoid clutter in the browse # what I indexed in Solr given we have## multiple call numbers for a single bib record ## multiple bib records for a single call number# Solr configuration, requests and responses to get call numbers before and after a given starting point as well as the desired information for display.# Other code needed to implement this feature in Blacklight (concepts easily ported to other UIs).umich edu
This virtual shelf is not only browsable across locations, but includes any item code4lib started with a call number half dozen library hackers and a list and it ain't like that anymore. I come to code4lib with strong opinions about why it's a positive force in our collection (digital or physical materials)my professional and personal life, but they're probably different from your opinions. I will share these opinions rudely yet succinctly to challenge everyone to think and argue about why code4lib works and what we need to do to keep it working.
All code is available, or will be by Code4Lib 2010.
----== Cloud4lib ==
'''Talk Title:'''* Jeremy Frumkin, University of Arizona, frumkinj at u library arizona edu* Terry Reese, Oregon State University, terry.reese at oregonstate edu
A Better Advanced Search?Major library vendors are creating proprietary platforms for libraries. We will propose that the code4lib community pursue the cloud4lib, a open digital library platform based on open source software and open services. This platform would provide common service layers for libraries, not only via code, but also allow libraries to easily utilize tools and systems through cloud services. Instead of a variety of competing cloud services and proprietary platforms, cloud4lib will attempt to be a unifying force that will allow libraries to be consumer of the services built on top of it as well as allow developers / researchers / code4lib'ers to hack, extend, and enhance the platform as it matures.
'''Speaker name, affiliation, and email address:'''
Naomi Dushay, Stanford University, ndushay@stanford.edu<br />Jessie Keck, Stanford University, jkeck@stanford.edu== Iterative development done simply ==
'''Abstract:'''* Emily Lynema, North Carolina State University Libraries, emily_lynema@ncsu.edu
Even though we’d like to get basic searches working so well that advanced search wouldn’t be necessary, there will always be With a small set IT unit and a wide array of users that want itprojects to support, and there will always be some requests for development from business stakeholders in the library searching needs that basic searching can’t servecan quickly spiral out of control. Our user interface designer was dissatisfied with many aspects To help make sense of advanced search as currently available in most library discovery software; the form she designed was excellent but challenging to implement. See http://searchworks.stanford.edu/advancedWe’ll share details chaos, increase the transparency of how the IT "black box," and shorten time lag between requirements definition and functional releases, we have implemented Advanced Search in Blacklight:# thoughtfully designed html form for a modified Agile/SCRUM methodology within the user (NOT done by techies!)# boolean syntax while using Solr dismax magic (dismax does not speak Boolean)# checkbox facets (multiple facet value selection)# fielded searching while using Solr dismax magic (dismax allows complex weighting formulae across multiple author/title/subject/… fields, but does not allow “fielded” searching development group in the way lucene does)## easily configured in solrconfig.xml# manipulating user entered queries before sending them to Solr# making advanced search results look like other search results: breadcrumbs, selectable facets, and other funIT department at NCSU Libraries.
This presentation will provide a brief overview of the Agile methodology as an introduction to our simplified approach to iteratively handling multiple projects across a small team. This iterative approach allows us to regularly re-evaluate requested enhancements against institutional priorities and more accurately estimate timelines for specific units of functionality. The presentation will highlight how we approach each development cycle (from planning to estimating to re-aligning) as well as some of the actual tools and techniques we use to manage work (like JIRA and Greenhopper). It will identify some challenges faced in applying an established development methodology to a small team of multi--'''Talk Title:'tasking developers, the outcomes we've seen, and the areas we'd like to continue improving. These types of iterative planning/development techniques could be adapted by even a single developer to help manage a chaotic workplace.
Scholarly annotation services using AtomPub and Fedora
'''Speaker name, affiliation, and email address:'''== Public Datasets in the Cloud ==
Andrew Ashton* Rosalyn Metz, Brown Wheaton College, metz_rosalyn@wheatoncollege.edu* Michael B. Klein, Oregon State University, andrew_ashtonMichael.Klein@brownoregonstate.edu
'''AbstractWhen most people think about cloud computing (if they think about it at all), it usually takes one of two forms:'''Infrastructure Services, such as Amazon EC2 and GoGrid, which provide raw, elastic computing capacity in the form of virtual servers, and Platform Services, such as Google App Engine and Heroku, which provide preconfigured application stacks and specialized deployment tools.
We are building a framework Several providers, however, offer access to large public datasets that would be impractical for doing granular annotations of objects housed in Brown’s Digital Repository. Beginning most organizations to download and work with our TEIlocally. From a 67-encoded text collectionsgigabyte dump of DBpedia's structured information store to the 180-gigabyte snapshot of astronomical data from the Sloan Digital Sky Survey, chemistry and eventually expanding biology to other mediaeconomic and geographic data, these scholarly annotations datasets are themselves objects stored available instantly and preserved in the repository. They are linked to other resources via URI references, and deployed using AtomPub services backed by enough pay-as part -you-go server capacity to make good use of Fedora’s Service/Dissemination modelthem.
This effort stems from the recognition that standard web annotation techniques (e.g. tagging, Google Sidebar, page-level commenting, etc.) are not flexible or persistent enough to handle scholarly annotations as an organic part of natively digital research collections. We are developing solutions to several challenges that arise with this approach; particularly, how do we address highly granular portions of digital objects in a way that is applicable to different types of media (encoded texts, images, video, etc.). This presentation will provide present an overview of the architecturecurrently-available datasets, a discussion what it takes to create and use snapshots of the possibilities and problems we face in implementing this frameworkdata, and a demo explore how the library community might push some of a live project using Atom annotations with a digital research collectionits own large stores of data and metadata into the cloud.
----
'''Talk Title:'''
With Great Power... Managing an Open-Source ILS in a state-wide consortium. '''Speaker name(s), affiliation(s), and email address(es):'''== Codename Arctika ==
Emily A. Almond, Software Development Manager* Toke Eskildsen, PINES/Georgia Public The State and University Library Serviceof Denmark, ealmondte@georgialibrariesstatsbiblioteket.orgdk
There'''Abstracts something missing in the state of Denmark. Most of our web based copyright deposit material is trapped in a dark archive. After a successful pilot; money and time has been allocated to open part of the data. We tried NutchWAX and it worked well, but we wanted more. Proper integrated search with existing library material, extraction of names etc. Therefore we propose the following recipe:'''Take a slice of a dark archive with copyright deposit material. Get permission to publish it (the tricky bit). Add an ARC reader to get the bits, Tika to get the text and Summa to get large-scale index and faceting. We mixed it up and we will show what happened.
Using agile software development methodology + project management to achieve a balance of support and expertise. Lessons learned after implementation that inform how the consortium should evolve so that you can utilize your new ILS for the benefit of all stakeholders.
Topics covered:
-- troubleshooting and help desk support
-- development project plans
-- roles and responsibility shifts
-- re-branding the ILS and related organizations.
== JeromeDL ----an open source social semantic digital library == * Sebastian Ryszard Kruk, Knowledge Hives, sebastian.kruk@knowledgehives.com'''Talk Title:'''* Jodi Schneider, DERI NUI Galway, jschneider@pobox.com
JeromeDL is an open source e-library with semantics. A fully functional digital library, JeromeDL uses linked data: Data Modeling; Logical Versus Physical; Why Do I Care?using standard "Web3.0" vocabularies such as SIOC, FOAF, and WordNet, JeromeDL publishes RDF descriptions of the e-library contents. Jerome DL uses FOAF to manage users--meaning that access privileges can be naturally assigned to a social network, in addition to individuals or all WWW users. Users can also share annotations, promoting collaborative browsing and collaborative filtering. To encourage users to provide meaningful annotations (beyond just tags), JeromeDL uses a WordNet-based vocabulary service. The system also leverages full-text indexing with Lucene and allows filtering with the SIMILE project's Exhibit. In short, JeromeDL is a social semantic digital library--allowing users to collect, publish, and share their library with their social network on the semantic web.
'''Speaker name(s), affiliation(s), and email address(es)*[http:'''//www.jeromedl.org/ JeromeDL homepage]*[http://bleedingedge.jeromedl.org/preview?show=techreport JeromeDL demo site]
Steve Dressler, Georgia Public Library Services, sdressler@georgialibraries.org== Kill the search button ==
'''Abstract of no more than 500 words:'''* Michael Poltorak Nielsen, State and University Library, Denmark, mn@statsbiblioteket.dk* Jørn Thøgersen, State and University Library, Denmark, jt@statsbiblioteket.dk
I am sure we have all been in We demo three concepts that eliminate the situation of having mountains of data stored in our database, needing a piece of information and yet being unable to determine how to get what we need. Computerized databases have been around for decades now and there are several architectures available; however, the ability of a database developer, regardless of the architecture, to store data in a format that is comprehensible to a businessperson yet readily accessible through software applications remains an impossible challengesearch button.
Topics to be discussed includeo Components comprising a logical model, how it is developed and how is it used1. Instant search. Why wait for tiresome page reloads when searching?o Components comprising a physical model, Instant search updates the search result on every key-press. We will show how it is developed and how is it used?o What does a logical model look like?o What does a physical model look like?o Who works we integrated this feature into our own library search system with a logical model and why?o Who works with a physical model and why?o What is the relationship between the logical model and the physical model?o What kind of a time investment is required minimal changes to develop and maintain logical and physical models?o What are the challenges of keeping the two models in sync as the software application evolves?existing setup.
Although data modeling is a huge discipline and presents research topics for millions 2. Index lookup. Ever dreamed of theses your own inline instant index lookup?We demo an instant index lookup feature that requires no search button and dissertations, this twentyno page refreshes -minute snapshot view will allow anyone, technical or business, to sit through a development meeting and be able to grasp what is being discussed as well as gain a better understanding of logical and physical business flowswithout ever leaving the search field.
----3. Slide your data. Sliders are an alternative way to fit search results to the user's search context.'''Talk Title:'''Examples are sliders that move search results priorities between title and subject and between books by an author and books about the author.
Media, Blacklight, and viewers like you.
'''Speaker name, affiliation, and email address== Controlling the flood:'''Re-plumbing fittings between a New Titles List and other services with Yahoo! Pipes. ==
Chris Beer* Jon Gorman, WGBHUniversity of Illinois, chris_beerjtgorman@wgbhillinois.orgedu
'''AbstractAbout four years ago the University of Illinois decided to create a New Titles service (http:'''//www.library.illinois.edu/newtitles/) that could provide RSS feeds. At the time a balance was struck between complexity of options and limited development time. Currently a feed is created by adding options, each option narrowing the scope of a feed. Selecting a date range, Unit Library and a call number range will retrieve material that match all three of the criteria. It was hoped that at some point a generic tool would be able to further manipulate and combine feeds produced by the simple options to customize very specific feed. Yahoo! Pipes has emerged to fill that niche.
There are many shared problems (and solutions) The talk will cover pipes that range from filter for libraries and archives a keyword in one feed to combining the interest of helping New Titles List with services like the userLibraryThing API or Worldcat APIs. There are Examples will also many "new" developments be given in the archives world that the library communities have been working on for ages, including item-level cataloging, metadata standards, and asset management. Even with these similarities, media archives have additional issues that are less relevant how to libraries: integrate the choice output of video players, large file sizes, proprietary file formats, challenges of time-based media, etc. In developing a web presence, many archives, including the WGBH Media Library Yahoo! Pipes into webpages and Archives, how we have created custom digital library applications to expose material onlineput them into our CMS (OpenCMS). In 2008, we began a prototyping phase for developing scholarly interfaces by creating a custom-written PHP front-end The talk will make sure to our Fedora repositoryaddress areas where Yahoo! Pipes either fails or is cumbersome and simpler CSS and Javascript solutions have worked.
In late 2009, we finally saw the (black)light, and after some initial experimentation, decided to build a new, public website to support our IMLS-funded /Vietnam: A Television History/ archive (as well as existing legacy content). In this session, we will share our experience of and challenges with customizing Blacklight as an archival interface, including work in rights management, how we integrated existing Ruby on Rails user-generated content plugins, and the development of media components to support a rich user experience.
----'''Talk Title== Vampires vs. Werewolves:''' Ending the War Between Developers and Sysadmins with Puppet ==
DAMS PAS - Digital Asset Management System* Bess Sadler, Public Access SystemUniversity of Virginia, bess@virginia.edu
Developers need to be able to write software and deploy it, and often require cutting edge software tools and system libraries. Sysadmins arecharged with maintaining stability in the production environment, and so are often resistant to rapid upgrade cycles. This has traditionally pitted us against each other, but it doesn'''Speaker name(s)t have to be that way. Using tools like puppet for maintaining and testing server configuration, affiliation(s)nagios for monitoring, and email address(es):''hudson for continuous code integration, UVA has brokered a peace that has given us the ability to maintain stable production environment with a rapid upgrade cycle. I'll discuss both the individual tools, our server configuration, and the social engineering that got us here.
Declan Fleming, University of California, San Diego, dfleming@ucsd.edu
Esmé Cowles== Building customizable themes for DSpace == * Elias Tzoc, Miami University of CaliforniaOhio, San Diego, ecowlestzoce@ucsdmuohio.edu
'''Abstract The popularity of no more than 500 words:'''DSpace (should I say DuraSpace?) continues to grow!Many universities and research institutions are using DSpace to create and provide access to digital content &mdash; including documents, images, audio, and video. With the variety of content, one of the challenges is "how to create customizable themes for different types of content?"
After years of describing our DAMS with PowerpointIn 2007, we finally have Manakin was developed as a public access system that we can show our mothers. And code4lib! The UCSD Libraries DAMS is an RDF user interface for DSpace based asset repository containing over 250,000 items and their derivativeson themes. We describe Now users have the core systemability to customize the web interface for DSpace collections by editing CSS, the metadata and storage challenges involved in managing hundreds of thousands of itemsXML, and the interesting political aspects involved in releasing subsets to the publicXSLT files. We also describe the caching approach we used Best of all, a singular theme can be applied to ensure performance and access controlindividual communities, collections or items.
This talk will be based on my work creating themes for DSpace, as well as tips & tricks for customizing the look-and---feel for individual communities and collections.Who knows, maybe someday a group of code4lib developers can create a whole library of themes for DuraSpace &mdash; similar to the WordPress or Drupal theme idea!
'''Talk Title:'''
You Either Surf or You Fight== HIVE: Integrating Library Services a new tool for working with Google Wavevocabularies ==
'''Speaker name(s)* Ryan Scherle, affiliation(s)National Evolutionary Synthesis Center, and email address(es):'''rscherle@nescent.org* Jose Aguera, Universitty of North Carolina, jose.aguera@gmail.com
Sean HannanHIVE is a toolkit that assists users in selecting vocabulary and ontology terms to annotate digital content. HIVE combines the ease of folksonomies with the rigor of traditional vocabularies. By combining semantic web standards with text mining techniques, Sheridan LibariesHIVE will improve the effectiveness of subject metadata generation, Johns Hopkins University, shannan@jhuallowing users to search and browse terms from a variety of vocabularies and ontologies. Documents can be submitted to HIVE to automatically generate suggested vocabulary terms.edu
'''Abstract Your system can interact with common vocabularies such as LCSH and MESH via the central HIVE server, or you can install a local copy of no more than 500 words:'''HIVE with your own custom set of vocabularies. This talk will give an overview of the current features of HIVE and describe how to build tools that use the HIVE services.
So Google Wave is == Implementing Metasearch and a new shiny web toyUnified Index with Masterkey == * [[User:DataGazetteer|Peter Murray]], but did you know that it's also a great platform for collaboration and research? (I bet you did.) ...And what platform for collaboration and research would not be complete without some library tools to aid and abet that process? I will talk about how to take your library web services and integrate them with Google Wave to create bots that users can interact with to get at your resources as part of their social and collaborative workOhioLINK, peter@OhioLINK.edu
----Index Data'''Talk Title:'''The Linked Library s suite of metasearch and local indexing tools under the product name Masterkey are a powerful way to provide access to a diverse set of databases. In 2009, OhioLINK contracted with Index Data Cloud: to help build a new metasearch platform and a unified index of locally-loaded records. Stop talking and start doing
'''Speaker name, affiliationBy the time conference rolls around, the user interface and email address:'''Ross Singerthe metasearch infrastructure will be set up and live. This part of the presentation will dive into the innards of the AJAX-powered end-user interface, Talisthe configuration back-end, ross.singer@talisand possibly a view of the Gecko-driven Index Data Connector Framework.com
'''Abstract:'''A year later and how far has Linked Library Data come? Outside of It is hard to predict at the Swedish National Library's LIBRIS (which already existed), point this talk is being proposed what the return state of lcsh.info as http://id.loc.gov/authorities/ and LC's Chronicling America, not much. But entry to the Linked Data cloud might unified index will be easier than you think. This presentation will describe various projects that are out in At the wild that can bridge the gap between our legacy data and the semantic webvery least, incremental steps we can take modeling our data, why linked data matters there will be broad system diagrams and a demonstration description of how a small template changes can contribute intend to the Linked Data cloudeventually bring 250 million records into one index. With luck, there might even be running code to show.
== Adding Solr----'''Talk Title:''based Search to Evergreen'A code4lib Manifestos OPAC ==
* Alexander O'''Speaker name(s)Neill, affiliation(s), and email address(es):'''Dan ChudnovRobertson Library, No Fixed HairstyleUniversity of Prince Edward Island, dchud at umich eduaoneill@upei.ca
'''Abstract of no more than 500 words:'''code4lib started with a half dozen library hackers and a list and it ain't like that anymore. I come The current way the Evergreen OPAC searches records is to code4lib with strong opinions about why use it's a positive force in my professional and personal lifedatabase back-end's search system, but they're probably different from your opinions. I will share these opinions rudely yet succinctly with heavy use of caching layers to challenge everyone compensate for the relatively long wait to think and argue about why code4lib works and what we need to do to keep it workingperform a new search.
This is a personal project to adapt the Evergreen search results page to use the Solr and Lucene search engine stack -integrating the external search function as closely as possible with Evergreen's existing look and feel. This is a possible alternative to replacing an entire OPAC just to take advantage of the very desirable features offered by the Solr stack as Evergreen does offer a very well---'''Talk Title:'designed extensible JavaScript interface which we and others have already gotten great results customizing and adding features to such as integrated Google Books previews and incorporating LibraryThing's social features. Adapting the leading open source search technology into this very powerful stack is one more feature to add to Evergreen'Cloud4libs very compelling list of selling points.
It is still possible to use Evergreen's OpenSRF messaging system to get live information about each book''Speaker name(s), affiliation(s), and email address(es):'''Jeremy Frumkin, University current availability status without having to push all of Arizona, frumkinj at u library arizona edu<br/>Terry Reese, Oregon State University, terrythis information into the Solr index.reese at oregonstate edu
'''Abstract I will show how I used SolrMarc to import records from Evergreen, taking advantage of no more than 500 words:'''Major library vendors are creating proprietary platforms for libraries. We will propose the fact that the code4lib community pursue the cloud4lib, a open digital library platform based on open source software VuFind and open services. This platform would provide common service layers for libraries, not only via code, but also allow libraries to easily utilize tools and systems through cloud services. Instead of a variety of competing cloud services and proprietary platforms, cloud4lib will attempt Blacklight projects have collaborated to be create a unifying force general import utility that is usable by third-party projects. I will allow libraries to be consumer discuss some of the services built on top of it as well as allow developers / researchers / code4libhurdles I encountered while using SolrMarc and the resulting changes to SolrMarc'ers s design that this use case helped to hack, extend, and enhance the platform as it maturesmotivate.
----I'll also make an effort to take measurements of performance when hosting both Solr and Evergreen on the same server compared with putting Solr on a separate server. It will also be informative to see how much of an Evergreen server's system load is devoted to processing user searches.
'''Talk Title:'''==Matching Dirty Data - Yet another wheel==
Iterative development done simply* Anjanette Young, University of Washington Libraries, younga3 at u washington edu* Jeff Sherwood, University of Washington Libraries, jeffs3 at u washington edu
'''Speaker nameRegular expressions is a powerful tool to identify matching data between similar files. When one or both of these files has inconsistent data due to differing character encodings or miskeying, affiliation, and email address:'''the use of regular expressions to find matches becomes impractically complex.
Emily Lynema, North Carolina State University LibrariesThe Levenshtein distance (LD) algorithm is a basic sequence comparison technique that can be used to measure word similarity more flexibly. Employing the LD to calculate difference eliminates the need to identify and code into regex patterns all of the ways in which otherwise matching strings might be inconsistent. Instead, emily_lynema@ncsua similarity threshold is tuned to identify close matches while eliminating false positives.edu
Recently, the UW Libraries began an effort to store Electronic Theses and Dissertations (ETD) in our institutional repository which runs on DSpace. We received 6,756 PDFs along with a file of UMI-created MARC records which needed to be matched to our library'''Abstract:'''s custom MARC records (60,175 records). Once matched, merged information from both records would be used to create the dublin_core.xml file needed for batch ingest into DSpace. Unfortunately, records within the MARC data had no common unique identifiers to facilitate matching. Direct matching by title or author was impractical due to slight inconsistencies in data entry. Additionally, one of the files had "flattened" characters in title and author fields to ASCII. We successfully employed LD to match records between the two files before merging them.
With a small IT unit This talk demonstrates one method of matching sets of MARC records that lack common unique identifiers and a wide array might contain slight differences in the matching fields. It will cover basic usage of projects to supportseveral python tools. No large stack traces, requests for development from business stakeholders in just the library can quickly spiral out comfort of controlpure python and basic computational algorithms in a step-by-step presentation on dealing with an old library task: matching dirty data. To help make sense While much literature exists on matching/merging duplicate bibliographic records, most of this literature does not specify how to accomplish the chaostask, increase just reports on the transparency efficiency of the IT “black box,” and shorten time lag between requirements definition and functional releasestools used to accomplish the task, we have implemented a modified Agile/SCRUM methodology often within the development group in the IT department at NCSU Librariesa larger system such as an ILS.
This presentation will provide a brief overview of the Agile methodology as an introduction ==Automating Git to our simplified approach to iteratively handling multiple projects across a small team. This iterative approach allows us to regularly recreate your own open-evaluate requested enhancements against institutional priorities and more accurately estimate timelines for specific units of functionality. The presentation will highlight how we approach each development cycle (from planning to estimating to re-aligning) as well as some of the actual tools and techniques we use to manage work (like JIRA and Greenhopper). It will identify some challenges faced in applying an established development methodology to a small team of multi-tasking developers, the outcomes we’ve seen, and the areas we’d like to continue improving. These types of iterative planning/development techniques could be adapted by even a single developer to help manage a chaotic workplace.source Dropbox clone==
----* Ian Walls, System Integration Librarian, NYU Health Sciences Libraries, Ian.Walls at med.nyu.edu
'''Talk Title'''Dropbox is a great tool for synchronizing files across pretty much any machine you’re working on. Unfortunately, it has some drawbacks:# Monthly fees for more than 2GB# The server isn’t yours# The server-side scripting isn’t open sourceHowever, using the [http://git-scm.com/ Git distributed version control system], file event APIs, and your favourite scripting language, it is possible to create a file synchronization system (with full replication and multiple histories) that connects all your computers to your own server.
Public Datasets in These scripts would allow library developers to collaborate and work on multiple machines with ease, while benefiting from the Cloudrobust version control of Git. An active internet connection is not required to have access to the full history of the repository, making it easier to work on the go. This also keeps your data more private and secure by only hosting it on machines you trust (important if you’re dealing with sensitive patron information).
'''Speaker name, affiliation and email address== Becoming Truly Innovative:'''Migrating from Millennium to Koha==
Rosalyn Metz* Ian Walls, Wheaton CollegeSystem Integration Librarian, metz_rosalyn@wheatoncollegeNYU Health Sciences Libraries, Ian.Walls at med.nyu.edu
Michael BOn Sept. Klein1st, Oregon State University2009, Michaelthe NYU Health Sciences Libraries made the unprecedented move from their Millennium ILS to Koha.Klein@oregonstate The migration was done over the course of 3 months, without assistance from either Innovative Interfaces, Inc. or any Koha vendor. The in-house script, written in Perl and XSLT, can be used with any Millennium installation, regardless of which modules have been purchased, and can be adapted to work for migration to systems other than Koha. Helper scripts were also developed to capture the current circulation state (checkouts, holds and fines), and do minor data cleanup.edu
'''Abstract'''This presentation will cover the planning and scheduling of the migration, as well as an overview of the code that was written for it. Opportunities for systems integration and development made newly available by having an open source platform are also discussed.
When most people think about cloud computing (if they think about it at all), it usually takes one of two forms: Infrastructure == 7 Ways to Enhance Library Interfaces with OCLC Web Services== * Karen A. Coombs, such as Amazon EC2 and GoGrid, which provide raw, elastic computing capacity in the form of virtual servers, and Platform Services, such as Google App Engine and Heroku, which provide preconfigured application stacks and specialized deployment toolslibrarywebchic@gmail.com
Several providersOCLC Web Services such as xISSN, howeverWorldCat Search API, offer access to large public datasets that would WorldCat Identities, and the WorldCat Registry provide a variety of data which can be impractical for most organizations used to download enhance and work with locallyimprove current library interfaces. From a 67-gigabyte dump of DBpedia's structured information store This talk will discuss several simple ideas to the 180-gigabyte snapshot of astronomical improve current users interfaces using data from the Sloan Digital Sky Survey, chemistry and biology to economic and geographic data, these datasets are available instantly and backed by enough pay-as-you-go server capacity to make good use of themservices.
We will present an overview Javascript and PHP code to add journal of table of currentlycontents information, peer-available datasetsreviewed journal designation, what it takes links to create and use snapshots of other libraries in the dataarea with a book, also available ..., and explore how the library community might push some of its own large stores of data and metadata into the cloudinfo about this author will be discussed.
---- '''Talk Title:'''== Adventures with Facebook Open Platform ==
Codename Arctika* Kenny Ketner, Texas Tech University Libraries, kenny.ketner@ttu.edu
'''Speaker nameDeveloping with the facebook platform can be both exciting and something that you wouldn’t wish on your worst enemy. This talk will chronicle the Texas Tech Libraries Development Team experimentation with Facebook Open Platform (sfbOpen)as we attempt to create a facebook-like social media application Texas Tech University Libraries, affiliation(s), and email addresshopefully expanding to the Texas Digital Library (esTDL):'''.
Toke EskildsenMore than just a facebook app or page, The State and University Library fbOpen is a complete implementation of Denmarkthe facebook system on a LAMP stack – Linux, te@statsbiblioteketApache, MySQL, PHP – which must be maintained by the institution itself. This project is at an early stage, so emphasis will be placed on the challenges of installation, configuration, and testing, as well as the pros and cons for institutions that are considering taking on a similar project.dk
'''Abstract:'''== Kurrently Kochief ==
There's something missing in the state of Denmark. Most of our web based copyright deposit material is trapped in a dark archive. After a successful pilot; money and time has been allocated to open part of the data. We tried NutchWAX and it worked well* Gabriel Farrell, but we wanted more. Proper integrated search with existing library materialDrexel University Libraries, extraction of names etc. Therefore we propose the following recipe: Take a slice of a dark archive with copyright deposit material. Get permission to publish it (the tricky bit). Add an ARC reader to get the bits, Tika to get the text and Summa to get large-scale index and faceting. We mixed it up and we will show what happenedgsf24@drexel.edu
'''Talk Title:'''Kochief is a discovery interface and catalogue manager. It rests on Solr and a'''Speaker name(s)Python stack including Django, affiliation(s)pymarc, and email address(es)rdflib. We're using it to highlighta few collections at Drexel. They live at http:'''//sets.library.drexel.edu.
I'll talk about the latest and greatest, including advances in the install and
configuration, details considered in the searcher's experience, and the
sourcing and exposing of Linked Data.
'''Abstract of no more than 500 words:'''== Fedora Commons Repository Workflow with Drupal 6 and SCXML ==
Place your submission at the bottom of the page below this line:* Scott Hammel, Clemson University, scott@clemson.edu
----Clemson is building an enterprise architecture repository to support the Medicaid Information Technology Architecture framework. Using Drupal 6 and Fedora Commons Repository and inspired by Islandora, we've written a module for Drupal that supports artifact governance workflow. Workflow is represented as a state machine stored as SCXML in datastreams on digital objects.
'''Talk Title:'''I will talk about the solution, challenges, standards and how workflow, governance, state, and policy are stored and manipulated as content on digital objects.
JeromeDL - an open source social semantic digital library '''Speaker name(s), affiliation(s), and email address(es)== Forging Connections:'''Current uses of SRU ==
* Sebastian Ryszard KrukT. Michael Silver, Knowledge HivesMLIS Student at the University of Alberta, sebastianmichael.kruk@knowledgehives.com* Jodi Schneider, DERI NUI Galway, jschneidersilver@poboxualberta.comca
Search / Retrieve via URL (SRU) has been touted as the next generation of the Z39.50 protocol. Its use of HTTP communication and XML data formats were designed to allow greater integration with other online resources. In October and November 2009, I interviewed seven SRU administrators from libraries, not-for-profit and for-profit organizations to gain insights into their experiences with the protocol.
'''Abstract of no The results from this small study show that SRU is being used as more than 500 words:'''a replacement for Z39.50. Instead, it is also being used to create connections between information resources and users by leveraging the protocol’s use of web standards. My presentation will focus on reporting the topics which emerged during the interviews, ranging from the history and future of information retrieval to differing views on SRU’s relationship with federated search, OpenSearch and other web protocols.
We will tell about the idea of binding together semantics coming from two sources: legacy, well-crafted annotations provided by librarians, ==Extending EZProxy for Fun and less organized/structured annotations provided by the community of library users. We will present <a hrefProfit=="http://www.jeromedl.org/">JeromeDL system</a> that enables users to provide and manage such annotations; it also implements a number of information discovery solutions that utilize these combined annotations, including collaborative browsing, natural language query templates and collaborative filtering. We will also talk about a vocabulary service used by JeromeDL that encourages users to provide more meaningful annotations than just tags. Finally, we will show how JeromeDL-based libraries contribute to the Web 3.0 linked data by utilizing standard vocabularies, such as SIOC, FOAF, and WordNet, and publishing RDF description of library content.
----* Brice Stacey, University of Massachusetts Boston, brice.stacey@umb.edu
'''Talk Title:'''EZProxy is much more than just an authentication tool for remote access to library resources. As middleware between electronic resources and patrons, EZProxy is the the backbone from which many applications may be built. Potential uses include monitoring resource use to enhance collection development decisions, injecting context-sensitive information and links to tutorials in a branded toolbar for the duration of a session, and using EZProxy as a single sign-on server. These three ideas alone could streamline the user experience, allow for more granular library instruction and increase awareness of what is actually important to users.
Kill In this session I'd also like to initiate a discussion about the search buttoncreation of a collaborative site for EZProxy administrators. The proposed site would feature a private workspace to manage EZProxy configurations, drawn from a public repository of database definitions and authentication schemes. Additionally, the site would be an ideal environment for developing additional applications as described above.
'''Speaker name(s), affiliation(s), and email address(es)== Micro Library Apps:'''Building library functionality into the Google Gadget platform ==
* Michael Poltorak NielsenJason A. Clark, State Head of Digital Access and University Library, Denmark, mn@statsbiblioteket.dk* Jørn ThøgersenWeb Services, Montana State and University Library, DenmarkLibraries, jtjaclark@statsbiblioteketmontana.dkedu
With implementations of the OpenSocial standard, complete functionality within Google Wave, and a huge user base actively using iGoogle, Google Gadgets and the Gadgets API can be used as an emerging platform for bite-sized pieces of library services and applications.
'''Abstract MSU Libraries has applied Google Gadget API technology to allow users to create their own dashboards or waves filled with library content modules. In this session we will demonstrate a wide range of no more than 500 wordsgadgetry including, but not limited to:'''tabbed gateway searching of catalogs and databases, flash-animated library subject maps, a customized database gateway, a digital collections app gadget, a feed aggregator for library data streams, and a gadget for campus maps and street views.
We demo three concepts that eliminate the search button[http://www.lib.montana.edu/tools/gadgets.php http://www.lib.montana.edu/tools/gadgets.php]
1. Instant search. Why wait We'll talk through the anatomy of a Google Gadget, the possibilities for tiresome page reloads when searching? Instant search updates the search result on every key-press. We will show how we integrated this feature into our own API and its use in library search system with minimal changes to settings, and the existing setupXML, Javascript, HTML, and occasional PHP that make it go.
2. Index lookup. Ever dreamed of your own inline instant index lookup?== Can't We demo an instant index lookup feature that requires no search button and no page refreshes - and without ever leaving the search field.All Just Get Along? ==
3. Slide your data. Sliders are an alternative way to fit search results to the user's search context.Examples are sliders that move search results priorities between title and subject and between books by an author and books about the author* Ryan Scherle, National Evolutionary Synthesis Center, rscherle@nescent. org
One of the greatest challenges of a large project is bringing together people from different traditions and getting them to work together. Most Code4Lib attendees are accustomed to working with a team of librarians, technologists, and subject specialists. Working with teams from multiple institutions and multiple disciplines increases the level of complexity, particularly when some teams have a history of maintaining their own discipline-specific technology solutions.
----[http://dataone.org DataONE] is a collaborative repository of scientific data being developed by a group of more than 20 organizations. It will combine contents from a diverse set of scientific repositories, covering many disciplines, metadata schemes, and usage policies.
I will give an overview of the DataONE project and its technical architecture, focusing on the architectural design process and techniques for overcoming the differences between the participating repositories. I will also outline the steps required if you want to connect a new repository to the DataONE system.
'''Talk Title== Data for all:'''Controlling the flood: Refacilitating access to reference transaction data using web-plumbing fittings between a New Titles List and other services with Yahoo! Pipes.based tools ==
'''Speaker name(s)* David Dahl, affiliation(s), and email address(es):'''Jon GormanEmerging Technologies Librarian, Towson University of Illinois, jtgormanddahl@illinoistowson.edu
'''Abstract of no more than 500 words:'''About four years ago the University of Illinois decided Like many libraries, Towson University’s Albert S. Cook Library uses a homegrown web application to create record reference transaction statistics into a New Titles service Microsoft Access database. (http://wwwOurs is informally called StatsTracker.library.illinois.edu/newtitles/) that could provide RSS feeds Previously this collected data was only available in a raw format within the database, limiting its usefulness to just 1 or 2 staff with knowledge of querying an Access database. At These individuals were frequently asked to compile data to aid in the time a balance was struck between complexity of options and limited development timedepartment’s decision-making. Currently a feed is created by adding options, each option narrowing A recent initiative to make this data more publicly accessible (to internal staff) motivated the scope creation of a feedsuite of web-based tools that aggregate and analyze collected data in order to make up-to-the-minute statistics available for use by the Reference Department. Selecting Using a date rangecombination of ASP.net, SQL, Microsoft Chart Controls, Unit Library and a call number range will retrieve material that match all three of the criteria. It was hoped that at some point a generic tool would be able to further manipulate Visual Web Developer (VWD) application for development, the StatsTracker Analysis Toolkit makes reference transaction data accessible and combine feeds produced usable by any member of the simple options to customize very specific feeddepartment. Yahoo! Pipes has emerged to fill that niche.
The talk This session will cover pipes that range from filter for a keyword in one feed to combining the New Titles List with services like the LibraryThing API or Worldcat APIs. Examples will also be given in development process, demonstrate how to integrate the output VWD facilitated development, and present possibilities for further use of Yahoo! Pipes into webpages and how we have put them into our CMS (OpenCMS). The talk will make sure to address areas where Yahoo! Pipes either fails or is cumbersome and simpler CSS and Javascript solutions have workedthis combination of tools[[Category: Code4Lib2010]]

Navigation menu