D-Lib: Facing the Challenge of Web Archives Preservation Collaboratively: The Role and Work of the IIPC Preservation Working Group
D-Lib: Statistical Translation of Hierarchical Classifications from Dewey Decimal Classification to the Regensburger Verbundklassifikation
D-Lib: Helping Members of the Community Manage Their Digital Lives: Developing a Personal Digital Archiving Workshop
Part ee of Amazon crawl..
This item belongs to: data/ol_data.
This item has files of the following types: Data, Data, Metadata, Text
Join a panel of Book Industry Study Group (BISG) and American Library Association (ALA) leaders at this year’s 2015 ALA Annual Conference in San Francisco when they discuss the results of a newly-released study on public library patrons’ use of digital content.
During the conference session “Digital Content in Public Libraries: What Do Patrons Think?” panelists will discuss the results of a new study by the BISG and ALA that was designed to provide invaluable insight into how readers interact with e-books in a library environment. The session takes place from 3:00 to 4:00 p.m. on Sunday, June 28, 2015, at the Moscone Convention Center in room 131 of the North Building.
The digital content survey was developed to understand the behavior of library patrons, including their use of digital resources and other services offered by public libraries. The study examined the impact of digital consumption behaviors, including the adoption of new business models, on library usage across America.Speakers
- Kathy Rosa, director, Office for Research and Statistics, American Library Association
- Carrie Russell, program director, Public Access to Information, Office for Information Technology Policy, American Library Association
- Nadine Vassallo, project manager, Research & Information, Book Industry Study Group
The post How do library patrons feel about digital content? appeared first on District Dispatch.
A previous post I made reviewing the Ithaka report “Streamlining access to Scholarly Resources” got a lot of attention. Thanks!
The primary issue I’m interested in there: Getting our patrons from a paywalled scholarly citation on the open unauthenticated web, to an authenticated library-licensed copy, or other library services. “Bridging the gap”.
Here, we use Umlaut to turn our “link resolver” into a full-service landing page offering library services for both books and articles: Licensed online copies, local print copies, and other library services.
This means we’ve got the “receiving” end taken care of — here’s a book and an article example of an Umlaut landing page — the problem reduces to getting the user from the open unauthenticated web to an Umlaut page for the citation in question.
Which is still a tricky problem. In this post, brief discussion of two things: 1) The new “Google Scholar Button” browser extension from Google, which is interesting in this area, but I think ultimately not enough of a solution to keep me from looking for more, and 2) Possibilities of Zotero open source code toward our end.The Google Scholar Button
This plugin will extract the title of an article from a page (either text you’ve selected on the page first, or it will try to scrape a title from HTML markup), and give you search results for that article title from Google Scholar, in a little popup window.
Interestingly, this is essentially the same thing a couple of third party software packages have done for a while: The LibX “Magic Button”, and Lazy Scholar. But now we get it in an official Google release, instead of hacky workarounds to Google’s lack of API from open source.
The Google Scholar Button is basically trying to bridge the same gap we are; it provides a condensed version of google scholar search results, with a link to an open access PDF if Google knows about one (I am still curious how many of these open access PDF’s are not-entirely-licensed copies put up by authors or professors without publisher permissions);
And it in some cases provides an OpenURL link to a library link resolver, which is just what we’re looking for.
However, it’s got some limitations that keep me from considering it a satisfactory ‘Bridging the Gap’ solution:
- In order to get the OpenURL link to your local library link resolver while you are off campus, you have to set your Google Scholar preferences in your browser, which is pretty confusing to do.
- The title has to match in Google Scholar’s index of course. Which is definitely extensive enough to still be hugely useful, as evidenced by the open source predecessors to Google Scholar Button trying to do the same thing.
- But most problematically at all, Google Scholar Button results will only show the local library link resolver link for some citations: The ones that have been registered as having institutional fulltext access in your institutional holdings registered with Google. I want to get users to the Umlaut landing page for any citation they want, even if we don’t have licensed fulltext (and we might even if Google doesn’t think we do, the holdings registrations are not always entirely accurate), I want to show them local physical copies (especially for books), and ILL and other document delivery services.
- The full Google Scholar gives a hard-to-find but at least it’s there OpenURL link for “no local fulltext” under a ‘more’ link, but the Google Scholar Button version doesn’t offer even this.
- Books/monographs might not be the primary use case, but I really want a solution that works for books too — and books are something users may be especially interested in a physical copy instead of online fulltext for, and books are also something that our holdings registration with Google pretty much doesn’t include, even ebooks. And book titles are a lot less likely to return hits in Google Scholar at all.
I really want a solution that works all or almost all of the time to get the patron to our library landing page, not just some of the time, and my experiments with Google Scholar Button revealed more of a ‘sometimes’ experience.
I’m not sure if the LibX or Lazy Scholar solutions can provide an OpenURL link in all cases, regardless of Google institutional holdings registration. They are both worth further inquiry for sure. But Lazy Scholar isn’t open source and I find it’s UI not great for our purposes. And I find LibX a bit too heavy weight for solving this problem, and have some other concerns about it.
So let’s consider another avenue for “Bridging the Gap”….Zotero’s scraping logic
Instead of trying to take a title and find a hit in a mega-corpus of scholarly citations like the Google Scholar Button approach, another approach would be to try to extract the full citation details from the source page, and construct an OpenURL to send straight to our landing page.
And, hey, it has occurred to me, there’s some software that already can scrape citation data elements from quite a long list of web sites our patrons might want to start from. Zotero. (And Mendeley too for that matter).
In fact, you could use Zotero as a method of ‘Bridging the Gap’ right now. Sign up for a Zotero account, install the Zotero extension. When you are on a paywalled citation page on the unauthenticated open web (or a search results page on Google Scholar, Amazon, or other places Zotero can scrape from), first import your citation into Zotero. Then go into your Zotero library, find the citation, and — if you’ve properly set up your OpenURL preferences in Zotero — it’ll give you a link to click on that will take you to your institutional OpenURL resolver. In our case, our Umlaut landing page.
We know from some faculty interviews that some faculty definitely use Zotero, hard to say if a majority do or not. I do not know how many have managed to set up their OpenURL preferences in Zotero, if this is part of their use of it.
Even of those who have, I wonder how many have figured out on their own that they can use Zotero to “bridge the gap” in this way. But even if we undertook an education campaign, it is a somewhat cumbersome process. You might not want to actually import into your Zotero library, you might want to take a look at the article first. And not everyone chooses to use Zotero, and we don’t want to require them to for a ‘briding the gap’ solution.
But that logic is there in Zotero, the pretty tricky task of compiling and maintaining ‘scraping’ rules for a huge list of sites likely to be desirable as ‘Bridging the Gap’ sources. And Zotero is open source, hmm.
We could imagine adding a feature to Zotero that let the user choose to go right to an institutional OpenURL link after scraping, instead of having to import and navigate to their Zotero library first. But I’m not sure such a feature would match the goals of the Zotero project, or how to integrate it into the UX in a clear way without disturbing from Zotero’s core functionality.
But again, it’s open source. We could imagine ‘forking’ Zotero, or extracting just the parts of Zotero that matter for our goal, into our own product that did exactly what we wanted. I’m not sure I have the local resources to maintain a ‘forked’ version of plugins for several browsers.
But Zotero also offers a bookmarklet. Which doesn’t have as good a UI as the browser plugins, and which doesn’t support all of the scrapers. But which unlike a browser plugin you can install on iOS and Android mobile browsers (although it’s a bit confusing to do so, at least it’s possible). And which it’s probably ‘less expensive’ for a developer to maintain a ‘fork’ of — we really just want to take Zotero’s scraping behavior, implemented via bookmarklet, and completely replace what you do with it after it’s scraped. Send it to our institutional OpenURL resolver.
I am very intrigued by this possibility, it seems at least worth some investigatory prototypes to have patrons test. But I haven’t yet figured out how where to actually find the bookmarklet code, and related code in Zotero that may be triggered by it, let alone the next step of figuring out if it can be extracted into a ‘fork’. I’ve tried looking around on the Zotero repo, but I can’t figure out what’s what. (I think all of Zotero is open source?).
Anyone know the Zotero devs, and want to see if they want to talk to me about it with any advice or suggestions? Or anyone familiar with the Zotero source code themselves and want to talk to me about it?
Filed under: General
Last updated May 14, 2015. Created by Peter Murray on May 14, 2015.
Log in to edit this page.
Binder is an open source digital repository management application, designed
to meet the needs and complex digital preservation requirements of museum
collections. Binder was created by
Artefactual Systems and the
Museum of Modern Art.
Binder aims to facilitate digital collections care, management, and
preservation for time-based media and born-digital artworks and is built
from integrating functionality of the
A presentation on Binder's functionality (Binder was formerly known as the
DRMC during development) can be found here:
Slides from a presentation at Code4LibBC 2014, including screenshots from the
application, can be found here:
Further resourcesArchival Record Manager and EditorLicense: GPLv3 Package Links In DevelopmentOperating System: Browser/Cross-PlatformTechnologies Used: XSLTProgramming Language: PHPDatabase: MySQLworks well with: Archivematica
Tonight, the House of Representatives will vote on the USA FREEDOM Act of 2015, H.R. 2048 to finally ban the “bulk collection” of Americans’ personal communications records (library, telephone and otherwise) under Section 215. Critically, it also would preclude the use of other surveillance laws (related to “PEN registers”) and NSLs to get around that prohibition and would bring the “gag order” provisions of the USA PATRIOT Act into compliance with the First Amendment by permitting them to be meaningfully challenged in court.
The bill, not incidentally, also permits phone and internet companies to publish information (in a sufficiently specific form to be useful) about the number of requests they receive from the government to produce personal subscriber information. It also, for the first time, would create opportunities for specially cleared civil liberties advocates to appear before the secret Foreign Intelligence Surveillance Act (FISA) court that authorizes surveillance activities. The bill also makes important “first step” reforms to privacy-hostile provisions, including Section 702, of the FISA Amendments Act.
ALA and its many public and private sector coalition partners strongly support passage of H.R. 2048. That message was underscored by the more than 400 librarian lobbyists who took to Capitol Hill on May 5, during the American Library Association’s (ALA) National Library Legislative Day. They carried with them a stirring and emphatic OpEd urging real reform entitled “Long Lines for Freedom” by ALA President Courtney Young, which was published that morning in The Hill, a Congress-centric newspaper widely read by Members of Congress, their staffs and the national press.
While House passage of the USA FREEDOM Act is widely expected, its fate in the Senate is uncertain at best. Stay tuned for more on how you can help!
The post U.S. House poised to pass real reforms to USA PATRIOT Act appeared first on District Dispatch.