Thursday, June 5, 2008

Executive Summary of Our Final Report

LSC accepted our final report and is currently considering our recommendations!

Below is the executive summary of the final report. The entire report, including appendices, is available here: http://staff.library.wisc.edu/rdetf/RDETF-final-report.pdf.

Stay tuned for our final open forum. We haven't set a date yet, but it is likely to be in July due to vacation schedules.

Thanks again for all of your feedback and support! Your participation at the open forums and all of the articles and ideas you shared with us really enriched this experience.


Executive Summary and Conclusions

This report recommends that the Libraries decouple the discovery interface from the ILS and implement a discovery interface that is aligned with user behaviors and expectations. It also recommends investigating the feasibility of replacing WorldCat FirstSearch with WorldCat Local to facilitate resource discovery beyond local collections. Additional conclusions drawn from environmental, user and product scans include redoubling efforts toward developing a single sign-on across library and UW resources, implementing direct linking in SFX and supporting a culture of assessment in order to better understand our users and be present in the online and physical spaces in which they work and play.

The conclusions in this report are tuned toward the library catalog because the library has other projects underway for improving and implementing access to non-catalog data. It is expected that these recommendations will apply to additional types of data beyond the library catalog.

The new discovery environment must:

  1. Decouple the interface from the ILS so that it is sleek, lean, and enabled for rapid change.
  2. Maintain complete control over the discovery interface, data, and index. Nothing should be unchangeable.
  3. Emphasize simplicity in the interface. As Lorcan Dempsey noted: "'simple search' but supported by smart results and rich browse" (single search box, single sign on, clean layout).
  4. Include sophisticated search and result functionalities (faceted browsing and/or topical clustering, natural language, obvious relevancy ranking, searching within results, clarity via FRBRization).
  5. Seamlessly integrate and deliver UW collections and resources at the campus and at the system level (library catalogs, library web sites, digital collections, museums, archives).
  6. Adapt to user behaviors and expectations (personalization, recommendations, "did you mean?" functionality, internationalization).
  7. Encourage personalization and customization of the discovery environment in MyUW and course management systems, including Learn@UW and Moodle.
  8. Deliver library search functionality, links and services where our users work and play, including off-campus resources (Amazon, iGoogle, Facebook, WorldCat).
  9. Compare well in design and user experience to popular Internet destinations. Resource discovery in the libraries must become Fast, Smart, and Engaging to compete in the current and future information marketplace.
  10. Be staffed for excellence and continuous change (developers, graphic and interaction designers, and public services staff). This includes collaboration and leadership within the Open Source community.

Recommendations

  1. Implement a decoupled interface for resource discovery (library catalogs, library web sites, digital collections, museums, archives) that meets the requirements of our vision.
  2. Enhance current discovery environment by:
    1. Continuously assessing, analyzing, and developing new tools and functionality for discovery.
      1. Maintain current awareness of browser extensions and library toolbars (LibX, Conduit.com).
      2. Widgets for personalized web pages (iGoogle, Facebook, NetVibes, PageFlakes).
      3. Promote use of information management tools (such as Zotero, Google Notebook, deli.cio.us, RefWorks, EndNote, Papers (Mac only)).
    2. Promoting library data reuse by exposing all freely available library metadata to direct harvesting by indexers.
    3. Using NetID for My MadCat Account and Library Express instead of the eleven digit ID.
    4. Enabling a persistent sign-on into library and campus resources using NetID.
    5. Finding ways to be more social and expand beyond work and research needs to encourage inquisitive exploration of all types.
    6. Encouraging Ex Libris to fully realize the DLF ILS Discovery Task Force API recommendations (referred to as the "Berkeley Accord") to allow the development of local discovery applications using library data.
    7. Implementing direct linking to full-text content wherever possible; in particular, within FindIt and MadCat.
    8. Allocating staff time to analyze and improve the accuracy of FindIt linking. Determine when and why FindIt fails and if there is anything we can do to make this better. Are there highly requested journals which we should license or obtain faster?
    9. Identifying and addressing discovery needs via mobile devices as soon as possible.
    10. Putting a single search box to the Libraries Web site with target selections for MadCat, the Libraries Web site, QuickSearch for Articles (i.e., http://www.lib.virginia.edu/).
    11. Strongly considering implementing WorldCat Local as a public interface for the UW System OPAC and WorldCat FirstSearch.
    12. Improving MadCat now by:
      1. Including value-added information, such as book covers, sample passages of text, reviews, and RSS feeds of journals' tables of contents. Adding and enabling user-generated content, like LibraryThing for Libraries.
      2. Making relevancy ranking the default search results display for more than just a 'words anywhere' search.
      3. Investigating adding MadCat to the MetaLib General Resources QuickSearch set.
      4. Linking to Google Books through the Google Books API and/or linking to CIC Google Collections Archive (in progress).
      5. Adding icons to indicate format in results lists.
      6. Relabeling fields to make them meaningful to patrons (subject links become "Find more like this").
      7. Improving the call number browse.
      8. Providing direct export to RefWorks and other citation managers.
      9. Displaying persistent links on brief and full records.
      10. Exploring linking to Amazon, Wikipedia, etc. for contextual information.
      11. Highlighting searched keywords in results.
      12. Enabling automatic stemming/truncation, if possible.

Thursday, May 15, 2008

Two weeks of online library interaction

You Must See This

The Resource Discovery Task Force (RDTF) tested 28 library homepages for two weeks, using the online tool CrazyEgg (CE) to record user site interaction. We collected a lot of data, which you really *must see* to begin to understand the significance of the study:
How It Works

Using a tiny bit of javascript, CE records the operating system, browser, referrer, associated search terms, window size and time to click for every user-generated click on the webpage (technically, the 95% of the world with javascript enabled). This data is aggregated into multiple views to help you understand how users interact with your site--where they click, how long they take to find a link, etc.

The screen-captures of the heatmaps we've collected show which spots on the pages were most frequently clicked--the hotter the area the more clicks.

Analyzing the Results

Individually, libraries can learn a lot about which links on their site are popular and which links are not. It's simple to see the results, change a few links, reorganize your page a bit and retest. In a few iterations of your design, you'll greatly improve the usability of your site.

On our campus, we have a common library site template. Looking across all the libraries using the template, here's what I believe to be true:
  • Headers - institutionalized and standardized content works well. Our headers see very consistent use across all the implementations. I believe this means they are well designed and very effective.
  • Databases - if you look at the Business library's results, you'll see their users really want simple access to database links. Looking across all the homepages, it turns out that quick, homepage level access to subject specific database links is the right way to go.
  • Search - our library template buries the optional search box in the footer of the design. When elevated, such as Wendt Library's search box, users opt for search with much more frequency. Having a consistent and comprehensive search solution across the campus library websites would be a major boost to usability.
  • Serial Content - many library sites have "dynamic" content indicating news and events or recent additions to the collection. These links are not frequently clicked, which makes me think we need to consider the staff cost of generating this content. Certainly, we should strongly consider downsizing the footprint of this content on our homepages.
Usability and Maturity

What do we do with all this new information about user interaction on our sites? Answer: we begin to shape a better, more user-friendly web presence for our libraries. Jakob Nielsen wrote a classic pair of web posts on the 8 stages of usability maturity. These posts are a great read and help illustrate the difficulty of achieving great usability in any corporation.

The stages:
  1. Hostility Toward Usability
  2. Developer-Centered Usability
  3. Skunkworks Usability
  4. Dedicated Usability Budget
  5. Managed Usability
  6. Systematic Usability Process
  7. Integrated User-Centered Design
  8. User-Driven
I think our libraries are somewhere between stages 3 and 4 at the moment. We have a few large projects (such as the RDTF) in the works to measure and recommend usability enhancements. There is a formal staff-time commitment (LWS Web Site Committee) towards improved design and functionality across our library system.

Gathering data to improve user interaction is critical to improving usability. This study should be seriously considered by every person who is a webmaster inside our libraries. We've done some good work towards improving our sites, but we have a lot of work left to do to make them great. I hope this study leads to more and continued user data gathering across campus. I also hope our future LWS brownbags lead to a greater sense of web-development community within the libraries.

Your Turn

If you made it to the brownbag Wednesday, you saw me demo the "confetti" view CE provides. This is were the CE data truly shines. Unfortunately, I cannot give everyone on campus the password/login to the CE site itself to produce these reports... having that information would allow you to add/delete tests or cancel our account altogether.

This was a one-off study, so all we can make available are the screenshots and data collected during the testing phase. However, you and your library *should* strongly consider signing up for your own CE account (their product is amazing so be nice, purchase a paying account!). Running tests across many of your pages helps gain a better perspective on how your site could be improved to better service your patrons.

If you have any questions about buy or using CE, just let me know. BTW--I'm not paid by CE is any way, I am just a big fan.

Questions? Comments?

Please let me know what you think of these screenshots. I would love to see many people comment on their reactions to seeing this data.

Cheers,
- Eric for the RDTF

Monday, April 21, 2008

Recheduled May Open Forum

The Resource Discovery Exploratory Task Force open forum that was set to occur on Friday, May 2 is being RESCHEDULED for

Wednesday, May 14
Noon to 1:00 pm
Memorial 126

Please move the May 2nd forum in your personal calendars to this new May 14th date.

Much of this forum will be devoted to discussing information received through our online user survey, focus groups, and web site data gathering tests. A more detailed agenda will be forthcoming.

Thanks for your continued support!

Monday, April 7, 2008

Open Forum Recap 04/04

At last Friday's open forum, we showed several innovative tools and initiatives that are just getting started or are not quite scalable to our environment. We asked forum attendees to pay special attention to the features offered by each and to contrast them with the features offered by the commercial products demonstrated at the March open forum.

We began the by asking the following question: "Describe the best web-based service you've experienced. What made it excellent? What distinguished it from the crowd?" Leave a comment describing the best web-based service you've received. I'll post about the responses from the forum crowd a little later.

Next, Sue Dentinger demoed University of Virginia's Project Blacklight. Blacklight is a prototype of a faceted discovery tool for catalog data and beyond. So far, they have indexed 3.7 million MARC records, a 500 text object subset from their digital collections repository, and 320 Tang Dynasty Chinese poems. Blacklight, like VuFind, is based on Solr/Lucene.

Our second demonstration, presented by Allan Barclay, was LibraryThing for Libraries. LibraryThing for Libraries enriches the catalog by drawing on content contributed by the collective intelligence of LibraryThing members. LibraryThing for Libraries adds book recommendations, tag clouds, and links to other editions and translations of a work to the OPAC. Allan showed us LibraryThing for Libraries implementations at the Danbury Public Library and San Francisco State University.

Albert Quattrucci showed us Scriblio, an open source OPAC based on WordPress, a blog publishing platform. Scriblio includes several innovative features, like Google Book Search integration and a "text this to your cellphone" option. Albert showed us Plymouth State University's implementation of Scriblio.

Finally, I showed the demo version of the Open Library. The Open Library is an open source project of the Internet Archive and is financially supported by Brewster Kahle. Aaron Swartz, who co-authored RSS when he was 14 years old, is the project's leader. The goal of the Open Library is to create at least one wiki-like Web page for every book ever published. The Open Library will be truly free to the people in that everyone will have the ability to create, catalog, and contribute content.

The forum ended with a brief discussion of the scope of resource discovery. How can we make other UW-Madison collections, like museum holdings and departmental resources, more findable and accessible?

Monday, March 31, 2008

Generic View from Nowhere

I stumbled upon this article written by Andrew Abbott of the University of Chicago at Library Juice. So, as is my wont, I will share some of the passages that resonated with me. Again this isn't an exhaustive critique of the paper, just some of the passages that struck me as important to the Resource Discovery Task Force. Note that Abbott speaks only of humanist and social science researchers not scientific researchers. Quotes from the paper are italicized.

Central to that investigation [Future of the Library Task Force] was a study of digital versus physical use of library materials, an analysis which showed clearly what we should have guessed ahead of time -that students who are heavy physical users of the library are also heavy electronic users and vice versa. The idea that electronic research was actually replacing physical research - at least at the expert end of the scale- proved wrong.

I think that this is something to bear in mind. I often fall into the trap of digital versus physical, but perhaps I should really think about heavy versus light users of libraries. Is format even an issue? Will tools librarians' build help?

More broadly, that library researchers have projects with clear designs is a myth. A few library researchers may actually have such clear designs. And the rest of us pretend to have had them after the fact.

Abbott underlines the fact that humanistic research in libraries is a very organic endeavor. There is no clear path through the literature. Browsing and reading are part of the process. A part that librarians, for the most part, are not privy to.

Not only is known item searching a relatively minor part of expert library research, precisely structured research questions are also a relatively minor part of expert library research.

Again, Abbott points to the importance of the practice of browsing for any tools librarians provide.

Everything I could find out about stack behavior in the 1950s indicated that faculty and graduate students weren't using catalogs, even for known-item searches. Nor were they using most of the wonderful apparatus I had written about, built for them by Wilson and ALA and the library profession. They were just wandering into the stacks and trolling. They were indeed standing in the stacks and reading whole chapters, then pulling something else off the shelf and reading that.

Is there any chance researchers will use tools librarians build? If Abbott's research is any indication scholars disengaged from librarians in the 1920s for a variety of reasons. In a large part because librarians represent what Abbott calls a universalist approach as opposed to scholars inclination for a partial or specialty approach to subject access.

But the message was everywhere the same. Faculty and graduate students got their references either from hearsay or from other people's footnotes or reference lists, just as - in fact - I was doing myself.

Now if faculty and graduate students were getting their research bibliography via hearsay or other professionals' published work, why were they doing this? The answer, at least theoretically, seemed obvious. What these sources had that the general bibliographical tools lacked was selectivity.

In my opinion, this is a major problem with bibliographic tools. Quality isn't addressed in any but a cursory fashion. Catalogs don't tell a researcher what the best book on Joyce is. And that in many instances is exactly the information library researchers need.

Finding something is easy. It's knowing you ought to be looking for it that is hard.

It was the librarians' contention that there ought to be one master index, but the research scholars always want partial indexes, indexes slanted their way, organized by their way of seeing the world, not by a generic view from nowhere.

library researchers started withdrawing from this universalist project in the 1920s and gradually erected a system of specialty tools and a set of research practices that enabled them to bypass the hugely inefficient searches that were the only possibility under the universal bibliographical system.

That's all for now. Back to building the master index.


Thursday, March 27, 2008

Faceted or guiding searching question

I have a question for you. Back a bit more than 2 years ago North Carolina State University came out with a major new look for a library catalog interface. See: http://www.lib.ncsu.edu/catalog/ This interface is based some products from Endeca which facilitate ‘guided navigation’. While the look has changed a bit since it was first deployed, the ability to suggest to patrons ways they may want to focus or refine their search using a faceted display of key subject, format, or dates off to the side, was a major library catalog innovation. Now instead of having folks refine their search query up front, you quickly gave them many ways to go to continue their search.

Now just about every other major library citation and catalog software quickly developed this type of faceted browsing off to the side, including MetaLib and Primo from ExLibris, WorldCat Local from OCLC, Encore from Innovative Interfaces, etc.

I’m wondering how much you really see patrons using these guided navigation aids that are off to the side after they do a search. We have this now in our MetaLib quicksearches, available right on the main page of our library website at http://www.library.wisc.edu/ . What’s your take on how much these are used?

A nice overview of faceting is in Wikipedia.

Friday, March 14, 2008

Notes on Information behaviour of the researcher of the future

Notes on Information behaviour of the researcher of the future - Executive summary

As I read through the Executive summary of "Information Behaviour of the Researcher of the Future," I noted passages that would be of interest to the Resource Discovery Task Force. I offer these notes below with some explanation of why I believe they are important for the Task Force to consider. These notes are not exhaustive, and I encourage others to read the article and offer their take on the report in future blog posts. The passages I comment on below are only the passages that caught my eye, so to speak. Quotes from the report are in italics. A link to the study's project page is at the end of the post as well.

they [Google Generation] exhibit a strong preference for expressing themselves in natural language rather than analysing which key words might be more effective. (12)

I feel this finding is very significant. Many library workers, including myself, enjoy a powerful advanced search. That said, many research studies, the rise of Google, and my experience at public service desks, all point to the fact that I don't see much power searching--keywords reign. I remember one reference meeting at which a librarian unveiled the top ten search queries from a database. I only remember the top search query, but the other top searches were just as unimpressive. The top search query that fateful month was protein. Yes, a single ubiquitous word, at least in this engineering database, stole the honor. So much for sophisticated searching!

CIBER’s considered view is that the real issue that the library community should be concerned about is the rise of the e-book, not social networking. (17)

This is a timely finding with the release of the Google Book Search API . With more and more books being digitized by a variety of entities, a challenge for any resource discovery tool will be to point users to possible print as well as digitized versions of a text.

for library interfaces, there is evidence that multimedia can quickly lose its appeal, providing short-term novelty. (19)

I think we all know this fact, but boy is it hard to resist some of these bells and whistles. This brings to mind, to me at least, Aquabrowser. In my humble opinion, I don't think that the visual interface would prove useful to me as a searcher. The Resource Discovery Task Force has demo'd some implementations of Aquabrowser, if you are curious:

Aquabrowser
Columbus Metropolitan Library implementation
Oklahoma State University implementation
University of Chicago implementation


But there is no evidence in the serious literature that young people are expert searchers, nor that the search skills of young people has improved with time. (22)

This finding definitely bucks the trend of most media coverage of the Google Generation. That said, this finding does coincide with my experience in library instruction and public service. Libraries offer a complicated information landscape with unmarked borders. Students typically (I'm generalizing here, I know) don't have a firm understanding of what a library catalog IS, never mind how to search a catalog effectively, nor do students have an intimate understanding of the composition of the information landscape before them. An intimate understanding would include: the publication process, awareness of aggregators, licensed versus purchased content, etc. Without this understanding students and other users are at a distinct disadvantage compared to library workers. We are the insiders. I don't say all this to toot the library worker horn, but this "tacit knowledge" that we possess as library workers does, I believe, enrich our search behavior. Even simple tactics such as double-checking the accuracy of our systems give us library workers a leg up--I know I don't believe SFX all the time.

Students usually prefer the global searching of Google to more sophisticated but more time-consuming searching provided by the library, where students must make separate searches of the online catalog and every database of potential interest, after first
identifying which databases might be relevant. In addition, not all searches of library catalogues or databases yield full-text materials, and NetGen students want not just speedy answers, but full gratification of their information requests on the spot. (31)

The above quote reminds me of a presentation that Steve Frye gave at the Reference Retreat in January 2008. He showed QuickSearch sets in comparison to Google Scholar. This made me think, as does the above quote, that a fruitful avenue for the future would be to develop QuickSearch sets with certain users in mind (personalized search). The Library has already developed some QuickSearch sets, but if we could improve the variety and usefulness of the QuickSearch sets, I think this would be a helpful service to users. I realize there are technical issues and performance issues to consider, but for now I can dream.

From the report "power browsing" is an information seeking behavior that the new discovery tool should address in order to be useful. The authors define "power browsing" new form of online reading behaviour is beginning to emerge, one based on skimming titles, contents pages
and abstracts: we call this `power browsing’. (8, 19, 31)

The authors of the study seem to denigrate "power browsing" at least that is my initial impression. To my mind, power browsing is just efficient searching behavior. User want to quickly ascertain whether an article or book is relevant to their project. Nothing wrong with that. For the Resource Discovery Task Force this behavior underlines that a resource discovery tool should lend itself to power browsing. In other words, a searcher should quickly and easily access: digitized content, full-text, reviews, book covers, table of contents, indexes, tags, etc.


The significance of this for research libraries is threefold:

•they need to make their sites more highly visible in
cyberspace by opening them up to search engines

•they should abandon any hope of being a one-stop
shop

•they should accept that much content will seldom or
never be used, other than perhaps a place from which
to bounce (31)

making simplicity their core mission. (31)

personal/social searching guidance offered so successfully by Amazon for many years? (33)

Finally, the authors leave us with these conclusions. More food for thought. Simplicity is an elusive goal in my opinion. Resources change, interfaces change.... I do think whatever resource discovery tool we adopt it should have some sort of recommendation system akin to Amazon's: Customers Who Bought Items Like This Also Bought. Well, I'm running out of steam, but I'm anxious to here others' thoughts on this report and resource discovery.

Jon Udell offers some further analysis and criticism of the report at his blog.


Google Generation Project page

Tuesday, March 11, 2008

Vendor – Supplied Versus Open Source

Vendor – Supplied Versus Open Source

The products you saw Friday 3/7/08 and at our last meeting do not replace our library catalog MadCat. Instead they require us to export data from MadCat, process it, possibly merge it with data from one or more other sources, then feed that data on a regular basis into a powerful indexer and search environment for our patrons to use.

So a main goal is faster, easier, more powerful and more flexible searching and retrieving of data. Another goal is having the ability to change as new features and new ideas and new methods of presenting data become available.

Are we going to be stuck with a rigid look and feel that is simply ‘newer’? Or could we output our data in a way that, as new ideas come along or new mobile devices become more available, our data can be readily adapted to have a new look and work in a different way that suits our rapidly changing needs?

And can our output of data be pre-sorted and relevancy-ranked according to criteria we possibly have control over?

So, the question is, do we pay up front for a vendor-supplied solution where these “paths” of exporting data have already been set up for us, and the look and feel is only moderately under our control? Or do we use vendors who provide the infrastructure but use API’s to let us design the exact interface we want. An API application programming interface is a source code interface that an operating system or environment provides to support requests for services to be made of it by computer programs. (This is a wikipedia quota from a Computerworld article.)

AquaBrowser, Primo, Endeca and WorldCat Local, are all vendor-supplied, and all have the ability and are already tested in large institutions like ours. The Digital Library Federation is also working on a list of features any API from an ILS should be prepared to support.

Another option is to use WorldCat Local’s API (which gives us the ability to use WorldCat Local’s underlying structure, but write our own public interface using ‘calls’ back to the data. David Walker showed at Code4Lib very recently his interface code based on the WorldCat Local API, so this capacity is functioning at some level right now. Clearly OCLC is recognizing the importance of offering multiple options for differing types of organizations and the importance of allowing local control and innovation using a stable underlying base.)

Either way we choose, we’ll need staff to set up the processing from MadCat and other sources. And if we go the Open Source route, which could potentially offer us the most flexibility, we have to make a staff investment to be able to make the changes and implement new features. Some of this cost of change might be lessened with a vendor-supplied solution—but then, depending on the vendor, we could be right back to where we’re at right now which is running on a dinosaur catalog infrastructure while the web-world changes so rapidly around us.

If we go the Open Source route, Steve Meyer recently reminded me of a quote by Richard Stallman: “Think free as in free speech, not as in free beer.” (I should add to this that I am somewhat mis-using Stallman’s quote here. He considers the Open Source movement a very watered down version of the Free Software Movement and he really wasn’t a supporter of it—he wants software to really be free for ethical reasons, not just the practical reasons behind the Open Source movement.)

The main point I want to make is that whatever solution we take is going to cost people, $$, and time. So the most important decision I think we can make is to choose a platform and a path where we keep the doors open to make at least look and feel and even underlying structure changes very easily and very rapidly. The data needs to be in our control. I mean, aren’t you sick of, as you say ‘can’t we do xxxx with MadCat?’ someone like Curran Riley or Edie Dixon saying ‘No, we can only change yyyy, not xxxx.’? :-)

But one thing to keep in mind is our size. It’s far more work and effort to do this level of relevancy ranking ‘on the fly’ on large sets of data. And that’s why either a vendor or Open Source solution really needs us to export and pre-process the data. We have on order of 6 million bibliographic records and we want to mesh this with data from additional sources. Steve Meyer was recently telling me that he had learned from our very knowledgeable DoIT LIRA staff that once you get over about 1 million records, the processing and work needed to do the indexing of this number of records is a completely different beast and FAR more complicated.

VUfind, the only Open Source project we’ve demo’d so far in the last forum, at this point in time hasn’t handled this large -- millions of records -- technical issue yet. It is currently indexing well under 1 million records. However, it is built on an underlying Open Source structure (using Lucene and Solr from the Apache foundation) which has the ability to handle larger size databases. We do have excellent technical staff here at DoIT, but we’ll need more if we choose undertake this level of work.

And there are other Open Source projects are also coming along using an underlying structure that can support our needs. One of which is the eXtensible Project, as Karen pointed out.

Thank you.

What Should the "Catalog" of the Future Include?

I scanned the front sides of the blue sheets so you can see how your colleagues reacted to this question. I've also pulled out some common themes in case you prefer Cliff Notes. Is anything missing from this list?

Emphasis on Local Holdings:
  • "Somewhere in-between. Emphasis needs to be on authoritative - hard to find (not easily googlable)."

  • "Only stuff we license and own because no one starts a search at the library for things that google can find faster."

  • "Catalog is good for managing physical content...the electronic "stuff" should be discoverable through general search tools that normal people use."

Librarians Should Select:
  • "No, there should be values applied."
  • "Everything in the info universe should be eligible for inclusion. Subject experts should continue to decide what to include."

  • "It should include everything we - or our faculty - think is relevant/useful for research & instruction here."

It's Really Up to the Users:
  • "What do users expect of the Libraries catalog - that should drive the answer."

  • "I hope that it will include all types of information - or somehow seamlessly integrate campus and non-campus resources. Why? I think that users would appreciate 'one-stop shopping.'"
  • "I do wonder if patrons expect the catalog to be a finding aid for library items."

Access and Findability are More Important than Scope:
  • "Only link to things that students/faculty/staff could access readily."
  • "I think that the 'concept of scope' is not as important as the ease of finding information."
  • "The question should be: from the library's single search interface, what resources in addition to the things owned or leased should the searcher be able to access. That single interface should search multiple discovery tools, of which the catalog is only one of many."

Monday, March 10, 2008

Open Forum Recap 3/7

We started last Friday's Open Forum by asking attendees to respond to the following questions:

"In the future, what should be the scope of the Libraries' online catalog - that is, what kinds of information should it contain? Should it only include library owned or leased items or should everything in the information universe be included? Why?"

We received some very thoughtful responses and I will post about those a little later.

Next, we demonstrated two next generation resource discovery tools. They were Ex Libris' Primo and OCLC's WorldCat Local. Following our quick presentations of these products, we asked you to vote on which one you preferred. 18 were for WorldCat Local, 3 people liked Primo, and 3 of you didn't like either of them :).

You can try searching Primo and WorldCat Local on your own at the following sites:

Primo
University of Iowa Libraries
University of Minnesota Libraries
Vanderbilt University Libraries

WorldCat Local
San Mateo County Library
University of Washington Libraries
(Remember, anyone can create a WorldCat account to see personalization features.)

Karen also discussed an open source project centered at the University of Rochester called EXtensible Catalog. They do not have a product to demo at this point in the project.

Finally, Sue closed the session by reading a wonderful piece she wrote about why we might want to consider exploring open source options. She is going to post the text of that document in a separate entry.

It certainly was a very full hour. We hope you didn't find it too overwhelming! Thanks again for your continued support.

Tuesday, February 19, 2008

Why do people use a university's library catalog?

So what do people come to MadCat, our library catalog at the University of Wisconsin - Madison, looking for? Do they already have a book/journal or whole bibliography of items in mind? Are they very familiar with this topic and want to see if there's anything new or anything they missed? Did a friend mention an interesting book they had just read at last night's party? Or are they brand new to a topic and just want to educate themselves on it.

How do we make a catalog of what our library owns, can get you to the full text of, or has a license for, work the best for all these different needs. Our library catalog, although very impressive, is not comprehensive. We don't own everything or have access to everything available. However, unlike most online bookstores, it does go quite far back in time and provide a wealth of material no longer available on the open market. Even as books and newspapers become digitized and available electronically, we'll still need to identify and track what is physically present in our collections.

But what does our public need when they search for something, and what suits their needs best? How can we get them to what they want with the least amount of effort yet also suit our need to track and know where something is physically at any moment. And how can we allow our collection materials to easily mesh with all the other materials a researcher may have collected from other sources if that is what they want, or simply deliver the one item they need at that particular moment. Is there one interface that can suit a variety of needs whether you know exactly what you want or you just want to browse a topic?

As we search for the best software being developed to hopefully improve our patrons library catalog experience, we'll be asking patrons--what is it you like or dislike about online bookstores or catalogs like Amazon. And how does searching Amazon compare with doing the same search in a library catalog with new features, such as the catalog at the University of Washington using WorldCat Local or the University of Iowa using Primo.

So please join us determining some interface differences. Find a favorite title in all three sites above. Now that you've found that title, misspell or perhaps 'rearrange' the words of that book or item so that your query won't exactly match. How well does each interface do? Are there differences in handling something that doesn't quite match-up right? What do you like or dislike about each of these interfaces? Let us know!

Tuesday, February 5, 2008

Responses from Open Forum Questions

At last Thursday's open forum, we asked you to answer the following questions:
  • What will the library search experience be like in five years?
  • What are the first things you'd change?
We collected 27 cards from you. Here is a ranked list of your responses to our first question. In five years, we hope that our search experiences will include:

  1. Ability to search large numbers of databases at once– 17
  2. Improved online browsing (facets, related results) – 11
  3. User-enriched records (tags, reviews, comments) – 9
  4. Customization, different views of the same database (including the ability to create lists and store records) – 9
  5. Intuitive searching – 8
  6. Spell correction and suggestions for alternate search terms – 7
  7. Visual-based searching – 6
  8. Focus on user-centered design – 6
  9. Integration into users online work environment – 6
  10. Fast searching – 2
  11. Removal of library lingo – 2
  12. Full-text everything / full-text searching – 2
  13. Relevancy Ranking – 2
  14. All formats in a single record – 1
What are the first things you'd change?

Most of you didn't indicate what should be changed first. Those that did mainly suggested that the inclusion of spell checking, faceted browsing, and enriched records should be priorities.

Is there anything missing from this list? Is the ranked list an accurate depiction of your change priorities?

Thursday, January 31, 2008

Open Forum Recap 1/31

A special thanks to everyone who made it to the Resource Discovery Open Forum this afternoon!

For those of you who couldn't make it, here is what we covered. Allan Barclay demoed Aquabrowser and Eric Larson showed us VuFind.

Aquabrowser
Columbus Metropolitan Library implementation
Oklahoma State University implementation
University of Chicago implementation

VuFind
Live Demo
Brochure

We asked everyone to answer the following questions at the beginning of the session:

"What will the library search experience be like in five years?"
"What are the first things you'd change?"

If you weren't at the forum, leave a comment with your answers to these questions. I'll summarize the responses from forum attendees in an upcoming post.

Tuesday, January 22, 2008

Where Do We Start and What Do Our Users Really Want?

Thanks to everyone for commenting. It is important to acknowledge that people begin searches differently and that these processes change according to the task at hand. Searching for information is personal and so we definitely want to select tools that can be customized by our users.

I asked you to describe how you begin a search for information to stress the importance of focusing on the user and making decisions based on their needs. The comments revealed two common threads in how you search for information. Many start their searches with a search engine and many value recommendations from trusted peers in social networks. Let's take a look at some recent studies to give us a rough idea of how our students might compare.

According to OCLC's 2005 report, College Students' Perceptions of Libraries and Information Resources, 89 percent of college student information searches begin with a search engine. Library Web sites were selected by only 2 percent of students as a place to begin an information search. Search engines were rated higher than libraries in the areas of reliability, cost-effectiveness, ease of use, convenience, and speed. So it seems that many of our users might agree with Sue in that whatever they use "it's got to be simple and FAST."

Only 2 percent of college students are starting searches at library Web sites, but this doesn't mean that the aren't using libraries. The recent PEW Internet study, Information Searches that Solve Problems, found that young adults (18-29) are the heaviest users of libraries when looking for information to solve problems. They are also the most likely visitors to the library for any purpose and especially value access to computers and the Internet.

So, if we know we have young adults in our libraries (I sure know they are in College Library!), how do we get them to use library resources? Well, we can start by asking them how they would improve library tools and by paying attention to what they already use.

Researchers at Idaho University Libraries conducted focus groups with undergraduate library users and asked them to describe a "dream information machine." The students imagined a machine that was a "mind reader," that was "intuitive," and could determine information needs without them having to verbalize them. The "dream machine" would be able to solve all of their information needs by searching a comprehensive collection of information resources. The ideal information source would also be portable and always available. What would your ideal information source look like?

Where does social networking fit in to all of this? Many of you reported that you start a search for information by consulting with a respected peer either in person or online. Does this mean that librarians should be in social networking sites or that social networking should be in the catalog? There will probably never be a consensus about whether or not librarians should be in Facebook and MySpace. The recent University of Michigan Library Web Survey found that 23% of library users would be interested in contacting a librarian in Facebook or Myspace, nearly half wouldn't be interested in contacting a librarian this way, and the rest don't use social networking sites. Many librarians in the blogosphere took this to mean that we shouldn't be in social networking sites. To me, this means that I should be in social networking sites for the 23% interested in talking to me (as long as I'm not stalking those that don't). Let's wait and see how our users feel about enriching the catalog with user reviews and ratings before we make any assumptions there.

The Resource Discovery Task Force does plan to conduct user surveys and focus groups to get a better idea of what our users value. Let's talk more about user needs and how to fill them at our open forum on Thursday, January 31 from 12:30 - 1:30 in Memorial 126.

Happy first day of class!

Thursday, January 10, 2008

Next Generation Resource Discovery

What is the Resource Discovery Exploratory Task Force?

The Resource Discovery Exploratory Task Force is charged with developing a vision for information resource discovery in the Libraries that supports teaching, learning and research at UW-Madison. The Task Force will conduct an environmental scan of current options, including the opportunities and challenges of each. Our tasks include looking at available software options and querying our patrons and staff about their information seeking behaviors and how we can better satisfy their information and resource discovery needs.

The Resource Discovery Task Force members are:
  • Allan Barclay
  • Susan Barribeau
  • Sue Dentinger (co-chair)
  • Kelli Keclik (co-chair)
  • Eric Larson
  • Albert Quattrucci
  • Karen Rattunde
  • Curran Riley

The task force is slated to present its findings to UW Madison library management by the end of May 2008.


Why do we need a Resource Discovery Exploratory Task Force?

Resource discovery with traditional library tools is a frustrating and time-consuming process for many researchers. As a result, information seekers frequently bypass the library in favor of tools that provide quick and easy results and are fun to use, like Google. Google offers features like relevancy ranking, customization, spell checking and provides instant access to a variety of information resources (web sites, images, videos, etc) without asking you to select an index. Compare Google searching to the many steps and decisions one must make in order to locate a scholarly article on our web site and you start to get a sense of the problem.


What’s in the future?

Luckily, there are several commercial options, like Primo and Endeca, and several open source options, like VuFind and LibraryFind, that are working to close this gap. The task force will explore these different products and approaches and make a recommendation based on user needs, our vision for the future of resource discovery, and the current library technology infrastructure. We will also define the broad technological infrastructure required to be well-positioned to handle future information discovery changes. How can we create an environment where change happens more rapidly and easily?


Why do we need this blog?

The purpose of this blog is to promote discussion and learn about your vision for the future of resource discovery. We will ask big picture questions and solicit feedback on particular products and tools. This blog will function as a companion piece to our monthly open forums. Staff, students, and faculty are encouraged to contribute to the blog and attend open forums. Open forums will be held in Memorial Library 126 at the following times:

  • Thursday January 31, 2008, 12:30 – 1:30 pm
  • Friday March 7, 2008, 12:00 – 1:00 pm
  • Friday April 4, 2008, 12:00-1:00 pm
  • Friday May 2, 2008, 12:00 – 1:00 pm

A more complete description for each session is forthcoming.


A question for you!

Where do you typically begin your search for information on a particular topic? Why?