Upload
sanjeet-mann
View
67
Download
0
Embed Size (px)
DESCRIPTION
Results from an availability study with University of Redlands undergraduates conducted Spring 2014 and presented at the 29th Annual NASIG Conference in Fort Worth, TX.
Citation preview
Why can’t students get the sources they need?Results from a real availability study
Sanjeet Mann
Arts & Electronic Resources Librarian, University of Redlands
29th Annual NASIG Conference, Ft. Worth, Texas
May 2, 2014
How do you get full text?
Librarian Student
Paige Mann Carlos Puma
Research methods
Technical errors = availability study
• Quantitative
• Large sample for statistical validity
• Researcher tests access (“simulated availability”)
Human interaction = usability study
• Qualitative
• 5-7 users
• Researcher observes library users
My methodology• Cognitive walkthrough
• 7 students x 2 searches x 10 results ≈ 142 interactions
• Jing screen capture software
• Demographic survey and results spreadsheet
Results
Errors
System 31
User35
Both16
Severe systems errors• A&I database has no OpenURL link
• Target database refuses the OpenURL
• A&I database has bad/missing metadata• Can info be found in Google?
• Knowledge base doesn’t offer article-level linking • Is student willing to browse?
Severe human errors
• Didn’t test link
• Used system incorrectly
• Overlooked important information
• Got frustrated and gave up
Student experiences user error
Conceptual model
Questions availability studies can address
• How often do errors occur? Should we be satisfied with our technical infrastructure? (systems)
• How often do users need ILL? (interlibrary loan)
• Do we have enough full text in the collection? (collection development)
• Are we teaching users what they need to be successful at obtaining electronic resources? (instruction)
For discussion at University of Redlands
• 13% error rate in 2013 simulated study• Common problems = source metadata, KB support for OA titles
• 41% local availability (research libraries average 60%)• 2 of 3 items not held in local collection in 2013 study• 43% interactions result in ILL in 2014 study
• Threshold concepts vs. search/retrieval mechanics• How do you teach students to be thoughtful, resilient searchers?
What would I do differently?• Larger, personalized incentives
• Simplify research design
• Cognitive walkthrough as group activity
• Jing + Camtasia Studio worked OK
Further Reading• Selective bibliography: http://goo.gl/4fu47
• Slides: http://www.slideshare.net/sanjeetmann
• University of Redlands availability study datasets• 2012 (simulated) http://goo.gl/606us • 2013 (simulated) http://goo.gl/O5XK9A • 2014 (real) http://goo.gl/BaAm5T