DITF March 15, 2024 Kick-off meeting
Mar 15, 2024
Attendees
Here: Christian Ward, Jill Strykowski, Corinne Ornelas, Can Li, Tim Fluhr, Heather Cribbs
Absent: Christina Hennessey, Dolly Lopez
Agenda
Goal: Determine starting topics and plans
Topic | Notes |
---|---|
Views and relevancy recommendations | Looked over table created by CH (See spreadsheet) CW: how do we determine what relevancy is “good” for our users? CL, HC: when doing testing, be aware of other factors such as timing of indexing, CDI updates, availability settings Look at University of Washing(?) methodology for collecting example search data CW, CO: look at common issues from LibChat/Reference sessions that should be focused on as common problems and/or can be used as test cases HC: Break up chart by columns and address as different issues CL, CW: do others look at the json and pnx code? Is it helpful in understanding? For Group: How do we develop rubrics for both testing and persona creation? For CH: Can we ask each campus to provide quick explanation of what changes they recall making to view settings and ranking and what their reasoning was? CL emailed: I’d like to briefly share what I recently learned and have been experimenting with adjusting relevancy ranking. With Primo search API, I can retrieve JSON records with the PNX record score. Records are listed with the score from the highest to the lowest, and the order matches what we see on OneSearch UI. These scores are different from scores we see in a browser. When adjusting a boost factor, I can see how the score changes correspondingly. This provides a predictable pattern, based on which I can then know what boost factor would be most appropriate for the desired ranking. |
FRBR learning and improvements | Group agreed this topic should be reviewed a revisited for whole CSU Uncertain if it will improve result as desired, but would be good to know |
Title/portfolio matching in Primo | For CH, JS: do we have any updates in ExLibris tickets or release notes on this topic? Check in with Nikki DeMoville - ask if she can share her background information, examples and solution/development requests |
Resource types across system | JS: Can I give us the crazy task of documenting complete mappings for : Resource type vs. Material type vs. Physical item type in Alma → Primo → CDI → GTI → digital collections, etc. Is ExLibris going to do a normalization project on this data like they are doing with CDI subjects? |
Secondary resource type work | For CH/JS: check with RMFC |
OA resource tagging for Primo filter | For CH, JS: do we have any updates in ExLibris tickets or release notes on this topic? Not otherwise discussed |
Enhancement request to improve course name/ID as a filter in Primo | Currently indexes and sort as full string – should sort as Alpha and then as number |
Pendo??? | did not discuss |
Linked data and AI?? | did not discuss |
Action items
For CH/JS:
Recommended reading
Configuring the Ranking of Search Results in Primo VE Subscribe
Good methodology Adventures in Primo Boosting (Igelu)
Measuring and Predicting Search Engine Users' Satisfaction | Dan, O., & Davison, B. (2016). Measuring and Predicting Search Engine Users’ Satisfaction. ACM Computing Surveys (CSUR), 49(1), 1–35. https://doi.org/10.1145/2893486 |
Web Search Engines - Not Yet a Reliable Replacement for Bibliographic Databases | Hughes, E. (2018). Web Search Engines - Not Yet a Reliable Replacement for Bibliographic Databases. Evidence Based Library and Information Practice, 13(3), 85–87. https://doi.org/10.18438/eblip29378 |
Lessons Learned: A Primo Usability Study | Brett, K., Lierman, A., & Turner, C. (2016). Lessons Learned: A Primo Usability Study. Information Technology and Libraries, 35(1), 7–25. https://doi.org/10.6017/ital.v35i1.8965 |
A Framework for Measuring Relevancy in Discovery Environments | Galbreath, B. L., Merrill, A., & Johnson, C. M. (2021). A Framework for Measuring Relevancy in Discovery Environments. Information Technology and Libraries, 40(2), 1–17. https://doi.org/10.6017/ital.v40i2.12835 |