These notes are a bit disparate, but here’s what I got in terms of notes for the breakout session. Continue reading “Batch Record Editing Processes Breakout Group Notes”
Panel of 3 on linked open data.
Here are my notes from Code4LibBC Day 1 lightning talks! Continue reading “Code4LibBC Day 1: Lightning Talks”
Afternoon of lightning talks Continue reading “Code4Lib North: Day 2 Afternoon Notes”
Morning presentations Continue reading “Code4Lib North: Day 2 Morning Notes”
Morning presentations Continue reading “Code4Lib North: Day 1 Morning Notes”
- Doris Lovadina-Lee,
- Aileen Farray,
- Lies Weijs, Ryerson Library Continue reading “TRY 2013: RDA Matters: A Snapshot at Ryerson Library”
Mark Matienzo – Wielding the Whip: Affect, Archives, & Ontological Fusion
- need to talk about emotion – a lot of things going on in my life
- inspired by Archives by Emotion
- why do we not acknowledge that archives are based on emotion, stories
- facebook ‘like’ is not an emotion
- losing connections to materials and history when not thinking about the platform and how that effects our stories
- can we write stories about our collections?
- should be using existing linked data to make those narratives
- let’s build this
- Full Write-up
Jason Casden and Cory Lown – My #HuntLibrary
- student engagement platform
- how do students use the space?
- what do they choose to document? – using instagram
- student-drive archival selection
- making use, harvesting social media
- moderated, responsive
- use for public display, can interactive, including larger public display
- inspired by kitten war: built image battle
- calculate popularity score
- also about preservation, collect images
- and then when files are uploaded
- Demo: hydratest.bpl.org:3000
- On GitHub
Will Hicks – Metadata Entry Beyond Usability
- think back to volunteering to donate blood, etc. what value did it give you?
- ~300 individuals creating metadata
- stats have bumped up with a new interface
- but 10% of the stuff is hidden
- “usable” forms are great, but little “ownership” and lack of domain expertise
- what if we applied the ideas used in social media?
- invite with openID, personalize projects, badges, stats, visualization, etc.
Kelly Lucas – Drupal OPAC Recipe
- the poor man’s Blacklight or Vufind with Solr backend
- MARC record dump into Solr using SolrMARC
- install search API, Search API Solr, Facet API, Sarnia and Views from Drupal.org
- configure Sarnia + FacetAPI (boosts, enable facets)
- create a view, add some fields, create an exposed filter (full text search box)
- slap some facets to the side of the page
- issues: new/updated records requires direct connection to ILS
Karen Coyle – Nerd Poetry
- cowboy poetry – told around the camp fire
Mark Redar – Django Dublin Core App
- plugin app
- on Github
- RecordExpress – lightweight, easy to use esp. for those not familiar with XML
James Stuart – Taming Email
- a really big problem
- part of your job
- here’s how to tame email
- don’t organize: organizing your email is like alphabetizing your recycling
- turn keyboard shortcuts on
- don’t be distracting
- Also take a look at Making Thunderbird More Gmail-y
No Break! Have a Cute Animal Anyway
Granularity in Library Linked Open Data
- self-similar at all levels of granularity
- each circle represents of things that look very similar (snowflake looking pattern but of different sizes)
- characteristic of fractals
- cannot determine level: all levels are equal, some more equal than others
- What is described by a bibliographic record? or a single statement?
- What is the level of description? How complete is it? e.g. AACR2
- How detailed is the schema used? How dumb? – especially relevant right now. The more detailed, the higher level of granularity possible.
- Semantic constraints? Unconstrained?
Resource Description Framework – Linked Data
- Triple: This resource | has intended audience | Juvenile
- Subject / Predicated / Object
- do each of these parts have granularity?
- higher/lower level, but should talk about coarse or fine grained granularity
Subject: What is the Statement About?
- we can focus on description an article / resource / work, then think about coarser or finer granularity:
- coarser: consortium collection / RDF map
- library collection / digital collection
- super-aggregate journal title / jurnal index
- aggregate: issue / festschrift
- focus on description an article / resource / work
- component: section / graphics / page
- sub-component: paragraph / markup
- finer: word rdf/xml
- uri / node
Predicate: What is the Aspect Described?
- similar coarse/fine breakdown:
- membership category
- access to resource
- access to content
- suitability rating
- audience and usage
- audience of audio-visual material
- diagram: possible audience map (partial) – unconstrained version to avoid collisions of isbd/dct/schema/rda/m21/frbrer
- different links can be made while still retain proper semantic links
- currently constructing just one giant graph
What is the Aspect Described?
- coarse to fine:
- resource record
- manifestation record
- title and s.o.r
- title statement
- title of manifestation
- title word
- first word of title
- why do librarians need so many titles? Why not just use dublin core title and be done with it? Because we need it to do our work e.g. spine title to browse
- title = string identifier
- RDA: what to do with this? how do we apply these needs?
- possible semantic map (partial) – I won’t even try to reproduce this
- need to take into account names and ranges
- make it more difficult, but more powerful
Semantic Reasoning: The Sub-Property Ladder
- this is where the graph becomes useful and property
- machines can’t reason, so we’re demantic the semantics such that we can give the rules to machines to process our data
- semantic rule:
- if property1 sub-property of property2;
- then data triple: resource property1 “string”
- implies data triple: resource property2 “string”
- otherwise, data triple remains the same
- simple enough for computer to carry out
- doesn’t matter how complex the map actually is, because it can still do it in matters of seconds
- machine entailment: isbd” “hast title proper” (finer) -> dct: “has title” (coarser)
- might sound simple, but making a computer do interferance
- ‘dumb(ing)-up, data has been lost, but still meaningful – moved from one schema to another
Data Triples from Multiple Schema / Entailed from Sub-Property Map / rom Property Domains
- frbrer: “has intended audience” – “primary school”
- isbd: “has note on use or audience” – “for ages 5-9”
- rda: “intended audience (work)” – “for children aged 7-“
- m21: “target audience” 0> m21terms: -> “Juvenile”
- definition attached to the vocabulary
- also talking about granularity
- can map the sub-property to top level of unc: “has note on use or audience”
- “is a” frbrer: “work”, isbd: “resource”, rda: “work” – rda and frbr schema actually separate, not semantically linked even though vocabulary is similar and RDA is based on FRBR
- once stabalized can be drawn from each other
What is the Aspect Described?
- coarser to finer:
- animation screenwriting
- children’s cartoon screenwriting
- different controlled vocabulary
- graph of RDA for author/creator/screenwriting in relation to work and agent
- graph of same thing, but for dc for creator and agent
- what is the semantic relationship between the dct creator and the rda creator?
- marcrel author maps to dc contributor, not creator – what is the relationship between rda author and marcrel author?
- decision from 2005, needs to be reappraised and reviewed
- relationship between dc creator and dc contributor?
- how does lcsh “screenwriters” fit?
- also has issues
- e.g. full-text indexing: down to the word level
- BabelNet: A very large multilingual ontology
- can get quite complex and granular
- users can actually generate useful metadata
- can use statistical methods to remove extremes and come back with consensus
- going to cause granularity problems e.g. “OK for my kids (7 and 9)”, “Too childish for me (age 14)”
- keep it simple, stupid
- keep it simple and stupid?
- data model is very simple: triples!
- in terms of complexity, actually very simple
- but metadata content is complex
- and therefore, resource discovery is complex
- complex structure of application of simple rules, similar in the hard sciences and math
- simplicity is elegance
- Anyone can say anything about any thing
- someone will say something about every thing
- in every conceivable way
- and then constrained linguistically
- open world assumption: the absence of a statement is not a statement of non-existence
Will it get so granular that it becomes too complex?
And the rest is science