A quick look at JorumOpen

As anyone with even a passing intererest in UKOER will know, JorumOpen went live earlier this week and I, for one, was dying to see just what the good folk at Mimas and Edina have come up with with their customised DSpace installation (and possibly “borrow” one or two ideas for Leeds Met Open Search!).

JorumOpen Home is at http://open.jorum.ac.uk/xmlui/ and allows the user to browse OER by FE or HE subject; alternatively there are links to browse by Communities & collections/Issue date/Authors/Titles and Keyword.  There is also a simple search box and a link to an Advanced search form:

The results page comprises different functionality depending on the search – for example, browsing by subject heading displays “Recent Deposits” and allows the user a simple/advanced search, or browse by Titles/Authors/Dates within that subject heading (I like this hierarchical search functionality); also includes an RSS button to subscribe to updates within the collection.

Results themselves comprise a hyperlinked title, author/author affiliation and date of deposit as well as a thumbnail graphic where available:

The record page is worth looking at in detail (this item – http://open.jorum.ac.uk/xmlui/handle/123456789/567):

Show full item record (link) – Full Dublin Core metadata record

Share (AddThis button) – third part social network service allowing record to be emailed to a friend or posted to various social networking sites.

The simple record comprises:

Title/Author/Description/Keywords/Persistent Link/Date

Then there are three buttons:

“Export resource” that requires a valid email address “As some resources are quite large in size it can take some time to prepare them for download. Due to this we required you to supply a valid email address so that you can be notified when your download is ready.”  Then follows an email from support@jorum.ac.uk that informs that “The item export you requested from the repository is now ready for download.” and includes a link to download the compressed file which comprises all files associated with the resource.*

“Preview content package” which allows the user to quickly view the different files and components of the resource in their browser without downloading (though it doesn’t work for .zip files)

“Download original content package” does exactly what it says on the tin and downloads a compressed file of all files associated with the resource.*

* I’m not entirely sure what the difference is between “export” and “download” – though the exported zip is bigger and contains more files (dublin_core.xml as well as imsmanifest.xml for example) – may be someone can enlighted me?

CC Licence Note – briefly explains implications of CC and links to relevant anchor later in the record.

Files in this item – allows the user to expand a list of files and download them individually (this particular item comprises 16 .zip and 2 .docx)

Creative Commons Licence – Link to relevant CC licence (opens in a nifty little window.)

Terms of service – Link to Jorum terms of service (also opens in a nifty little window)

This item appears in the following collections – linked to appropriate search terms in browse tree

Show full item record (repeated link from top of page) – Full Dublin Core metadata record

This item has been viewed x times – presumably counts visits to the record page

All in all, first impressions are pretty favourable and there are certainly some ideas that I would like to explore for Leeds Met Open Search – I’ve already included the AddThis button on the development server and plan to go live with it as soon as it has been approved by the powers that be (there are one or two issues with user tracking by this third party service – Mike has disabled Flash tracking that the widget injects into the page by default but it will still track each click-through.)

I’m also keen to explore how we may manage packaged content in a similar way to JorumOpen (preview content and download options for individual files) – currently we have very little packaged content in the repository but the default download link is currently just for an individual file – I do know that intraLibrary is able to manage content packages, however, and that a package download link is exposed by SRU so I think we should be able to achieve this.

Browse by date (of deposit) should also be achievable I think but browse by author is a little more problematic by SRU (both for research and OER) as there is no authority file for authors.

I’m not sure about recording page visits – will need to speak to Mike.

Now I just need to figure out the most efficient way of getting our UniCycle resources into JorumOpen – I will look at the deposit process in a later post (depositors can log in from JorumOpen Home via UK Federation) and I think Jorum are still exploring harvesting RSS feeds from ukoer projects though, as discussed in a recent post, our feed is not currrently suitable for this.


End of a pipe dream?

I’ve been enjoying exploring Yahoo Pipes and have now managed to generate a feed for OER that also incorporates the author vCard; I’m sure it must also be possible to extract other data including rights information (though I’m not quite sure how!):


(vCard not visible in the Pipe but is displayed if the Pipe is rendered as RSS)

It also occurred to me that it may be possible to use a similar method to extract the abstract for research RSS feeds (which is in the second description field) and I’ve been grappling with Pipes to this end but am struggling at the moment – tantalisingly I am able to construct a “Path to item list” that returns an individual abstract but for the life of me, can’t figure out how to return abstracts for all records – SRW:records.SRW:record.1.SRW:recordData.lom:lom.lom:general.lom:description.1.lom:string.

I’ll certainly continue to use Pipes as it’s a very powerful tool, especially for a non-programmer like me, however, for the time being this mini-project is postponed. Nothing is ever wasted though and I’m sure what I’ve learned will come in useful somewhere down the line….

Pipe dream

This afternoon I’ve been grappling with Yahoo Pipes, trying to generate a feed that incorporates metadata from an SRU query.  I’ve made a modicum of progress and, as I’m now a bit stuck, this is just a quick post to document that modicum before it all leaks away over the weekend.

The Pipe module that seems most appropriate is “Fetch Data”: “This module retrieves any XML, JSON, iCal or KML file and tries to extract a list of elements using the provided path parameter.”

I input an appropriate URL to query “ukoer” which will return all OERs uploaded as part of UniCycle:


And read the instructions here which describe how to use the “Path to item list” field in order to extract just a portion of the data by listing the nested XML elements, separating each with a dot (“.”)

After a bit of trial and error I was able to zero in on the “Description” field for the first record by entering SRW:records.SRW:record.0.SRW:recordData.lom:lom.lom:general.lom:description.lom:string.content into “path to item list”

However, when I attempt to run the pipe, nothing is returned  – though the correct field IS displayed correctly in the debugger panel:

Also, what I would like to do of course, is to return the “Description” field for ALL records but I have no idea how to achieve this…I’ll have another look when I’m nice and fresh on Monday morning!

Return all records via Open Search

This primarily for my reference – in order to return all records from a given collection enter cql.allRecords=1 into the standard search at http://repository.leedsmet.ac.uk/main/index.php and select which token you wish to use (Research/Open Educational Resources)