Allow NULL "use restriction" fields for located URIs
The asset.uri.use_restriction field, which is really a sort of public notes
field for 856 fields, was grabbing the $u subfield (URL) as a sort of last-gasp
effort to give it some data. However, the effect was rather odd and led to
workarounds like Conifer's skin to avoid displaying the use restriction field
if its value was identical to the URL, etc.
Instead, stop grabbing $u and handle the case where use_restriction column is
NULL gracefully, just like the schema intended.
Delete ##URI## call numbers and uri_call_number_map entries on bib reingest
This approach will lead to some acn/auricnm ID inflation, but it works.
Addresses LP# 761130 (immortal ##URI## entries in asset.call_number) reported
by Ben Shum and LP# 761085 (cannot delete bib with ##URI## volumes) reported
by Jason Etheridge.
Protect dumb JavaScript engines from having to deal with actual Unicode
The holdings_xml format did not include an XML declaration, but adding that
as we do here still does not make the Firefox and Chromium JS engines capable
of consuming XML that contains Unicode content outside of the base ASCII
range.
So, we invoke entityize() to convert anything outside of the realm of
ASCII to XML entities. An alternative would be to invoke entityize() in
OpenILS::Application::SuperCat::unAPI::acn but it's not clear if that
would interfere with any other uses.
With this change, library names / copy location names with Unicode content
can be displayed correctly on the search results page.
At some point (r16750) we started doing a numeric comparison of
$flesh instead of just checking to see if $flesh was defined; this
returned false when $flesh == 'uris', preventing URIs from being
included in the marcxml-uris unAPI format.
This restores URIs to marcxml-uris and so we can revert the extra
BibTemplate call in rdetail_summary.xml.
Specify the holdings_xml unAPI format for URI calls
The unAPI marcxml-uris format is not returning URIs at the moment.
While we're getting that fixed, use the holdings_xml format to
get the URI job done; requires an extra JS call, but that's
better than not working at all.
Escape rather than filter SIMILAR TO metacharacters in patron crazy search
The filtering I introduced in r19983 was overly aggressive, and included
characters that weren't actually SIMILAR TO metacharacters. Instead, escape
each character, carefully going through the list of metacharacters listed at
http://www.postgresql.org/docs/8.4/interactive/functions-matching.html
Works for email addresses like "foo.bar+baz@example.com".
* used version from wiki, which provides same results as the
previous version but performs better on large databases
* now works without editing (a vacuum cannot run inside of a transaction)
* don't do vacuum full, just a regular vacuum analyze
particularly for running the catalog embedded in the staff client, which makes no visual indication of page progress, it's good to let the caller know something is happening w/ a search. after a 1-second search delay, show a small progress spinny icon
inisial staff client integration in record details page w/ new staff js file; move footer and other js loading to their own templates; hide top-nav pane (my account summary) for embedded mode; load slim version of marc html (no external css; no print button)
moving toward svf for mattype extraction; much media/material-type icon cleanup; icons are now accessed directly via code instead of inconsistent and map-requiring human names
Add support for Multi-Homed Items (aka Foreign Bibs, aka Linked Items)
Evergreen needs to support the ability to attach a barcoded item to more than one bibliographic record. Use cases include:
1. Barcoded E-Readers with preloaded content
* Readers would all be items attached to a single "master" bib record in the traditional way, through call numbers that define their ownership
* Each reader, as a barcoded item, can be attached through Multi-homed Items to records describing the list of preloaded content
* These attached Multi-homed Items can be added and removed as content is swapped out on each reader
2. Dual-language items
* Cataloger decides which of several alternate languages is the primary, and attaches the barcoded item to that record in the traditional way
* Alternate language records are attached to this item through Multi-homed Items
3. "Back-to-back" books -- two books printed upside down relative to one another, with two "front" covers
* Cataloger decides which of the two titles is the primary, and attaches the barcoded item to that record in the traditional way
* Alternate title record is attached to this item through Multi-homed Items
4. Bound Volumes -- Sets of individual works collected into a single barcoded package
* Cataloger decides which of the titles is the primary (or creates a record for the collection as a whole), and attaches the barcoded item to that record in the traditional way
* Remaining title records for the collected peices are attached to this item through Multi-homed Items
Functionality funded by Natural Resources Canada -- http://www.nrcan-rncan.gc.ca/com/
Please see http://git.esilibrary.com/?p=evergreen-equinox.git;a=shortlog;h=refs/heads/multi_home for the full commit history.
patch from Ben Ostrowsky (w/ input) to add support to the Apache redirect module to also optionally read redirect skin and domain from the library IP's configuration file.
after clearing out a stale ses cookie, return user to originally requested resource instead directing home. this prevents the case where going to 'my account' in presence of expired ses cookie redirects the user home
For consistency, make new sub-pages (payments history, preferences) load
as a separate page in the same path (prefs -> prefs_notify) instead of
either an ?expand= option or a sub-path page (prefs -> prefs/notify).
Let the my-account base page load the generic base page as a wrapper so
that my-account page are not required to load both and (for me, anyway)
make the inheritance more obvious.
Enable marc2sre.pl to run reasonably fast with a large set of bibs
Our previous iteration of marc2sre.pl used an ILIKE stanza
beginning with a wildcard to match system control numbers
without having to specify the institution's MARC code.
This worked, but was painfully slow in large bib sets as
the database needed to use a bitmap index scan to find matches.
By adding a --prefix flag, the user can specify the institutional
MARC code for the set of records and we can use an exact match
against metabib.full_rec.value, which is immeasurably faster.
This is, of course, a problem if there are multiple institutional
MARC codes in use for a given set of bibliographic records.
Improve error handling in marc2sre.pl when bib ID is not found
If we can't find a bibliographic record ID to use in our load, then
skip that MFHD record and move on to the next one. Using the counter
gives sites a chance to identify which record caused the problem.
Aside: bitmap index scans for leading '%' LIKE searches make the
--bibfield / --bibsubfield extremely slow in large datasets. If
at all possible, avoid this path!