← Previous day | Today | Next day → | Search | Index
All times shown according to UTC.
Time | Nick | Message |
---|---|---|
12:39 | jungle | hello people |
13:34 | owen | kados: is the plan to create an opac system preference for color stylesheet like the intranet has? |
13:52 | kados | owen: yep |
13:52 | owen: the syspref's already in rel_2_2 | |
13:52 | owen: not sure if support for it is built in to the scripts yet | |
13:56 | hey thd | |
13:58 | dweezil19 | can i ask a quick question about koha? |
14:04 | kados | dweezil19: sure |
14:20 | dweezil19 | Is it possible to authenticate via active directory? |
14:56 | kados | dweezil19: hypothetically, yes |
14:56 | dweezil19: Koha has support for LDAP | |
14:57 | dweezil19: can you use that with active directory? | |
14:57 | dweezil19 | i guess so |
14:57 | it is an ldap server so that should work (i think!) | |
15:00 | thanks, I've got to run now | |
15:19 | dce | has anyone here dealt with importing MARC data from Follett when ordering books from them? (Not migrating from Follett software.) |
15:24 | thd | kados: I am awake again, are you? |
15:24 | kados | thd: yes, but I have a conference call soon I must prepare for |
15:25 | thd | kados: after your call would be when? |
15:25 | kados | dce: could you expand on what you mean by that? |
15:25 | thd: probably around 3:00 or so | |
15:25 | thd: I'll ping you when I am available to talk | |
15:26 | thd | kados: I await your ping |
15:26 | kados | dce: if you mean integrating the order process within Koha's acquisition module, noone has done that to date |
15:26 | dce: but if we had a sponsor for such a feature it could be done very quickly | |
15:26 | dce: if you'd like to talk more about it later, let me know | |
15:27 | dce | kados, when ordering books from Follett - http://www.flr.follett.com/ - our library staff has the option of getting a disk with MARC records. |
15:27 | thd | dice: Do those records cost extra? |
15:27 | dce | We are looking at using Koha in a new school opening this fall and they want to know if they will still be able to import that info. |
15:27 | kados | dce: (In that case, you can import MARC directly into Koha from the disk ... ) |
15:27 | dce | thd, I can ask |
15:28 | kados | dce: (though it may require some customization of the import file) |
15:28 | dce: to map your holdings data to Koha's internal structure for holdings) | |
15:29 | dce | apparently they can provide the records in more than one format. I'll get more details |
15:30 | thd | dice: MARC records may always be imported at any time but you are not dependent upon Follett records. You could also copy catalogue over Z39.50. |
15:32 | dice: I am interested in knowing what the price differential is between purchasing books without records supplied and purchasing with records supplied. | |
15:32 | dce | thad: copy catalogue from where over Z39.50? I'm afraid that I have very little experience with the library side of things but a lot on the tech side. |
15:34 | thd | dice: Any other library with a Z39.50 server, which maybe any other library from a neighbour to the national library in another country. |
15:34 | dce | thad: cool |
15:35 | thanks. | |
15:36 | I'll lurk here for a while waiting for the librarian to give me more details then I'll let you know. | |
15:36 | thd | thd: There may still be the advantage of convenience, known quality standard, and actually having a record that is hard to find elsewhere by obtaining records from you vendor. |
15:37 | s/you/your/ | |
15:40 | dice: I am interested in knowing how that value is priced or if it is unavoidably built into the service. How does the service price differ from the publisher's suggested retail price. | |
15:44 | dce: my vision is focused now and I see that you are dce not dice as I had used. | |
15:44 | kados | cancelled taht is :-) |
15:44 | that even | |
15:45 | too much coffee today :-) | |
15:46 | thd | kados: just say no |
15:46 | kados | heh |
15:46 | I don't drink coffee every day | |
15:47 | thd: do you happen to know if it's possible to create frameworks within Koha that store fixed fields differently? | |
15:47 | thd: I haven't played with the frameworks much | |
15:53 | thd | kados: How differently do you want to store the fixed fields? |
15:55 | kados: I cannot see how anyone drinks coffee ever but I thought it tasted foul when I sipped it once and forever found it easy to say no. | |
16:28 | kados: what does your question mean? Paul does have a plugin for modifying the most used UNIMARC fixed field. UNIMARC 100 $a is a fixed subfield that along with other 1XX works similarly to MARC 21 005-008 fixed fields. | |
16:29 | kados | I just had someone ask about koha's ability to handle fixed-field itemtypes |
16:30 | material type and the like | |
16:31 | thd | kados: You could write a plugin for those to use in the Koha record editor, |
16:32 | kados: Look at paul's plugin for UNIMARC 100. | |
16:34 | kados | yea, suppose a plugin would work |
16:34 | thd | kados: look at the MOD mapping for material types also the complete MARC 21 Bibliographic manual hosted on TLC. |
16:40 | kados: This table is useful: http://www.itsmarc.com/crs/Bib0020.htm | |
16:41 | kados | yea, that is useful |
16:41 | I could easily create a framework for each material | |
16:41 | with a plugin that auto-filled the proper fixed field | |
16:43 | thd | kados: I had started writing some Perl code months ago for determining material types from the MARC record but I aborted that effort with the number of fields involved in either general or special material designation. The question is complex and needs to be stored in a separate local use MARC field for Koha 3.0. |
16:44 | kados: Proper material typing is too inefficient to be done at query time. | |
16:46 | kados: If you want a simple plugin for the material types that table and the related fixed field documentation will help you write a plugin. | |
16:46 | kados | thd: thanks |
16:48 | thd | kados: The problem shown by that table is that you need more than one fixed field for even distinguishing whether language material is a serial or a book. |
16:53 | chris | morning rosa |
16:53 | morning thd and kados | |
16:54 | rosa | morning |
16:54 | all | |
16:55 | how's the weather in elly? | |
16:55 | Welly, even | |
16:55 | It's very middling here | |
17:04 | thd | good morning chris and rosa |
17:04 | chris | sorry rosa, middling here too |
17:06 | rosa | bummer. I thought it was going to be beautiful |
17:06 | chris | it was yesterday |
17:06 | thd | kados: The real problems are reading the meaning of the coded and other values. Writing them to the individual fields is is easy for a plugin. |
17:06 | rosa | here too |
17:06 | chris | it might still get sunny .. its threatening too |
17:09 | thd | kados: A sophisticated plugin that added values to multiple fields at once would be nice though, that could be the reverse of a complex general and special material type reading script. |
17:14 | kados: It is much less work for the cataloguer to do things once for the whole record but much more work for the programmer. Yet, the parallel reverse script for reading is needed for even copy catalogued records. | |
17:19 | kados: A change is eventually needed in how the frameworks function so that plugins can operate on discrete parts of fixed fields that correspond to an individual value. Fixed fields would then be much easier to manage. | |
17:39 | dce | thd: The MARC records from Follett are $0.26/record. |
17:44 | thd | dce: That is probably a good price to even save you the trouble of sending a query to a Z39.50 server. What does Follett charge for the books themselves in relation to the publisher's suggested retail price? |
17:54 | jungle | hello |
17:54 | hi to all hope everything is good | |
17:54 | : ) | |
18:31 | kados | thd: you still around? |
18:31 | thd | yes kados |
18:32 | kados | did you see stephen balkits latest email? |
18:35 | thd: after you read it ping me | |
18:35 | thd: i'd like to talk about how best to respond | |
18:36 | thd | kados: still reading |
18:46 | kados: I have read. I have corresponded with Steven in the past his details about Koha are not current enough sometimes. | |
18:46 | kados | right |
18:47 | thd: any comments on his statement about the 650? | |
18:47 | thd: what is he referring to? | |
18:47 | thd | kados: He cares greatly about library standards as do I and every one working on Koha should. |
18:47 | kados | of course |
18:47 | MARC is a terrible standard | |
18:48 | but believe me, koha developers do care about it | |
18:49 | thd | kados: The problem with MARC is that in order to be a very meaningful standard LC has been very resistant to changes for mere data format fashion. |
18:50 | kados: Yet, after 4 decades that is a little problematic. | |
18:51 | kados | well ... the problem with MARC as I see it is that it doesn't live up to its name as a machine readable record format |
18:51 | for instance | |
18:51 | no indexer on the planet | |
18:51 | thd | kados: I looked for a supposed reply from paul but found none. |
18:51 | kados | can deal with MARCs notion of subfield order |
18:52 | if $k follows $b it's subtitle if it follows $a it's title | |
18:52 | it's absurd | |
18:52 | thd | kados: full text indexers have no terrible problem. |
18:52 | audrey | kados: do you know a website about Marc? with lots of details and examples |
18:53 | I am still searching for one. | |
18:53 | kados | audrey: sure ... just a sec |
18:53 | audrey | I don't have a card for KnowItNow at the moment |
18:53 | left the card in US | |
18:54 | thd | audrey: http://www.itsmarc.com/crs/CRS0000.htm |
18:54 | audrey: http://www.loc.gov/marc/ | |
18:55 | audrey | have the loc one already |
18:55 | kados | audrey: this is my favorite intro ... nice and simple: |
18:55 | audrey: http://www.loc.gov/marc/umb/um01to06.html | |
18:55 | thd | audrey: This is an excellent brief guide: http://www.loc.gov/marc/umb/ |
18:55 | audrey | I need explanation of reason for subfields in 852 field |
18:56 | they still confuse me | |
18:56 | more detailed than introductory | |
18:56 | kados | audrey: I can explain them to you |
18:56 | audrey | thanks for the recommendations:) |
18:56 | great | |
18:56 | kados | let me pull up a record |
18:57 | audrey | the classification number is created from 4/more bits of information |
18:57 | and goes into more than 4 MARC 852 subfields | |
18:57 | thd | audrey: The examples in the LC documentation are the best clue. |
18:57 | audrey | I've looked at them, and am still confused |
18:57 | kados | audrey: you're confused about classification? |
18:58 | audrey | no |
18:58 | kados | audrey: what then? |
18:58 | audrey | i want more information about why the classification bits go into each subfield |
18:58 | the dewey number is easy | |
18:59 | its subfield says its just for dewey | |
18:59 | kados | are you looking at Koha's MARC or MARC in general |
18:59 | cause there is a difference | |
18:59 | audrey | MARC in general, right now |
19:00 | how much difference is there? | |
19:00 | thd | audrey: http://www.itsmarc.com/crs/Bib1424.htm is actually the complete LC documentation for 852 with extra HTML. |
19:01 | kados | audrey: thd beat me too it with the link |
19:01 | audrey: the thing to keep in mind with LOC classification | |
19:02 | audrey: is that it was designed by the library of congress :-) | |
19:02 | audrey | looking at it now:) will see if it clears the fog:) |
19:02 | kados | audrey: have you ever been there? |
19:02 | audrey | no |
19:02 | in DC, but not at LC | |
19:02 | kados | audrey: well i have ... it's HUGE! |
19:02 | audrey: they have a serious amount of space | |
19:02 | thd | audrey: Koha does not use 852 internally. I have a plan but that would require quite a bit of work. |
19:03 | kados | thd: (even with zebra?) |
19:03 | audrey | so the suffix subfield is really because they have departments inside of buildings |
19:03 | kados | audrey: exactly |
19:03 | audrey: it's really an internal taxonomy for LOC | |
19:03 | audrey: based on their specific needs | |
19:04 | thd | audrey: prefix use is common. Suffix use is rare. |
19:04 | kados | audrey: academic libraries use it too ... because their collections are also often quite large |
19:04 | audrey: and spread across multiple buildings and floors | |
19:04 | audrey: etc etc. | |
19:05 | audrey | am looking at the reasons for prefixes, dewey, cutter--why split between different subfields? |
19:05 | why not in only one subfield? | |
19:05 | thd | kados: using 852 would require nontrivial changes to how Koha manages items. |
19:05 | audrey | koha already uses 852. i saw it yesterday |
19:05 | kados | thd: lets talk about that ... how nontrivial? |
19:06 | audrey: naw ... not for managing the items :( | |
19:06 | thd | audrey: Actually, Koha does it that way. |
19:06 | audrey | but is mapped and searchable, right? |
19:06 | kados | audrey: though we can nab stuff from the 852 and put it in local use fields in the 900s |
19:07 | audrey | hmm |
19:07 | kados | where Koha stores item info currently |
19:07 | thd | audrey: you can place the whole MARC record inside Koha but Koha uses a local use field for managing items. |
19:07 | kados | lots of systems do actually |
19:07 | audrey | what is "manage"? |
19:07 | kados | including Dynix |
19:08 | audrey: well ... there are a few bits of information that Koha pays attention to with items | |
19:08 | audrey: barcode, item type are two | |
19:08 | audrey | right |
19:08 | kados | audrey: managing basically means changing the status of an item |
19:08 | checked out, on reserve, lost, etc. | |
19:09 | audrey | not looking for that with classification #'s. |
19:09 | kados | nope |
19:09 | thd | audrey: This might help confuse you more but explains the origin of Koha item management. http://www.kohadocs.org/holdings.html |
19:09 | audrey | item management is not focus at moment |
19:10 | classification details are | |
19:10 | not for status purposes | |
19:10 | chris | ohh is that what its modelled on |
19:10 | cool | |
19:10 | i thought it was just something olwen and I thought up | |
19:11 | thd | chris: I have an unfinished revision for that document proposing changes. |
19:11 | audrey | why does marc divide the info on back of books (the sticker bits) into different subfields? |
19:12 | why not just one big, changeable field? | |
19:12 | thd | audrey: MARC Koha works with the whole call number in one local use subfield. |
19:13 | chris | thats a good explanation thd |
19:13 | audrey | thanks, will be looking at all recent information |
19:13 | and clearing my fogged brain:) | |
19:14 | thd | chris: That is also the most turgid writing that I have ever produced. At least it serves some useful purpose. |
19:14 | kados | thd: in your opinion, if we implemented your recommendations in the holdings.html doc, would folks like stephen stop complaining? |
19:16 | thd | kados: The holdings.html doc has a careless mistake in the mapping but I do have an almost finished revision that is much more comprehensive. |
19:17 | kados: To avoid the problems that Steven complains about a valid complete default MARC framework for MARC 21 would be needed. | |
19:18 | kados: Even the UNIMARC default framework is full of errors. | |
19:19 | kados: I could build one easily.. It would require a few days to carefully audit everything. | |
19:20 | kados: Steven complains about the non-repeatable subfields being set by default to repeatable for 650 etc. | |
19:21 | kados | thd: will that even be an issue now that we've got zebra? |
19:21 | thd: I don't think it will | |
19:21 | thd: at least not for searching | |
19:22 | thd | kados: It is still an issue for a record editor that uses the frameworks and also for the several months before 3.X is stable enough for production use.a |
19:22 | kados | right |
19:23 | hmmm | |
19:24 | thd: don't we still have problems with subfield ordering? | |
19:24 | thd | kados: The absence of 000-008 in the frameworks makes Koha look very poor to anyone who investigates and causes data loss when editing records. |
19:24 | kados | thd: ie not preserving it? |
19:25 | thd: but we _do_ have that now! | |
19:25 | thd | kados: exactly |
19:25 | kados | thd: so it's more complicated than just fixing the default framework |
19:25 | thd | kados: Only if the user audits the framework and validates everything to add what is missing. |
19:26 | kados | ok ... so we need to make sure that 2.2.6 has a better default framework |
19:26 | that fixes those issues | |
19:26 | the subfield ordering thing ... that's tricky | |
19:27 | thd | kados: Ordering has been mostly fixed. There are still correctable issues relating to presentation. |
19:27 | kados | ahh ... has it |
19:27 | wonderful | |
19:27 | thd: feel free to create a new valid framework on LibLime's demo | |
19:28 | :-) | |
19:28 | thd | kados: I would be eager to fix the default framework for MARC 21. Would it help me pay my bills? |
19:28 | kados | thd: sure :-) |
19:29 | thd: feel free to bill LibLime :-) | |
19:39 | walk even | |
20:05 | thd: in your new version of the holdings doc, it would be interesting to say what other systems base their holdings on recomendation 995 | |
20:05 | thd: some 'name dropping' would be nice there :-) | |
20:07 | thd | kados: French systems only. It would not be very meaningful to your customers. |
20:07 | kados | thd: I think Dynix does |
20:07 | thd: I could be wrong though | |
20:08 | thd | kados: Unless there is a French branch or subsidiary of a non-French company. |
20:09 | kados: The largest vendors support UNIMARC in some manner so they might have French localised versions. | |
20:10 | kados: I will see if paul has an idea tomorrow. | |
20:13 | kados: What might also be good is to know what software does not use MARC 21 holdings format or uses nonstandard holdings data in current versions. | |
20:14 | kados | yep |
20:15 | thd | kados: Of course Koha should head to support MARC standard holdings. Even recommendation 995 support is an adaptation rather than a fully compliant implementation. |
20:16 | kados | thd: well ... OCLC doesn't :-) |
20:16 | thd: I'm not 100% convinced we need standard MARC holdings | |
20:16 | I can see some advantanges | |
20:16 | well ... before I continue down that thread | |
20:17 | what would be involved in getting MARC compliant holdings into Koha? | |
20:17 | thd | kados: OCLC should now be converting to the MARC holdings format according to their intention announced at the beginning of last year. |
20:17 | kados | it can't really be that complicated |
20:17 | ahh ... they are converting _to_ MARC ... | |
20:17 | I thought they were converting away from it | |
20:18 | yea, so ... how hard could it be really? | |
20:18 | and what advantages would it afford us? | |
20:19 | thd | kados: They announced at a conference at the beginning of last year that they would be using the separate MARC 21 holdings format records. |
20:21 | kados: I had started a message to you about supporting MARC 21 and now also UNIMARC standard holdings when you were driving off to meet with a law library client.. | |
20:21 | kados | thd: still have it? |
20:22 | be right back | |
20:22 | thd | kados: Most of the text was still in my head. I could reconstitute it. |
20:53 | kados | we don't want Koha's holdings to be _tied_ to MARC standard holdings |
20:53 | because there are instances when libraries or other orgs will not want to use MARC bib records | |
20:55 | thd | kados: How many libraries are there in the world that do not want MARC? |
20:55 | kados: I do not mean the ones that want MARC to be hidden. | |
20:56 | kados: Hiding the complexity of MARC from the user can readily be done. | |
20:57 | kados | marc is evil |
20:57 | thd | kados: Even paul's record editor does much for that. |
20:57 | kados | there's no backing down on that point :-) |
20:57 | I want to support it | |
20:57 | in Koha | |
20:57 | but I don't want to be tied to it | |
20:58 | thd | kados: MARC is designed for the ease of record exchange and detailed careful data recording. |
20:58 | kados | thd: by a bunch of loony academics |
20:59 | thd | kados: Koha should support many different record types. |
20:59 | kados | thd: exactly |
20:59 | chris | i think you need to change that too |
20:59 | MARC was designed for the ease | |
21:00 | thd | kados: However, MARC has the most carefully thought out solutions to the most complex problems of information records. |
21:00 | chris | now people want to use it catalog in |
21:00 | kados | thd: that's only partly true |
21:00 | thd: it's not a very wel-though out storage format | |
21:00 | chris | ie its purpose has been overloaded and hence its no longer as easy |
21:01 | thd | kados: It was designed when every byte was precious. |
21:01 | chris | my main beef with MARC |
21:02 | thd | kados: MARC in XML is very extensible. |
21:02 | chris | is that everyone forgets the M is Machine |
21:02 | humans arent supposed to see it | |
21:02 | ever | |
21:02 | ever | |
21:02 | :) | |
21:02 | kados | well ... but also |
21:02 | chris | but because they do |
21:02 | kados | machines often don't know what to do with parts of it |
21:02 | chris | they broke it |
21:02 | thd | chris: That is exactly my point about hiding it. |
21:02 | kados | chris++ |
21:02 | chris | im sure it was a much better thing |
21:03 | when it was a standard | |
21:03 | not 12 different standards that arent standard | |
21:03 | kados | true |
21:03 | thd | chris: machines know what to do if th programmers tell them how. |
21:03 | chris | yep |
21:04 | its still broken | |
21:04 | kados | so ... more thinking out loud |
21:04 | we need a way to abstract our holdings format | |
21:04 | chris | im not saying it was always broken, just it is now :) |
21:04 | thd | chris: there are over 50 variants but there has been significant unification of the variant formats so that is much less of a problem now. |
21:04 | kados | in the same way we're abstracting our searching format with Zebra |
21:05 | chris | yeah |
21:05 | kados | ie, I can put in dublin core records into Koha RIGHT NOW |
21:05 | and Koha will find them | |
21:05 | chris | yep |
21:05 | kados | so we've made a huge leap forward there |
21:06 | we need to also abstract our record editor's underlying format | |
21:06 | thd | kados: Most of you customers and the potential ones with the most money to spend will want MARC holdings. |
21:06 | kados | thd: so we'll give them those |
21:06 | thd: but I don't want to make the mistake we made last time with MARC support | |
21:07 | thd: which is to hardcode it into the scripts, templates and database | |
21:07 | we need to compile some general purpose assumptions about holdings | |
21:07 | then have a way to configure the holdings mappings to a specific holdings format | |
21:08 | thd | kados: Serials holdings can be very complex. |
21:09 | kados: Certainly, a fresh approach to any problem with the advantage of knowing the work that has gone before can produce a better solution. | |
21:11 | kados: Yet, without spending a significant amount of time how will you find a general problem that includes all the issues for MARC and everything else. | |
21:12 | kados | well, as I see it |
21:12 | thd | kados: Karen Coyle wants to kill MARC but many of her potential allies are intimidated by the LC superpower with the veto. |
21:13 | kados | holdings are used mainly for one thing |
21:13 | keeping track of the movement of items | |
21:13 | right? | |
21:13 | chris | thats my understanding |
21:14 | thd | kados: Karen speaks about the possibility of a system designed for supporting FRBR efficiently but she has no design herself. |
21:14 | kados | thd: we're not talkinng about FRBR yet :-) |
21:15 | i didn't think FRBR addressed managing the movement of items | |
21:15 | I thought it was mainly for searching | |
21:15 | thd | kados: That seems close. |
21:15 | kados | and FRBR's what, 11 years old? |
21:15 | with not a single app using it? | |
21:15 | chris | thats only about 2 weeks old in library time tho :-) |
21:16 | kados | true |
21:16 | thd | kados: Implementing FRBR is very inefficient. Variant data problems in legacy records and too many CPU cycles for processing. |
21:16 | kados | anyway, I digress |
21:16 | the point is | |
21:16 | we need to improve Koha's internal methods for managing items | |
21:16 | that much is clear | |
21:17 | for instance, we should be able to place reserves on specific items | |
21:17 | chris | yep |
21:17 | kados | we should also allow itemtypes at the item level |
21:17 | chris | this is what i was saying about biblioitems .. the point of them was missed |
21:17 | kados | so what we need, is a list of things that the MARC standard for holdings can _DO_ |
21:18 | thd | kados: I have a solution to some of that for 2.X. |
21:18 | chris | ohh good point kados |
21:18 | kados | then, it's a simple matter of doing it better inside Koha |
21:18 | chris | yep |
21:18 | kados | then mapping it to MARC for those that want it |
21:20 | chris | you using wifi at a bar? |
21:20 | kados | hehe yea, casa actually :-) |
21:20 | chris | cool :) |
21:20 | kados | where I do my best thinking :-) |
21:20 | thd | kados: I see more clearly, searching complex MARC serials holdings is very inefficient. |
21:23 | kados | thd: exactly |
21:23 | thd: are you up to writing some on what MARC holdings (serial and otherwise) allow a library to do? | |
21:24 | thd | kados: You need a means of addressing the same complexity as MARC can. |
21:24 | kados | thd: can you explain what you mean by 'complexity'? |
21:25 | thd | kados: complex publication patterns, special issues, supplements, missing issues, title changes of the same publication. |
21:27 | kados: The problems are worst when the information has been recorded as a machine readability aid. | |
21:28 | kados | thd: Kokha can do much of that already |
21:28 | thd | kados: There is are NISO and ISO standards for serials. |
21:28 | kados | interesting |
21:29 | thd | kados: yet what Koha can do already is as difficult to use as MARC and does not interface well with items although I know hdl partly fixed that. |
21:31 | kados: there are also part, whole, etc. record relationships for linking between sets of holdings data and bibliographic records. | |
21:32 | kados | I need to digest that sentence |
21:33 | I _have_ afterall been drinking :-) | |
21:33 | thd | kados: Unfortunately, the international standards are insufficiently detailed. They exist as a mere starting point constructed after the fact. |
21:35 | kados: The hierarchal relations are the most troublesome for encoding, decoding, and searching. | |
21:36 | kados | thd: what kinds of searching would we need to do? |
21:37 | thd: what do the hierarchal relationships allow the system to do? | |
21:39 | thd | kados: The system has to manage the relationships between serial supplement 5B and a monograph after a title change. Display whole runs with links to the records and missing issues. Etc. |
21:39 | kados | the title change should be easy |
21:39 | with authorities | |
21:40 | links are cake too | |
21:40 | already does missing issues | |
21:40 | thd | kados: manage overlapping, embargoed, and gaps in electronic database coverage. |
21:41 | kados | coudl you expand on that last bit? |
21:41 | what's supplement 5B? | |
21:42 | thd | kados: holdings can store information about indexing and abstracting database coverage as well as full text databases. |
21:42 | kados | k ... holdings can store info about where full text databases are |
21:42 | and it can store information 'about indexing and abstracting database coverage'? | |
21:43 | I don't quite understand what you mean by that | |
21:43 | does it say _how_ to index? | |
21:43 | thd | kados: vendors will have gaps for single issues, spans of time, exclusion and coverage that changes on a sliding basis over time. |
21:44 | kados: Holdings can link to electronic databases, or hard copy, electronic databases can link back. | |
21:44 | kados | thd: linking is easy |
21:44 | thd: we just need openurl | |
21:44 | thd: about the gaps, spans of time, etc. | |
21:45 | thd: have you seen any really good examples of serials managment functionality in a proprietary system? | |
21:45 | thd: ie, do you understand how they work? | |
21:45 | thd | kados: OpenURL yes but OpenURL needs detailed holdings for the resolver to use. |
21:46 | kados | thd: isn't that what we're talking about providing? |
21:47 | thd | kados: I have seen examples where the catalogues point to multiple hard copy and not yet come back next month when the embargo period is over electronic databases. |
21:47 | kados: I understand how OpenURL works. | |
21:48 | kados: Karen Coyle did a fine job rewriting the documentation. | |
21:49 | kados: There are published papers on link resolvers. | |
21:50 | kados: The big secrets are keeping the information about electronic resources current when your vendor is constantly changing the coverage without overwhel;ming the system maintenance staff. | |
21:51 | kados | right |
21:53 | thd | kads: 5B was an arbitrary example for something like a serial back cover pocket insert that goes inside a volume of a law book. |
21:54 | a supplement with the latest text revisions and case law updates. | |
21:55 | kados: Supplements are theft targets and vital to track for law libraries. | |
21:56 | kados: Not that they are replaced when stolen at the poor libraries. | |
21:56 | kados | thd: before I forget: http://opactest.liblime.com/cg[…]ha/search-test.pl |
21:56 | thd: that's the CQL test that chris built | |
21:56 | thd: probably not 100% CQL yet but it does a pretty good job | |
21:56 | for two days of programming :-) | |
21:57 | thd: Koha supports summpements right now? | |
21:57 | thd: I've seen them in serials | |
21:59 | thd | kados: The issue is not for some support for supplements but supporting the complex part/whole relationships for supplements, special issues, etc. |
21:59 | kados | I don't really understand what that entails |
21:59 | could you explain it? | |
22:02 | thd | kados: Koha support for serials is based on regular publication patterns that are predetermined. It is not very flexible for accommodating changes in publication patterns in the midst of a period or changing and irregular patterns. |
22:03 | kados | that's true |
22:03 | so let's focus on one thing at a time | |
22:03 | thd | kados: also, there is poor linking between bibliographic records where one record might cover the title as a whole, another an issue, another an article. |
22:04 | kados | ??? |
22:04 | one serial record you mean? | |
22:04 | I dont' quite follow | |
22:05 | thd | kados: holdings records can also be some arbitrary level below the title level. |
22:07 | kados | that's a little too abstract for me |
22:07 | I need some examples | |
22:08 | thd | kados: You have the annotated edition of the state laws with many volumes and a monograph record for the set. |
22:10 | kados: The set itself has holdings record for the set as whole. | |
22:11 | kados: The criminal code part has a record for its several volumes both bibliographic and holdings. | |
22:11 | kados: | |
22:12 | kados: The criminal code may have its own index record linked to the other volumes. | |
22:12 | kados | and there are links between the different records? is that what you're getting at? |
22:12 | thd | kados: And each one can have separate holdings. |
22:13 | kados | meaning what exactly ... you can interact with the records at each level in the tree? |
22:14 | what kinds of things can you do with the separate holdings records? | |
22:14 | thd | kados: furthermore, the books may be monograph records and the back pocket supplements may be serial records. |
22:15 | kados: separate records allow for easier management, however, there could be one hyper-complex record managing it all. | |
22:16 | kados: Either way the system has to bring all the relations together. | |
22:17 | kados | that's not too hard to do |
22:17 | we could tackle an indefinite heirarchy using either an adjacency list or a nested set | |
22:17 | thd | kados: Separate bibliographic records for various parts allow the inclusion of details about parts that would not have a proper place in a general record. |
22:18 | kados | you could assign ids for each node in the tree |
22:18 | nodes could be ordered by date | |
22:18 | peer nodes that is | |
22:18 | if we used nested sets | |
22:18 | thd | kados: the difficulty is bringing a complex search path together when the records were not created in a neat top down manner. |
22:18 | kados | actually, with nested sets the search path is quite simple |
22:19 | you can traverse the heirarchy fairly easily | |
22:19 | thd | kados: Also, searching hierarchies can be inefficient. |
22:19 | kados | thd: nested sets are ver efficient |
22:19 | very even | |
22:20 | I've been doing some work with them for PINES | |
22:20 | what we need are some specific case examples | |
22:20 | very specific examples | |
22:20 | like actual serial records | |
22:20 | for real serials | |
22:20 | thd | kados: I mean that only relatively as compared to flat data. Extra CPU cycles add up in large collections. |
22:23 | kados: Obtaining bibliographic records is easy, finding complex holdings records may be more difficult.. | |
22:23 | kados | a well designed heirarchy can be very fast |
22:23 | thd | kados: Koha has historically used the simplest set of records as examples. |
22:24 | kados: You need to have access to the most troublesome records for testing that challenge all aspects of a system. | |
22:26 | kados: Fast is merely relative but part of my point is that the records may have tortured relations with one way mappings. Koha could find the implied relationships for easier access. | |
22:28 | kados: A large problem will be that accessible records will make extensive use of human readable relation encoding that some system needs to parse in order to work well. | |
22:30 | kados: Koha uses machine readable encodings for serials now, most serials holdings data was actually meant for human readability because the machine parsable holdings fields had not been designed. | |
22:31 | kados: The most high demand holdings would have been redone as new records were created and systems supported their creation. | |
22:33 | kados: There is a wide variation in actual practise. If you can parse textual holdings fields even partly with regular expressions you will have a migration advantage. | |
22:36 | kados: After all the nonsense to obtain a client, I assume that data migration is the next most time consuming part of your services. Am I correct? | |
22:37 | kados | of course :-) |
22:38 | thd | kados: The better your support for mapping MARC holdings onto whatever the lower your costs will be. |
22:39 | kados: Of course, you need a two way map in order not to be an evil trap for your customer like so many proprietary systems are. | |
22:42 | kados: Half the problem with Koha serials holdings now is the user interface requires careful application of machine encoded values, and is fairly brittle. | |
22:42 | kados | true |
22:43 | thd | kados: Record creation and modification is of comparable complexity to migration. |
22:45 | kados: Currently the migration path would be tedious except that hdl designed holdings for libraries that usually had simple reliable publication patterns. | |
22:46 | kados: Who is the intended audience for Koha 'fact sheets'. | |
22:47 | ? | |
22:50 | kados: I think it will be easier for me to change a default MARC 21 framework on my own system and send you an SQL file. I will have better access to the data than if I work on yours. | |
22:53 | kados: The ISBD mapping took about ten hours and I never audited it when I was done but I had to follow some ambiguous ISBD rules. At least there was no consequence for data loss in case of error there. | |
22:56 | kados | ok |
22:56 | that's fine | |
22:57 | thd | kados: "For whom are 'fact sheets' intended? |
22:58 | kados | potential clients of course |
23:00 | thd | kados: The bean counters who do not know MARC from Dublin Core and make all the decisions or the cataloguers who dream MARC and have no influence? |
10:12 | kados | paul: http://opactest.liblime.com/cg[…]a/search-test.pl? |
10:12 | paul | paul greets kados |
10:12 | kados | import of one of my client's data seems to have finished |
10:13 | and chris's CQL query is working nicely | |
10:13 | hi paul :-) | |
10:13 | paul | christian library it seems. |
10:13 | The Book of Ezekiel / | |
10:13 | kados | yep :-) |
10:13 | paul | The Book of Daniel / |
10:13 | The Book of Isaiah / | |
10:14 | (don't ask me what I entered as search term ;-) ) | |
10:14 | kados | try 'book not daniel' |
10:14 | paul | works nicely |
10:15 | kados | I'm still not sure all the records made it in |
10:15 | as the last record that loaded threw an error | |
10:15 | about invalid ZXML | |
10:15 | XML that is | |
10:15 | I was seeing similar errors after importing about 200 items | |
10:16 | the import would crash | |
10:16 | but I think that was because we wern't destroying the zebra connection | |
10:16 | and zebra was getting overloaded with too many connections | |
10:17 | import is quite slow the connect - import one record - disconnect way | |
10:17 | I wonder if we can have just one connection for the entire process | |
10:17 | might be a speed improvement? | |
10:18 | i also noticed that it gets slower as more items are added | |
10:18 | so at the beginning, it was about 10-20 per second ... but after 30,000 it was about 1 per second | |
10:18 | just some things I've noticed :-) | |
10:30 | paul: in your opinion, should we use the create()/connect() method to maintain the same object while updating records? | |
10:30 | http://search.cpan.org/~mirk/N[…]b/ZOOM.pod#___top | |
10:31 | paul | ??? not sure to understand what you mean. |
10:31 | kados | right now |
10:31 | we do: | |
10:31 | $Zconn = new ZOOM::Connection(C4::Context->config("zebradb")); | |
10:31 | update a record | |
10:32 | $Zconn->destroy(); | |
10:32 | I was hoping it would be faster to do: | |
10:32 | $Zconn = new ZOOM::Connection(C4::Context->config("zebradb")); | |
10:32 | update many records (all in fact) | |
10:32 | $Zconn->destroy(); | |
10:32 | does that make sense? | |
10:35 | paul | maybe you're right, but with mod_perl that will be automatic behaviour I think (without destroy), while without mod_perl, we would have a destroy when the process ends |
10:36 | (writing a mail on koha-devel to give -quite good- news from Ouest-provence) | |
10:37 | kados | woo hoo! |
10:47 | paul: did you write the zebra_create method in Biblio.pm? or did chris? | |
10:47 | paul | I did |
10:47 | (not sure the commited one is the last I have) | |
10:48 | kados | ok |
10:50 | I'm thinking we should change the name | |
10:51 | there are several methods in the ZOOM::Package module | |
10:51 | one of them allows 'createdb' | |
10:51 | if we had a generic ZOOM::Package wrapper we could pass in options that we want | |
10:54 | it looks like we could pass in a hash object for $options | |
10:54 | and then we need an 'operation' for the $p->send method | |
10:55 | I may work on this routine a bit | |
10:56 | paul | (createdb is already in some code I wrote. note also that there is a dropdb... that is buggy and do nothing) |
10:56 | mail sent to koha-devel, should arrive in 4 hours ;-) | |
10:57 | kados | heh |
10:57 | hmmm, dropdb is buggy eh? | |
10:57 | did you notify Mike Taylor? | |
10:57 | paul | yep. does nothing. |
10:57 | it's mike that warned me ;-) | |
10:57 | kados | ahh :-) |
10:58 | do you like my idea for having one ZOOM::Package method in Koha | |
10:58 | and passing it options, operation, etc.? | |
10:59 | paul | something like the ->dbh in Context iiuc |
10:59 | yes. | |
10:59 | kados | yes I am speaking of two things |
← Previous day | Today | Next day → | Search | Index