← Previous day | Today | Next day → | Search | Index
All times shown according to UTC.
Time | Nick | Message |
---|---|---|
12:08 | thd | tim: I was most interested in the MARC record issue. I have no knowledge of circulation issues, they are not formally part of my problem space. I had been pleased to see the code that you had posted on the list for helping me think about the MARC records issue. |
12:19 | kados | tim: yes ... thanks very much ... I'm still waiting for the client to export his data |
12:19 | tim | hdl: did you see my post about the bug in the original script? |
12:20 | kados: I'm not sure how good I did at setting up the marc/db links, but could send them to you if you want them. | |
12:21 | They would be in marc_tag_structure and marc_subfield_structure wouldn't they? | |
12:21 | hdl | no. |
12:21 | tim | hdl: are you converting from Sagebrush Athena? |
12:22 | hdl | I am french ;) |
12:23 | tim | I'm talking to the wrong person :) |
12:23 | hdl | no problem :) |
12:23 | tim | I seem to do that a lot. |
12:25 | kados | tim: did you have to redefine the marc links to match your records? |
12:26 | tim: (asside from holdings) | |
12:58 | tim | If I remember right, most of the changes were to try to get it to display more like the Athena easy edit screens. |
12:58 | To make the switch easier on staff. | |
12:59 | I did that early on and then took so long working on holdings. I need to check my work and see if it still makes sense to me. | |
15:11 | Gonna try talking to the person I mean to talk to this time. | |
15:12 | thd: did you see my post about the bug in the original marc coversion script? | |
15:12 | And I was also wondering if you're moving from Athena. | |
15:21 | thd | tim: I had seen your second post, which I believe was the one I had linked with the URL, where you corrected your code from your first message. |
15:22 | tim: I am not moving from Athena, but am just generally interested in the problems involved in moring MARC data between systems and how best to overcome them. | |
15:24 | tim: s/moring/moving | |
15:54 | tim | It's the first and only perl script I wrote except for a few test scripts. |
15:55 | I hope it's helpful. | |
16:12 | thd | tim: It has helped me to think about the issues in migrating MARC records. Thankfully, I do not have to migrate a system immediately. |
17:40 | chris | morning |
17:47 | kados | morning chris |
17:47 | chris | heya kados |
17:49 | kados | composing a summary of our progress with zebra integration atm |
17:50 | chris | cool |
18:03 | kados | mason: so ... is it ok if I list you as a MARC expert? ;-) |
18:03 | mason: I'm putting together a summary of where we're at with Zebra integration for koha-zebra and I'm listing roles as well | |
18:08 | mason | hi kados, my marc is thin and rusty ;) |
18:08 | im just googling zebra | |
18:22 | kados | :-) |
18:23 | mason | tell be a bit about zebra |
18:24 | thd | kados: My MARC familiarity is reasonably good, although, I gained much of my experience with MARC outside a library setting. Do not let my confession about spending years as a bookseller mislead you. William Moen, leading research into MARC at UNT, also spent may years as a bookseller before adopting library science as a profession. That is usually done the other way round when librarians retire to take up bookselling. |
18:25 | kados: I have not had much occasion to study UNIMARC though. I have been learning much more about UNIMARC in the past few days :) | |
18:26 | mason | this zebra? |
18:26 | A system with Zebra installed acts as a dedicated router. With Zebra, your machine exchanges routing information with other routers using routing protocols. Zebra uses this information to update the kernel routing table so that the right data goes to the right place. You can dynamically change the configuration and you may view routing table information from the Zebra terminal interface. | |
18:28 | chris | nope :) |
18:28 | http://indexdata.dk/zebra/ <-- that zebra | |
18:30 | kados | mason: sorry ... writing a summary ... |
18:31 | mason: there's a new Koha list: koha-zebra | |
18:31 | mason: we're going to be moving from MARC-in-SQL to using Zebra as the primary storage and retrieval engine for Koha | |
18:31 | chris | for the marc data that is |
18:32 | marc in sql makes little sense really, when i think about it | |
18:32 | you lose all the advantages of a RDBMS | |
18:33 | and end up with a ton of redundant data | |
18:33 | kados | yep |
18:34 | thd | I am having difficulty getting a question through on either the MARC or AUTOCAT listserves. I never had trouble on the MARC list before a few months ago. I think my *.com address is being identified as a potential source of spam. Maybe I need to register a *.org address. : / |
18:35 | mason | looks nice |
18:35 | esp. the pic of the zebra | |
18:36 | chris | heh |
18:36 | so our main task as i see it | |
18:36 | is integrating with zebra in such a way that we can still offer our 2 views of the data | |
18:37 | i counted 11 clients of ours, who use koha and dont ever want to to see MARC | |
18:37 | and 2 who do | |
18:38 | kados | chris: wow ... I didn't know you had that many koha clients! |
18:38 | thd | Green, Rebecca. Design of a relational database for large-scale bibliographic retrieval. Information technology and libraries v. 15, no. 4 (Dec. 1996) p. 207-221. |
18:38 | chris | storing bibliographical data in a relational db is easy |
18:38 | storing MARC isnt | |
18:39 | anyway, dont get me started on marc .. we are stuck with it :) | |
18:39 | kados | not necessarily |
18:39 | chris | i think in the near future anyway |
18:39 | kados | with zebra you can change the underlying record format |
18:39 | chris | yeah |
18:40 | kados | without changing the api at all |
18:40 | chris | some ppl just love to see 245a |
18:40 | kados | which is why it's so nice |
18:40 | chris | instead of title |
18:40 | kados | right |
18:40 | well it makes a difference ;-) | |
18:40 | chris | :) |
18:40 | kados | because if you're talking about 245a or 246b ;-) |
18:40 | chris | unido |
18:40 | rach | do they? mad buggers |
18:41 | chris | and the refugee woman was quite keen (student librarian from vic) |
18:41 | rach | ah right |
18:41 | thd | the article has the same conclusions koha developers have discovered for themselves with more effort |
18:42 | chris | yep kados, so as long as we can offer both views, im happy |
18:42 | both = lots of | |
18:42 | :) | |
18:42 | thd: i didnt even attempt to get MARC in .. i left that for the more masochistic programmers to try :-) | |
18:43 | thd | :) |
18:47 | There has been a discussion on the MARC listserv just recently about RDB and MARC with the usual negative conclusions. The addition of XML was considered to be no advantage for the databse design problem. | |
18:48 | chris | yeah |
18:49 | it gets worse when you try to support more than one flavour | |
18:50 | i think moving marc out of the db will be the best thing we can do | |
18:51 | thd | chris: I was wrestling with magic tricks to overcome the MARC21 UNICODE divide. I have not found the right spell book :) |
18:52 | chris | yeah, thats not even worrying about the other 20 thousand marcs (chris might be exaggerating slightly) |
18:53 | kados | ok ... summary sent |
18:53 | chris | cool |
18:53 | thd | chris: fortunately, we are following after a significant degree of format conversion to MARC21 and UNIMARC |
18:53 | kados | I spoke to sebastian this morning |
18:53 | he's interested in expanding the zebra api | |
18:53 | perl api that is | |
18:53 | chris | cool |
18:54 | kados | so that's something to keep in mind |
18:54 | thd | kados: what is missing form the api currently? |
18:55 | kados | thd: the perl api for zebra is not fully developed |
18:55 | thd: it was developed by a third party who's vanished ;-) | |
18:56 | thd | kados: you mean it does not support all the features that the C or whatever API does? |
18:56 | kados | thd: there are other ways to tap into the api but a perl-specific interface would be ideal |
18:56 | thd: right | |
20:56 | thd | chris: are you stiil there? |
21:49 | chris | am now |
22:02 | thd | chris: I saw your post about 11 of your 13 customers prefer to never see MARC. |
22:02 | chris: Did I interpret that correctly? | |
22:02 | chris | thats the one |
22:03 | most of them are corporate or special interest libraries | |
22:05 | thd | chris: If you read paul's migration email carefully, he seems to intend to keep the database structure along with blob storage or whatever for MARC records that zebra can index. |
22:06 | chris | yep, i dont imagine it will be a problem, as long as we make sure that then non-MARC cataloguing pages update the files zebra indexes |
22:07 | then=the | |
22:08 | thd | chris: I would imagine, there will always be ways to hide MARC from the user who may be frightened by 650 $a$x. |
22:10 | chris | yep |
22:10 | thd | chris: However, proper MARC support implemented well should give you the possibilty of serving a whole new class of potential customers. |
22:10 | chris | its just that when marc was introduced, the non-marc cataloguing features lagged behind and werent tested as well as the marc ones (as much from my fault as anyone elses) |
22:11 | so just want to make sure we dont repeat our mistakes :) | |
22:11 | thd: indeed and thats great, as long as we dont lose HLT (the original client and reason Koha exists) | |
22:12 | (one of the 11) | |
22:13 | thd | chris: The potential new class of customers, would be much larger and potentially a much better source of business than the instutuions for which Koha had previously offered a solution. |
22:14 | chris | yep, as i said thats great |
22:14 | as long as we dont turn into one of the big ILS vendors and forget about all the little ppl who made it possible | |
22:14 | thd | chris: We will hide MARC from HLT. |
22:15 | chris | im sure we will :) i was just saying that thats what my main focus with the zebra integration will be |
22:15 | in making sure it works for non-MARC .. ill leave the MARC bits too the people who know more | |
22:16 | thd | chris: My background is in bookselling, where I hid LIS stuff from the user while taking advantage of MARC records. |
22:16 | chris | cool |
22:17 | thd | chris: I am keen to make the system work well for the end user who should never need to know MARC. |
22:17 | chris | cool |
22:21 | thd | chris: MANY ILS systems require knowing how to encode 650##$a$x$z$y to use them at all or get much advantage from them. Paul's work already goes a significant way to correct that. I like to have data views that support the value of MARC but guide the user with meaningful labels where something is not self-evident. No code numbers should trouble or frighten the user. |
22:33 | chris: There may be a standards compliant way of identifying the lesser degree of record detail that Koha had originally used with Dublin Core. | |
22:36 | chris: Dublin Core actually scares me because it does not use code numbers and ignores the degree of detail found in full MARC. | |
22:36 | :) | |
22:36 | chris | that would be cool |
03:43 | osmoze | hello |
09:44 | kados | morning owen |
← Previous day | Today | Next day → | Search | Index