← Previous day | Today | Next day → | Search | Index
All times shown according to UTC.
Time | Nick | Message |
---|---|---|
00:59 | BobB | @wunder Sydney, Australia |
00:59 | huginn | BobB: The current temperature in Sydney, New South Wales is 16.0°C (10:30 AM EST on June 09, 2014). Conditions: Scattered Clouds. Humidity: 59%. Dew Point: 8.0°C. Pressure: 30.36 in 1028 hPa (Steady). |
00:59 | ibeardslee joined #koha | |
01:00 | mtompset joined #koha | |
01:00 | mtompset | Greetings, #koha. |
01:01 | @seen dcook | |
01:01 | huginn | mtompset: dcook was last seen in #koha 2 days, 18 hours, 8 minutes, and 3 seconds ago: <dcook> yo reiveune |
01:03 | eythian | mtompset: Class::ISA doesn't seem to be a dependency. |
01:04 | mtompset | It was on an install attempt for me under Ubuntu 14.04 |
01:04 | (if I recall correctly) | |
01:04 | eythian | robinzarathud:~/catalyst/koha$ grep -r Class::ISA * | wc -l |
01:04 | 0 | |
01:05 | mtompset | Just a second... I'll attempt something. |
01:06 | * mtompset | grumbles about 2 minute boot time, because of wanting to be flexible. |
01:08 | mtompset | Oh... it probably isn't a koha dependency. It's probably a package up a perl library issue. |
01:58 | pastebot | "mtompset" at 127.0.0.1 pasted "What does the suggested command mean?" (6 lines) at http://paste.koha-community.org/34 |
01:58 | mtompset | eythian: Any ideas? |
01:58 | wahanui | Any ideas are welcome :) |
01:59 | eythian | yeah, it's a branch that became a directory with the same name |
01:59 | try running git remote prune origin, though I don't know what it actually does. | |
02:16 | mtompset | Well, it worked, but I got a whole bunch of things "pruned". |
02:17 | eythian | It's probably for the best... |
02:24 | rangi | if you havent run that in a long time, or git gc, you will have had all the old branches that were archived |
02:25 | http://git.koha-community.org/[…]chive.git;a=heads all of those | |
02:51 | jcamins | wizzyrea: when you're the organizer and had months to plan, you don't get a bye that easily. |
03:04 | mtompset | Yes, that looks like it, rangi. |
03:12 | wizzyrea | jcamins: actually, if it's what I'm thinking of, and it was a live demo, I've actually done that. But I didn't have months to plan. |
03:12 | jcamins | wizzyrea: oh, I missed some important context. |
03:12 | wizzyrea | more like "oh by the way we want you to show this giant room the mobile app" |
03:13 | yeah context would be helpful ^.^ | |
03:13 | jcamins | It was a hackathon. The entire point was that a bunch of people would be presenting interesting hacks they'd done. |
03:14 | The subject was a website. | |
03:14 | They announced after everyone was there that they were using a document camera. | |
03:15 | And it wasn't a cost thing... I think they spent over $100k. | |
03:15 | wizzyrea | we are talking about a thing that is kind of like an overhead projector, right? |
03:15 | jcamins | Yes. |
03:15 | wizzyrea | but digital |
03:15 | jcamins | Which they brought in especially to project pictures of computer screens. |
03:16 | Exactly. | |
03:16 | Ever pointed a digital video camera at a computer screen? | |
03:16 | wizzyrea | ok, then what is the proper way to put a phone interface on a projector? |
03:16 | eythian | in bad cases, you'd get flicker |
03:16 | wizzyrea | yes, actually |
03:16 | jcamins | wizzyrea: that's fine if you're presenting on a phone. |
03:16 | wizzyrea | the time I did this, there was no flicker, which may be why I'm a bit sympathetic to this |
03:17 | jcamins | But there was no indication up-front that this was a mobile hackathon. |
03:17 | eythian | it's ideal for a phone, because you can show ui. It seems dumb for a computer though. |
03:17 | jcamins | I wouldn't have gone had I known. |
03:17 | eythian: exactly. | |
03:18 | Fortunately when a sizable portion of the participants said "oh well, I guess we won't present... when's dinner?" they decided to set up the projector hookup. | |
03:18 | wizzyrea | wait they were pointing a document camera at a computer screen? |
03:18 | jcamins | wizzyrea: yes! |
03:18 | wizzyrea | oh I misunderstood I thought you were upset that they were pointing it at a phone. |
03:19 | :) | |
03:19 | jcamins | No, that's a good idea. |
03:19 | wizzyrea | I was like, well that seems sensible. |
03:19 | jcamins | What's dumb is pointing it at a computer. |
03:19 | They wanted two people... one to hold the laptop upside down, the other to type backwards. | |
03:19 | wizzyrea | uhhhh |
03:19 | ok yeah, that's not sensible. | |
03:20 | mtj | my brain hurts just thinking about that :0) |
03:21 | jcamins | :) |
03:21 | Good night! | |
03:21 | wahanui | If you feel like someone is looking through your window, it's OK, it's just me. |
03:21 | mtj | cya jcamins |
03:36 | mtompset | wizzyrea: That's another way of thinking about bug 6874. :) |
03:36 | huginn | 04Bug http://bugs.koha-community.org[…]w_bug.cgi?id=6874 enhancement, P3, ---, julian.maurice, Needs Signoff , Attach a file to a MARC record (Was: File upload in MARC) |
04:33 | wizzyrea | nobody could find the bug with the old title. |
04:38 | eythian | really it should be a biblio record I suppose |
05:04 | @later tell tcohen how do you mean? | |
05:04 | huginn | eythian: The operation succeeded. |
05:18 | jseplae joined #koha | |
05:35 | cait joined #koha | |
05:46 | paxed | *grmbl* fun, just noticed Yet Another Conversion Problem. the due dates were set with time 00:00:00, when they should've been 23:59:00 |
05:48 | laurence joined #koha | |
06:01 | mtompset | And with that horrible grammar moment, I think it is best to head to sleep. |
06:01 | dcook: 11592 is ready for testing. :) | |
06:02 | @later tell dcook bug 11592 is ready for testing now. | |
06:02 | huginn | mtompset: The operation succeeded. |
06:02 | mtompset | Have a great day, #koha eythian wizzyrea mtj |
06:15 | cait | @wunder Konstanz |
06:15 | huginn | cait: The current temperature in Taegerwilen, Taegerwilen, Germany is 20.0°C (8:15 AM CEST on June 09, 2014). Conditions: Clear. Humidity: 83%. Dew Point: 17.0°C. Pressure: 30.12 in 1020 hPa (Rising). |
06:36 | reiveune joined #koha | |
06:36 | reiveune | hello |
06:37 | cait | hi reiveune |
06:37 | holiday here today :) | |
06:37 | reiveune | here too, hi cait |
06:37 | cait | oh nice :) |
06:47 | ashimema joined #koha | |
06:54 | Joubu | Hi #koha |
06:54 | ashimema | Morning Joubu |
06:55 | alex_a joined #koha | |
07:00 | yohann joined #koha | |
07:00 | yohann | salut |
07:01 | cait | hi yohann and ashimema |
07:01 | oh and hi Joubu | |
07:01 | *reads back* | |
07:01 | ashimema | moring cait, morning yohann |
07:02 | * ashimema | wishes google chrome would 'just work' the way it used to on this Ubuntu box. |
07:05 | * cait | offers cookies |
07:12 | ashimema | brb.. |
07:16 | sophie_m joined #koha | |
07:20 | gaetan_B joined #koha | |
07:20 | ashimema joined #koha | |
07:20 | gaetan_B | hello |
07:21 | cait | hi gaetan_B |
07:21 | everyone around on a holiday? | |
07:22 | gaetan_B | gaetan_B: hmmm, it's not really a holiday in France anymore but it used to be so a lot of people take a day off i guess |
07:22 | cait | ah |
07:22 | gaetan_B | not so many here at biblibre though |
07:23 | fridolin joined #koha | |
07:26 | fridolin | hie all |
07:27 | cait | hi fridolin |
07:52 | Tony joined #koha | |
07:53 | Tony | Greetings. I'm getting errors on staging marc for import: stage-marc-import.pl: Filehandle STDOUT reopened as FH only for input at /usr/lib/perl5/Template/Provider.pm |
07:53 | This is a clean install and we are trying to populate the db for the first time | |
07:54 | On Manage staged MARC records, the citations are null. Any thoughts? | |
08:09 | tony123 joined #koha | |
08:10 | tony123 | Greetings. I'm trying to do bulk import (file generated from MarcEdit) an on Stage MARC for import, I'm getting this error: stage-marc-import.pl: Filehandle STDOUT reopened as FH only for input at /usr/lib/perl5/Template/Provider.pm |
08:10 | paul_p joined #koha | |
08:11 | tony123 | In Manage staged MARC records, the Citation field is null for each record. The staging results shows items as 0 |
08:11 | Any thoughts? This is a clean install and trying to import he data for the first time | |
08:28 | paxed | does anyone else find advance_notices.pl formats the date-fields in wrong format? |
08:30 | cait | tony123: it#s hard to tell |
08:30 | are you importing a valid marc file? | |
08:30 | in iso format? | |
08:30 | paxed: there is a bug for that - the problem was the the proposed solution slowed down the notice generation quite a bit, so it got stuck I think | |
08:31 | paxed | cait: ugh. and one of the most visible notices for patrons, at that. sucks. |
08:31 | cait | i can find the bug for you |
08:32 | tony123 | I assume so... I used MarcEdit to build the file. I used the Z39.50 client to retrieve the ISBN data. It built file and decompiled file (back to the mrk file) |
08:32 | cait | bug 11244 |
08:32 | huginn | 04Bug http://bugs.koha-community.org[…]_bug.cgi?id=11244 normal, P5 - low, ---, kyle, In Discussion , notices ignoring the dateformat preference |
08:32 | cait | tony123: did you convert it to an mrc file after using marcedit? |
08:32 | or with marcedit? | |
08:33 | tony123 | I used marcedit to convert it to the mrc file |
08:33 | cait | tony123: sorry to ask so many questions, but it's hard to tell - the message from thelogs probably is just a warning and not a problem |
08:33 | ok | |
08:33 | tony123 | no worries :) |
08:33 | cait | your system is set up to use marc21? :) |
08:33 | tony123 | Here is the results log: |
08:33 | MARC staging results : 2 records in file 0 records not staged because of MARC error 2 records staged Did not check for matches with existing records in catalog 0 item records found and staged Manage staged records Back | |
08:34 | cait | that looks not bad |
08:34 | but you say the title does not show up in the table below? | |
08:34 | on manage staged? | |
08:34 | tony123 | the values in Staged MARC record management are all null |
08:34 | cait | hm |
08:34 | what do you mean by values? | |
08:34 | does the batch not show up at all? | |
08:35 | tony123 | the values in the the Citation field are all null. Status: Stages. Match details and Record and blank |
08:35 | the text is "null" with a hyperlink | |
08:36 | When clicked, the javascript error is: TypeError: $(...).html(...).dialog is not a function title: _("MARC Preview") | |
08:37 | BobB | tony123, you could go back to MarcEdit and run the marc validator - that will show if your records are valid; |
08:37 | tony123 | sure, thanks. |
08:38 | cait | packages? |
08:38 | wahanui | somebody said packages was at http://wiki.koha-community.org/wiki/Debian |
08:38 | BobB | do that on the .mrk file, then recompile it to .mrc |
08:38 | cait | tony123: just for a test - can you catalog a reacord manually? |
08:38 | BobB | are you sure you did that before? |
08:38 | cait | tony123: i had this happen when someone deleted the default framework once |
08:38 | BobB | Anyway, if it compiles ok, then it should mean your marc file is ok and your problem is a Koha one |
08:38 | cait | so there were no definitions for the fields and subfields in Koha - it's in the mandatory part of the web installer, but it's possible to uncheck it |
08:39 | and also what BobB says :) | |
08:39 | BobB | :) |
08:39 | tony123 | MarcEdit's MARCValidator results: No errors were reported using the specified rules file. |
08:40 | Koha is installed on Ubuntu 12.04 | |
08:42 | BobB | by what method? how did you install Koha? |
08:42 | tony123 | apt-get via the instructions on the website |
08:42 | If you think using a different distro will fix it, or reinstalling, I'm happy to do that | |
08:42 | BobB | hmm ... |
08:42 | cait | tony123: when you go into cataloguing |
08:43 | can you add a record there? | |
08:43 | tony123 | hang on, one sec. Let me check |
08:43 | cait | and when you open your file in a editor... it should be basically unreadable, then it's right :) |
08:52 | BobB | diinner time, I'm off |
08:56 | tony123 | Thanks BobB for your help! :) |
08:57 | I did create a record. I realized that I didn't create an Item type... I'll retry the import now. | |
08:58 | cait:: yes, the mrc file is largely unreadable. | |
09:00 | Same result: Citation field reads "null" Status reads "Staged". Match details and Records and both blank. | |
09:00 | Do you recommend doing a clean install? | |
09:00 | If you recommend a different distro, I can do that too | |
09:05 | cait | didyou try catalouging manually? |
09:05 | if you haven't made any configuration yet | |
09:06 | tony123 | I did a manual cataluging. It went in fine. |
09:06 | cait | i am not sure reinstalling would make a difference |
09:06 | what's your marcflavour system preference set to? | |
09:08 | tony123 | Good question... where do I find it? I think I just selected the default flavour |
09:08 | cait | administration > system preferences |
09:08 | and search for marcfl | |
09:08 | wahanui joined #koha | |
09:11 | tony123 | thanks. UNIMARC |
09:14 | cait | oh |
09:14 | where are you located? | |
09:15 | tony123 | Central Asia |
09:15 | cait | hm |
09:15 | my guess is that your database is unimarc but your data is marc21 | |
09:15 | tony123 | ahhh |
09:15 | cait | for example, the title in unimarc is 200 and the title in marc21 is in 245 |
09:15 | so there are quite some important differences | |
09:16 | i am not sure which is the most common format in central asia | |
09:16 | tony123 | We have Cyrillic titles, so we need to run unicode |
09:16 | Ok, this gives me something to go on | |
09:16 | cait | it's a misunderstanding that happens often |
09:16 | unimarc has nothing to do with unicode | |
09:16 | tony123 | oh, ok. wrong assumption. :) |
09:16 | cait | marc21 can be unicode too :) we are using it with hebrew for example |
09:17 | ok, now we got the problem | |
09:17 | just switching the pref won't fix it all | |
09:17 | it mght be best to redo the installatoin - did you install from packages? | |
09:17 | tony123 | I did |
09:18 | cait | that's cool then |
09:18 | tony123 | No problem. At least I have something to go on now :D |
09:18 | cait | in the conf file for creating the instance |
09:18 | check what marcflavour is given there | |
09:18 | it should be marc21 | |
09:18 | tony123 | ok |
09:18 | cait | then drop your instance or create a new one, make sure in the web installer, that you select marc21 as well |
09:18 | and then try reimporting | |
09:19 | tony123 | Great, thanks! I'll make the changes and reinstall. |
09:20 | cait | hope it all works out on second try :) |
09:20 | Viktor joined #koha | |
09:20 | cait | making sure the instance is marc21 is important so your data can be searched correctly /indexing configuration |
09:21 | it's a lot easier that way then trying to fix everything manually | |
09:21 | hi Viktor :) | |
09:21 | Viktor | Hi cait :) |
10:06 | * paxed | boggles at advance_notices.pl |
10:06 | paxed | $titles .= join("\t",@item_info) . "\n"; |
10:06 | 'items.content' => $titles, | |
10:07 | so, <<items.content>> in the notice will contain tab-separated lines of crap that's mostly useless to the patron? | |
10:08 | cait | it contains the fields you define |
10:08 | to generate an item list | |
10:08 | but yeah, it's tab separated, only overdues can currently do the nicer <item></item> formatting i think (haven't tested lately) | |
10:09 | there is a command line option for the fields it will output for items.content | |
10:10 | paxed | also, not very html-firendly, that. |
10:11 | cait | paxed: i might remember incorrectly, but i thought there was some code to bulid a table from it |
10:11 | for the html, but that might be another notice... hm | |
10:16 | sophie_m | paxed: cait it used to be an array formated, I don't know when it disappeared |
10:17 | paxed | *grmbl* |
10:41 | bug 11607 | |
10:41 | huginn | 04Bug http://bugs.koha-community.org[…]_bug.cgi?id=11607 enhancement, P5 - low, ---, koha-bugs, NEW , items.content does not contain any formatting when HTML message is selected. |
11:32 | Viktor joined #koha | |
11:45 | collum joined #koha | |
11:46 | talljoy joined #koha | |
11:47 | NateC joined #koha | |
11:49 | meliss joined #koha | |
12:13 | oleonard joined #koha | |
12:19 | BobB joined #koha | |
12:31 | nengard joined #koha | |
12:56 | Dyrcona joined #koha | |
12:58 | fridolin joined #koha | |
13:00 | druthb | o/ |
13:00 | fridolin left #koha | |
13:03 | edveal joined #koha | |
13:05 | oleonard | Hi druthb, and a belated happy birthday to you |
13:05 | I hope it was a nice one | |
13:06 | rambutan joined #koha | |
13:08 | druthb | It was lovely! Had a great weekend—busy busy, but good. |
13:20 | kmlussier joined #koha | |
13:48 | ashimema | hmmm. no cait today? |
13:54 | oleonard | So the deletedborrowers table doesn't show a timestamp for when they were deleted, or am I missing it? |
14:10 | BigRig joined #koha | |
14:11 | BigRig joined #koha | |
14:16 | ashimema | There was a bit of a debate around that oleonard.. |
14:16 | I'll did out the bug number for when it was all worked out. | |
14:16 | the data is there.. somewhere. | |
14:20 | oleonard | Oh, in action_logs it seems ashimema? |
14:20 | ashimema | There's one.. http://bugs.koha-community.org[…]w_bug.cgi?id=8926 |
14:20 | huginn | 04Bug 8926: enhancement, P5 - low, ---, gmcharlt, NEW , deletedborrowers should have a timestamp |
14:20 | ashimema | but thats not the one I was thinking of.. |
14:20 | maybe it's deleted bibs I was tihnking of.. | |
14:20 | there was a bug about updating the timestamp in a dleeted table.. which I QA'd | |
14:21 | must be getting my wires crossed | |
14:23 | I can't find the bug I was thinking of.. so must have wires crossed somewhere.. sorry. | |
14:28 | oleonard | Oh nice. You can't use a saved SQL report to query action_logs for 'DELETE' transactions because the query has the word DELETE in it :P |
14:29 | paxed | use concat? |
14:29 | or some other way to split the searchable text | |
14:38 | ashimema | oh dear.. how silly |
14:52 | rocio joined #koha | |
14:53 | NateC joined #koha | |
15:02 | tgoat joined #koha | |
15:13 | yohann left #koha | |
15:19 | cait joined #koha | |
15:31 | * cait | waves |
15:56 | jcamins | @later tell gmcharlt It occurs to me that it might be a good idea to have something in the code of conduct about cultural sensitivity, given the different traditions and cultures of members of the community. |
15:56 | huginn | jcamins: The operation succeeded. |
15:56 | gmcharlt | jcamins: indeed |
15:57 | jcamins | I guess it's later than I realized. |
15:57 | barton | oleonard: are you looking to run the report as a cron job or from the reports interface? |
15:59 | (... reports interface => koha's reports page) | |
15:59 | oleonard | From the reports interface. I ended up getting good results by using SUBSTRING |
15:59 | barton | hmm. tricky. |
16:01 | Do you mind posting that? I'd like to see what you did. | |
16:04 | pastebot | "oleonard" at 127.0.0.1 pasted "Querying deletedborrowers based on DELETE action in action_logs" (1 line) at http://paste.koha-community.org/36 |
16:04 | oleonard | ...of course one would want to limit that query in other ways on a production system |
16:15 | laurence left #koha | |
16:25 | mtompset joined #koha | |
16:25 | mtompset | Greetings, #koha. |
16:32 | trendynick joined #koha | |
16:55 | gaetan_B | bye ! |
17:04 | reiveune | bye |
17:04 | reiveune left #koha | |
18:09 | jcamins | nengard: are you still on the constant lookout for video suggestions? You may already have this one, but if not a video showing how to mark items lost on recent versions of Koha might be an idea. |
18:36 | rambutan joined #koha | |
18:52 | Viktor joined #koha | |
18:58 | cait | gmcharlt++ |
19:15 | paul_p joined #koha | |
19:23 | cait | @later tell tcohen ping me when you got some time to talk? |
19:23 | huginn | cait: The operation succeeded. |
19:30 | nengard_ joined #koha | |
20:10 | nengard | jcamins - haven't had time to do videos in forever |
20:10 | what what has changed about marking items lost? | |
20:10 | jcamins | nengard: you used to do it in the additem screen but now you do it on the moredetails screen. |
20:11 | nengard | that happened a LONG time ago so I think I have that covered - not in a video but in a tutorial i wrote |
20:14 | rangi | morning |
20:14 | jcamins | nengard: yeah, I only thought of it because someone who learned from 3.2 tutorial videos and didn't need to mark anything lost since then (!) mentioned it. |
20:14 | nengard | hehe |
20:14 | got it | |
20:45 | jce joined #koha | |
20:50 | nengard left #koha | |
20:51 | jce | I'm working on a Koha 3.08 server that's having problems with OPAC. When I do a search under "Library Catalog" I can get plausible results. But if I select 'Author', 'Title', 'Subject', 'ISBN' or any of the other options from the pull-down list, I get "No results found!" I've tried rebuilding Zebra many times. Are the "Library Catalog" results not related to zebra? |
20:52 | kmlussier left #koha | |
20:53 | cait | jce: hmmm are you using opacsuppression? |
20:53 | are you using searchmylibraryfirst or a similar setting? | |
20:54 | jce | cait: I'm not aware of having any of those options set. Where are they found? |
20:55 | I have the same problem on the original server that I'm trying to fix, and on a server running in a VM that I loaded an SQL backup into. | |
20:55 | cait | yeah, looks like something in your configuration |
20:55 | check administration > system preferences > opac suppression first | |
20:55 | well opacsuppresson | |
20:55 | opacsuppression | |
20:55 | wahanui | well, opacsuppression is different. |
20:57 | jce | I know there are some bad biblio records that I need to fix, but I don't think this problem is related. |
20:57 | cait | did you check the setting? |
20:58 | turn the feature off ( don't hide records) if it's on | |
20:58 | jce | Just getting the VM fired up again. |
21:09 | OpacSuppression is set to "Don't hide". It also says" "Note that you must have the Suppress index set up in Zebra and at least one suppressed item, or your searches will be broken." Does that pertain only if you have it set to 'hide'? | |
21:13 | SearchMyLibraryFirst is set to "Don't limit". | |
21:15 | Could it be a problem of items (holdings?) vs. biblio records? Is that the difference between the "Library catalog" search and the others? | |
21:18 | cait | hm |
21:18 | what does your search url for a non keyword search look like? | |
21:19 | jce | You mean a search under "Library catalog" in the pull-down? |
21:20 | localhost/cgi-bin/koha/opac-search.pl?q=searchterm | |
21:21 | cait | so that works |
21:21 | how does the other url look like? for a title search or similar? | |
21:23 | jce | Hmm. It does look different: bcmckoha/cgi-bin/koha/opac-search.pl?idx=ti&q=searchterm [bcmckoha would probably be an instance name, and may not be in /etc/hosts] |
21:27 | cait | hm that doesn't look quite right |
21:27 | really one is localhost and the other bcmckoha? | |
21:28 | jce | Actually, as I look at it again, I don't think that's the problem. On the 'real' server, the URL reads: localhost/cgi-bin/koha/opac-search.pl?idx=au&q=searchterm |
21:30 | Sorry, it reads differently between the 'real' server and the VM, but is consistent on each. | |
21:30 | mtj | morning all... |
21:31 | jce: if i were you, i would load your 3.0.8 db into a new 3.14.x koha | |
21:32 | ...and see if your search problems are fixed | |
21:33 | cait | jce: running out of ideas sorry |
21:33 | jce | Well, I've tried that and not had good results. I'm trying to fix 3.08 first, before upgrading. |
21:33 | cait | it's certainly weird... |
21:33 | you didn't change any settings? when did it stop working? | |
21:33 | jce | cait: I |
21:33 | cait | you could check the action logs table for configration changes done in that time span |
21:33 | mtj | yeah, it is a bit weird :/ |
21:34 | jce, perhaps your mappings or frameworks are a bit wonky? | |
21:34 | jce | 'm not exactly sure. I'm an outside consultant who comes in to maintain a small church library's Koha server when problems arise. |
21:34 | mjt: that's entirely possible. | |
21:35 | cait | jce: action_logs should hav elogged changes to system preferneces |
21:35 | so if the problem is in the configuration... might be a chance to find it there | |
21:35 | if it happened out of nowhere without updating | |
21:36 | jce | cait: I doubt there was a change in the config, but I guess it could be. |
21:36 | mtj | Tools -> Koha to MARC mapping |
21:37 | jce: if you really get stuck - you usually need to set up a fresh/clean koha, and start comparing settings | |
21:38 | jce | The system was originally set up in about 2007 or 2008 and has had several OS and Koha version upgrades. I think that one problem came in when updating to 3.08. Some of the MARC records have the infamous frey50 infection. But I don't think that pertains to this problem. |
21:39 | cait | it's like i have seen something like this somewhere, but can't figure out where and what it was :( |
21:39 | mtompset | Is it 3.0.8 or 3.8? |
21:39 | jce | The VM is a fresh install that has the most recent SQL backup loaded. |
21:39 | mtompset | There is that extra upgrade step if it is prior to 3.4 :) |
21:41 | mtj | jce, if its a really old koha, the zebra config files might be way out of date too |
21:41 | jce | It's 3.8. The package download file is 'koha-common_3.08.16.1-1_all.deb'. |
21:42 | mtj | ah ok... not too many zebra changes since 3.8 - thats good |
21:43 | jce | mtj: I did a "Complete Removal" of Koha in Synaptic, which I understand removes config files as well. Is it possible that old zebra config could have survived? |
21:43 | mtj: but it did start out life long before 3.8. | |
21:43 | mtj | hmm, yeah, its possible |
21:44 | pro-tip: i use git to track /etc, before i do any upgrade | |
21:44 | ...then you can always do a nice before/after diff, to see what really changed | |
21:45 | jce | Part of this long, sad saga is that I did upgrade from 3.8 to 3.14, but it didn't go well, so I put it back to 3.8 to fix it. I'm now going to upgrade a VM first, then migrate that data to the production server. |
21:46 | Guess that's another reason to learn how to use git. :) | |
21:46 | mtj | yeah, i use git for general sysadmin tasks too |
21:47 | jce | On the other hand, stale config files aren't likely to be plaguing the VM. |
21:48 | mtj | yeah, correct |
21:48 | jce | That's a fresh install of 3.8. |
21:48 | So it must be something in the SQL. | |
21:49 | cait | jce: it would be my guess too... but really hard to tell :( |
21:49 | youcould check the error logs | |
21:49 | or run the zebrasrv in the foreground (can you say that in english?) and look at the queries | |
21:49 | see if that gives you some kind of hint | |
21:50 | mtj | you really need to start testing from scratch - against a clean koha, with some clean test records |
21:51 | prove that an ISBN search works on your clean koha, and work backwards | |
21:52 | here is a dir of handy test records from LOC -> http://www.loc.gov/catdir/cpso/RDAtest/ | |
21:53 | jce | mtj: by clean koha, you mean a fresh instance? |
21:54 | cait | a fresh instance... and some records, see if that can be searched |
21:54 | both instances use the same koha code.. might give a clue? | |
21:54 | but probably I'd take a look at the logs first | |
21:55 | mtj | jce: yeah, a fresh instance |
21:56 | load these bibs into your fresh instance -> www.loc.gov/catdir/cpso/RDAtest/extra_bib_marc.zip | |
21:57 | jce | Ok, on the VM, the last error in /var/log/koha/BCMCKoha/opac-error.log is: opac-search.pl: Use of uninitialized value $sort_by[0] in join or string at /usr/share/koha/opac/cgi-bin/opac/opac-search.pl line 698. |
21:57 | mtj | ...then index, and confirm that the seaching behaves 'sanely' |
21:58 | ...then load those bibs into your problem koha, and confirm that you get different search results | |
21:58 | cait | ah |
21:58 | jce | mtj: I'll have to add items against these bibs, correct? |
21:58 | cait | hm what is the default sort option set to in this instance? there are prefs for that |
21:58 | look for search | |
21:59 | mtj | jce: no need for items |
21:59 | pianohacker joined #koha | |
21:59 | pianohacker | hallo |
22:02 | http://devopsreactions.tumblr.[…]sting-my-own-code | |
22:03 | cait | aaw |
22:03 | that kid is cool | |
22:03 | rangi | http://kohadevreactions.tumblr[…]-wrote-months-ago |
22:04 | pianohacker | rangi: I was looking at the last attempt I did at ajaxcirc back in 2009 the other day like that... |
22:04 | rangi | :) |
22:05 | pianohacker | "What a dumbass idea. Who thought that was reasonable? ... Oh..." |
22:05 | cait | itshows your development :) |
22:07 | jce | Ok, I searched 'search' in preferences and found that NoZebra was set to "Don't use". Hmm. I set it to 'Use' and saved Searching preferences. Didn't seem to help immediately. Do I need to restart apache to make Koha see that change? |
22:08 | pianohacker | jce: which version of koha are you using? |
22:09 | cait | jce: undo that |
22:09 | jce | 3.8 |
22:09 | mtj | jce: that is very very suspicious |
22:09 | cait | jce: Don't use NoZebra is correct |
22:09 | it#s double negated... it means use zebra then | |
22:09 | in other words: you don't want to use NoZebra :) | |
22:09 | pianohacker | jce: NoZebra should be set to don't use, it's old code that was removed not long after the version you were using |
22:10 | mtj | jce: your search should not work at all, without zebra |
22:11 | jce | It's worded very poorly. The pull-down says [Use|Don't use] the Zebra search engine. But I'll set it back the way I found it. It didn't appear to make any difference. |
22:11 | mtj | is your koha config pointing to another db? |
22:12 | cait | jce: hm the wording is irritating me too indeed |
22:12 | pianohacker | jce: yeah, sorry about that. That and several other things were why it's since been eviscerated |
22:12 | jce | mtj: No. And again, I do get results when I don't specify Title, Author, etc. |
22:13 | pianohacker | jce: Do you have zebra set up for unimarc and are using MARC21, or vice versa? |
22:14 | jce | pianohacker: No, it's MARC21 all the way. |
22:14 | pianohacker | hmm, okay. what flags are you passing to rebuild_zebra.pl? |
22:16 | jce | cait: I see defaultSortField and defaultSortOrder and OPAC versions thereof. Both are set to relevance and ascending. |
22:16 | pianohacker | and wait a second. From what I'm reading in the preferences file, shouldn't NoZebra be set to "use"? The .pref file accounts for the negation |
22:16 | at least as of 8626a5bebb3f900d852a3d98c5f45d95cea5272d | |
22:17 | cait | pianohacker: i was wondering that |
22:17 | jce | pianohacker: -f -v {instance} |
22:17 | cait | maybe you are right |
22:17 | i am glad we finally removed that | |
22:17 | end of confusion | |
22:17 | pianohacker | amen! |
22:18 | there are several .pref descriptions that I wrote that only partially removed confusion... things like gist, for instance | |
22:18 | jce | So to clarify, the NoZebra preference is already do-nothing cruft by 3.8, or do I need to make sure it is set right? |
22:19 | pianohacker | that's a good question. |
22:19 | jce | It didn't seem to have an effect. |
22:19 | mtj | looks like its still live on 3.8.16 |
22:20 | the search code in opac-seach.pl refers to it | |
22:20 | pianohacker | yup |
22:21 | ace | |
22:21 | mtj | jebus |
22:21 | pianohacker | jce: the only thing I can think of is to also pass -x to rebuild-zebra |
22:21 | mtj | according to searching.pref, nozebra = yes *enables* zebra?! |
22:22 | jce | Would I need to rebuild zebra's index to get the NoZebra preference to take effect? |
22:22 | mtj | oops, i take that back ^^ :) |
22:22 | pianohacker | the wording is confusing, I hated describing the No* style prefs... |
22:22 | jce | mtj: But in the web interface, the wording is "Use" or "Don't use" zebra search engine. |
22:23 | mtj | yes = 'dont use' |
22:23 | pianohacker | jce: It should be set to "Use", from what we can tell |
22:23 | jce | Mine was set to "Don't use." But setting it to "Use" didn't immediately change anything. |
22:23 | Ok, I'll set it back, restart apache, reindex (with -x) and see what that does. | |
22:24 | pianohacker | mtj: since "Use" should set NoZebra to no or 0 |
22:24 | jce: also restart the zebra server, just to be safe | |
22:29 | mtj | yeah, nozebra needs to be set to 'use' - thanks pianohacker, cait :) |
22:29 | jce | Ok, after I set it to "Use" I got no results in "Library catalog". When I put it back to "Don't use" "Library catalog" searches work. |
22:30 | cait | hmmmm - but only library catalog searched, not the other options |
22:30 | jce | This suggests to me that my zebra is broken and "Library catalog" searches aren't using zebra. Is this possible? |
22:30 | cait: right. | |
22:30 | cait | they normally use zebra |
22:30 | but that is really what the pref des | |
22:31 | does | |
22:31 | it might be zebra is generally broken for some reason... and that only nozebra partially works | |
22:31 | mtj | your koha has not been using zebra, perhap jce? |
22:31 | eythian | hi |
22:31 | jce | mtj: It was at one time. It may be broken now. |
22:31 | cait | does the rebuld give positive feedback? you could try adding the -x and a second -v |
22:33 | mtj | jce: if you stop zebra, and get search results... you have some problem :) |
22:34 | pianohacker | yeah, that's a good idea. Stop zebrasrv/koha-zebra if it's running, then try again |
22:34 | mtj | nozebra = 'dont use' should *not* give you any results, on 3.8.16 |
22:34 | jce | cait: The syntax on koha-rebuild-zebra: -x -x -v -v, or -xx -vv? |
22:34 | cait | one x to v |
22:34 | 2 v | |
22:34 | i think | |
22:34 | -x -v -v for more output | |
22:35 | also the -f i think | |
22:37 | jce | It exports 3295 biblios, which is plausible. It gives a warning: Record didn't contain match fields in (bib1,Local-number) |
22:37 | cait | ah |
22:37 | that's not good | |
22:37 | jce | Is that a problem with mapping or frameworks? |
22:39 | mtj | hmm, mapping (i think?) |
22:40 | jce | mtj: pianohacker: I did 'koha-stop-zebra' and I can still search by "Library catalog" and get results. Keyword searches still don't work. Clearly Zebra is not working. |
22:40 | cait | can you still search when you flip the nozebra? |
22:40 | mtj | jce: double check |
22:40 | $ ps -ef | grep zebra | |
22:41 | tcohen joined #koha | |
22:42 | mtj | jce: my hunch is you are not actually using zebra |
22:43 | jce | mtj: assuming koha-stop-zebra actually shut Zebra down, that's what's happening. So how do I get Zebra running again? |
22:44 | cait | i thnk it's a data problem |
22:44 | something about the biblionumber? | |
22:45 | mtj | yeah, a 999/mapping problem |
22:45 | jce | cait: It must be a data problem that has moved from the original server to the VM in the SQL. A mapping problem would make sense. |
22:46 | tcohen | #koha: i'm on a small trip to Ecuador, i'll try to push stuff to master soon, but not today |
22:47 | eythian | jce: koha-start-zebra |
22:47 | jce: though, I recommend 'sudo service koha-common start/stop/restart' instead | |
22:48 | jce | I ran a bibliographic framework test a couple days ago and it said I had tabs 6 and 3 in use in tag 696, but when I went to try and fix it, I couldn't figure out how to get rid of one. |
22:49 | eythian: the real problem is not how to start the server, but how to get it working correctly. | |
22:49 | cait | tcohen: have fun :) |
22:49 | jce: the problem might be that your records are missing the 999 tag | |
22:49 | with the bilbionumber | |
22:50 | tcohen | eythian: don't know what was about |
22:50 | mtj | tcohen: Ecuador sounds nice :0) |
22:50 | eythian | tcohen: ES API |
22:50 | tcohen | it is :-D |
22:50 | eythian: oh | |
22:50 | eythian | I didn't understand your question |
22:50 | tcohen | we all know Zebra's code needs to be refactored |
22:51 | eythian | s/refactored/taken out and shot/ but yes. |
22:51 | tcohen | and also, that in the mid term, Zebra might even dissapear |
22:51 | I was thinking, that if there was an API for implementing the ES code | |
22:51 | (like there was for SolR) | |
22:51 | some people might like to work on making Zebra code a bit better | |
22:52 | following that hypothetical API | |
22:52 | eythian | tcohen: there is, sorta. |
22:52 | pianohacker | It's going through K::SearchEngine, no? |
22:52 | tcohen | pianohacker: that's SolR |
22:52 | eythian | I'm writing a compatibility layer so it can get requests from the existing stuff |
22:52 | it is also mostly using K::SE | |
22:52 | wahanui | okay, eythian. |
22:53 | tcohen | ok, so sticking to K::SE is a good way to do it |
22:53 | thanks eythian | |
22:53 | eythian | It also has its own API, though it's a bit under-defined at the moment as I'm sorting out use cases. |
22:53 | Basically you feed it a lucene search string. | |
22:53 | tcohen | instead of PQF |
22:54 | eythian | Yeah. |
22:54 | tcohen | so the search is not built byQueryParser |
22:54 | eythian | It is |
22:54 | cait | oh |
22:54 | eythian | If you're using the zebra-like code, it builds a lucene-type search string, and then turns that into an ES query. |
22:54 | cait | interesting |
22:54 | wahanui | somebody said interesting was sometimes good and sometimes bad |
22:55 | eythian | this also means that you can type lucene stuff straight into the keyword box and it should work. |
22:55 | e.g. title:foo AND subject:bar | |
22:55 | pianohacker | wait, does that mean the existing code allows raw PQF for zebra? |
22:56 | eythian | probably |
22:56 | possible | |
22:56 | wahanui | i heard possible was everything, but I think it's probably easy to miss something here |
22:56 | eythian | y |
22:57 | jce | cait: I'm looking through a MARC XML file I exported and it appears that you're right. No 999 tags. Could it be that we were never using Zebra even though we thought we were? Seems more likely than that all the 999 tags got stripped out. Is it possible that there was another tag that was being used instead. I have vague thoughts about 7xx-something. |
22:58 | tcohen | eythian: will there be something like K::SE::ES ? |
22:58 | jce | I see 700 tags that contain what look like author and title information. |
22:58 | eythian | tcohen: yep |
22:58 | tcohen | awesome |
22:59 | cait | jce: you ar elooking basically for 2 identical numbers |
22:59 | eythian | that aspect of it needs a bit of work right now, but it's really just a bit of rearranging of files and such. |
23:00 | mtj | jce, you are prolly missing 999 framework mappings - thats why your bibs have no 999 fields |
23:00 | Home › Administration › Koha to MARC mapping | |
23:00 | tcohen | eythian: thanks for the update |
23:01 | mtj | ..its a common-ish bug for upgraded old kohas |
23:01 | cait | tcohen: when will you be back? :) |
23:01 | eythian | some time soon I hope to squash patches appropriately and publish something for testing. |
23:01 | well, super-alpha tesing | |
23:01 | tcohen | saturday actually |
23:02 | eythian: looking forward for it, thanks | |
23:03 | mtj | jce: based on your info - you probably have not been using zebra for a while |
23:03 | * tcohen | notices that hotel hasn't filtered port 22 :-D |
23:03 | jce | mtj: That could be. |
23:04 | mtj | ..if ever? |
23:05 | jce | So, I'm looking in Koha to MARC mapping, and I don't find any reference to 999. I'll need to add it then? Should this be under biblio, biblioitems, or items? |
23:05 | cait | hmm |
23:05 | eythian | that sounds like it would be very problematic. |
23:06 | cait | i'd recommend looking at a standard default framework |
23:06 | eythian | I'd be inclined to reload them from default |
23:06 | cait | probably even load a standard framework |
23:06 | yeah | |
23:06 | i agree | |
23:07 | mtj | me... i would load your bib, bibitems and items tables into a fresh koha, and test from there |
23:07 | cait | the problem is, even after reloading the frameworks |
23:07 | rangi | mtj: that wont work if those records have no 999 on them |
23:07 | mtj | or the other way around, too :) |
23:07 | cait | you'd need to add bilbionumbers i guess |
23:07 | rangi | yep |
23:08 | the best thng to do | |
23:08 | * eythian | wonders if there's a fixbiblionumbers type script around |
23:08 | rangi | before worrying about frameworks |
23:08 | mtj | aah, yeah - those arent automagically created :/ |
23:08 | rangi | is do a select marcxml from biblioitems limit 2; |
23:08 | mtj | me forgot that |
23:08 | rangi | and have a look |
23:11 | papa joined #koha | |
23:11 | jce | rangi: I'll need a little more context for your suggestion. I've exported all the records in marcxml. Are you just saying to check if I have 999 in the records. I can already confirm I don't. |
23:16 | dcook | pianohacker: If you prefix your query with "pqf=" you can send straight PQF to Zebra |
23:16 | I don't think it would work without that prefix though as it would just interpret it as a kw search, me thinks | |
23:20 | * pianohacker | has a geeky trick |
23:20 | pianohacker | dcook++ |
23:20 | dcook | hehe |
23:27 | jce | Ok all, here's another bit of information. When I last tried to re-index Zebra, it reported the following: BIBLIONUMBER in: 090$c BIBLIOITEMNUMBER in: 090$d. I take it that these should be in 999$c and 999$d? |
23:28 | rangi | isnt 090 unimarc |
23:28 | ? | |
23:28 | hmm and old old koha too i think | |
23:29 | jce | www.loc.gov/marc/bibliographic/bd09x.html says 090 is "Local Call Number [OBSOLETE, 1982] |
23:30 | So maybe I simply need to run a script to upgrade the schema? [hoping for something easy like that] | |
23:30 | cait | you should change the records |
23:30 | if you can do that, i think it might work | |
23:31 | loc is not relevant here - this is koha specific | |
23:31 | dcook | rangi: It looks like 090$9 is unimarc |
23:31 | So I'm guessing old Koha? | |
23:32 | jce | Would it work to do a global text replace in a marcxml dump, changing the 090 tag to 999? |
23:32 | rangi | id try it on 2 or 3 records |
23:32 | reindex | |
23:32 | jce | dcook: I think this is still leftover mess from years ago. |
23:32 | rangi | and see if it works, then do all of them (after a backup of course) |
23:32 | dcook | jce: I think you'll also need to change the place that is telling Zebra to look in 090$c and 090$d, I suspect |
23:33 | rangi | dcook: zebra is looking in 999 |
23:33 | jce | rangi: sounds like a good idea. I have a working backup, the original server, and 2 virtual machines. |
23:33 | dcook | "GetMarcFromKohaField" |
23:33 | rangi | the problem is the are in 090 |
23:33 | dcook | rangi: BIBLIONUMBER in: 090$c BIBLIOITEMNUMBER in: 090$d is based on GetMarcFromKohaField |
23:33 | rangi | you will need to fix up your frameworks too jce |
23:33 | yes | |
23:33 | thats the problem | |
23:34 | the data and the frameworks say 090 | |
23:34 | zebra looks for 999 | |
23:34 | need to fix the data and fraemworks to say 999 | |
23:34 | dcook | That's what I meant. Bad communication on my part :p |
23:34 | Also want to update Marc => Koha mappings, as those get used quite a bit in Koha | |
23:34 | rangi | (without fixing the frameworks, the problem will recreate itself over time) |
23:35 | dcook | Err "Koha to MARC Mapping" |
23:36 | jce | dcook: is GetMarcFromKohaField in system preferences? I'm not finding it there. |
23:37 | cait | it's in administration |
23:37 | koha 2 marc mapping | |
23:37 | dcook | /cgi-bin/koha/admin/koha2marclinks.pl |
23:37 | wahanui | /cgi-bin/koha/admin/koha2marclinks.pl is the place to check 1st |
23:38 | dcook | Whoa. Wahanui and I are in sync. |
23:38 | cait | you'd want to make sure that the mappings there are correct - they get applied to all frameworks |
23:38 | dcook | Not to be confused with nsync... |
23:38 | Yep. If those have 090 for the biblionumber/biblioitemnumber, new records and mods will get the number changed back to 090. | |
23:39 | pianohacker | cait: that's a question I recently ran into, as some places use the marc mapping of the relevant framework and others just use the default |
23:39 | is there a best practice? | |
23:40 | (or at least an accepted one) | |
23:40 | I've been looking at circulation code, so my head hasn't exactly been in the clean, recently architected part of Koha lately :P | |
23:41 | cait | i think it doesn't seem to make sense fo rme for different frameworks having different mapping right now |
23:41 | there might be a use case i can't think of... | |
23:41 | you never know with libraries | |
23:41 | pianohacker | always is in the library world, yeah... |
23:42 | * dcook | supposes it might depend on the case as well |
23:42 | turns back to his Friday work before he starts thinking about other things.. | |
23:42 | dcook | But yay for pianohacker looking at circulation! :D |
23:42 | pianohacker | well, from a very unscientific ag -i marctokoha | grep -v framework | wc -l |
23:43 | about 80% of the instances use it | |
23:43 | eythian | what is 'ag'? |
23:43 | * dcook | was also wondering this |
23:44 | pianohacker | sorry, 80% of instances in Koha don't pass frameworkcode |
23:44 | jce | dcook: I don't find any reference to GetMarcFromKohaField in /usr/share/koha/intranet/cgi-bin/admin/koha2marclinks.pl Is it a variable name? A function? A Perl module? |
23:44 | dcook | jce: a function. It's referred to in rebuild_zebra.pl |
23:44 | pianohacker | dcook, eythian: It's a superfast recursive grep that automatically ignores certain directories |
23:44 | dcook | What you want to do is go to the page koha2marclinks.pl |
23:44 | eythian | pianohacker: you mean ack |
23:45 | ? | |
23:45 | pianohacker | eythian: nope, even faster than ack, though not quite as full featured |
23:45 | cait | jce: i think before thinking furhter, just try changing the data for 2 records or so and see if you can search them - might save you more headache, doing it step by step |
23:45 | jce | dcook: Ah, I was looking in the code. |
23:45 | eythian | pianohacker: ah, interesting |
23:46 | pianohacker | eythian: see http://geoff.greer.fm/2012/09/[…]g-my-own-scandir/ and http://geoff.greer.fm/2011/12/[…]-better-than-ack/ |
23:46 | dude is a bit performance crazy | |
23:46 | cait | night |
23:46 | :) | |
23:46 | cait left #koha | |
23:46 | pianohacker | night cait |
23:46 | rangi | jce: id just try editing a couple of records, reindexing and see if it fixes the search |
23:46 | jce: everything else is moot if it doesnt | |
23:47 | dcook | jce: Both cait and rangi have been doing this much longer than me, so I'd probably listen to them :p |
23:48 | rangi | once we are sure that shifting the 090 to 999 fixes it, then we can worry about the mapping/frameworks :) |
23:51 | mtj | ahh, this script might do the 090->999 fixup jce |
23:51 | http://git.kohaaloha.com/?p=mi[…]fix-090-to-999.pl | |
23:51 | jce | Ok, so I should edit a couple records by changing their 090 tags to 999, or copying them to 999 tags in marcxml, import those records, re-index Zebra, and see if I can search those records on Title, Author, etc. If that works, I should fix the framework and Koha to MARC mapping, mass-edit the rest of the marcxml records, import them, re-index, and there's a 100% guarantee that everything will be happy in the world. :) |
23:51 | mtj: Ooh, I may check into that. | |
23:51 | eythian | pianohacker: it's around half the speed of ack for me. |
23:51 | pianohacker | never a 100% guarantee |
23:51 | eythian | though perhaps there are differences in options |
23:52 | pianohacker | eythian: that's interesting. running a simple ag/ack MarcToKoha in my kohaclone is 0.47 vs 22.27 seconds |
23:52 | jce | pianohacker: :O shock and disbelief! |
23:52 | eythian | pianohacker: make sure you have a warm cache though |
23:52 | pianohacker | (on my machine) |
23:53 | eythian | i.e. run each one twice ant least |
23:53 | pianohacker | jce: heh. Just making sure :) |
23:54 | eythian: Just did, same result. What test are you running/are you on an SSD? | |
23:54 | jce | May have to take a supper break before diving into this. Thanks all for your help. |
23:54 | eythian | > time ack GetMarcFromKohaField |
23:54 | not on SSD | |
23:54 | mtj | jce: you are making good progress :) |
23:55 | eythian | pianohacker: OK, when I make a .agignore from my .ackrc, it's a huge lot faster. |
23:55 | jce | mtj: I do think this sounds like a plausible explanation for what is going on. |
23:56 | eythian | I had ack tuned to drop a large amount of unnecessary Koha directories |
23:56 | mtj | jce, yeah - ive travelled your path a few times, myself :) |
23:57 | rangi | pianohacker: you did the overdrive work eh? |
23:57 | mtj | a well worn road |
23:58 | pianohacker | eythian: Ah, yeah, even just making ack ignore .po files like I did for ag reduced the difference a lot |
23:58 | rangi: Yup. That reminds me, though | |
23:58 | rangi | pianohacker: http://ils.stdc.govt.nz/cgi-bi[…]?q=harry%20potter |
23:59 | pianohacker | gmcharlt: do you have a second to discuss a dependency problem that affects 3.16? |
23:59 | rangi: cool, didn't know NZ libraries used overdrive |
← Previous day | Today | Next day → | Search | Index