← Previous day | Today | Next day → | Search | Index
All times shown according to UTC.
| Time | Nick | Message |
|---|---|---|
| 00:08 | chris_n joined #koha | |
| 00:11 | kathryn joined #koha | |
| 02:11 | kathryn joined #koha | |
| 03:09 | KSorbo joined #koha | |
| 05:42 | fridolin joined #koha | |
| 05:43 | fridolin | hi there |
| 05:51 | caroline joined #koha | |
| 05:52 | caroline | hello #koha! |
| 05:53 | is anybody around? | |
| 05:53 | What is the active time zone right now? | |
| 05:53 | NZ? | |
| 05:57 | cait joined #koha | |
| 06:00 | cait1 joined #koha | |
| 06:01 | KeithS joined #koha | |
| 06:08 | laurence joined #koha | |
| 06:14 | cait2 joined #koha | |
| 06:17 | cait2 left #koha | |
| 06:21 | z0k joined #koha | |
| 06:25 | alex_a joined #koha | |
| 06:26 | alex_a | bonjour |
| 06:29 | reiveune joined #koha | |
| 06:29 | reiveune | hello |
| 06:49 | paul_p joined #koha | |
| 06:55 | andrew_ joined #koha | |
| 07:26 | gaetan_B joined #koha | |
| 07:27 | gaetan_B | hello |
| 07:30 | sophie_m joined #koha | |
| 07:30 | caroline left #koha | |
| 07:44 | ashimema | guten morgen #koha |
| 08:01 | z0k joined #koha | |
| 08:13 | sophie_m left #koha | |
| 08:18 | cait joined #koha | |
| 08:27 | irma_ joined #koha | |
| 08:27 | LeeJ joined #koha | |
| 08:30 | paul_p joined #koha | |
| 08:42 | cait1 joined #koha | |
| 08:44 | alexbuckley joined #koha | |
| 08:58 | cait2 joined #koha | |
| 09:18 | fridolin left #koha | |
| 09:19 | fridolin joined #koha | |
| 09:24 | KeithS joined #koha | |
| 09:25 | KSorbo joined #koha | |
| 09:50 | alex_a | kidclamp, around? |
| 09:50 | * kidclamp | waves |
| 09:50 | alex_a | \o |
| 09:51 | kidclamp, bug 18213 does not apply. Probably because 18235 has been reverted | |
| 09:51 | huginn` | Bug http://bugs.koha-community.org[…]_bug.cgi?id=18213 enhancement, P5 - low, ---, nick, Signed Off , Add language facets to Elasticsearch |
| 09:51 | alex_a | What is the plan? Waiting for 18235 being pushed again? |
| 09:52 | Or i can try a rebase (if your are ok) | |
| 09:52 | kidclamp | I think you mean 20073? 18235 was never pushed |
| 09:52 | yeah, go ahaed, I can also take a look if you wantr | |
| 09:52 | alex_a | mmh |
| 10:03 | barneyjobs joined #koha | |
| 10:03 | barneyjobs | hello po |
| 10:04 | z0k joined #koha | |
| 10:04 | barneyjobs | hello? |
| 10:49 | kidclamp joined #koha | |
| 11:07 | cait joined #koha | |
| 11:09 | cait1 joined #koha | |
| 11:23 | Dyrcona joined #koha | |
| 11:56 | ncbaratta joined #koha | |
| 12:05 | z0k joined #koha | |
| 12:10 | Hugo joined #koha | |
| 12:10 | Hugo | morning |
| 12:10 | I have this error at zebra zebrasrv(1617) [log] dict_lookup_grep: (\x01}\x01\x04)\x01\x14\"\x01\x1C\x01\x1A\x01\x12\x01\x01\x01\x09\x01\x01\x01\x09 | |
| 12:10 | any idea what it could be? | |
| 12:49 | ashimema | whose the best person to ask about the oai pmh code these days? |
| 12:53 | Joubu | hi |
| 12:53 | ashimema: shoot the question and you will see | |
| 12:54 | ashimema | haha.. wasn't a question.. |
| 12:54 | I was going to ask them to take a look at bug 20271 from the OAI perspective ;) | |
| 12:54 | huginn` | Bug http://bugs.koha-community.org[…]_bug.cgi?id=20271 major, P1 - high, ---, oha, Failed QA , Merge deleted* tables with their "alive" cousins |
| 12:55 | ashimema | as I've not dabbled there at all I don't feel qualified to comment on what the syspref should be changed to say (or whether it is indeed still relevant with deletedbiblio table going away 😉) |
| 12:58 | tcohen | ashimema: OAI-PMH repositories need to specify how they deal with deleted records, which yields different harvsting strategies. I think we should keep it configurable to keep the current behaviour. If that's what you're asking |
| 13:00 | ashimema | I agree with keeping it.. |
| 13:01 | but the pref description needs updating at the very least to not reference deletedbiblio as a table ;) | |
| 13:06 | with that very minor comment I think that bug is pretty much good to go.. I'm just running it through the test suit a few times | |
| 13:16 | irma_ joined #koha | |
| 13:17 | KeithS joined #koha | |
| 13:18 | KSorbo joined #koha | |
| 13:20 | kmlussier joined #koha | |
| 13:21 | tcohen | ashimema: great job! |
| 13:25 | ashimema | mercelr did all the hard work really.. ;) |
| 13:25 | I'm jsut giving it the second eye ;) | |
| 13:25 | I'm so going to regret passing qa on it soon enough though aren't I.. bet it's going to lead to all sorts of backporting fun this cycle ;) | |
| 13:35 | tcohen | it needs to happen at some point anyway, better sooner than later |
| 13:39 | ashimema | indeed |
| 13:40 | * LeeJ | waves |
| 13:40 | LeeJ | hi #koha |
| 13:40 | even though I technically never left :) | |
| 13:46 | ashimema | tcohen.. am I right in thinking atomic updates should always attempt to be idempotent these days? |
| 13:46 | Joubu | yes |
| 13:46 | ashimema | i.e. the update in bug 20271 should check for the existence of the deleted* tables before attempting to migrate and delete them |
| 13:46 | huginn` | Bug http://bugs.koha-community.org[…]_bug.cgi?id=20271 major, P1 - high, ---, oha, Failed QA , Merge deleted* tables with their "alive" cousins |
| 13:46 | Joubu | s/should/must |
| 13:47 | ashimema | or is that something the RMaints, RManagers expected to do as part of the RM process |
| 13:47 | okies.. | |
| 13:47 | thanks Joubu | |
| 13:47 | Joubu | there is a TableExist sub in updatedatabase.pl |
| 13:47 | ashimema | :) |
| 13:47 | Joubu | TableExists even |
| 13:48 | ashimema | brill |
| 13:54 | tuxayo | Hi :) |
| 13:54 | Those who use `delete_patrons.pl` , how do you blacklist borrowers categories? (to avoid collateral damages on "strange" accounts) Do you call `delete_patrons.pl` multiple times with each `--category_code` that needs cleaning? | |
| 14:55 | Joubu | Thanks ashimema for your quick signoffs! |
| 14:56 | ashimema | haha.. no worries.. they were nice easy one's to test ;) |
| 14:56 | and you caught me on one of my community days ;) | |
| 15:01 | ncbaratta joined #koha | |
| 15:06 | reiveune | bye |
| 15:06 | reiveune left #koha | |
| 15:24 | lavamind | another trivial patch https://bugs.koha-community.or[…]_bug.cgi?id=20920 |
| 15:24 | huginn` | Bug 20920: normal, P5 - low, ---, koha-bugs, Needs Signoff , Plack timeout because of missing CGI::Compile Perl dependency |
| 15:25 | ncbaratta joined #koha | |
| 15:31 | indradg joined #koha | |
| 15:50 | z0k joined #koha | |
| 15:57 | lisettelatah joined #koha | |
| 16:11 | fridolin left #koha | |
| 16:35 | cait joined #koha | |
| 16:39 | cait1 joined #koha | |
| 16:41 | indradg joined #koha | |
| 17:20 | lavamind | Joubu: once I installed libcgip-compile-perl, I restarted plack and it started to work (intranet) |
| 17:21 | Joubu | lavamind: ok hopefully Mirko will take a look soon |
| 17:22 | lavamind | running 'aptitude why libcgi-compile-perl' : i A libplack-perl Recommends libcgi-compile-perl |
| 17:22 | so if the goal if to make the package hard-depend on plack, it should be added as a Dependency | |
| 17:23 | otherwise, it should be documented | |
| 17:25 | Joubu | lavamind: yes sure. But it's what my jessie install tells too. No idea why it has not be reported before |
| 17:25 | been* | |
| 17:25 | lavamind | Joubu: well recommends are installed by default |
| 17:26 | I think I may have run apt --no-install-recommends at some point | |
| 17:27 | ideally there would be a seperate package eg. koha-common-plack with pulls in all that stuff | |
| 17:28 | I'm not a fan of the monolithic "one package for everything" | |
| 17:34 | laurence joined #koha | |
| 17:35 | Joubu | lavamind: if you have time and knowledge, feel free to help. We clearly lack people on this side |
| 17:35 | laurence left #koha | |
| 17:35 | lavamind | Joubu: well I'm a DD :P |
| 17:36 | but time is more or less lacking these days | |
| 17:36 | kidclamp | cait++ |
| 17:36 | lavamind | eventually it would be best to upload these packages to the main archive |
| 17:37 | I think with a little work it could be made to be policy compliant | |
| 17:37 | the difficult part is ensuring a smooth transition | |
| 17:38 | it owuld probably help if there were sprints organized to work on this | |
| 17:38 | I might be able to attend during $work hours | |
| 17:39 | (Eastern timezone) | |
| 17:39 | well, north-america eastern :P | |
| 17:39 | the puppet module is about 75% done (for a 1.0 release) | |
| 17:40 | Joubu | lavamind: you could start a discussion on koha-devel, at least to tell people you are willing to help and explain what you have in mind/how things can be enhanced |
| 17:40 | lavamind | Joubu: good idea, I'll put it in my TODO list |
| 17:43 | Joubu | lavamind: tcohen (Tomas), droft (Mirko) and mtj (Mason) are the ones who knows more about this topic |
| 17:46 | lavamind | Joubu: thanks |
| 19:10 | when creating an instance with a default database, is it normal to require "koha-rebuild-zebra -f" ? | |
| 19:11 | my instance gets created normally, and the indexer daemon is started, but the index seems empty | |
| 19:11 | (searching doesn't work) | |
| 19:11 | once I run the full rebuild, search starts working | |
| 19:15 | Joubu | the indexer daemon is watching the zebraqueue table, which is filled on import/create/delete/update of biblio or auth records |
| 19:15 | so yes it's expected to launch a full reindex if your record tables already contain something | |
| 19:17 | lavamind | Joubu: hrmm |
| 19:19 | in the default.sql.gz I use when creating my instance, zebraqueue is empty | |
| 19:21 | my biblio table is full of records however | |
| 19:21 | I think the way my default SQL is set up, it may not trigger a full reindex | |
| 19:21 | I'm not sure how I can trigger it | |
| 19:28 | Joubu | koha-rebuild-zebra -f will trigger a full reindex |
| 19:32 | lavamind | Joubu: yeah, I know, I'm wondering whether I can shape my default.sql.gz to avoid having to run it |
| 19:35 | Joubu | you could fill zebraqueue |
| 19:38 | lavamind | yeah, however that seems like more of a hassle, especially since I'll have to update the sql from time to time |
| 19:38 | each time I'd have to regenerate those records | |
| 19:41 | Joubu | the easiest method is to add the rebuild -f command after you inserted your dump |
| 19:52 | lavamind | agreed |
| 19:53 | magnuse joined #koha | |
| 19:54 | magnuse | anyone got a hunch about the max number of records the stage/import tools can handle? |
| 19:55 | ..per file, i mean | |
| 19:57 | i have a file with 33869 records | |
| 20:25 | indradg joined #koha | |
| 20:29 | indradg_ joined #koha | |
| 20:35 | lavamind | where can I find a description of the database tables |
| 20:36 | I have a bunch of import_* tables with some data in it, I'm not sure I need it in there | |
| 20:44 | rangi | lavamind: http://schema.koha-community.org/ |
| 20:44 | lavamind | rangi: thank you |
| 20:44 | I also found https://github.com/Koha-Commun[…]eanup_database.pl | |
| 20:45 | I think import_* doesn't have much use in a default sql | |
| 20:45 | since the cleanup cronjob remove it anyway, no sense in putting it in the database in the first place | |
| 20:45 | rangi | hmm? |
| 20:45 | the cron job only removes data from it | |
| 20:46 | not the tables themselves | |
| 20:47 | lavamind | yes of course, only the data hehe |
| 20:47 | for example there are many many records in the import_record table | |
| 20:47 | I don't think these are used for anything inside Koha ? | |
| 20:47 | rangi | yes they are |
| 20:48 | acquisitions | |
| 20:48 | and staged marc record management (under tools) | |
| 20:48 | use them | |
| 20:49 | its where records sit outside of the catalogue, that you can either bring into the catalogue via the staged marc record management, or via acquisitions | |
| 20:49 | also if you search in cataloguing | |
| 20:49 | * lavamind | looks |
| 20:50 | rangi | you will find them too |
| 20:50 | (cataloguing, not search the catalogue ;)) | |
| 20:50 | it's called the reservoir there | |
| 20:51 | * lavamind | is just the koha admin, not a koha user |
| 20:51 | rangi | whenever you do a z3950 search, the results are stored in there too |
| 20:51 | but you are probably ok to wipe them out on a fresh koha | |
| 20:52 | lavamind | well it's not supposed to be a fresh koha |
| 20:52 | rangi | one acquisition workflow (that your library probably doesn't use) is that book sellers give your the catalogue data for their inventory |
| 20:52 | you can stage that in tools | |
| 20:52 | lavamind | there's supposed to be a bunch of borrowers, items in the catalogue, etc |
| 20:52 | rangi | then you can order from that acquisitions |
| 20:52 | yep the import_ stuff sits outside the catalogue | |
| 20:52 | lavamind | it's for classes, so that each student gets their own koha instance from a template |
| 20:53 | rangi | librarians can see it, other users can't |
| 20:53 | lavamind | well one of the classes teaches acquisitions |
| 20:53 | rangi | they could always load a marc file via stage marc records to test that |
| 20:53 | that would be testing the whole workflow | |
| 20:53 | lavamind | I'm not sure if they do it, probably though, I guess I should find out |
| 20:55 | rangi | if its for teaching you could zero those tables out. THey can populate them by using the stage marc records if they need |
| 20:57 | lavamind | I'm pretty sure that they do, because they've never complained even though the cleanup script deletes them |
| 20:57 | the imports were recorded as being made 7 years ago | |
| 20:57 | rangi | yeah you wont need them then |
| 20:57 | lavamind | ack |
| 21:13 | kathryn joined #koha | |
| 21:16 | rangi joined #koha | |
| 21:19 | alexbuckley joined #koha | |
| 21:20 | rangi joined #koha | |
| 21:21 | wahanui joined #koha | |
| 21:22 | logbot joined #koha | |
| 21:23 | pastebot joined #koha | |
| 22:06 | cait joined #koha | |
| 22:47 | lukeG left #koha |
← Previous day | Today | Next day → | Search | Index