1 package Koha
::MetaSearcher
;
3 # Copyright 2014 ByWater Solutions
5 # This file is part of Koha.
7 # Koha is free software; you can redistribute it and/or modify it
8 # under the terms of the GNU General Public License as published by
9 # the Free Software Foundation; either version 3 of the License, or
10 # (at your option) any later version.
12 # Koha is distributed in the hope that it will be useful, but
13 # WITHOUT ANY WARRANTY; without even the implied warranty of
14 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
15 # GNU General Public License for more details.
17 # You should have received a copy of the GNU General Public License
18 # along with Koha; if not, see <http://www.gnu.org/licenses>.
22 use base
'Class::Accessor';
24 use C4
::Charset
qw( MarcToUTF8Record );
25 use C4
::Search
qw(); # Purely for new_record_from_zebra
26 use DBIx
::Class
::ResultClass
::HashRefInflator
;
30 use Koha
::MetadataRecord
;
32 use Storable
qw( store_fd fd_retrieve );
33 use Time
::HiRes
qw( clock_gettime CLOCK_MONOTONIC );
39 __PACKAGE__
->mk_accessors( qw( fetch offset on_error resultset ) );
42 my ( $class, $options ) = @_;
44 my ( $uuid, $uuidstring );
45 UUID
::generate
($uuid);
46 UUID
::unparse
( $uuid, $uuidstring );
53 resultset
=> $uuidstring,
59 my ( $self, $index, $server, $marcrecord ) = @_;
61 my $record = Koha
::MetadataRecord
->new( { schema
=> 'marc', record
=> $marcrecord } );
64 title
=> 'biblio.title',
65 seriestitle
=> 'biblio.seriestitle',
66 author
=> 'biblio.author',
67 isbn
=>'biblioitems.isbn',
68 issn
=>'biblioitems.issn',
69 lccn
=>'biblioitems.lccn', #LC control number (not call number)
70 edition
=>'biblioitems.editionstatement',
71 date
=> 'biblio.copyrightdate', #MARC21
72 date2
=> 'biblioitems.publicationyear', #UNIMARC
76 while ( my ( $key, $kohafield ) = each %fetch ) {
77 $metadata->{$key} = $record->getKohaField($kohafield);
79 $metadata->{date
} //= $metadata->{date2
};
81 push @
{ $self->{results
} }, {
84 record
=> $marcrecord,
85 metadata
=> $metadata,
90 my ( $self, $server_ids, $query ) = @_;
92 my $resultset_expiry = 300;
95 eval { $cache = Koha
::Cache
->new(); };
96 my $schema = Koha
::Database
->new->schema;
99 map { $_ => 0 } @
$server_ids
102 map { $_ => 0 } @
$server_ids
107 my $start = clock_gettime
( CLOCK_MONOTONIC
);
108 my $select = IO
::Select
->new;
113 foreach my $server_id ( @
$server_ids ) {
114 if ( $server_id =~ /^\d+$/ ) {
116 my $server = $schema->resultset('Z3950server')->find(
117 { id
=> $server_id },
118 { result_class
=> 'DBIx::Class::ResultClass::HashRefInflator' },
120 $server->{type
} = 'z3950';
122 push @servers, $server;
123 } elsif ( $server_id =~ /(\w+)(?::(\w+))?/ ) {
135 # HashRefInflator is used so that the information will survive into the fork
136 foreach my $server ( @servers ) {
138 my $set = $cache->get_from_cache( 'z3950-resultset-' . $self->resultset . '-' . $server->{id
} );
139 if ( ref($set) eq 'HASH' ) {
140 $set->{server
} = $server;
141 push @cached_sets, $set;
146 $select->add( $self->_start_worker( $server, $query ) );
149 # Handle these while the servers are searching
150 foreach my $set ( @cached_sets ) {
151 $self->_handle_hits( $stats, $set );
154 while ( $select->count ) {
155 foreach my $readfh ( $select->can_read() ) {
156 my $result = fd_retrieve
( $readfh );
158 $select->remove( $readfh );
162 next if ( ref $result ne 'HASH' );
164 if ( $result->{error
} ) {
165 $self->{on_error
}->( $result->{server
}, $result->{error
} );
169 $self->_handle_hits( $stats, $result );
172 $cache->set_in_cache( 'z3950-resultset-' . $self->resultset . '-' . $result->{server
}->{id
}, {
173 hits
=> $result->{hits
},
174 num_fetched
=> $result->{num_fetched
},
175 num_hits
=> $result->{num_hits
},
176 }, $resultset_expiry );
181 $stats->{time} = clock_gettime
( CLOCK_MONOTONIC
) - $start;
187 my ( $self, $server, $query ) = @_;
188 pipe my $readfh, my $writefh;
190 # Accessing the cache or Koha database after the fork is risky, so get any resources we need
193 my $marcflavour = C4
::Context
->preference('marcflavour');
195 if ( ( $pid = fork ) ) {
200 } elsif ( !defined $pid ) {
203 $self->{on_error
}->( $server, 'Failed to fork' );
209 my ( $num_hits, $num_fetched, $hits, $results );
212 if ( $server->{type
} eq 'z3950' ) {
213 my $zoptions = ZOOM
::Options
->new();
214 $zoptions->option( 'elementSetName', 'F' );
215 $zoptions->option( 'databaseName', $server->{db
} );
216 $zoptions->option( 'user', $server->{userid
} ) if $server->{userid
};
217 $zoptions->option( 'password', $server->{password
} ) if $server->{password
};
218 $zoptions->option( 'preferredRecordSyntax', $server->{syntax
} );
219 $zoptions->option( 'timeout', $server->{timeout
} ) if $server->{timeout
};
221 $connection = ZOOM
::Connection
->create($zoptions);
223 $connection->connect( $server->{host
}, $server->{port
} );
224 $results = $connection->search_pqf( $query ); # Starts the search
225 } elsif ( $server->{type
} eq 'koha' ) {
226 $connection = C4
::Context
->Zconn( $server->{extra
} );
227 $results = $connection->search_pqf( $query ); # Starts the search
228 } elsif ( $server->{type
} eq 'batch' ) {
229 $server->{encoding
} = 'utf-8';
234 error
=> $connection ?
$connection->exception() : $@
,
240 if ( $server->{type
} eq 'batch' ) {
241 # TODO: actually handle PQF
242 $query =~ s/@\w+ (?:\d+=\d+ )?//g;
245 my $schema = Koha
::Database
->new->schema;
246 $schema->storage->debug(1);
247 my $match_condition = [ map +{ -like
=> '%' . $_ . '%' }, split( /\s+/, $query ) ];
248 $hits = [ $schema->resultset('ImportRecord')->search(
250 import_batch_id
=> $server->{extra
},
252 { 'import_biblios.title' => $match_condition },
253 { 'import_biblios.author' => $match_condition },
254 { 'import_biblios.isbn' => $match_condition },
255 { 'import_biblios.issn' => $match_condition },
259 join => [ qw( import_biblios ) ],
260 rows
=> $self->{fetch
},
262 )->get_column( 'marc' )->all ];
264 $num_hits = $num_fetched = scalar @
$hits;
266 $num_hits = $results->size;
267 $num_fetched = ( $self->{offset
} + $self->{fetch
} ) < $num_hits ?
$self->{fetch
} : $num_hits;
269 $hits = [ map { $_->raw() } @
{ $results->records( $self->{offset
}, $num_fetched, 1 ) } ];
272 if ( !@
$hits && $connection && $connection->exception() ) {
274 error
=> $connection->exception(),
280 if ( $server->{type
} eq 'koha' ) {
281 $hits = [ map { C4
::Search
::new_record_from_zebra
( $server->{extra
}, $_ ) } @
$hits ];
283 $hits = [ map { $self->_import_record( $_, $marcflavour, $server->{encoding
} ?
$server->{encoding
} : "iso-5426" ) } @
$hits ];
288 num_fetched
=> $num_fetched,
289 num_hits
=> $num_hits,
297 my ( $self, $raw, $marcflavour, $encoding ) = @_;
299 my ( $marcrecord ) = MarcToUTF8Record
( $raw, $marcflavour, $encoding ); #ignores charset return values
305 my ( $self, $stats, $set ) = @_;
307 my $server = $set->{server
};
309 my $num_hits = $stats->{num_hits
}->{ $server->{id
} } = $set->{num_hits
};
310 my $num_fetched = $stats->{num_fetched
}->{ $server->{id
} } = $set->{num_fetched
};
312 $stats->{total_hits
} += $num_hits;
313 $stats->{total_fetched
} += $num_fetched;
315 foreach my $j ( 0..$#{ $set->{hits} } ) {
316 $self->handle_hit( $self->{offset
} + $j, $server, $set->{hits
}->[$j] );
321 my ( $self, $key, $direction ) = @_;
323 my $empty_flip = -1; # Determines the flip of ordering for records with empty sort keys.
325 foreach my $hit ( @
{ $self->{results
} } ) {
326 ( $hit->{sort_key
} = $hit->{metadata
}->{$key} || '' ) =~ s/\W//g;
329 $self->{results
} = [ sort {
330 # Sort empty records at the end
331 return -$empty_flip unless $a->{sort_key
};
332 return $empty_flip unless $b->{sort_key
};
334 $direction * ( $a->{sort_key
} cmp $b->{sort_key
} );
335 } @
{ $self->{results
} } ];
339 my ( $self, $offset, $length ) = @_;
343 foreach my $i ( $offset..( $offset + $length - 1 ) ) {
344 push @subset, $self->{results
}->[$i] if $self->{results
}->[$i];