WHATSNEW: add SmartCard/PKINIT improvements
[Samba.git] / ctdb / server / ctdb_daemon.c
blob816766796b4e4aaee0b8a6d0f2b27715faedce79
1 /*
2 ctdb daemon code
4 Copyright (C) Andrew Tridgell 2006
6 This program is free software; you can redistribute it and/or modify
7 it under the terms of the GNU General Public License as published by
8 the Free Software Foundation; either version 3 of the License, or
9 (at your option) any later version.
11 This program is distributed in the hope that it will be useful,
12 but WITHOUT ANY WARRANTY; without even the implied warranty of
13 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
14 GNU General Public License for more details.
16 You should have received a copy of the GNU General Public License
17 along with this program; if not, see <http://www.gnu.org/licenses/>.
20 #include "replace.h"
21 #include "system/network.h"
22 #include "system/filesys.h"
23 #include "system/wait.h"
24 #include "system/time.h"
26 #include <talloc.h>
27 /* Allow use of deprecated function tevent_loop_allow_nesting() */
28 #define TEVENT_DEPRECATED
29 #include <tevent.h>
30 #include <tdb.h>
32 #include "lib/tdb_wrap/tdb_wrap.h"
33 #include "lib/util/dlinklist.h"
34 #include "lib/util/debug.h"
35 #include "lib/util/samba_util.h"
36 #include "lib/util/blocking.h"
38 #include "ctdb_version.h"
39 #include "ctdb_private.h"
40 #include "ctdb_client.h"
42 #include "common/rb_tree.h"
43 #include "common/reqid.h"
44 #include "common/system.h"
45 #include "common/common.h"
46 #include "common/logging.h"
48 struct ctdb_client_pid_list {
49 struct ctdb_client_pid_list *next, *prev;
50 struct ctdb_context *ctdb;
51 pid_t pid;
52 struct ctdb_client *client;
55 const char *ctdbd_pidfile = NULL;
57 static void daemon_incoming_packet(void *, struct ctdb_req_header *);
59 static void print_exit_message(void)
61 if (debug_extra != NULL && debug_extra[0] != '\0') {
62 DEBUG(DEBUG_NOTICE,("CTDB %s shutting down\n", debug_extra));
63 } else {
64 DEBUG(DEBUG_NOTICE,("CTDB daemon shutting down\n"));
66 /* Wait a second to allow pending log messages to be flushed */
67 sleep(1);
73 static void ctdb_time_tick(struct tevent_context *ev, struct tevent_timer *te,
74 struct timeval t, void *private_data)
76 struct ctdb_context *ctdb = talloc_get_type(private_data, struct ctdb_context);
78 if (getpid() != ctdb->ctdbd_pid) {
79 return;
82 tevent_add_timer(ctdb->ev, ctdb,
83 timeval_current_ofs(1, 0),
84 ctdb_time_tick, ctdb);
87 /* Used to trigger a dummy event once per second, to make
88 * detection of hangs more reliable.
90 static void ctdb_start_time_tickd(struct ctdb_context *ctdb)
92 tevent_add_timer(ctdb->ev, ctdb,
93 timeval_current_ofs(1, 0),
94 ctdb_time_tick, ctdb);
97 static void ctdb_start_periodic_events(struct ctdb_context *ctdb)
99 /* start monitoring for connected/disconnected nodes */
100 ctdb_start_keepalive(ctdb);
102 /* start periodic update of tcp tickle lists */
103 ctdb_start_tcp_tickle_update(ctdb);
105 /* start listening for recovery daemon pings */
106 ctdb_control_recd_ping(ctdb);
108 /* start listening to timer ticks */
109 ctdb_start_time_tickd(ctdb);
112 static void ignore_signal(int signum)
114 struct sigaction act;
116 memset(&act, 0, sizeof(act));
118 act.sa_handler = SIG_IGN;
119 sigemptyset(&act.sa_mask);
120 sigaddset(&act.sa_mask, signum);
121 sigaction(signum, &act, NULL);
126 send a packet to a client
128 static int daemon_queue_send(struct ctdb_client *client, struct ctdb_req_header *hdr)
130 CTDB_INCREMENT_STAT(client->ctdb, client_packets_sent);
131 if (hdr->operation == CTDB_REQ_MESSAGE) {
132 if (ctdb_queue_length(client->queue) > client->ctdb->tunable.max_queue_depth_drop_msg) {
133 DEBUG(DEBUG_ERR,("CTDB_REQ_MESSAGE queue full - killing client connection.\n"));
134 talloc_free(client);
135 return -1;
138 return ctdb_queue_send(client->queue, (uint8_t *)hdr, hdr->length);
142 message handler for when we are in daemon mode. This redirects the message
143 to the right client
145 static void daemon_message_handler(uint64_t srvid, TDB_DATA data,
146 void *private_data)
148 struct ctdb_client *client = talloc_get_type(private_data, struct ctdb_client);
149 struct ctdb_req_message_old *r;
150 int len;
152 /* construct a message to send to the client containing the data */
153 len = offsetof(struct ctdb_req_message_old, data) + data.dsize;
154 r = ctdbd_allocate_pkt(client->ctdb, client->ctdb, CTDB_REQ_MESSAGE,
155 len, struct ctdb_req_message_old);
156 CTDB_NO_MEMORY_VOID(client->ctdb, r);
158 talloc_set_name_const(r, "req_message packet");
160 r->srvid = srvid;
161 r->datalen = data.dsize;
162 memcpy(&r->data[0], data.dptr, data.dsize);
164 daemon_queue_send(client, &r->hdr);
166 talloc_free(r);
170 this is called when the ctdb daemon received a ctdb request to
171 set the srvid from the client
173 int daemon_register_message_handler(struct ctdb_context *ctdb, uint32_t client_id, uint64_t srvid)
175 struct ctdb_client *client = reqid_find(ctdb->idr, client_id, struct ctdb_client);
176 int res;
177 if (client == NULL) {
178 DEBUG(DEBUG_ERR,("Bad client_id in daemon_request_register_message_handler\n"));
179 return -1;
181 res = srvid_register(ctdb->srv, client, srvid, daemon_message_handler,
182 client);
183 if (res != 0) {
184 DEBUG(DEBUG_ERR,(__location__ " Failed to register handler %llu in daemon\n",
185 (unsigned long long)srvid));
186 } else {
187 DEBUG(DEBUG_INFO,(__location__ " Registered message handler for srvid=%llu\n",
188 (unsigned long long)srvid));
191 return res;
195 this is called when the ctdb daemon received a ctdb request to
196 remove a srvid from the client
198 int daemon_deregister_message_handler(struct ctdb_context *ctdb, uint32_t client_id, uint64_t srvid)
200 struct ctdb_client *client = reqid_find(ctdb->idr, client_id, struct ctdb_client);
201 if (client == NULL) {
202 DEBUG(DEBUG_ERR,("Bad client_id in daemon_request_deregister_message_handler\n"));
203 return -1;
205 return srvid_deregister(ctdb->srv, srvid, client);
208 int daemon_check_srvids(struct ctdb_context *ctdb, TDB_DATA indata,
209 TDB_DATA *outdata)
211 uint64_t *ids;
212 int i, num_ids;
213 uint8_t *results;
215 if ((indata.dsize % sizeof(uint64_t)) != 0) {
216 DEBUG(DEBUG_ERR, ("Bad indata in daemon_check_srvids, "
217 "size=%d\n", (int)indata.dsize));
218 return -1;
221 ids = (uint64_t *)indata.dptr;
222 num_ids = indata.dsize / 8;
224 results = talloc_zero_array(outdata, uint8_t, (num_ids+7)/8);
225 if (results == NULL) {
226 DEBUG(DEBUG_ERR, ("talloc failed in daemon_check_srvids\n"));
227 return -1;
229 for (i=0; i<num_ids; i++) {
230 if (srvid_exists(ctdb->srv, ids[i]) == 0) {
231 results[i/8] |= (1 << (i%8));
234 outdata->dptr = (uint8_t *)results;
235 outdata->dsize = talloc_get_size(results);
236 return 0;
240 destroy a ctdb_client
242 static int ctdb_client_destructor(struct ctdb_client *client)
244 struct ctdb_db_context *ctdb_db;
246 ctdb_takeover_client_destructor_hook(client);
247 reqid_remove(client->ctdb->idr, client->client_id);
248 client->ctdb->num_clients--;
250 if (client->num_persistent_updates != 0) {
251 DEBUG(DEBUG_ERR,(__location__ " Client disconnecting with %u persistent updates in flight. Starting recovery\n", client->num_persistent_updates));
252 client->ctdb->recovery_mode = CTDB_RECOVERY_ACTIVE;
254 ctdb_db = find_ctdb_db(client->ctdb, client->db_id);
255 if (ctdb_db) {
256 DEBUG(DEBUG_ERR, (__location__ " client exit while transaction "
257 "commit active. Forcing recovery.\n"));
258 client->ctdb->recovery_mode = CTDB_RECOVERY_ACTIVE;
261 * trans3 transaction state:
263 * The destructor sets the pointer to NULL.
265 talloc_free(ctdb_db->persistent_state);
268 return 0;
273 this is called when the ctdb daemon received a ctdb request message
274 from a local client over the unix domain socket
276 static void daemon_request_message_from_client(struct ctdb_client *client,
277 struct ctdb_req_message_old *c)
279 TDB_DATA data;
280 int res;
282 if (c->hdr.destnode == CTDB_CURRENT_NODE) {
283 c->hdr.destnode = ctdb_get_pnn(client->ctdb);
286 /* maybe the message is for another client on this node */
287 if (ctdb_get_pnn(client->ctdb)==c->hdr.destnode) {
288 ctdb_request_message(client->ctdb, (struct ctdb_req_header *)c);
289 return;
292 /* its for a remote node */
293 data.dptr = &c->data[0];
294 data.dsize = c->datalen;
295 res = ctdb_daemon_send_message(client->ctdb, c->hdr.destnode,
296 c->srvid, data);
297 if (res != 0) {
298 DEBUG(DEBUG_ERR,(__location__ " Failed to send message to remote node %u\n",
299 c->hdr.destnode));
304 struct daemon_call_state {
305 struct ctdb_client *client;
306 uint32_t reqid;
307 struct ctdb_call *call;
308 struct timeval start_time;
310 /* readonly request ? */
311 uint32_t readonly_fetch;
312 uint32_t client_callid;
316 complete a call from a client
318 static void daemon_call_from_client_callback(struct ctdb_call_state *state)
320 struct daemon_call_state *dstate = talloc_get_type(state->async.private_data,
321 struct daemon_call_state);
322 struct ctdb_reply_call_old *r;
323 int res;
324 uint32_t length;
325 struct ctdb_client *client = dstate->client;
326 struct ctdb_db_context *ctdb_db = state->ctdb_db;
328 talloc_steal(client, dstate);
329 talloc_steal(dstate, dstate->call);
331 res = ctdb_daemon_call_recv(state, dstate->call);
332 if (res != 0) {
333 DEBUG(DEBUG_ERR, (__location__ " ctdbd_call_recv() returned error\n"));
334 CTDB_DECREMENT_STAT(client->ctdb, pending_calls);
336 CTDB_UPDATE_LATENCY(client->ctdb, ctdb_db, "call_from_client_cb 1", call_latency, dstate->start_time);
337 return;
340 length = offsetof(struct ctdb_reply_call_old, data) + dstate->call->reply_data.dsize;
341 /* If the client asked for readonly FETCH, we remapped this to
342 FETCH_WITH_HEADER when calling the daemon. So we must
343 strip the extra header off the reply data before passing
344 it back to the client.
346 if (dstate->readonly_fetch
347 && dstate->client_callid == CTDB_FETCH_FUNC) {
348 length -= sizeof(struct ctdb_ltdb_header);
351 r = ctdbd_allocate_pkt(client->ctdb, dstate, CTDB_REPLY_CALL,
352 length, struct ctdb_reply_call_old);
353 if (r == NULL) {
354 DEBUG(DEBUG_ERR, (__location__ " Failed to allocate reply_call in ctdb daemon\n"));
355 CTDB_DECREMENT_STAT(client->ctdb, pending_calls);
356 CTDB_UPDATE_LATENCY(client->ctdb, ctdb_db, "call_from_client_cb 2", call_latency, dstate->start_time);
357 return;
359 r->hdr.reqid = dstate->reqid;
360 r->status = dstate->call->status;
362 if (dstate->readonly_fetch
363 && dstate->client_callid == CTDB_FETCH_FUNC) {
364 /* client only asked for a FETCH so we must strip off
365 the extra ctdb_ltdb header
367 r->datalen = dstate->call->reply_data.dsize - sizeof(struct ctdb_ltdb_header);
368 memcpy(&r->data[0], dstate->call->reply_data.dptr + sizeof(struct ctdb_ltdb_header), r->datalen);
369 } else {
370 r->datalen = dstate->call->reply_data.dsize;
371 memcpy(&r->data[0], dstate->call->reply_data.dptr, r->datalen);
374 res = daemon_queue_send(client, &r->hdr);
375 if (res == -1) {
376 /* client is dead - return immediately */
377 return;
379 if (res != 0) {
380 DEBUG(DEBUG_ERR, (__location__ " Failed to queue packet from daemon to client\n"));
382 CTDB_UPDATE_LATENCY(client->ctdb, ctdb_db, "call_from_client_cb 3", call_latency, dstate->start_time);
383 CTDB_DECREMENT_STAT(client->ctdb, pending_calls);
384 talloc_free(dstate);
387 struct ctdb_daemon_packet_wrap {
388 struct ctdb_context *ctdb;
389 uint32_t client_id;
393 a wrapper to catch disconnected clients
395 static void daemon_incoming_packet_wrap(void *p, struct ctdb_req_header *hdr)
397 struct ctdb_client *client;
398 struct ctdb_daemon_packet_wrap *w = talloc_get_type(p,
399 struct ctdb_daemon_packet_wrap);
400 if (w == NULL) {
401 DEBUG(DEBUG_CRIT,(__location__ " Bad packet type '%s'\n", talloc_get_name(p)));
402 return;
405 client = reqid_find(w->ctdb->idr, w->client_id, struct ctdb_client);
406 if (client == NULL) {
407 DEBUG(DEBUG_ERR,(__location__ " Packet for disconnected client %u\n",
408 w->client_id));
409 talloc_free(w);
410 return;
412 talloc_free(w);
414 /* process it */
415 daemon_incoming_packet(client, hdr);
418 struct ctdb_deferred_fetch_call {
419 struct ctdb_deferred_fetch_call *next, *prev;
420 struct ctdb_req_call_old *c;
421 struct ctdb_daemon_packet_wrap *w;
424 struct ctdb_deferred_fetch_queue {
425 struct ctdb_deferred_fetch_call *deferred_calls;
428 struct ctdb_deferred_requeue {
429 struct ctdb_deferred_fetch_call *dfc;
430 struct ctdb_client *client;
433 /* called from a timer event and starts reprocessing the deferred call.*/
434 static void reprocess_deferred_call(struct tevent_context *ev,
435 struct tevent_timer *te,
436 struct timeval t, void *private_data)
438 struct ctdb_deferred_requeue *dfr = (struct ctdb_deferred_requeue *)private_data;
439 struct ctdb_client *client = dfr->client;
441 talloc_steal(client, dfr->dfc->c);
442 daemon_incoming_packet(client, (struct ctdb_req_header *)dfr->dfc->c);
443 talloc_free(dfr);
446 /* the referral context is destroyed either after a timeout or when the initial
447 fetch-lock has finished.
448 at this stage, immediately start reprocessing the queued up deferred
449 calls so they get reprocessed immediately (and since we are dmaster at
450 this stage, trigger the waiting smbd processes to pick up and aquire the
451 record right away.
453 static int deferred_fetch_queue_destructor(struct ctdb_deferred_fetch_queue *dfq)
456 /* need to reprocess the packets from the queue explicitely instead of
457 just using a normal destructor since we want, need, to
458 call the clients in the same oder as the requests queued up
460 while (dfq->deferred_calls != NULL) {
461 struct ctdb_client *client;
462 struct ctdb_deferred_fetch_call *dfc = dfq->deferred_calls;
463 struct ctdb_deferred_requeue *dfr;
465 DLIST_REMOVE(dfq->deferred_calls, dfc);
467 client = reqid_find(dfc->w->ctdb->idr, dfc->w->client_id, struct ctdb_client);
468 if (client == NULL) {
469 DEBUG(DEBUG_ERR,(__location__ " Packet for disconnected client %u\n",
470 dfc->w->client_id));
471 continue;
474 /* process it by pushing it back onto the eventloop */
475 dfr = talloc(client, struct ctdb_deferred_requeue);
476 if (dfr == NULL) {
477 DEBUG(DEBUG_ERR,("Failed to allocate deferred fetch requeue structure\n"));
478 continue;
481 dfr->dfc = talloc_steal(dfr, dfc);
482 dfr->client = client;
484 tevent_add_timer(dfc->w->ctdb->ev, client, timeval_zero(),
485 reprocess_deferred_call, dfr);
488 return 0;
491 /* insert the new deferral context into the rb tree.
492 there should never be a pre-existing context here, but check for it
493 warn and destroy the previous context if there is already a deferral context
494 for this key.
496 static void *insert_dfq_callback(void *parm, void *data)
498 if (data) {
499 DEBUG(DEBUG_ERR,("Already have DFQ registered. Free old %p and create new %p\n", data, parm));
500 talloc_free(data);
502 return parm;
505 /* if the original fetch-lock did not complete within a reasonable time,
506 free the context and context for all deferred requests to cause them to be
507 re-inserted into the event system.
509 static void dfq_timeout(struct tevent_context *ev, struct tevent_timer *te,
510 struct timeval t, void *private_data)
512 talloc_free(private_data);
515 /* This function is used in the local daemon to register a KEY in a database
516 for being "fetched"
517 While the remote fetch is in-flight, any futher attempts to re-fetch the
518 same record will be deferred until the fetch completes.
520 static int setup_deferred_fetch_locks(struct ctdb_db_context *ctdb_db, struct ctdb_call *call)
522 uint32_t *k;
523 struct ctdb_deferred_fetch_queue *dfq;
525 k = ctdb_key_to_idkey(call, call->key);
526 if (k == NULL) {
527 DEBUG(DEBUG_ERR,("Failed to allocate key for deferred fetch\n"));
528 return -1;
531 dfq = talloc(call, struct ctdb_deferred_fetch_queue);
532 if (dfq == NULL) {
533 DEBUG(DEBUG_ERR,("Failed to allocate key for deferred fetch queue structure\n"));
534 talloc_free(k);
535 return -1;
537 dfq->deferred_calls = NULL;
539 trbt_insertarray32_callback(ctdb_db->deferred_fetch, k[0], &k[0], insert_dfq_callback, dfq);
541 talloc_set_destructor(dfq, deferred_fetch_queue_destructor);
543 /* if the fetch havent completed in 30 seconds, just tear it all down
544 and let it try again as the events are reissued */
545 tevent_add_timer(ctdb_db->ctdb->ev, dfq, timeval_current_ofs(30, 0),
546 dfq_timeout, dfq);
548 talloc_free(k);
549 return 0;
552 /* check if this is a duplicate request to a fetch already in-flight
553 if it is, make this call deferred to be reprocessed later when
554 the in-flight fetch completes.
556 static int requeue_duplicate_fetch(struct ctdb_db_context *ctdb_db, struct ctdb_client *client, TDB_DATA key, struct ctdb_req_call_old *c)
558 uint32_t *k;
559 struct ctdb_deferred_fetch_queue *dfq;
560 struct ctdb_deferred_fetch_call *dfc;
562 k = ctdb_key_to_idkey(c, key);
563 if (k == NULL) {
564 DEBUG(DEBUG_ERR,("Failed to allocate key for deferred fetch\n"));
565 return -1;
568 dfq = trbt_lookuparray32(ctdb_db->deferred_fetch, k[0], &k[0]);
569 if (dfq == NULL) {
570 talloc_free(k);
571 return -1;
575 talloc_free(k);
577 dfc = talloc(dfq, struct ctdb_deferred_fetch_call);
578 if (dfc == NULL) {
579 DEBUG(DEBUG_ERR, ("Failed to allocate deferred fetch call structure\n"));
580 return -1;
583 dfc->w = talloc(dfc, struct ctdb_daemon_packet_wrap);
584 if (dfc->w == NULL) {
585 DEBUG(DEBUG_ERR,("Failed to allocate deferred fetch daemon packet wrap structure\n"));
586 talloc_free(dfc);
587 return -1;
590 dfc->c = talloc_steal(dfc, c);
591 dfc->w->ctdb = ctdb_db->ctdb;
592 dfc->w->client_id = client->client_id;
594 DLIST_ADD_END(dfq->deferred_calls, dfc);
596 return 0;
601 this is called when the ctdb daemon received a ctdb request call
602 from a local client over the unix domain socket
604 static void daemon_request_call_from_client(struct ctdb_client *client,
605 struct ctdb_req_call_old *c)
607 struct ctdb_call_state *state;
608 struct ctdb_db_context *ctdb_db;
609 struct daemon_call_state *dstate;
610 struct ctdb_call *call;
611 struct ctdb_ltdb_header header;
612 TDB_DATA key, data;
613 int ret;
614 struct ctdb_context *ctdb = client->ctdb;
615 struct ctdb_daemon_packet_wrap *w;
617 CTDB_INCREMENT_STAT(ctdb, total_calls);
618 CTDB_INCREMENT_STAT(ctdb, pending_calls);
620 ctdb_db = find_ctdb_db(client->ctdb, c->db_id);
621 if (!ctdb_db) {
622 DEBUG(DEBUG_ERR, (__location__ " Unknown database in request. db_id==0x%08x",
623 c->db_id));
624 CTDB_DECREMENT_STAT(ctdb, pending_calls);
625 return;
628 if (ctdb_db->unhealthy_reason) {
630 * this is just a warning, as the tdb should be empty anyway,
631 * and only persistent databases can be unhealthy, which doesn't
632 * use this code patch
634 DEBUG(DEBUG_WARNING,("warn: db(%s) unhealty in daemon_request_call_from_client(): %s\n",
635 ctdb_db->db_name, ctdb_db->unhealthy_reason));
638 key.dptr = c->data;
639 key.dsize = c->keylen;
641 w = talloc(ctdb, struct ctdb_daemon_packet_wrap);
642 CTDB_NO_MEMORY_VOID(ctdb, w);
644 w->ctdb = ctdb;
645 w->client_id = client->client_id;
647 ret = ctdb_ltdb_lock_fetch_requeue(ctdb_db, key, &header,
648 (struct ctdb_req_header *)c, &data,
649 daemon_incoming_packet_wrap, w, true);
650 if (ret == -2) {
651 /* will retry later */
652 CTDB_DECREMENT_STAT(ctdb, pending_calls);
653 return;
656 talloc_free(w);
658 if (ret != 0) {
659 DEBUG(DEBUG_ERR,(__location__ " Unable to fetch record\n"));
660 CTDB_DECREMENT_STAT(ctdb, pending_calls);
661 return;
665 /* check if this fetch request is a duplicate for a
666 request we already have in flight. If so defer it until
667 the first request completes.
669 if (ctdb->tunable.fetch_collapse == 1) {
670 if (requeue_duplicate_fetch(ctdb_db, client, key, c) == 0) {
671 ret = ctdb_ltdb_unlock(ctdb_db, key);
672 if (ret != 0) {
673 DEBUG(DEBUG_ERR,(__location__ " ctdb_ltdb_unlock() failed with error %d\n", ret));
675 CTDB_DECREMENT_STAT(ctdb, pending_calls);
676 return;
680 /* Dont do READONLY if we don't have a tracking database */
681 if ((c->flags & CTDB_WANT_READONLY) && !ctdb_db->readonly) {
682 c->flags &= ~CTDB_WANT_READONLY;
685 if (header.flags & CTDB_REC_RO_REVOKE_COMPLETE) {
686 header.flags &= ~CTDB_REC_RO_FLAGS;
687 CTDB_INCREMENT_STAT(ctdb, total_ro_revokes);
688 CTDB_INCREMENT_DB_STAT(ctdb_db, db_ro_revokes);
689 if (ctdb_ltdb_store(ctdb_db, key, &header, data) != 0) {
690 ctdb_fatal(ctdb, "Failed to write header with cleared REVOKE flag");
692 /* and clear out the tracking data */
693 if (tdb_delete(ctdb_db->rottdb, key) != 0) {
694 DEBUG(DEBUG_ERR,(__location__ " Failed to clear out trackingdb record\n"));
698 /* if we are revoking, we must defer all other calls until the revoke
699 * had completed.
701 if (header.flags & CTDB_REC_RO_REVOKING_READONLY) {
702 talloc_free(data.dptr);
703 ret = ctdb_ltdb_unlock(ctdb_db, key);
705 if (ctdb_add_revoke_deferred_call(ctdb, ctdb_db, key, (struct ctdb_req_header *)c, daemon_incoming_packet, client) != 0) {
706 ctdb_fatal(ctdb, "Failed to add deferred call for revoke child");
708 CTDB_DECREMENT_STAT(ctdb, pending_calls);
709 return;
712 if ((header.dmaster == ctdb->pnn)
713 && (!(c->flags & CTDB_WANT_READONLY))
714 && (header.flags & (CTDB_REC_RO_HAVE_DELEGATIONS|CTDB_REC_RO_HAVE_READONLY)) ) {
715 header.flags |= CTDB_REC_RO_REVOKING_READONLY;
716 if (ctdb_ltdb_store(ctdb_db, key, &header, data) != 0) {
717 ctdb_fatal(ctdb, "Failed to store record with HAVE_DELEGATIONS set");
719 ret = ctdb_ltdb_unlock(ctdb_db, key);
721 if (ctdb_start_revoke_ro_record(ctdb, ctdb_db, key, &header, data) != 0) {
722 ctdb_fatal(ctdb, "Failed to start record revoke");
724 talloc_free(data.dptr);
726 if (ctdb_add_revoke_deferred_call(ctdb, ctdb_db, key, (struct ctdb_req_header *)c, daemon_incoming_packet, client) != 0) {
727 ctdb_fatal(ctdb, "Failed to add deferred call for revoke child");
730 CTDB_DECREMENT_STAT(ctdb, pending_calls);
731 return;
734 dstate = talloc(client, struct daemon_call_state);
735 if (dstate == NULL) {
736 ret = ctdb_ltdb_unlock(ctdb_db, key);
737 if (ret != 0) {
738 DEBUG(DEBUG_ERR,(__location__ " ctdb_ltdb_unlock() failed with error %d\n", ret));
741 DEBUG(DEBUG_ERR,(__location__ " Unable to allocate dstate\n"));
742 CTDB_DECREMENT_STAT(ctdb, pending_calls);
743 return;
745 dstate->start_time = timeval_current();
746 dstate->client = client;
747 dstate->reqid = c->hdr.reqid;
748 talloc_steal(dstate, data.dptr);
750 call = dstate->call = talloc_zero(dstate, struct ctdb_call);
751 if (call == NULL) {
752 ret = ctdb_ltdb_unlock(ctdb_db, key);
753 if (ret != 0) {
754 DEBUG(DEBUG_ERR,(__location__ " ctdb_ltdb_unlock() failed with error %d\n", ret));
757 DEBUG(DEBUG_ERR,(__location__ " Unable to allocate call\n"));
758 CTDB_DECREMENT_STAT(ctdb, pending_calls);
759 CTDB_UPDATE_LATENCY(ctdb, ctdb_db, "call_from_client 1", call_latency, dstate->start_time);
760 return;
763 dstate->readonly_fetch = 0;
764 call->call_id = c->callid;
765 call->key = key;
766 call->call_data.dptr = c->data + c->keylen;
767 call->call_data.dsize = c->calldatalen;
768 call->flags = c->flags;
770 if (c->flags & CTDB_WANT_READONLY) {
771 /* client wants readonly record, so translate this into a
772 fetch with header. remember what the client asked for
773 so we can remap the reply back to the proper format for
774 the client in the reply
776 dstate->client_callid = call->call_id;
777 call->call_id = CTDB_FETCH_WITH_HEADER_FUNC;
778 dstate->readonly_fetch = 1;
781 if (header.dmaster == ctdb->pnn) {
782 state = ctdb_call_local_send(ctdb_db, call, &header, &data);
783 } else {
784 state = ctdb_daemon_call_send_remote(ctdb_db, call, &header);
785 if (ctdb->tunable.fetch_collapse == 1) {
786 /* This request triggered a remote fetch-lock.
787 set up a deferral for this key so any additional
788 fetch-locks are deferred until the current one
789 finishes.
791 setup_deferred_fetch_locks(ctdb_db, call);
795 ret = ctdb_ltdb_unlock(ctdb_db, key);
796 if (ret != 0) {
797 DEBUG(DEBUG_ERR,(__location__ " ctdb_ltdb_unlock() failed with error %d\n", ret));
800 if (state == NULL) {
801 DEBUG(DEBUG_ERR,(__location__ " Unable to setup call send\n"));
802 CTDB_DECREMENT_STAT(ctdb, pending_calls);
803 CTDB_UPDATE_LATENCY(ctdb, ctdb_db, "call_from_client 2", call_latency, dstate->start_time);
804 return;
806 talloc_steal(state, dstate);
807 talloc_steal(client, state);
809 state->async.fn = daemon_call_from_client_callback;
810 state->async.private_data = dstate;
814 static void daemon_request_control_from_client(struct ctdb_client *client,
815 struct ctdb_req_control_old *c);
817 /* data contains a packet from the client */
818 static void daemon_incoming_packet(void *p, struct ctdb_req_header *hdr)
820 struct ctdb_client *client = talloc_get_type(p, struct ctdb_client);
821 TALLOC_CTX *tmp_ctx;
822 struct ctdb_context *ctdb = client->ctdb;
824 /* place the packet as a child of a tmp_ctx. We then use
825 talloc_free() below to free it. If any of the calls want
826 to keep it, then they will steal it somewhere else, and the
827 talloc_free() will be a no-op */
828 tmp_ctx = talloc_new(client);
829 talloc_steal(tmp_ctx, hdr);
831 if (hdr->ctdb_magic != CTDB_MAGIC) {
832 ctdb_set_error(client->ctdb, "Non CTDB packet rejected in daemon\n");
833 goto done;
836 if (hdr->ctdb_version != CTDB_PROTOCOL) {
837 ctdb_set_error(client->ctdb, "Bad CTDB version 0x%x rejected in daemon\n", hdr->ctdb_version);
838 goto done;
841 switch (hdr->operation) {
842 case CTDB_REQ_CALL:
843 CTDB_INCREMENT_STAT(ctdb, client.req_call);
844 daemon_request_call_from_client(client, (struct ctdb_req_call_old *)hdr);
845 break;
847 case CTDB_REQ_MESSAGE:
848 CTDB_INCREMENT_STAT(ctdb, client.req_message);
849 daemon_request_message_from_client(client, (struct ctdb_req_message_old *)hdr);
850 break;
852 case CTDB_REQ_CONTROL:
853 CTDB_INCREMENT_STAT(ctdb, client.req_control);
854 daemon_request_control_from_client(client, (struct ctdb_req_control_old *)hdr);
855 break;
857 default:
858 DEBUG(DEBUG_CRIT,(__location__ " daemon: unrecognized operation %u\n",
859 hdr->operation));
862 done:
863 talloc_free(tmp_ctx);
867 called when the daemon gets a incoming packet
869 static void ctdb_daemon_read_cb(uint8_t *data, size_t cnt, void *args)
871 struct ctdb_client *client = talloc_get_type(args, struct ctdb_client);
872 struct ctdb_req_header *hdr;
874 if (cnt == 0) {
875 talloc_free(client);
876 return;
879 CTDB_INCREMENT_STAT(client->ctdb, client_packets_recv);
881 if (cnt < sizeof(*hdr)) {
882 ctdb_set_error(client->ctdb, "Bad packet length %u in daemon\n",
883 (unsigned)cnt);
884 return;
886 hdr = (struct ctdb_req_header *)data;
887 if (cnt != hdr->length) {
888 ctdb_set_error(client->ctdb, "Bad header length %u expected %u\n in daemon",
889 (unsigned)hdr->length, (unsigned)cnt);
890 return;
893 if (hdr->ctdb_magic != CTDB_MAGIC) {
894 ctdb_set_error(client->ctdb, "Non CTDB packet rejected\n");
895 return;
898 if (hdr->ctdb_version != CTDB_PROTOCOL) {
899 ctdb_set_error(client->ctdb, "Bad CTDB version 0x%x rejected in daemon\n", hdr->ctdb_version);
900 return;
903 DEBUG(DEBUG_DEBUG,(__location__ " client request %u of type %u length %u from "
904 "node %u to %u\n", hdr->reqid, hdr->operation, hdr->length,
905 hdr->srcnode, hdr->destnode));
907 /* it is the responsibility of the incoming packet function to free 'data' */
908 daemon_incoming_packet(client, hdr);
912 static int ctdb_clientpid_destructor(struct ctdb_client_pid_list *client_pid)
914 if (client_pid->ctdb->client_pids != NULL) {
915 DLIST_REMOVE(client_pid->ctdb->client_pids, client_pid);
918 return 0;
922 static void ctdb_accept_client(struct tevent_context *ev,
923 struct tevent_fd *fde, uint16_t flags,
924 void *private_data)
926 struct sockaddr_un addr;
927 socklen_t len;
928 int fd;
929 struct ctdb_context *ctdb = talloc_get_type(private_data, struct ctdb_context);
930 struct ctdb_client *client;
931 struct ctdb_client_pid_list *client_pid;
932 pid_t peer_pid = 0;
934 memset(&addr, 0, sizeof(addr));
935 len = sizeof(addr);
936 fd = accept(ctdb->daemon.sd, (struct sockaddr *)&addr, &len);
937 if (fd == -1) {
938 return;
941 set_blocking(fd, false);
942 set_close_on_exec(fd);
944 DEBUG(DEBUG_DEBUG,(__location__ " Created SOCKET FD:%d to connected child\n", fd));
946 client = talloc_zero(ctdb, struct ctdb_client);
947 if (ctdb_get_peer_pid(fd, &peer_pid) == 0) {
948 DEBUG(DEBUG_INFO,("Connected client with pid:%u\n", (unsigned)peer_pid));
951 client->ctdb = ctdb;
952 client->fd = fd;
953 client->client_id = reqid_new(ctdb->idr, client);
954 client->pid = peer_pid;
956 client_pid = talloc(client, struct ctdb_client_pid_list);
957 if (client_pid == NULL) {
958 DEBUG(DEBUG_ERR,("Failed to allocate client pid structure\n"));
959 close(fd);
960 talloc_free(client);
961 return;
963 client_pid->ctdb = ctdb;
964 client_pid->pid = peer_pid;
965 client_pid->client = client;
967 DLIST_ADD(ctdb->client_pids, client_pid);
969 client->queue = ctdb_queue_setup(ctdb, client, fd, CTDB_DS_ALIGNMENT,
970 ctdb_daemon_read_cb, client,
971 "client-%u", client->pid);
973 talloc_set_destructor(client, ctdb_client_destructor);
974 talloc_set_destructor(client_pid, ctdb_clientpid_destructor);
975 ctdb->num_clients++;
981 create a unix domain socket and bind it
982 return a file descriptor open on the socket
984 static int ux_socket_bind(struct ctdb_context *ctdb)
986 struct sockaddr_un addr;
988 ctdb->daemon.sd = socket(AF_UNIX, SOCK_STREAM, 0);
989 if (ctdb->daemon.sd == -1) {
990 return -1;
993 memset(&addr, 0, sizeof(addr));
994 addr.sun_family = AF_UNIX;
995 strncpy(addr.sun_path, ctdb->daemon.name, sizeof(addr.sun_path)-1);
997 /* First check if an old ctdbd might be running */
998 if (connect(ctdb->daemon.sd,
999 (struct sockaddr *)&addr, sizeof(addr)) == 0) {
1000 DEBUG(DEBUG_CRIT,
1001 ("Something is already listening on ctdb socket '%s'\n",
1002 ctdb->daemon.name));
1003 goto failed;
1006 /* Remove any old socket */
1007 unlink(ctdb->daemon.name);
1009 set_close_on_exec(ctdb->daemon.sd);
1010 set_blocking(ctdb->daemon.sd, false);
1012 if (bind(ctdb->daemon.sd, (struct sockaddr *)&addr, sizeof(addr)) == -1) {
1013 DEBUG(DEBUG_CRIT,("Unable to bind on ctdb socket '%s'\n", ctdb->daemon.name));
1014 goto failed;
1017 if (chown(ctdb->daemon.name, geteuid(), getegid()) != 0 ||
1018 chmod(ctdb->daemon.name, 0700) != 0) {
1019 DEBUG(DEBUG_CRIT,("Unable to secure ctdb socket '%s', ctdb->daemon.name\n", ctdb->daemon.name));
1020 goto failed;
1024 if (listen(ctdb->daemon.sd, 100) != 0) {
1025 DEBUG(DEBUG_CRIT,("Unable to listen on ctdb socket '%s'\n", ctdb->daemon.name));
1026 goto failed;
1029 DEBUG(DEBUG_NOTICE, ("Listening to ctdb socket %s\n",
1030 ctdb->daemon.name));
1031 return 0;
1033 failed:
1034 close(ctdb->daemon.sd);
1035 ctdb->daemon.sd = -1;
1036 return -1;
1039 static void initialise_node_flags (struct ctdb_context *ctdb)
1041 if (ctdb->pnn == -1) {
1042 ctdb_fatal(ctdb, "PNN is set to -1 (unknown value)");
1045 ctdb->nodes[ctdb->pnn]->flags &= ~NODE_FLAGS_DISCONNECTED;
1047 /* do we start out in DISABLED mode? */
1048 if (ctdb->start_as_disabled != 0) {
1049 DEBUG(DEBUG_NOTICE, ("This node is configured to start in DISABLED state\n"));
1050 ctdb->nodes[ctdb->pnn]->flags |= NODE_FLAGS_DISABLED;
1052 /* do we start out in STOPPED mode? */
1053 if (ctdb->start_as_stopped != 0) {
1054 DEBUG(DEBUG_NOTICE, ("This node is configured to start in STOPPED state\n"));
1055 ctdb->nodes[ctdb->pnn]->flags |= NODE_FLAGS_STOPPED;
1059 static void ctdb_setup_event_callback(struct ctdb_context *ctdb, int status,
1060 void *private_data)
1062 if (status != 0) {
1063 ctdb_die(ctdb, "Failed to run setup event");
1065 ctdb_run_notification_script(ctdb, "setup");
1067 /* tell all other nodes we've just started up */
1068 ctdb_daemon_send_control(ctdb, CTDB_BROADCAST_ALL,
1069 0, CTDB_CONTROL_STARTUP, 0,
1070 CTDB_CTRL_FLAG_NOREPLY,
1071 tdb_null, NULL, NULL);
1073 /* Start the recovery daemon */
1074 if (ctdb_start_recoverd(ctdb) != 0) {
1075 DEBUG(DEBUG_ALERT,("Failed to start recovery daemon\n"));
1076 exit(11);
1079 ctdb_start_periodic_events(ctdb);
1081 ctdb_wait_for_first_recovery(ctdb);
1084 static struct timeval tevent_before_wait_ts;
1085 static struct timeval tevent_after_wait_ts;
1087 static void ctdb_tevent_trace(enum tevent_trace_point tp,
1088 void *private_data)
1090 struct timeval diff;
1091 struct timeval now;
1092 struct ctdb_context *ctdb =
1093 talloc_get_type(private_data, struct ctdb_context);
1095 if (getpid() != ctdb->ctdbd_pid) {
1096 return;
1099 now = timeval_current();
1101 switch (tp) {
1102 case TEVENT_TRACE_BEFORE_WAIT:
1103 if (!timeval_is_zero(&tevent_after_wait_ts)) {
1104 diff = timeval_until(&tevent_after_wait_ts, &now);
1105 if (diff.tv_sec > 3) {
1106 DEBUG(DEBUG_ERR,
1107 ("Handling event took %ld seconds!\n",
1108 (long)diff.tv_sec));
1111 tevent_before_wait_ts = now;
1112 break;
1114 case TEVENT_TRACE_AFTER_WAIT:
1115 if (!timeval_is_zero(&tevent_before_wait_ts)) {
1116 diff = timeval_until(&tevent_before_wait_ts, &now);
1117 if (diff.tv_sec > 3) {
1118 DEBUG(DEBUG_CRIT,
1119 ("No event for %ld seconds!\n",
1120 (long)diff.tv_sec));
1123 tevent_after_wait_ts = now;
1124 break;
1126 default:
1127 /* Do nothing for future tevent trace points */ ;
1131 static void ctdb_remove_pidfile(void)
1133 /* Only the main ctdbd's PID matches the SID */
1134 if (ctdbd_pidfile != NULL && getsid(0) == getpid()) {
1135 if (unlink(ctdbd_pidfile) == 0) {
1136 DEBUG(DEBUG_NOTICE, ("Removed PID file %s\n",
1137 ctdbd_pidfile));
1138 } else {
1139 DEBUG(DEBUG_WARNING, ("Failed to Remove PID file %s\n",
1140 ctdbd_pidfile));
1145 static void ctdb_create_pidfile(pid_t pid)
1147 if (ctdbd_pidfile != NULL) {
1148 FILE *fp;
1150 fp = fopen(ctdbd_pidfile, "w");
1151 if (fp == NULL) {
1152 DEBUG(DEBUG_ALERT,
1153 ("Failed to open PID file %s\n", ctdbd_pidfile));
1154 exit(11);
1157 fprintf(fp, "%d\n", pid);
1158 fclose(fp);
1159 DEBUG(DEBUG_NOTICE, ("Created PID file %s\n", ctdbd_pidfile));
1160 atexit(ctdb_remove_pidfile);
1164 static void ctdb_initialise_vnn_map(struct ctdb_context *ctdb)
1166 int i, j, count;
1168 /* initialize the vnn mapping table, skipping any deleted nodes */
1169 ctdb->vnn_map = talloc(ctdb, struct ctdb_vnn_map);
1170 CTDB_NO_MEMORY_FATAL(ctdb, ctdb->vnn_map);
1172 count = 0;
1173 for (i = 0; i < ctdb->num_nodes; i++) {
1174 if ((ctdb->nodes[i]->flags & NODE_FLAGS_DELETED) == 0) {
1175 count++;
1179 ctdb->vnn_map->generation = INVALID_GENERATION;
1180 ctdb->vnn_map->size = count;
1181 ctdb->vnn_map->map = talloc_array(ctdb->vnn_map, uint32_t, ctdb->vnn_map->size);
1182 CTDB_NO_MEMORY_FATAL(ctdb, ctdb->vnn_map->map);
1184 for(i=0, j=0; i < ctdb->vnn_map->size; i++) {
1185 if (ctdb->nodes[i]->flags & NODE_FLAGS_DELETED) {
1186 continue;
1188 ctdb->vnn_map->map[j] = i;
1189 j++;
1193 static void ctdb_set_my_pnn(struct ctdb_context *ctdb)
1195 int nodeid;
1197 if (ctdb->address == NULL) {
1198 ctdb_fatal(ctdb,
1199 "Can not determine PNN - node address is not set\n");
1202 nodeid = ctdb_ip_to_nodeid(ctdb, ctdb->address);
1203 if (nodeid == -1) {
1204 ctdb_fatal(ctdb,
1205 "Can not determine PNN - node address not found in node list\n");
1208 ctdb->pnn = ctdb->nodes[nodeid]->pnn;
1209 DEBUG(DEBUG_NOTICE, ("PNN is %u\n", ctdb->pnn));
1213 start the protocol going as a daemon
1215 int ctdb_start_daemon(struct ctdb_context *ctdb, bool do_fork)
1217 int res, ret = -1;
1218 struct tevent_fd *fde;
1220 /* create a unix domain stream socket to listen to */
1221 res = ux_socket_bind(ctdb);
1222 if (res!=0) {
1223 DEBUG(DEBUG_ALERT,("Cannot continue. Exiting!\n"));
1224 exit(10);
1227 if (do_fork && fork()) {
1228 return 0;
1231 tdb_reopen_all(false);
1233 if (do_fork) {
1234 if (setsid() == -1) {
1235 ctdb_die(ctdb, "Failed to setsid()\n");
1237 close(0);
1238 if (open("/dev/null", O_RDONLY) != 0) {
1239 DEBUG(DEBUG_ALERT,(__location__ " Failed to setup stdin on /dev/null\n"));
1240 exit(11);
1243 ignore_signal(SIGPIPE);
1244 ignore_signal(SIGUSR1);
1246 ctdb->ctdbd_pid = getpid();
1247 DEBUG(DEBUG_ERR, ("Starting CTDBD (Version %s) as PID: %u\n",
1248 CTDB_VERSION_STRING, ctdb->ctdbd_pid));
1249 ctdb_create_pidfile(ctdb->ctdbd_pid);
1251 /* Make sure we log something when the daemon terminates.
1252 * This must be the first exit handler to run (so the last to
1253 * be registered.
1255 atexit(print_exit_message);
1257 if (ctdb->do_setsched) {
1258 /* try to set us up as realtime */
1259 if (!set_scheduler()) {
1260 exit(1);
1262 DEBUG(DEBUG_NOTICE, ("Set real-time scheduler priority\n"));
1265 ctdb->ev = tevent_context_init(NULL);
1266 tevent_loop_allow_nesting(ctdb->ev);
1267 tevent_set_trace_callback(ctdb->ev, ctdb_tevent_trace, ctdb);
1268 ret = ctdb_init_tevent_logging(ctdb);
1269 if (ret != 0) {
1270 DEBUG(DEBUG_ALERT,("Failed to initialize TEVENT logging\n"));
1271 exit(1);
1274 /* set up a handler to pick up sigchld */
1275 if (ctdb_init_sigchld(ctdb) == NULL) {
1276 DEBUG(DEBUG_CRIT,("Failed to set up signal handler for SIGCHLD\n"));
1277 exit(1);
1280 ctdb_set_child_logging(ctdb);
1282 TALLOC_FREE(ctdb->srv);
1283 if (srvid_init(ctdb, &ctdb->srv) != 0) {
1284 DEBUG(DEBUG_CRIT,("Failed to setup message srvid context\n"));
1285 exit(1);
1288 /* initialize statistics collection */
1289 ctdb_statistics_init(ctdb);
1291 /* force initial recovery for election */
1292 ctdb->recovery_mode = CTDB_RECOVERY_ACTIVE;
1294 ctdb_set_runstate(ctdb, CTDB_RUNSTATE_INIT);
1295 ret = ctdb_event_script(ctdb, CTDB_EVENT_INIT);
1296 if (ret != 0) {
1297 ctdb_die(ctdb, "Failed to run init event\n");
1299 ctdb_run_notification_script(ctdb, "init");
1301 if (strcmp(ctdb->transport, "tcp") == 0) {
1302 ret = ctdb_tcp_init(ctdb);
1304 #ifdef USE_INFINIBAND
1305 if (strcmp(ctdb->transport, "ib") == 0) {
1306 ret = ctdb_ibw_init(ctdb);
1308 #endif
1309 if (ret != 0) {
1310 DEBUG(DEBUG_ERR,("Failed to initialise transport '%s'\n", ctdb->transport));
1311 return -1;
1314 if (ctdb->methods == NULL) {
1315 DEBUG(DEBUG_ALERT,(__location__ " Can not initialize transport. ctdb->methods is NULL\n"));
1316 ctdb_fatal(ctdb, "transport is unavailable. can not initialize.");
1319 /* Initialise the transport. This sets the node address if it
1320 * was not set via the command-line. */
1321 if (ctdb->methods->initialise(ctdb) != 0) {
1322 ctdb_fatal(ctdb, "transport failed to initialise");
1325 ctdb_set_my_pnn(ctdb);
1327 initialise_node_flags(ctdb);
1329 if (ctdb->public_addresses_file) {
1330 ret = ctdb_set_public_addresses(ctdb, true);
1331 if (ret == -1) {
1332 DEBUG(DEBUG_ALERT,("Unable to setup public address list\n"));
1333 exit(1);
1337 ctdb_initialise_vnn_map(ctdb);
1339 /* attach to existing databases */
1340 if (ctdb_attach_databases(ctdb) != 0) {
1341 ctdb_fatal(ctdb, "Failed to attach to databases\n");
1344 /* start frozen, then let the first election sort things out */
1345 if (!ctdb_blocking_freeze(ctdb)) {
1346 ctdb_fatal(ctdb, "Failed to get initial freeze\n");
1349 /* now start accepting clients, only can do this once frozen */
1350 fde = tevent_add_fd(ctdb->ev, ctdb, ctdb->daemon.sd, TEVENT_FD_READ,
1351 ctdb_accept_client, ctdb);
1352 if (fde == NULL) {
1353 ctdb_fatal(ctdb, "Failed to add daemon socket to event loop");
1355 tevent_fd_set_auto_close(fde);
1357 /* Start the transport */
1358 if (ctdb->methods->start(ctdb) != 0) {
1359 DEBUG(DEBUG_ALERT,("transport failed to start!\n"));
1360 ctdb_fatal(ctdb, "transport failed to start");
1363 /* Recovery daemon and timed events are started from the
1364 * callback, only after the setup event completes
1365 * successfully.
1367 ctdb_set_runstate(ctdb, CTDB_RUNSTATE_SETUP);
1368 ret = ctdb_event_script_callback(ctdb,
1369 ctdb,
1370 ctdb_setup_event_callback,
1371 ctdb,
1372 CTDB_EVENT_SETUP,
1373 "%s",
1374 "");
1375 if (ret != 0) {
1376 DEBUG(DEBUG_CRIT,("Failed to set up 'setup' event\n"));
1377 exit(1);
1380 lockdown_memory(ctdb->valgrinding);
1382 /* go into a wait loop to allow other nodes to complete */
1383 tevent_loop_wait(ctdb->ev);
1385 DEBUG(DEBUG_CRIT,("event_loop_wait() returned. this should not happen\n"));
1386 exit(1);
1390 allocate a packet for use in daemon<->daemon communication
1392 struct ctdb_req_header *_ctdb_transport_allocate(struct ctdb_context *ctdb,
1393 TALLOC_CTX *mem_ctx,
1394 enum ctdb_operation operation,
1395 size_t length, size_t slength,
1396 const char *type)
1398 int size;
1399 struct ctdb_req_header *hdr;
1401 length = MAX(length, slength);
1402 size = (length+(CTDB_DS_ALIGNMENT-1)) & ~(CTDB_DS_ALIGNMENT-1);
1404 if (ctdb->methods == NULL) {
1405 DEBUG(DEBUG_INFO,(__location__ " Unable to allocate transport packet for operation %u of length %u. Transport is DOWN.\n",
1406 operation, (unsigned)length));
1407 return NULL;
1410 hdr = (struct ctdb_req_header *)ctdb->methods->allocate_pkt(mem_ctx, size);
1411 if (hdr == NULL) {
1412 DEBUG(DEBUG_ERR,("Unable to allocate transport packet for operation %u of length %u\n",
1413 operation, (unsigned)length));
1414 return NULL;
1416 talloc_set_name_const(hdr, type);
1417 memset(hdr, 0, slength);
1418 hdr->length = length;
1419 hdr->operation = operation;
1420 hdr->ctdb_magic = CTDB_MAGIC;
1421 hdr->ctdb_version = CTDB_PROTOCOL;
1422 hdr->generation = ctdb->vnn_map->generation;
1423 hdr->srcnode = ctdb->pnn;
1425 return hdr;
1428 struct daemon_control_state {
1429 struct daemon_control_state *next, *prev;
1430 struct ctdb_client *client;
1431 struct ctdb_req_control_old *c;
1432 uint32_t reqid;
1433 struct ctdb_node *node;
1437 callback when a control reply comes in
1439 static void daemon_control_callback(struct ctdb_context *ctdb,
1440 int32_t status, TDB_DATA data,
1441 const char *errormsg,
1442 void *private_data)
1444 struct daemon_control_state *state = talloc_get_type(private_data,
1445 struct daemon_control_state);
1446 struct ctdb_client *client = state->client;
1447 struct ctdb_reply_control_old *r;
1448 size_t len;
1449 int ret;
1451 /* construct a message to send to the client containing the data */
1452 len = offsetof(struct ctdb_reply_control_old, data) + data.dsize;
1453 if (errormsg) {
1454 len += strlen(errormsg);
1456 r = ctdbd_allocate_pkt(ctdb, state, CTDB_REPLY_CONTROL, len,
1457 struct ctdb_reply_control_old);
1458 CTDB_NO_MEMORY_VOID(ctdb, r);
1460 r->hdr.reqid = state->reqid;
1461 r->status = status;
1462 r->datalen = data.dsize;
1463 r->errorlen = 0;
1464 memcpy(&r->data[0], data.dptr, data.dsize);
1465 if (errormsg) {
1466 r->errorlen = strlen(errormsg);
1467 memcpy(&r->data[r->datalen], errormsg, r->errorlen);
1470 ret = daemon_queue_send(client, &r->hdr);
1471 if (ret != -1) {
1472 talloc_free(state);
1477 fail all pending controls to a disconnected node
1479 void ctdb_daemon_cancel_controls(struct ctdb_context *ctdb, struct ctdb_node *node)
1481 struct daemon_control_state *state;
1482 while ((state = node->pending_controls)) {
1483 DLIST_REMOVE(node->pending_controls, state);
1484 daemon_control_callback(ctdb, (uint32_t)-1, tdb_null,
1485 "node is disconnected", state);
1490 destroy a daemon_control_state
1492 static int daemon_control_destructor(struct daemon_control_state *state)
1494 if (state->node) {
1495 DLIST_REMOVE(state->node->pending_controls, state);
1497 return 0;
1501 this is called when the ctdb daemon received a ctdb request control
1502 from a local client over the unix domain socket
1504 static void daemon_request_control_from_client(struct ctdb_client *client,
1505 struct ctdb_req_control_old *c)
1507 TDB_DATA data;
1508 int res;
1509 struct daemon_control_state *state;
1510 TALLOC_CTX *tmp_ctx = talloc_new(client);
1512 if (c->hdr.destnode == CTDB_CURRENT_NODE) {
1513 c->hdr.destnode = client->ctdb->pnn;
1516 state = talloc(client, struct daemon_control_state);
1517 CTDB_NO_MEMORY_VOID(client->ctdb, state);
1519 state->client = client;
1520 state->c = talloc_steal(state, c);
1521 state->reqid = c->hdr.reqid;
1522 if (ctdb_validate_pnn(client->ctdb, c->hdr.destnode)) {
1523 state->node = client->ctdb->nodes[c->hdr.destnode];
1524 DLIST_ADD(state->node->pending_controls, state);
1525 } else {
1526 state->node = NULL;
1529 talloc_set_destructor(state, daemon_control_destructor);
1531 if (c->flags & CTDB_CTRL_FLAG_NOREPLY) {
1532 talloc_steal(tmp_ctx, state);
1535 data.dptr = &c->data[0];
1536 data.dsize = c->datalen;
1537 res = ctdb_daemon_send_control(client->ctdb, c->hdr.destnode,
1538 c->srvid, c->opcode, client->client_id,
1539 c->flags,
1540 data, daemon_control_callback,
1541 state);
1542 if (res != 0) {
1543 DEBUG(DEBUG_ERR,(__location__ " Failed to send control to remote node %u\n",
1544 c->hdr.destnode));
1547 talloc_free(tmp_ctx);
1551 register a call function
1553 int ctdb_daemon_set_call(struct ctdb_context *ctdb, uint32_t db_id,
1554 ctdb_fn_t fn, int id)
1556 struct ctdb_registered_call *call;
1557 struct ctdb_db_context *ctdb_db;
1559 ctdb_db = find_ctdb_db(ctdb, db_id);
1560 if (ctdb_db == NULL) {
1561 return -1;
1564 call = talloc(ctdb_db, struct ctdb_registered_call);
1565 call->fn = fn;
1566 call->id = id;
1568 DLIST_ADD(ctdb_db->calls, call);
1569 return 0;
1575 this local messaging handler is ugly, but is needed to prevent
1576 recursion in ctdb_send_message() when the destination node is the
1577 same as the source node
1579 struct ctdb_local_message {
1580 struct ctdb_context *ctdb;
1581 uint64_t srvid;
1582 TDB_DATA data;
1585 static void ctdb_local_message_trigger(struct tevent_context *ev,
1586 struct tevent_timer *te,
1587 struct timeval t, void *private_data)
1589 struct ctdb_local_message *m = talloc_get_type(
1590 private_data, struct ctdb_local_message);
1592 srvid_dispatch(m->ctdb->srv, m->srvid, CTDB_SRVID_ALL, m->data);
1593 talloc_free(m);
1596 static int ctdb_local_message(struct ctdb_context *ctdb, uint64_t srvid, TDB_DATA data)
1598 struct ctdb_local_message *m;
1599 m = talloc(ctdb, struct ctdb_local_message);
1600 CTDB_NO_MEMORY(ctdb, m);
1602 m->ctdb = ctdb;
1603 m->srvid = srvid;
1604 m->data = data;
1605 m->data.dptr = talloc_memdup(m, m->data.dptr, m->data.dsize);
1606 if (m->data.dptr == NULL) {
1607 talloc_free(m);
1608 return -1;
1611 /* this needs to be done as an event to prevent recursion */
1612 tevent_add_timer(ctdb->ev, m, timeval_zero(),
1613 ctdb_local_message_trigger, m);
1614 return 0;
1618 send a ctdb message
1620 int ctdb_daemon_send_message(struct ctdb_context *ctdb, uint32_t pnn,
1621 uint64_t srvid, TDB_DATA data)
1623 struct ctdb_req_message_old *r;
1624 int len;
1626 if (ctdb->methods == NULL) {
1627 DEBUG(DEBUG_INFO,(__location__ " Failed to send message. Transport is DOWN\n"));
1628 return -1;
1631 /* see if this is a message to ourselves */
1632 if (pnn == ctdb->pnn) {
1633 return ctdb_local_message(ctdb, srvid, data);
1636 len = offsetof(struct ctdb_req_message_old, data) + data.dsize;
1637 r = ctdb_transport_allocate(ctdb, ctdb, CTDB_REQ_MESSAGE, len,
1638 struct ctdb_req_message_old);
1639 CTDB_NO_MEMORY(ctdb, r);
1641 r->hdr.destnode = pnn;
1642 r->srvid = srvid;
1643 r->datalen = data.dsize;
1644 memcpy(&r->data[0], data.dptr, data.dsize);
1646 ctdb_queue_packet(ctdb, &r->hdr);
1648 talloc_free(r);
1649 return 0;
1654 struct ctdb_client_notify_list {
1655 struct ctdb_client_notify_list *next, *prev;
1656 struct ctdb_context *ctdb;
1657 uint64_t srvid;
1658 TDB_DATA data;
1662 static int ctdb_client_notify_destructor(struct ctdb_client_notify_list *nl)
1664 int ret;
1666 DEBUG(DEBUG_ERR,("Sending client notify message for srvid:%llu\n", (unsigned long long)nl->srvid));
1668 ret = ctdb_daemon_send_message(nl->ctdb, CTDB_BROADCAST_CONNECTED, (unsigned long long)nl->srvid, nl->data);
1669 if (ret != 0) {
1670 DEBUG(DEBUG_ERR,("Failed to send client notify message\n"));
1673 return 0;
1676 int32_t ctdb_control_register_notify(struct ctdb_context *ctdb, uint32_t client_id, TDB_DATA indata)
1678 struct ctdb_notify_data_old *notify = (struct ctdb_notify_data_old *)indata.dptr;
1679 struct ctdb_client *client = reqid_find(ctdb->idr, client_id, struct ctdb_client);
1680 struct ctdb_client_notify_list *nl;
1682 DEBUG(DEBUG_INFO,("Register srvid %llu for client %d\n", (unsigned long long)notify->srvid, client_id));
1684 if (indata.dsize < offsetof(struct ctdb_notify_data_old, notify_data)) {
1685 DEBUG(DEBUG_ERR,(__location__ " Too little data in control : %d\n", (int)indata.dsize));
1686 return -1;
1689 if (indata.dsize != (notify->len + offsetof(struct ctdb_notify_data_old, notify_data))) {
1690 DEBUG(DEBUG_ERR,(__location__ " Wrong amount of data in control. Got %d, expected %d\n", (int)indata.dsize, (int)(notify->len + offsetof(struct ctdb_notify_data_old, notify_data))));
1691 return -1;
1695 if (client == NULL) {
1696 DEBUG(DEBUG_ERR,(__location__ " Could not find client parent structure. You can not send this control to a remote node\n"));
1697 return -1;
1700 for(nl=client->notify; nl; nl=nl->next) {
1701 if (nl->srvid == notify->srvid) {
1702 break;
1705 if (nl != NULL) {
1706 DEBUG(DEBUG_ERR,(__location__ " Notification for srvid:%llu already exists for this client\n", (unsigned long long)notify->srvid));
1707 return -1;
1710 nl = talloc(client, struct ctdb_client_notify_list);
1711 CTDB_NO_MEMORY(ctdb, nl);
1712 nl->ctdb = ctdb;
1713 nl->srvid = notify->srvid;
1714 nl->data.dsize = notify->len;
1715 nl->data.dptr = talloc_memdup(nl, notify->notify_data,
1716 nl->data.dsize);
1717 CTDB_NO_MEMORY(ctdb, nl->data.dptr);
1719 DLIST_ADD(client->notify, nl);
1720 talloc_set_destructor(nl, ctdb_client_notify_destructor);
1722 return 0;
1725 int32_t ctdb_control_deregister_notify(struct ctdb_context *ctdb, uint32_t client_id, TDB_DATA indata)
1727 uint64_t srvid = *(uint64_t *)indata.dptr;
1728 struct ctdb_client *client = reqid_find(ctdb->idr, client_id, struct ctdb_client);
1729 struct ctdb_client_notify_list *nl;
1731 DEBUG(DEBUG_INFO,("Deregister srvid %llu for client %d\n", (unsigned long long)srvid, client_id));
1733 if (client == NULL) {
1734 DEBUG(DEBUG_ERR,(__location__ " Could not find client parent structure. You can not send this control to a remote node\n"));
1735 return -1;
1738 for(nl=client->notify; nl; nl=nl->next) {
1739 if (nl->srvid == srvid) {
1740 break;
1743 if (nl == NULL) {
1744 DEBUG(DEBUG_ERR,(__location__ " No notification for srvid:%llu found for this client\n", (unsigned long long)srvid));
1745 return -1;
1748 DLIST_REMOVE(client->notify, nl);
1749 talloc_set_destructor(nl, NULL);
1750 talloc_free(nl);
1752 return 0;
1755 struct ctdb_client *ctdb_find_client_by_pid(struct ctdb_context *ctdb, pid_t pid)
1757 struct ctdb_client_pid_list *client_pid;
1759 for (client_pid = ctdb->client_pids; client_pid; client_pid=client_pid->next) {
1760 if (client_pid->pid == pid) {
1761 return client_pid->client;
1764 return NULL;
1768 /* This control is used by samba when probing if a process (of a samba daemon)
1769 exists on the node.
1770 Samba does this when it needs/wants to check if a subrecord in one of the
1771 databases is still valied, or if it is stale and can be removed.
1772 If the node is in unhealthy or stopped state we just kill of the samba
1773 process holding htis sub-record and return to the calling samba that
1774 the process does not exist.
1775 This allows us to forcefully recall subrecords registered by samba processes
1776 on banned and stopped nodes.
1778 int32_t ctdb_control_process_exists(struct ctdb_context *ctdb, pid_t pid)
1780 struct ctdb_client *client;
1782 if (ctdb->nodes[ctdb->pnn]->flags & (NODE_FLAGS_BANNED|NODE_FLAGS_STOPPED)) {
1783 client = ctdb_find_client_by_pid(ctdb, pid);
1784 if (client != NULL) {
1785 DEBUG(DEBUG_NOTICE,(__location__ " Killing client with pid:%d on banned/stopped node\n", (int)pid));
1786 talloc_free(client);
1788 return -1;
1791 return kill(pid, 0);
1794 int ctdb_control_getnodesfile(struct ctdb_context *ctdb, uint32_t opcode, TDB_DATA indata, TDB_DATA *outdata)
1796 struct ctdb_node_map_old *node_map = NULL;
1798 CHECK_CONTROL_DATA_SIZE(0);
1800 node_map = ctdb_read_nodes_file(ctdb, ctdb->nodes_file);
1801 if (node_map == NULL) {
1802 DEBUG(DEBUG_ERR, ("Failed to read nodes file\n"));
1803 return -1;
1806 outdata->dptr = (unsigned char *)node_map;
1807 outdata->dsize = talloc_get_size(outdata->dptr);
1809 return 0;
1812 void ctdb_shutdown_sequence(struct ctdb_context *ctdb, int exit_code)
1814 if (ctdb->runstate == CTDB_RUNSTATE_SHUTDOWN) {
1815 DEBUG(DEBUG_NOTICE,("Already shutting down so will not proceed.\n"));
1816 return;
1819 DEBUG(DEBUG_NOTICE,("Shutdown sequence commencing.\n"));
1820 ctdb_set_runstate(ctdb, CTDB_RUNSTATE_SHUTDOWN);
1821 ctdb_stop_recoverd(ctdb);
1822 ctdb_stop_keepalive(ctdb);
1823 ctdb_stop_monitoring(ctdb);
1824 ctdb_release_all_ips(ctdb);
1825 ctdb_event_script(ctdb, CTDB_EVENT_SHUTDOWN);
1826 if (ctdb->methods != NULL && ctdb->methods->shutdown != NULL) {
1827 ctdb->methods->shutdown(ctdb);
1830 DEBUG(DEBUG_NOTICE,("Shutdown sequence complete, exiting.\n"));
1831 exit(exit_code);
1834 /* When forking the main daemon and the child process needs to connect
1835 * back to the daemon as a client process, this function can be used
1836 * to change the ctdb context from daemon into client mode. The child
1837 * process must be created using ctdb_fork() and not fork() -
1838 * ctdb_fork() does some necessary housekeeping.
1840 int switch_from_server_to_client(struct ctdb_context *ctdb, const char *fmt, ...)
1842 int ret;
1843 va_list ap;
1845 /* Add extra information so we can identify this in the logs */
1846 va_start(ap, fmt);
1847 debug_extra = talloc_strdup_append(talloc_vasprintf(NULL, fmt, ap), ":");
1848 va_end(ap);
1850 /* get a new event context */
1851 ctdb->ev = tevent_context_init(ctdb);
1852 tevent_loop_allow_nesting(ctdb->ev);
1854 /* Connect to main CTDB daemon */
1855 ret = ctdb_socket_connect(ctdb);
1856 if (ret != 0) {
1857 DEBUG(DEBUG_ALERT, (__location__ " Failed to init ctdb client\n"));
1858 return -1;
1861 ctdb->can_send_controls = true;
1863 return 0;