ndr.pc: needs samba-util not samba-hostconfig
[Samba/bjacke.git] / source4 / wrepl_server / wrepl_out_helpers.c
blob660e1badb7e02b8786baf9fe85902b05bb46a78a
1 /*
2 Unix SMB/CIFS implementation.
4 WINS Replication server
6 Copyright (C) Stefan Metzmacher 2005
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3 of the License, or
11 (at your option) any later version.
13 This program is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
18 You should have received a copy of the GNU General Public License
19 along with this program. If not, see <http://www.gnu.org/licenses/>.
22 #include "includes.h"
23 #include "lib/events/events.h"
24 #include "lib/socket/socket.h"
25 #include "smbd/service_task.h"
26 #include "smbd/service_stream.h"
27 #include "librpc/gen_ndr/winsrepl.h"
28 #include "wrepl_server/wrepl_server.h"
29 #include "nbt_server/wins/winsdb.h"
30 #include "libcli/composite/composite.h"
31 #include "libcli/wrepl/winsrepl.h"
32 #include "libcli/resolve/resolve.h"
33 #include "param/param.h"
35 enum wreplsrv_out_connect_stage {
36 WREPLSRV_OUT_CONNECT_STAGE_WAIT_SOCKET,
37 WREPLSRV_OUT_CONNECT_STAGE_WAIT_ASSOC_CTX,
38 WREPLSRV_OUT_CONNECT_STAGE_DONE
41 struct wreplsrv_out_connect_state {
42 enum wreplsrv_out_connect_stage stage;
43 struct composite_context *c;
44 struct wrepl_associate assoc_io;
45 enum winsrepl_partner_type type;
46 struct wreplsrv_out_connection *wreplconn;
47 struct tevent_req *subreq;
50 static void wreplsrv_out_connect_handler_treq(struct tevent_req *subreq);
52 static NTSTATUS wreplsrv_out_connect_wait_socket(struct wreplsrv_out_connect_state *state)
54 NTSTATUS status;
56 status = wrepl_connect_recv(state->subreq);
57 TALLOC_FREE(state->subreq);
58 NT_STATUS_NOT_OK_RETURN(status);
60 state->subreq = wrepl_associate_send(state,
61 state->wreplconn->service->task->event_ctx,
62 state->wreplconn->sock, &state->assoc_io);
63 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
65 tevent_req_set_callback(state->subreq,
66 wreplsrv_out_connect_handler_treq,
67 state);
69 state->stage = WREPLSRV_OUT_CONNECT_STAGE_WAIT_ASSOC_CTX;
71 return NT_STATUS_OK;
74 static NTSTATUS wreplsrv_out_connect_wait_assoc_ctx(struct wreplsrv_out_connect_state *state)
76 NTSTATUS status;
78 status = wrepl_associate_recv(state->subreq, &state->assoc_io);
79 TALLOC_FREE(state->subreq);
80 NT_STATUS_NOT_OK_RETURN(status);
82 state->wreplconn->assoc_ctx.peer_ctx = state->assoc_io.out.assoc_ctx;
83 state->wreplconn->assoc_ctx.peer_major = state->assoc_io.out.major_version;
85 if (state->type == WINSREPL_PARTNER_PUSH) {
86 if (state->wreplconn->assoc_ctx.peer_major >= 5) {
87 state->wreplconn->partner->push.wreplconn = state->wreplconn;
88 talloc_steal(state->wreplconn->partner, state->wreplconn);
89 } else {
90 state->type = WINSREPL_PARTNER_NONE;
92 } else if (state->type == WINSREPL_PARTNER_PULL) {
93 state->wreplconn->partner->pull.wreplconn = state->wreplconn;
94 talloc_steal(state->wreplconn->partner, state->wreplconn);
97 state->stage = WREPLSRV_OUT_CONNECT_STAGE_DONE;
99 return NT_STATUS_OK;
102 static void wreplsrv_out_connect_handler(struct wreplsrv_out_connect_state *state)
104 struct composite_context *c = state->c;
106 switch (state->stage) {
107 case WREPLSRV_OUT_CONNECT_STAGE_WAIT_SOCKET:
108 c->status = wreplsrv_out_connect_wait_socket(state);
109 break;
110 case WREPLSRV_OUT_CONNECT_STAGE_WAIT_ASSOC_CTX:
111 c->status = wreplsrv_out_connect_wait_assoc_ctx(state);
112 c->state = COMPOSITE_STATE_DONE;
113 break;
114 case WREPLSRV_OUT_CONNECT_STAGE_DONE:
115 c->status = NT_STATUS_INTERNAL_ERROR;
118 if (!NT_STATUS_IS_OK(c->status)) {
119 c->state = COMPOSITE_STATE_ERROR;
122 if (c->state >= COMPOSITE_STATE_DONE && c->async.fn) {
123 c->async.fn(c);
127 static void wreplsrv_out_connect_handler_treq(struct tevent_req *subreq)
129 struct wreplsrv_out_connect_state *state = tevent_req_callback_data(subreq,
130 struct wreplsrv_out_connect_state);
131 wreplsrv_out_connect_handler(state);
132 return;
135 static struct composite_context *wreplsrv_out_connect_send(struct wreplsrv_partner *partner,
136 enum winsrepl_partner_type type,
137 struct wreplsrv_out_connection *wreplconn)
139 struct composite_context *c = NULL;
140 struct wreplsrv_service *service = partner->service;
141 struct wreplsrv_out_connect_state *state = NULL;
142 struct wreplsrv_out_connection **wreplconnp = &wreplconn;
143 bool cached_connection = false;
145 c = talloc_zero(partner, struct composite_context);
146 if (!c) goto failed;
148 state = talloc_zero(c, struct wreplsrv_out_connect_state);
149 if (!state) goto failed;
150 state->c = c;
151 state->type = type;
153 c->state = COMPOSITE_STATE_IN_PROGRESS;
154 c->event_ctx = service->task->event_ctx;
155 c->private_data = state;
157 if (type == WINSREPL_PARTNER_PUSH) {
158 cached_connection = true;
159 wreplconn = partner->push.wreplconn;
160 wreplconnp = &partner->push.wreplconn;
161 } else if (type == WINSREPL_PARTNER_PULL) {
162 cached_connection = true;
163 wreplconn = partner->pull.wreplconn;
164 wreplconnp = &partner->pull.wreplconn;
167 /* we have a connection already, so use it */
168 if (wreplconn) {
169 if (wrepl_socket_is_connected(wreplconn->sock)) {
170 state->stage = WREPLSRV_OUT_CONNECT_STAGE_DONE;
171 state->wreplconn= wreplconn;
172 composite_done(c);
173 return c;
174 } else if (!cached_connection) {
175 state->stage = WREPLSRV_OUT_CONNECT_STAGE_DONE;
176 state->wreplconn= NULL;
177 composite_done(c);
178 return c;
179 } else {
180 talloc_free(wreplconn);
181 *wreplconnp = NULL;
185 wreplconn = talloc_zero(state, struct wreplsrv_out_connection);
186 if (!wreplconn) goto failed;
188 wreplconn->service = service;
189 wreplconn->partner = partner;
190 wreplconn->sock = wrepl_socket_init(wreplconn, service->task->event_ctx);
191 if (!wreplconn->sock) goto failed;
193 state->stage = WREPLSRV_OUT_CONNECT_STAGE_WAIT_SOCKET;
194 state->wreplconn= wreplconn;
195 state->subreq = wrepl_connect_send(state,
196 service->task->event_ctx,
197 wreplconn->sock,
198 partner->our_address?partner->our_address:wrepl_best_ip(service->task->lp_ctx, partner->address),
199 partner->address);
200 if (!state->subreq) goto failed;
202 tevent_req_set_callback(state->subreq,
203 wreplsrv_out_connect_handler_treq,
204 state);
206 return c;
207 failed:
208 talloc_free(c);
209 return NULL;
212 static NTSTATUS wreplsrv_out_connect_recv(struct composite_context *c, TALLOC_CTX *mem_ctx,
213 struct wreplsrv_out_connection **wreplconn)
215 NTSTATUS status;
217 status = composite_wait(c);
219 if (NT_STATUS_IS_OK(status)) {
220 struct wreplsrv_out_connect_state *state = talloc_get_type(c->private_data,
221 struct wreplsrv_out_connect_state);
222 if (state->wreplconn) {
223 *wreplconn = talloc_reference(mem_ctx, state->wreplconn);
224 if (!*wreplconn) status = NT_STATUS_NO_MEMORY;
225 } else {
226 status = NT_STATUS_CONNECTION_DISCONNECTED;
230 talloc_free(c);
231 return status;
235 struct wreplsrv_pull_table_io {
236 struct {
237 struct wreplsrv_partner *partner;
238 uint32_t num_owners;
239 struct wrepl_wins_owner *owners;
240 } in;
241 struct {
242 uint32_t num_owners;
243 struct wrepl_wins_owner *owners;
244 } out;
247 enum wreplsrv_pull_table_stage {
248 WREPLSRV_PULL_TABLE_STAGE_WAIT_CONNECTION,
249 WREPLSRV_PULL_TABLE_STAGE_WAIT_TABLE_REPLY,
250 WREPLSRV_PULL_TABLE_STAGE_DONE
253 struct wreplsrv_pull_table_state {
254 enum wreplsrv_pull_table_stage stage;
255 struct composite_context *c;
256 struct wrepl_pull_table table_io;
257 struct wreplsrv_pull_table_io *io;
258 struct composite_context *creq;
259 struct wreplsrv_out_connection *wreplconn;
260 struct tevent_req *subreq;
263 static void wreplsrv_pull_table_handler_treq(struct tevent_req *subreq);
265 static NTSTATUS wreplsrv_pull_table_wait_connection(struct wreplsrv_pull_table_state *state)
267 NTSTATUS status;
269 status = wreplsrv_out_connect_recv(state->creq, state, &state->wreplconn);
270 NT_STATUS_NOT_OK_RETURN(status);
272 state->table_io.in.assoc_ctx = state->wreplconn->assoc_ctx.peer_ctx;
273 state->subreq = wrepl_pull_table_send(state,
274 state->wreplconn->service->task->event_ctx,
275 state->wreplconn->sock, &state->table_io);
276 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
278 tevent_req_set_callback(state->subreq,
279 wreplsrv_pull_table_handler_treq,
280 state);
282 state->stage = WREPLSRV_PULL_TABLE_STAGE_WAIT_TABLE_REPLY;
284 return NT_STATUS_OK;
287 static NTSTATUS wreplsrv_pull_table_wait_table_reply(struct wreplsrv_pull_table_state *state)
289 NTSTATUS status;
291 status = wrepl_pull_table_recv(state->subreq, state, &state->table_io);
292 TALLOC_FREE(state->subreq);
293 NT_STATUS_NOT_OK_RETURN(status);
295 state->stage = WREPLSRV_PULL_TABLE_STAGE_DONE;
297 return NT_STATUS_OK;
300 static void wreplsrv_pull_table_handler(struct wreplsrv_pull_table_state *state)
302 struct composite_context *c = state->c;
304 switch (state->stage) {
305 case WREPLSRV_PULL_TABLE_STAGE_WAIT_CONNECTION:
306 c->status = wreplsrv_pull_table_wait_connection(state);
307 break;
308 case WREPLSRV_PULL_TABLE_STAGE_WAIT_TABLE_REPLY:
309 c->status = wreplsrv_pull_table_wait_table_reply(state);
310 c->state = COMPOSITE_STATE_DONE;
311 break;
312 case WREPLSRV_PULL_TABLE_STAGE_DONE:
313 c->status = NT_STATUS_INTERNAL_ERROR;
316 if (!NT_STATUS_IS_OK(c->status)) {
317 c->state = COMPOSITE_STATE_ERROR;
320 if (c->state >= COMPOSITE_STATE_DONE && c->async.fn) {
321 c->async.fn(c);
325 static void wreplsrv_pull_table_handler_creq(struct composite_context *creq)
327 struct wreplsrv_pull_table_state *state = talloc_get_type(creq->async.private_data,
328 struct wreplsrv_pull_table_state);
329 wreplsrv_pull_table_handler(state);
330 return;
333 static void wreplsrv_pull_table_handler_treq(struct tevent_req *subreq)
335 struct wreplsrv_pull_table_state *state = tevent_req_callback_data(subreq,
336 struct wreplsrv_pull_table_state);
337 wreplsrv_pull_table_handler(state);
338 return;
341 static struct composite_context *wreplsrv_pull_table_send(TALLOC_CTX *mem_ctx, struct wreplsrv_pull_table_io *io)
343 struct composite_context *c = NULL;
344 struct wreplsrv_service *service = io->in.partner->service;
345 struct wreplsrv_pull_table_state *state = NULL;
347 c = talloc_zero(mem_ctx, struct composite_context);
348 if (!c) goto failed;
350 state = talloc_zero(c, struct wreplsrv_pull_table_state);
351 if (!state) goto failed;
352 state->c = c;
353 state->io = io;
355 c->state = COMPOSITE_STATE_IN_PROGRESS;
356 c->event_ctx = service->task->event_ctx;
357 c->private_data = state;
359 if (io->in.num_owners) {
360 state->table_io.out.num_partners = io->in.num_owners;
361 state->table_io.out.partners = io->in.owners;
362 state->stage = WREPLSRV_PULL_TABLE_STAGE_DONE;
363 composite_done(c);
364 return c;
367 state->stage = WREPLSRV_PULL_TABLE_STAGE_WAIT_CONNECTION;
368 state->creq = wreplsrv_out_connect_send(io->in.partner, WINSREPL_PARTNER_PULL, NULL);
369 if (!state->creq) goto failed;
371 state->creq->async.fn = wreplsrv_pull_table_handler_creq;
372 state->creq->async.private_data = state;
374 return c;
375 failed:
376 talloc_free(c);
377 return NULL;
380 static NTSTATUS wreplsrv_pull_table_recv(struct composite_context *c, TALLOC_CTX *mem_ctx,
381 struct wreplsrv_pull_table_io *io)
383 NTSTATUS status;
385 status = composite_wait(c);
387 if (NT_STATUS_IS_OK(status)) {
388 struct wreplsrv_pull_table_state *state = talloc_get_type(c->private_data,
389 struct wreplsrv_pull_table_state);
390 io->out.num_owners = state->table_io.out.num_partners;
391 io->out.owners = talloc_reference(mem_ctx, state->table_io.out.partners);
394 talloc_free(c);
395 return status;
398 struct wreplsrv_pull_names_io {
399 struct {
400 struct wreplsrv_partner *partner;
401 struct wreplsrv_out_connection *wreplconn;
402 struct wrepl_wins_owner owner;
403 } in;
404 struct {
405 uint32_t num_names;
406 struct wrepl_name *names;
407 } out;
410 enum wreplsrv_pull_names_stage {
411 WREPLSRV_PULL_NAMES_STAGE_WAIT_CONNECTION,
412 WREPLSRV_PULL_NAMES_STAGE_WAIT_SEND_REPLY,
413 WREPLSRV_PULL_NAMES_STAGE_DONE
416 struct wreplsrv_pull_names_state {
417 enum wreplsrv_pull_names_stage stage;
418 struct composite_context *c;
419 struct wrepl_pull_names pull_io;
420 struct wreplsrv_pull_names_io *io;
421 struct composite_context *creq;
422 struct wreplsrv_out_connection *wreplconn;
423 struct tevent_req *subreq;
426 static void wreplsrv_pull_names_handler_treq(struct tevent_req *subreq);
428 static NTSTATUS wreplsrv_pull_names_wait_connection(struct wreplsrv_pull_names_state *state)
430 NTSTATUS status;
432 status = wreplsrv_out_connect_recv(state->creq, state, &state->wreplconn);
433 NT_STATUS_NOT_OK_RETURN(status);
435 state->pull_io.in.assoc_ctx = state->wreplconn->assoc_ctx.peer_ctx;
436 state->pull_io.in.partner = state->io->in.owner;
437 state->subreq = wrepl_pull_names_send(state,
438 state->wreplconn->service->task->event_ctx,
439 state->wreplconn->sock,
440 &state->pull_io);
441 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
443 tevent_req_set_callback(state->subreq,
444 wreplsrv_pull_names_handler_treq,
445 state);
447 state->stage = WREPLSRV_PULL_NAMES_STAGE_WAIT_SEND_REPLY;
449 return NT_STATUS_OK;
452 static NTSTATUS wreplsrv_pull_names_wait_send_reply(struct wreplsrv_pull_names_state *state)
454 NTSTATUS status;
456 status = wrepl_pull_names_recv(state->subreq, state, &state->pull_io);
457 TALLOC_FREE(state->subreq);
458 NT_STATUS_NOT_OK_RETURN(status);
460 state->stage = WREPLSRV_PULL_NAMES_STAGE_DONE;
462 return NT_STATUS_OK;
465 static void wreplsrv_pull_names_handler(struct wreplsrv_pull_names_state *state)
467 struct composite_context *c = state->c;
469 switch (state->stage) {
470 case WREPLSRV_PULL_NAMES_STAGE_WAIT_CONNECTION:
471 c->status = wreplsrv_pull_names_wait_connection(state);
472 break;
473 case WREPLSRV_PULL_NAMES_STAGE_WAIT_SEND_REPLY:
474 c->status = wreplsrv_pull_names_wait_send_reply(state);
475 c->state = COMPOSITE_STATE_DONE;
476 break;
477 case WREPLSRV_PULL_NAMES_STAGE_DONE:
478 c->status = NT_STATUS_INTERNAL_ERROR;
481 if (!NT_STATUS_IS_OK(c->status)) {
482 c->state = COMPOSITE_STATE_ERROR;
485 if (c->state >= COMPOSITE_STATE_DONE && c->async.fn) {
486 c->async.fn(c);
490 static void wreplsrv_pull_names_handler_creq(struct composite_context *creq)
492 struct wreplsrv_pull_names_state *state = talloc_get_type(creq->async.private_data,
493 struct wreplsrv_pull_names_state);
494 wreplsrv_pull_names_handler(state);
495 return;
498 static void wreplsrv_pull_names_handler_treq(struct tevent_req *subreq)
500 struct wreplsrv_pull_names_state *state = tevent_req_callback_data(subreq,
501 struct wreplsrv_pull_names_state);
502 wreplsrv_pull_names_handler(state);
503 return;
506 static struct composite_context *wreplsrv_pull_names_send(TALLOC_CTX *mem_ctx, struct wreplsrv_pull_names_io *io)
508 struct composite_context *c = NULL;
509 struct wreplsrv_service *service = io->in.partner->service;
510 struct wreplsrv_pull_names_state *state = NULL;
511 enum winsrepl_partner_type partner_type = WINSREPL_PARTNER_PULL;
513 if (io->in.wreplconn) partner_type = WINSREPL_PARTNER_NONE;
515 c = talloc_zero(mem_ctx, struct composite_context);
516 if (!c) goto failed;
518 state = talloc_zero(c, struct wreplsrv_pull_names_state);
519 if (!state) goto failed;
520 state->c = c;
521 state->io = io;
523 c->state = COMPOSITE_STATE_IN_PROGRESS;
524 c->event_ctx = service->task->event_ctx;
525 c->private_data = state;
527 state->stage = WREPLSRV_PULL_NAMES_STAGE_WAIT_CONNECTION;
528 state->creq = wreplsrv_out_connect_send(io->in.partner, partner_type, io->in.wreplconn);
529 if (!state->creq) goto failed;
531 state->creq->async.fn = wreplsrv_pull_names_handler_creq;
532 state->creq->async.private_data = state;
534 return c;
535 failed:
536 talloc_free(c);
537 return NULL;
540 static NTSTATUS wreplsrv_pull_names_recv(struct composite_context *c, TALLOC_CTX *mem_ctx,
541 struct wreplsrv_pull_names_io *io)
543 NTSTATUS status;
545 status = composite_wait(c);
547 if (NT_STATUS_IS_OK(status)) {
548 struct wreplsrv_pull_names_state *state = talloc_get_type(c->private_data,
549 struct wreplsrv_pull_names_state);
550 io->out.num_names = state->pull_io.out.num_names;
551 io->out.names = talloc_reference(mem_ctx, state->pull_io.out.names);
554 talloc_free(c);
555 return status;
559 enum wreplsrv_pull_cycle_stage {
560 WREPLSRV_PULL_CYCLE_STAGE_WAIT_TABLE_REPLY,
561 WREPLSRV_PULL_CYCLE_STAGE_WAIT_SEND_REPLIES,
562 WREPLSRV_PULL_CYCLE_STAGE_WAIT_STOP_ASSOC,
563 WREPLSRV_PULL_CYCLE_STAGE_DONE
566 struct wreplsrv_pull_cycle_state {
567 enum wreplsrv_pull_cycle_stage stage;
568 struct composite_context *c;
569 struct wreplsrv_pull_cycle_io *io;
570 struct wreplsrv_pull_table_io table_io;
571 uint32_t current;
572 struct wreplsrv_pull_names_io names_io;
573 struct composite_context *creq;
574 struct wrepl_associate_stop assoc_stop_io;
575 struct tevent_req *subreq;
578 static void wreplsrv_pull_cycle_handler_creq(struct composite_context *creq);
579 static void wreplsrv_pull_cycle_handler_treq(struct tevent_req *subreq);
581 static NTSTATUS wreplsrv_pull_cycle_next_owner_do_work(struct wreplsrv_pull_cycle_state *state)
583 struct wreplsrv_owner *current_owner=NULL;
584 struct wreplsrv_owner *local_owner;
585 uint32_t i;
586 uint64_t old_max_version = 0;
587 bool do_pull = false;
589 for (i=state->current; i < state->table_io.out.num_owners; i++) {
590 current_owner = wreplsrv_find_owner(state->io->in.partner->service,
591 state->io->in.partner->pull.table,
592 state->table_io.out.owners[i].address);
594 local_owner = wreplsrv_find_owner(state->io->in.partner->service,
595 state->io->in.partner->service->table,
596 state->table_io.out.owners[i].address);
598 * this means we are ourself the current owner,
599 * and we don't want replicate ourself
601 if (!current_owner) continue;
604 * this means we don't have any records of this owner
605 * so fetch them
607 if (!local_owner) {
608 do_pull = true;
610 break;
614 * this means the remote partner has some new records of this owner
615 * fetch them
617 if (current_owner->owner.max_version > local_owner->owner.max_version) {
618 do_pull = true;
619 old_max_version = local_owner->owner.max_version;
620 break;
623 state->current = i;
625 if (do_pull) {
626 state->names_io.in.partner = state->io->in.partner;
627 state->names_io.in.wreplconn = state->io->in.wreplconn;
628 state->names_io.in.owner = current_owner->owner;
629 state->names_io.in.owner.min_version = old_max_version + 1;
630 state->creq = wreplsrv_pull_names_send(state, &state->names_io);
631 NT_STATUS_HAVE_NO_MEMORY(state->creq);
633 state->creq->async.fn = wreplsrv_pull_cycle_handler_creq;
634 state->creq->async.private_data = state;
636 return STATUS_MORE_ENTRIES;
639 return NT_STATUS_OK;
642 static NTSTATUS wreplsrv_pull_cycle_next_owner_wrapper(struct wreplsrv_pull_cycle_state *state)
644 NTSTATUS status;
646 status = wreplsrv_pull_cycle_next_owner_do_work(state);
647 if (NT_STATUS_IS_OK(status)) {
648 state->stage = WREPLSRV_PULL_CYCLE_STAGE_DONE;
649 } else if (NT_STATUS_EQUAL(STATUS_MORE_ENTRIES, status)) {
650 state->stage = WREPLSRV_PULL_CYCLE_STAGE_WAIT_SEND_REPLIES;
651 status = NT_STATUS_OK;
654 if (state->stage == WREPLSRV_PULL_CYCLE_STAGE_DONE && state->io->in.wreplconn) {
655 state->assoc_stop_io.in.assoc_ctx = state->io->in.wreplconn->assoc_ctx.peer_ctx;
656 state->assoc_stop_io.in.reason = 0;
657 state->subreq = wrepl_associate_stop_send(state,
658 state->io->in.wreplconn->service->task->event_ctx,
659 state->io->in.wreplconn->sock,
660 &state->assoc_stop_io);
661 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
663 tevent_req_set_callback(state->subreq,
664 wreplsrv_pull_cycle_handler_treq,
665 state);
667 state->stage = WREPLSRV_PULL_CYCLE_STAGE_WAIT_STOP_ASSOC;
670 return status;
673 static NTSTATUS wreplsrv_pull_cycle_wait_table_reply(struct wreplsrv_pull_cycle_state *state)
675 NTSTATUS status;
676 uint32_t i;
678 status = wreplsrv_pull_table_recv(state->creq, state, &state->table_io);
679 NT_STATUS_NOT_OK_RETURN(status);
681 /* update partner table */
682 for (i=0; i < state->table_io.out.num_owners; i++) {
683 status = wreplsrv_add_table(state->io->in.partner->service,
684 state->io->in.partner,
685 &state->io->in.partner->pull.table,
686 state->table_io.out.owners[i].address,
687 state->table_io.out.owners[i].max_version);
688 NT_STATUS_NOT_OK_RETURN(status);
691 status = wreplsrv_pull_cycle_next_owner_wrapper(state);
692 NT_STATUS_NOT_OK_RETURN(status);
694 return status;
697 static NTSTATUS wreplsrv_pull_cycle_apply_records(struct wreplsrv_pull_cycle_state *state)
699 NTSTATUS status;
701 status = wreplsrv_apply_records(state->io->in.partner,
702 &state->names_io.in.owner,
703 state->names_io.out.num_names,
704 state->names_io.out.names);
705 NT_STATUS_NOT_OK_RETURN(status);
707 talloc_free(state->names_io.out.names);
708 ZERO_STRUCT(state->names_io);
710 return NT_STATUS_OK;
713 static NTSTATUS wreplsrv_pull_cycle_wait_send_replies(struct wreplsrv_pull_cycle_state *state)
715 NTSTATUS status;
717 status = wreplsrv_pull_names_recv(state->creq, state, &state->names_io);
718 NT_STATUS_NOT_OK_RETURN(status);
721 * TODO: this should maybe an async call,
722 * because we may need some network access
723 * for conflict resolving
725 status = wreplsrv_pull_cycle_apply_records(state);
726 NT_STATUS_NOT_OK_RETURN(status);
728 status = wreplsrv_pull_cycle_next_owner_wrapper(state);
729 NT_STATUS_NOT_OK_RETURN(status);
731 return status;
734 static NTSTATUS wreplsrv_pull_cycle_wait_stop_assoc(struct wreplsrv_pull_cycle_state *state)
736 NTSTATUS status;
738 status = wrepl_associate_stop_recv(state->subreq, &state->assoc_stop_io);
739 TALLOC_FREE(state->subreq);
740 NT_STATUS_NOT_OK_RETURN(status);
742 state->stage = WREPLSRV_PULL_CYCLE_STAGE_DONE;
744 return status;
747 static void wreplsrv_pull_cycle_handler(struct wreplsrv_pull_cycle_state *state)
749 struct composite_context *c = state->c;
751 switch (state->stage) {
752 case WREPLSRV_PULL_CYCLE_STAGE_WAIT_TABLE_REPLY:
753 c->status = wreplsrv_pull_cycle_wait_table_reply(state);
754 break;
755 case WREPLSRV_PULL_CYCLE_STAGE_WAIT_SEND_REPLIES:
756 c->status = wreplsrv_pull_cycle_wait_send_replies(state);
757 break;
758 case WREPLSRV_PULL_CYCLE_STAGE_WAIT_STOP_ASSOC:
759 c->status = wreplsrv_pull_cycle_wait_stop_assoc(state);
760 break;
761 case WREPLSRV_PULL_CYCLE_STAGE_DONE:
762 c->status = NT_STATUS_INTERNAL_ERROR;
765 if (state->stage == WREPLSRV_PULL_CYCLE_STAGE_DONE) {
766 c->state = COMPOSITE_STATE_DONE;
769 if (!NT_STATUS_IS_OK(c->status)) {
770 c->state = COMPOSITE_STATE_ERROR;
773 if (c->state >= COMPOSITE_STATE_DONE && c->async.fn) {
774 c->async.fn(c);
778 static void wreplsrv_pull_cycle_handler_creq(struct composite_context *creq)
780 struct wreplsrv_pull_cycle_state *state = talloc_get_type(creq->async.private_data,
781 struct wreplsrv_pull_cycle_state);
782 wreplsrv_pull_cycle_handler(state);
783 return;
786 static void wreplsrv_pull_cycle_handler_treq(struct tevent_req *subreq)
788 struct wreplsrv_pull_cycle_state *state = tevent_req_callback_data(subreq,
789 struct wreplsrv_pull_cycle_state);
790 wreplsrv_pull_cycle_handler(state);
791 return;
794 struct composite_context *wreplsrv_pull_cycle_send(TALLOC_CTX *mem_ctx, struct wreplsrv_pull_cycle_io *io)
796 struct composite_context *c = NULL;
797 struct wreplsrv_service *service = io->in.partner->service;
798 struct wreplsrv_pull_cycle_state *state = NULL;
800 c = talloc_zero(mem_ctx, struct composite_context);
801 if (!c) goto failed;
803 state = talloc_zero(c, struct wreplsrv_pull_cycle_state);
804 if (!state) goto failed;
805 state->c = c;
806 state->io = io;
808 c->state = COMPOSITE_STATE_IN_PROGRESS;
809 c->event_ctx = service->task->event_ctx;
810 c->private_data = state;
812 state->stage = WREPLSRV_PULL_CYCLE_STAGE_WAIT_TABLE_REPLY;
813 state->table_io.in.partner = io->in.partner;
814 state->table_io.in.num_owners = io->in.num_owners;
815 state->table_io.in.owners = io->in.owners;
816 state->creq = wreplsrv_pull_table_send(state, &state->table_io);
817 if (!state->creq) goto failed;
819 state->creq->async.fn = wreplsrv_pull_cycle_handler_creq;
820 state->creq->async.private_data = state;
822 return c;
823 failed:
824 talloc_free(c);
825 return NULL;
828 NTSTATUS wreplsrv_pull_cycle_recv(struct composite_context *c)
830 NTSTATUS status;
832 status = composite_wait(c);
834 talloc_free(c);
835 return status;
838 enum wreplsrv_push_notify_stage {
839 WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_CONNECT,
840 WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_UPDATE,
841 WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_INFORM,
842 WREPLSRV_PUSH_NOTIFY_STAGE_DONE
845 struct wreplsrv_push_notify_state {
846 enum wreplsrv_push_notify_stage stage;
847 struct composite_context *c;
848 struct wreplsrv_push_notify_io *io;
849 enum wrepl_replication_cmd command;
850 bool full_table;
851 struct wrepl_send_ctrl ctrl;
852 struct wrepl_packet req_packet;
853 struct wrepl_packet *rep_packet;
854 struct composite_context *creq;
855 struct wreplsrv_out_connection *wreplconn;
856 struct tevent_req *subreq;
859 static void wreplsrv_push_notify_handler_creq(struct composite_context *creq);
860 static void wreplsrv_push_notify_handler_treq(struct tevent_req *subreq);
862 static NTSTATUS wreplsrv_push_notify_update(struct wreplsrv_push_notify_state *state)
864 struct wreplsrv_service *service = state->io->in.partner->service;
865 struct wrepl_packet *req = &state->req_packet;
866 struct wrepl_replication *repl_out = &state->req_packet.message.replication;
867 struct wrepl_table *table_out = &state->req_packet.message.replication.info.table;
868 NTSTATUS status;
870 /* prepare the outgoing request */
871 req->opcode = WREPL_OPCODE_BITS;
872 req->assoc_ctx = state->wreplconn->assoc_ctx.peer_ctx;
873 req->mess_type = WREPL_REPLICATION;
875 repl_out->command = state->command;
877 status = wreplsrv_fill_wrepl_table(service, state, table_out,
878 service->wins_db->local_owner, state->full_table);
879 NT_STATUS_NOT_OK_RETURN(status);
881 /* queue the request */
882 state->subreq = wrepl_request_send(state,
883 state->wreplconn->service->task->event_ctx,
884 state->wreplconn->sock, req, NULL);
885 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
887 tevent_req_set_callback(state->subreq,
888 wreplsrv_push_notify_handler_treq,
889 state);
891 state->stage = WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_UPDATE;
893 return NT_STATUS_OK;
896 static NTSTATUS wreplsrv_push_notify_inform(struct wreplsrv_push_notify_state *state)
898 struct wreplsrv_service *service = state->io->in.partner->service;
899 struct wrepl_packet *req = &state->req_packet;
900 struct wrepl_replication *repl_out = &state->req_packet.message.replication;
901 struct wrepl_table *table_out = &state->req_packet.message.replication.info.table;
902 NTSTATUS status;
904 req->opcode = WREPL_OPCODE_BITS;
905 req->assoc_ctx = state->wreplconn->assoc_ctx.peer_ctx;
906 req->mess_type = WREPL_REPLICATION;
908 repl_out->command = state->command;
910 status = wreplsrv_fill_wrepl_table(service, state, table_out,
911 service->wins_db->local_owner, state->full_table);
912 NT_STATUS_NOT_OK_RETURN(status);
914 /* we won't get a reply to a inform message */
915 state->ctrl.send_only = true;
917 state->subreq = wrepl_request_send(state,
918 state->wreplconn->service->task->event_ctx,
919 state->wreplconn->sock, req, &state->ctrl);
920 NT_STATUS_HAVE_NO_MEMORY(state->subreq);
922 tevent_req_set_callback(state->subreq,
923 wreplsrv_push_notify_handler_treq,
924 state);
926 state->stage = WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_INFORM;
928 return NT_STATUS_OK;
931 static NTSTATUS wreplsrv_push_notify_wait_connect(struct wreplsrv_push_notify_state *state)
933 NTSTATUS status;
935 status = wreplsrv_out_connect_recv(state->creq, state, &state->wreplconn);
936 NT_STATUS_NOT_OK_RETURN(status);
938 /* is the peer doesn't support inform fallback to update */
939 switch (state->command) {
940 case WREPL_REPL_INFORM:
941 if (state->wreplconn->assoc_ctx.peer_major < 5) {
942 state->command = WREPL_REPL_UPDATE;
944 break;
945 case WREPL_REPL_INFORM2:
946 if (state->wreplconn->assoc_ctx.peer_major < 5) {
947 state->command = WREPL_REPL_UPDATE2;
949 break;
950 default:
951 break;
954 switch (state->command) {
955 case WREPL_REPL_UPDATE:
956 state->full_table = true;
957 return wreplsrv_push_notify_update(state);
958 case WREPL_REPL_UPDATE2:
959 state->full_table = false;
960 return wreplsrv_push_notify_update(state);
961 case WREPL_REPL_INFORM:
962 state->full_table = true;
963 return wreplsrv_push_notify_inform(state);
964 case WREPL_REPL_INFORM2:
965 state->full_table = false;
966 return wreplsrv_push_notify_inform(state);
967 default:
968 return NT_STATUS_INTERNAL_ERROR;
972 static NTSTATUS wreplsrv_push_notify_wait_update(struct wreplsrv_push_notify_state *state)
974 struct wreplsrv_in_connection *wrepl_in;
975 struct tstream_context *stream;
976 NTSTATUS status;
978 status = wrepl_request_recv(state->subreq, state, NULL);
979 TALLOC_FREE(state->subreq);
980 NT_STATUS_NOT_OK_RETURN(status);
983 * now we need to convert the wrepl_socket (client connection)
984 * into a wreplsrv_in_connection (server connection), because
985 * we'll act as a server on this connection after the WREPL_REPL_UPDATE*
986 * message is received by the peer.
989 status = wrepl_socket_split_stream(state->wreplconn->sock, state, &stream);
990 NT_STATUS_NOT_OK_RETURN(status);
993 * now create a wreplsrv_in_connection,
994 * on which we act as server
996 * NOTE: stream will be stolen by
997 * wreplsrv_in_connection_merge()
999 status = wreplsrv_in_connection_merge(state->io->in.partner,
1000 state->wreplconn->assoc_ctx.peer_ctx,
1001 &stream,
1002 &wrepl_in);
1003 NT_STATUS_NOT_OK_RETURN(status);
1005 /* now we can free the wreplsrv_out_connection */
1006 TALLOC_FREE(state->wreplconn);
1008 state->stage = WREPLSRV_PUSH_NOTIFY_STAGE_DONE;
1009 return NT_STATUS_OK;
1012 static NTSTATUS wreplsrv_push_notify_wait_inform(struct wreplsrv_push_notify_state *state)
1014 NTSTATUS status;
1016 status = wrepl_request_recv(state->subreq, state, NULL);
1017 TALLOC_FREE(state->subreq);
1018 NT_STATUS_NOT_OK_RETURN(status);
1020 state->stage = WREPLSRV_PUSH_NOTIFY_STAGE_DONE;
1021 return status;
1024 static void wreplsrv_push_notify_handler(struct wreplsrv_push_notify_state *state)
1026 struct composite_context *c = state->c;
1028 switch (state->stage) {
1029 case WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_CONNECT:
1030 c->status = wreplsrv_push_notify_wait_connect(state);
1031 break;
1032 case WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_UPDATE:
1033 c->status = wreplsrv_push_notify_wait_update(state);
1034 break;
1035 case WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_INFORM:
1036 c->status = wreplsrv_push_notify_wait_inform(state);
1037 break;
1038 case WREPLSRV_PUSH_NOTIFY_STAGE_DONE:
1039 c->status = NT_STATUS_INTERNAL_ERROR;
1042 if (state->stage == WREPLSRV_PUSH_NOTIFY_STAGE_DONE) {
1043 c->state = COMPOSITE_STATE_DONE;
1046 if (!NT_STATUS_IS_OK(c->status)) {
1047 c->state = COMPOSITE_STATE_ERROR;
1050 if (c->state >= COMPOSITE_STATE_DONE && c->async.fn) {
1051 c->async.fn(c);
1055 static void wreplsrv_push_notify_handler_creq(struct composite_context *creq)
1057 struct wreplsrv_push_notify_state *state = talloc_get_type(creq->async.private_data,
1058 struct wreplsrv_push_notify_state);
1059 wreplsrv_push_notify_handler(state);
1060 return;
1063 static void wreplsrv_push_notify_handler_treq(struct tevent_req *subreq)
1065 struct wreplsrv_push_notify_state *state = tevent_req_callback_data(subreq,
1066 struct wreplsrv_push_notify_state);
1067 wreplsrv_push_notify_handler(state);
1068 return;
1071 struct composite_context *wreplsrv_push_notify_send(TALLOC_CTX *mem_ctx, struct wreplsrv_push_notify_io *io)
1073 struct composite_context *c = NULL;
1074 struct wreplsrv_service *service = io->in.partner->service;
1075 struct wreplsrv_push_notify_state *state = NULL;
1076 enum winsrepl_partner_type partner_type;
1078 c = talloc_zero(mem_ctx, struct composite_context);
1079 if (!c) goto failed;
1081 state = talloc_zero(c, struct wreplsrv_push_notify_state);
1082 if (!state) goto failed;
1083 state->c = c;
1084 state->io = io;
1086 if (io->in.inform) {
1087 /* we can cache the connection in partner->push->wreplconn */
1088 partner_type = WINSREPL_PARTNER_PUSH;
1089 if (io->in.propagate) {
1090 state->command = WREPL_REPL_INFORM2;
1091 } else {
1092 state->command = WREPL_REPL_INFORM;
1094 } else {
1095 /* we can NOT cache the connection */
1096 partner_type = WINSREPL_PARTNER_NONE;
1097 if (io->in.propagate) {
1098 state->command = WREPL_REPL_UPDATE2;
1099 } else {
1100 state->command = WREPL_REPL_UPDATE;
1104 c->state = COMPOSITE_STATE_IN_PROGRESS;
1105 c->event_ctx = service->task->event_ctx;
1106 c->private_data = state;
1108 state->stage = WREPLSRV_PUSH_NOTIFY_STAGE_WAIT_CONNECT;
1109 state->creq = wreplsrv_out_connect_send(io->in.partner, partner_type, NULL);
1110 if (!state->creq) goto failed;
1112 state->creq->async.fn = wreplsrv_push_notify_handler_creq;
1113 state->creq->async.private_data = state;
1115 return c;
1116 failed:
1117 talloc_free(c);
1118 return NULL;
1121 NTSTATUS wreplsrv_push_notify_recv(struct composite_context *c)
1123 NTSTATUS status;
1125 status = composite_wait(c);
1127 talloc_free(c);
1128 return status;