2 Unix SMB/CIFS implementation.
6 Copyright (C) Andrew Tridgell 2006
8 This program is free software; you can redistribute it and/or modify
9 it under the terms of the GNU General Public License as published by
10 the Free Software Foundation; either version 3 of the License, or
11 (at your option) any later version.
13 This program is distributed in the hope that it will be useful,
14 but WITHOUT ANY WARRANTY; without even the implied warranty of
15 MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
16 GNU General Public License for more details.
18 You should have received a copy of the GNU General Public License
19 along with this program. If not, see <http://www.gnu.org/licenses/>.
23 #include "torture/torture.h"
24 #include "libcli/raw/libcliraw.h"
25 #include "system/time.h"
26 #include "system/filesys.h"
27 #include "libcli/libcli.h"
28 #include "torture/util.h"
29 #include "lib/events/events.h"
30 #include "lib/cmdline/popt_common.h"
31 #include "libcli/composite/composite.h"
32 #include "libcli/smb_composite/smb_composite.h"
33 #include "libcli/resolve/resolve.h"
34 #include "param/param.h"
36 #define BASEDIR "\\benchlock"
37 #define FNAME BASEDIR "\\lock.dat"
40 static int lock_failed;
41 static int num_connected;
43 enum lock_stage {LOCK_INITIAL, LOCK_LOCK, LOCK_UNLOCK};
45 struct benchlock_state {
46 struct torture_context *tctx;
47 struct event_context *ev;
48 struct smbcli_tree *tree;
52 enum lock_stage stage;
57 struct smbcli_request *req;
58 struct smb_composite_connect reconnect;
59 struct timed_event *te;
61 /* these are used for reconnections */
62 const char **dest_ports;
63 const char *dest_host;
64 const char *called_name;
65 const char *service_type;
68 static void lock_completion(struct smbcli_request *);
71 send the next lock request
73 static void lock_send(struct benchlock_state *state)
76 struct smb_lock_entry lock;
78 switch (state->stage) {
80 io.lockx.in.ulock_cnt = 0;
81 io.lockx.in.lock_cnt = 1;
82 state->lock_offset = 0;
83 state->unlock_offset = 0;
84 lock.offset = state->lock_offset;
87 io.lockx.in.ulock_cnt = 0;
88 io.lockx.in.lock_cnt = 1;
89 state->lock_offset = (state->lock_offset+1)%(nprocs+1);
90 lock.offset = state->lock_offset;
93 io.lockx.in.ulock_cnt = 1;
94 io.lockx.in.lock_cnt = 0;
95 lock.offset = state->unlock_offset;
96 state->unlock_offset = (state->unlock_offset+1)%(nprocs+1);
101 lock.pid = state->tree->session->pid;
103 io.lockx.level = RAW_LOCK_LOCKX;
104 io.lockx.in.mode = LOCKING_ANDX_LARGE_FILES;
105 io.lockx.in.timeout = 100000;
106 io.lockx.in.locks = &lock;
107 io.lockx.in.file.fnum = state->fnum;
109 state->req = smb_raw_lock_send(state->tree, &io);
110 if (state->req == NULL) {
111 DEBUG(0,("Failed to setup lock\n"));
114 state->req->async.private = state;
115 state->req->async.fn = lock_completion;
118 static void reopen_connection(struct event_context *ev, struct timed_event *te,
119 struct timeval t, void *private_data);
122 static void reopen_file(struct event_context *ev, struct timed_event *te,
123 struct timeval t, void *private_data)
125 struct benchlock_state *state = (struct benchlock_state *)private_data;
127 /* reestablish our open file */
128 state->fnum = smbcli_open(state->tree, FNAME, O_RDWR|O_CREAT, DENY_NONE);
129 if (state->fnum == -1) {
130 printf("Failed to open %s on connection %d\n", FNAME, state->client_num);
136 DEBUG(0,("reconnect to %s finished (%u connected)\n", state->dest_host,
139 state->stage = LOCK_INITIAL;
144 complete an async reconnect
146 static void reopen_connection_complete(struct composite_context *ctx)
148 struct benchlock_state *state = (struct benchlock_state *)ctx->async.private_data;
150 struct smb_composite_connect *io = &state->reconnect;
152 status = smb_composite_connect_recv(ctx, state->mem_ctx);
153 if (!NT_STATUS_IS_OK(status)) {
154 talloc_free(state->te);
155 state->te = event_add_timed(state->ev, state->mem_ctx,
156 timeval_current_ofs(1,0),
157 reopen_connection, state);
161 talloc_free(state->tree);
162 state->tree = io->out.tree;
164 /* do the reopen as a separate event */
165 event_add_timed(state->ev, state->mem_ctx, timeval_zero(), reopen_file, state);
173 static void reopen_connection(struct event_context *ev, struct timed_event *te,
174 struct timeval t, void *private_data)
176 struct benchlock_state *state = (struct benchlock_state *)private_data;
177 struct composite_context *ctx;
178 struct smb_composite_connect *io = &state->reconnect;
183 if (!torture_get_conn_index(state->client_num, state->mem_ctx, state->tctx, &host, &share)) {
184 DEBUG(0,("Can't find host/share for reconnect?!\n"));
188 io->in.dest_host = state->dest_host;
189 io->in.dest_ports = state->dest_ports;
190 io->in.called_name = state->called_name;
191 io->in.service = share;
192 io->in.service_type = state->service_type;
193 io->in.credentials = cmdline_credentials;
194 io->in.fallback_to_anonymous = false;
195 io->in.workgroup = lp_workgroup(state->tctx->lp_ctx);
196 io->in.max_xmit = lp_max_xmit(state->tctx->lp_ctx);
197 io->in.max_mux = lp_maxmux(state->tctx->lp_ctx);
198 io->in.ntstatus_support = lp_nt_status_support(state->tctx->lp_ctx);
199 io->in.max_protocol = lp_cli_maxprotocol(state->tctx->lp_ctx);
200 io->in.unicode = lp_unicode(state->tctx->lp_ctx);
201 io->in.use_spnego = lp_use_spnego(state->tctx->lp_ctx) && lp_nt_status_support(state->tctx->lp_ctx);
202 io->in.signing = lp_client_signing(state->tctx->lp_ctx);
204 /* kill off the remnants of the old connection */
205 talloc_free(state->tree);
208 ctx = smb_composite_connect_send(io, state->mem_ctx,
209 lp_resolve_context(state->tctx->lp_ctx),
212 DEBUG(0,("Failed to setup async reconnect\n"));
216 ctx->async.fn = reopen_connection_complete;
217 ctx->async.private_data = state;
222 called when a lock completes
224 static void lock_completion(struct smbcli_request *req)
226 struct benchlock_state *state = (struct benchlock_state *)req->async.private;
227 NTSTATUS status = smbcli_request_simple_recv(req);
229 if (!NT_STATUS_IS_OK(status)) {
230 if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
231 NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT)) {
232 talloc_free(state->tree);
235 DEBUG(0,("reopening connection to %s\n", state->dest_host));
236 talloc_free(state->te);
237 state->te = event_add_timed(state->ev, state->mem_ctx,
238 timeval_current_ofs(1,0),
239 reopen_connection, state);
241 DEBUG(0,("Lock failed - %s\n", nt_errstr(status)));
247 switch (state->stage) {
249 state->stage = LOCK_LOCK;
252 state->stage = LOCK_UNLOCK;
255 state->stage = LOCK_LOCK;
264 static void echo_completion(struct smbcli_request *req)
266 struct benchlock_state *state = (struct benchlock_state *)req->async.private;
267 NTSTATUS status = smbcli_request_simple_recv(req);
268 if (NT_STATUS_EQUAL(status, NT_STATUS_END_OF_FILE) ||
269 NT_STATUS_EQUAL(status, NT_STATUS_LOCAL_DISCONNECT)) {
270 talloc_free(state->tree);
273 DEBUG(0,("reopening connection to %s\n", state->dest_host));
274 talloc_free(state->te);
275 state->te = event_add_timed(state->ev, state->mem_ctx,
276 timeval_current_ofs(1,0),
277 reopen_connection, state);
281 static void report_rate(struct event_context *ev, struct timed_event *te,
282 struct timeval t, void *private_data)
284 struct benchlock_state *state = talloc_get_type(private_data,
285 struct benchlock_state);
287 for (i=0;i<nprocs;i++) {
288 printf("%5u ", (unsigned)(state[i].count - state[i].lastcount));
289 state[i].lastcount = state[i].count;
293 event_add_timed(ev, state, timeval_current_ofs(1, 0), report_rate, state);
295 /* send an echo on each interface to ensure it stays alive - this helps
297 for (i=0;i<nprocs;i++) {
299 struct smbcli_request *req;
301 if (!state[i].tree) {
305 p.in.repeat_count = 1;
308 req = smb_raw_echo_send(state[i].tree->session->transport, &p);
309 req->async.private = &state[i];
310 req->async.fn = echo_completion;
315 benchmark locking calls
317 bool torture_bench_lock(struct torture_context *torture)
320 TALLOC_CTX *mem_ctx = talloc_new(torture);
322 int timelimit = torture_setting_int(torture, "timelimit", 10);
324 struct event_context *ev = event_context_find(mem_ctx);
325 struct benchlock_state *state;
326 int total = 0, minops=0;
327 struct smbcli_state *cli;
330 progress = torture_setting_bool(torture, "progress", true);
332 nprocs = torture_setting_int(torture, "nprocs", 4);
334 state = talloc_zero_array(mem_ctx, struct benchlock_state, nprocs);
336 printf("Opening %d connections\n", nprocs);
337 for (i=0;i<nprocs;i++) {
338 state[i].tctx = torture;
339 state[i].mem_ctx = talloc_new(state);
340 state[i].client_num = i;
342 if (!torture_open_connection_ev(&cli, i, torture, ev)) {
345 talloc_steal(mem_ctx, state);
346 state[i].tree = cli->tree;
347 state[i].dest_host = talloc_strdup(state[i].mem_ctx,
348 cli->tree->session->transport->socket->hostname);
349 state[i].dest_ports = talloc_array(state[i].mem_ctx,
351 state[i].dest_ports[0] = talloc_asprintf(state[i].dest_ports,
353 cli->tree->session->transport->socket->port);
354 state[i].dest_ports[1] = NULL;
355 state[i].called_name = talloc_strdup(state[i].mem_ctx,
356 cli->tree->session->transport->called.name);
357 state[i].service_type = talloc_strdup(state[i].mem_ctx,
363 if (!torture_setup_dir(cli, BASEDIR)) {
367 for (i=0;i<nprocs;i++) {
368 state[i].fnum = smbcli_open(state[i].tree,
370 O_RDWR|O_CREAT, DENY_NONE);
371 if (state[i].fnum == -1) {
372 printf("Failed to open %s on connection %d\n", FNAME, i);
376 state[i].stage = LOCK_INITIAL;
377 lock_send(&state[i]);
380 tv = timeval_current();
383 event_add_timed(ev, state, timeval_current_ofs(1, 0), report_rate, state);
386 printf("Running for %d seconds\n", timelimit);
387 while (timeval_elapsed(&tv) < timelimit) {
391 DEBUG(0,("locking failed\n"));
396 printf("%.2f ops/second\n", total/timeval_elapsed(&tv));
397 minops = state[0].count;
398 for (i=0;i<nprocs;i++) {
399 printf("[%d] %u ops\n", i, state[i].count);
400 if (state[i].count < minops) minops = state[i].count;
402 if (minops < 0.5*total/nprocs) {
403 printf("Failed: unbalanced locking\n");
407 for (i=0;i<nprocs;i++) {
408 talloc_free(state[i].req);
409 smb_raw_exit(state[i].tree->session);
412 smbcli_deltree(state[0].tree, BASEDIR);
413 talloc_free(mem_ctx);
418 talloc_free(mem_ctx);