FreeBSD/Linux Kernel Cross Reference
sys/rpc/clnt_vc.c
1 /* $NetBSD: clnt_vc.c,v 1.4 2000/07/14 08:40:42 fvdl Exp $ */
2
3 /*
4 * Sun RPC is a product of Sun Microsystems, Inc. and is provided for
5 * unrestricted use provided that this legend is included on all tape
6 * media and as a part of the software program in whole or part. Users
7 * may copy or modify Sun RPC without charge, but are not authorized
8 * to license or distribute it to anyone else except as part of a product or
9 * program developed by the user.
10 *
11 * SUN RPC IS PROVIDED AS IS WITH NO WARRANTIES OF ANY KIND INCLUDING THE
12 * WARRANTIES OF DESIGN, MERCHANTIBILITY AND FITNESS FOR A PARTICULAR
13 * PURPOSE, OR ARISING FROM A COURSE OF DEALING, USAGE OR TRADE PRACTICE.
14 *
15 * Sun RPC is provided with no support and without any obligation on the
16 * part of Sun Microsystems, Inc. to assist in its use, correction,
17 * modification or enhancement.
18 *
19 * SUN MICROSYSTEMS, INC. SHALL HAVE NO LIABILITY WITH RESPECT TO THE
20 * INFRINGEMENT OF COPYRIGHTS, TRADE SECRETS OR ANY PATENTS BY SUN RPC
21 * OR ANY PART THEREOF.
22 *
23 * In no event will Sun Microsystems, Inc. be liable for any lost revenue
24 * or profits or other special, indirect and consequential damages, even if
25 * Sun has been advised of the possibility of such damages.
26 *
27 * Sun Microsystems, Inc.
28 * 2550 Garcia Avenue
29 * Mountain View, California 94043
30 */
31
32 #if defined(LIBC_SCCS) && !defined(lint)
33 static char *sccsid2 = "@(#)clnt_tcp.c 1.37 87/10/05 Copyr 1984 Sun Micro";
34 static char *sccsid = "@(#)clnt_tcp.c 2.2 88/08/01 4.0 RPCSRC";
35 static char sccsid3[] = "@(#)clnt_vc.c 1.19 89/03/16 Copyr 1988 Sun Micro";
36 #endif
37 #include <sys/cdefs.h>
38 __FBSDID("$FreeBSD: releng/8.2/sys/rpc/clnt_vc.c 217617 2011-01-20 00:54:12Z rmacklem $");
39
40 /*
41 * clnt_tcp.c, Implements a TCP/IP based, client side RPC.
42 *
43 * Copyright (C) 1984, Sun Microsystems, Inc.
44 *
45 * TCP based RPC supports 'batched calls'.
46 * A sequence of calls may be batched-up in a send buffer. The rpc call
47 * return immediately to the client even though the call was not necessarily
48 * sent. The batching occurs if the results' xdr routine is NULL (0) AND
49 * the rpc timeout value is zero (see clnt.h, rpc).
50 *
51 * Clients should NOT casually batch calls that in fact return results; that is,
52 * the server side should be aware that a call is batched and not produce any
53 * return message. Batched calls that produce many result messages can
54 * deadlock (netlock) the client and the server....
55 *
56 * Now go hang yourself.
57 */
58
59 #include <sys/param.h>
60 #include <sys/systm.h>
61 #include <sys/lock.h>
62 #include <sys/malloc.h>
63 #include <sys/mbuf.h>
64 #include <sys/mutex.h>
65 #include <sys/pcpu.h>
66 #include <sys/proc.h>
67 #include <sys/protosw.h>
68 #include <sys/socket.h>
69 #include <sys/socketvar.h>
70 #include <sys/syslog.h>
71 #include <sys/time.h>
72 #include <sys/uio.h>
73
74 #include <net/vnet.h>
75
76 #include <netinet/tcp.h>
77
78 #include <rpc/rpc.h>
79 #include <rpc/rpc_com.h>
80
81 #define MCALL_MSG_SIZE 24
82
83 struct cmessage {
84 struct cmsghdr cmsg;
85 struct cmsgcred cmcred;
86 };
87
88 static enum clnt_stat clnt_vc_call(CLIENT *, struct rpc_callextra *,
89 rpcproc_t, struct mbuf *, struct mbuf **, struct timeval);
90 static void clnt_vc_geterr(CLIENT *, struct rpc_err *);
91 static bool_t clnt_vc_freeres(CLIENT *, xdrproc_t, void *);
92 static void clnt_vc_abort(CLIENT *);
93 static bool_t clnt_vc_control(CLIENT *, u_int, void *);
94 static void clnt_vc_close(CLIENT *);
95 static void clnt_vc_destroy(CLIENT *);
96 static bool_t time_not_ok(struct timeval *);
97 static int clnt_vc_soupcall(struct socket *so, void *arg, int waitflag);
98
99 static struct clnt_ops clnt_vc_ops = {
100 .cl_call = clnt_vc_call,
101 .cl_abort = clnt_vc_abort,
102 .cl_geterr = clnt_vc_geterr,
103 .cl_freeres = clnt_vc_freeres,
104 .cl_close = clnt_vc_close,
105 .cl_destroy = clnt_vc_destroy,
106 .cl_control = clnt_vc_control
107 };
108
109 /*
110 * A pending RPC request which awaits a reply. Requests which have
111 * received their reply will have cr_xid set to zero and cr_mrep to
112 * the mbuf chain of the reply.
113 */
114 struct ct_request {
115 TAILQ_ENTRY(ct_request) cr_link;
116 uint32_t cr_xid; /* XID of request */
117 struct mbuf *cr_mrep; /* reply received by upcall */
118 int cr_error; /* any error from upcall */
119 char cr_verf[MAX_AUTH_BYTES]; /* reply verf */
120 };
121
122 TAILQ_HEAD(ct_request_list, ct_request);
123
124 struct ct_data {
125 struct mtx ct_lock;
126 int ct_threads; /* number of threads in clnt_vc_call */
127 bool_t ct_closing; /* TRUE if we are closing */
128 bool_t ct_closed; /* TRUE if we are closed */
129 struct socket *ct_socket; /* connection socket */
130 bool_t ct_closeit; /* close it on destroy */
131 struct timeval ct_wait; /* wait interval in milliseconds */
132 struct sockaddr_storage ct_addr; /* remote addr */
133 struct rpc_err ct_error;
134 uint32_t ct_xid;
135 char ct_mcallc[MCALL_MSG_SIZE]; /* marshalled callmsg */
136 size_t ct_mpos; /* pos after marshal */
137 const char *ct_waitchan;
138 int ct_waitflag;
139 struct mbuf *ct_record; /* current reply record */
140 size_t ct_record_resid; /* how much left of reply to read */
141 bool_t ct_record_eor; /* true if reading last fragment */
142 struct ct_request_list ct_pending;
143 int ct_upcallrefs; /* Ref cnt of upcalls in prog. */
144 };
145
146 static void clnt_vc_upcallsdone(struct ct_data *);
147
148 static const char clnt_vc_errstr[] = "%s : %s";
149 static const char clnt_vc_str[] = "clnt_vc_create";
150 static const char clnt_read_vc_str[] = "read_vc";
151 static const char __no_mem_str[] = "out of memory";
152
153 /*
154 * Create a client handle for a connection.
155 * Default options are set, which the user can change using clnt_control()'s.
156 * The rpc/vc package does buffering similar to stdio, so the client
157 * must pick send and receive buffer sizes, 0 => use the default.
158 * NB: fd is copied into a private area.
159 * NB: The rpch->cl_auth is set null authentication. Caller may wish to
160 * set this something more useful.
161 *
162 * fd should be an open socket
163 */
164 CLIENT *
165 clnt_vc_create(
166 struct socket *so, /* open file descriptor */
167 struct sockaddr *raddr, /* servers address */
168 const rpcprog_t prog, /* program number */
169 const rpcvers_t vers, /* version number */
170 size_t sendsz, /* buffer recv size */
171 size_t recvsz) /* buffer send size */
172 {
173 CLIENT *cl; /* client handle */
174 struct ct_data *ct = NULL; /* client handle */
175 struct timeval now;
176 struct rpc_msg call_msg;
177 static uint32_t disrupt;
178 struct __rpc_sockinfo si;
179 XDR xdrs;
180 int error, interrupted, one = 1;
181 struct sockopt sopt;
182
183 if (disrupt == 0)
184 disrupt = (uint32_t)(long)raddr;
185
186 cl = (CLIENT *)mem_alloc(sizeof (*cl));
187 ct = (struct ct_data *)mem_alloc(sizeof (*ct));
188
189 mtx_init(&ct->ct_lock, "ct->ct_lock", NULL, MTX_DEF);
190 ct->ct_threads = 0;
191 ct->ct_closing = FALSE;
192 ct->ct_closed = FALSE;
193 ct->ct_upcallrefs = 0;
194
195 if ((so->so_state & (SS_ISCONNECTED|SS_ISCONFIRMING)) == 0) {
196 error = soconnect(so, raddr, curthread);
197 SOCK_LOCK(so);
198 interrupted = 0;
199 while ((so->so_state & SS_ISCONNECTING)
200 && so->so_error == 0) {
201 error = msleep(&so->so_timeo, SOCK_MTX(so),
202 PSOCK | PCATCH | PBDRY, "connec", 0);
203 if (error) {
204 if (error == EINTR || error == ERESTART)
205 interrupted = 1;
206 break;
207 }
208 }
209 if (error == 0) {
210 error = so->so_error;
211 so->so_error = 0;
212 }
213 SOCK_UNLOCK(so);
214 if (error) {
215 if (!interrupted)
216 so->so_state &= ~SS_ISCONNECTING;
217 rpc_createerr.cf_stat = RPC_SYSTEMERROR;
218 rpc_createerr.cf_error.re_errno = error;
219 goto err;
220 }
221 }
222
223 CURVNET_SET(so->so_vnet);
224 if (!__rpc_socket2sockinfo(so, &si)) {
225 CURVNET_RESTORE();
226 goto err;
227 }
228
229 if (so->so_proto->pr_flags & PR_CONNREQUIRED) {
230 bzero(&sopt, sizeof(sopt));
231 sopt.sopt_dir = SOPT_SET;
232 sopt.sopt_level = SOL_SOCKET;
233 sopt.sopt_name = SO_KEEPALIVE;
234 sopt.sopt_val = &one;
235 sopt.sopt_valsize = sizeof(one);
236 sosetopt(so, &sopt);
237 }
238
239 if (so->so_proto->pr_protocol == IPPROTO_TCP) {
240 bzero(&sopt, sizeof(sopt));
241 sopt.sopt_dir = SOPT_SET;
242 sopt.sopt_level = IPPROTO_TCP;
243 sopt.sopt_name = TCP_NODELAY;
244 sopt.sopt_val = &one;
245 sopt.sopt_valsize = sizeof(one);
246 sosetopt(so, &sopt);
247 }
248 CURVNET_RESTORE();
249
250 ct->ct_closeit = FALSE;
251
252 /*
253 * Set up private data struct
254 */
255 ct->ct_socket = so;
256 ct->ct_wait.tv_sec = -1;
257 ct->ct_wait.tv_usec = -1;
258 memcpy(&ct->ct_addr, raddr, raddr->sa_len);
259
260 /*
261 * Initialize call message
262 */
263 getmicrotime(&now);
264 ct->ct_xid = ((uint32_t)++disrupt) ^ __RPC_GETXID(&now);
265 call_msg.rm_xid = ct->ct_xid;
266 call_msg.rm_direction = CALL;
267 call_msg.rm_call.cb_rpcvers = RPC_MSG_VERSION;
268 call_msg.rm_call.cb_prog = (uint32_t)prog;
269 call_msg.rm_call.cb_vers = (uint32_t)vers;
270
271 /*
272 * pre-serialize the static part of the call msg and stash it away
273 */
274 xdrmem_create(&xdrs, ct->ct_mcallc, MCALL_MSG_SIZE,
275 XDR_ENCODE);
276 if (! xdr_callhdr(&xdrs, &call_msg)) {
277 if (ct->ct_closeit) {
278 soclose(ct->ct_socket);
279 }
280 goto err;
281 }
282 ct->ct_mpos = XDR_GETPOS(&xdrs);
283 XDR_DESTROY(&xdrs);
284 ct->ct_waitchan = "rpcrecv";
285 ct->ct_waitflag = 0;
286
287 /*
288 * Create a client handle which uses xdrrec for serialization
289 * and authnone for authentication.
290 */
291 sendsz = __rpc_get_t_size(si.si_af, si.si_proto, (int)sendsz);
292 recvsz = __rpc_get_t_size(si.si_af, si.si_proto, (int)recvsz);
293 error = soreserve(ct->ct_socket, sendsz, recvsz);
294 if (error != 0) {
295 if (ct->ct_closeit) {
296 soclose(ct->ct_socket);
297 }
298 goto err;
299 }
300 cl->cl_refs = 1;
301 cl->cl_ops = &clnt_vc_ops;
302 cl->cl_private = ct;
303 cl->cl_auth = authnone_create();
304
305 SOCKBUF_LOCK(&ct->ct_socket->so_rcv);
306 soupcall_set(ct->ct_socket, SO_RCV, clnt_vc_soupcall, ct);
307 SOCKBUF_UNLOCK(&ct->ct_socket->so_rcv);
308
309 ct->ct_record = NULL;
310 ct->ct_record_resid = 0;
311 TAILQ_INIT(&ct->ct_pending);
312 return (cl);
313
314 err:
315 if (cl) {
316 if (ct) {
317 mtx_destroy(&ct->ct_lock);
318 mem_free(ct, sizeof (struct ct_data));
319 }
320 if (cl)
321 mem_free(cl, sizeof (CLIENT));
322 }
323 return ((CLIENT *)NULL);
324 }
325
326 static enum clnt_stat
327 clnt_vc_call(
328 CLIENT *cl, /* client handle */
329 struct rpc_callextra *ext, /* call metadata */
330 rpcproc_t proc, /* procedure number */
331 struct mbuf *args, /* pointer to args */
332 struct mbuf **resultsp, /* pointer to results */
333 struct timeval utimeout)
334 {
335 struct ct_data *ct = (struct ct_data *) cl->cl_private;
336 AUTH *auth;
337 struct rpc_err *errp;
338 enum clnt_stat stat;
339 XDR xdrs;
340 struct rpc_msg reply_msg;
341 bool_t ok;
342 int nrefreshes = 2; /* number of times to refresh cred */
343 struct timeval timeout;
344 uint32_t xid;
345 struct mbuf *mreq = NULL, *results;
346 struct ct_request *cr;
347 int error;
348
349 cr = malloc(sizeof(struct ct_request), M_RPC, M_WAITOK);
350
351 mtx_lock(&ct->ct_lock);
352
353 if (ct->ct_closing || ct->ct_closed) {
354 mtx_unlock(&ct->ct_lock);
355 free(cr, M_RPC);
356 return (RPC_CANTSEND);
357 }
358 ct->ct_threads++;
359
360 if (ext) {
361 auth = ext->rc_auth;
362 errp = &ext->rc_err;
363 } else {
364 auth = cl->cl_auth;
365 errp = &ct->ct_error;
366 }
367
368 cr->cr_mrep = NULL;
369 cr->cr_error = 0;
370
371 if (ct->ct_wait.tv_usec == -1) {
372 timeout = utimeout; /* use supplied timeout */
373 } else {
374 timeout = ct->ct_wait; /* use default timeout */
375 }
376
377 call_again:
378 mtx_assert(&ct->ct_lock, MA_OWNED);
379
380 ct->ct_xid++;
381 xid = ct->ct_xid;
382
383 mtx_unlock(&ct->ct_lock);
384
385 /*
386 * Leave space to pre-pend the record mark.
387 */
388 MGETHDR(mreq, M_WAIT, MT_DATA);
389 mreq->m_data += sizeof(uint32_t);
390 KASSERT(ct->ct_mpos + sizeof(uint32_t) <= MHLEN,
391 ("RPC header too big"));
392 bcopy(ct->ct_mcallc, mreq->m_data, ct->ct_mpos);
393 mreq->m_len = ct->ct_mpos;
394
395 /*
396 * The XID is the first thing in the request.
397 */
398 *mtod(mreq, uint32_t *) = htonl(xid);
399
400 xdrmbuf_create(&xdrs, mreq, XDR_ENCODE);
401
402 errp->re_status = stat = RPC_SUCCESS;
403
404 if ((! XDR_PUTINT32(&xdrs, &proc)) ||
405 (! AUTH_MARSHALL(auth, xid, &xdrs,
406 m_copym(args, 0, M_COPYALL, M_WAITOK)))) {
407 errp->re_status = stat = RPC_CANTENCODEARGS;
408 mtx_lock(&ct->ct_lock);
409 goto out;
410 }
411 mreq->m_pkthdr.len = m_length(mreq, NULL);
412
413 /*
414 * Prepend a record marker containing the packet length.
415 */
416 M_PREPEND(mreq, sizeof(uint32_t), M_WAIT);
417 *mtod(mreq, uint32_t *) =
418 htonl(0x80000000 | (mreq->m_pkthdr.len - sizeof(uint32_t)));
419
420 cr->cr_xid = xid;
421 mtx_lock(&ct->ct_lock);
422 /*
423 * Check to see if the other end has already started to close down
424 * the connection. The upcall will have set ct_error.re_status
425 * to RPC_CANTRECV if this is the case.
426 * If the other end starts to close down the connection after this
427 * point, it will be detected later when cr_error is checked,
428 * since the request is in the ct_pending queue.
429 */
430 if (ct->ct_error.re_status == RPC_CANTRECV) {
431 if (errp != &ct->ct_error) {
432 errp->re_errno = ct->ct_error.re_errno;
433 errp->re_status = RPC_CANTRECV;
434 }
435 stat = RPC_CANTRECV;
436 goto out;
437 }
438 TAILQ_INSERT_TAIL(&ct->ct_pending, cr, cr_link);
439 mtx_unlock(&ct->ct_lock);
440
441 /*
442 * sosend consumes mreq.
443 */
444 error = sosend(ct->ct_socket, NULL, NULL, mreq, NULL, 0, curthread);
445 mreq = NULL;
446 if (error == EMSGSIZE) {
447 SOCKBUF_LOCK(&ct->ct_socket->so_snd);
448 sbwait(&ct->ct_socket->so_snd);
449 SOCKBUF_UNLOCK(&ct->ct_socket->so_snd);
450 AUTH_VALIDATE(auth, xid, NULL, NULL);
451 mtx_lock(&ct->ct_lock);
452 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
453 goto call_again;
454 }
455
456 reply_msg.acpted_rply.ar_verf.oa_flavor = AUTH_NULL;
457 reply_msg.acpted_rply.ar_verf.oa_base = cr->cr_verf;
458 reply_msg.acpted_rply.ar_verf.oa_length = 0;
459 reply_msg.acpted_rply.ar_results.where = NULL;
460 reply_msg.acpted_rply.ar_results.proc = (xdrproc_t)xdr_void;
461
462 mtx_lock(&ct->ct_lock);
463 if (error) {
464 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
465 errp->re_errno = error;
466 errp->re_status = stat = RPC_CANTSEND;
467 goto out;
468 }
469
470 /*
471 * Check to see if we got an upcall while waiting for the
472 * lock. In both these cases, the request has been removed
473 * from ct->ct_pending.
474 */
475 if (cr->cr_error) {
476 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
477 errp->re_errno = cr->cr_error;
478 errp->re_status = stat = RPC_CANTRECV;
479 goto out;
480 }
481 if (cr->cr_mrep) {
482 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
483 goto got_reply;
484 }
485
486 /*
487 * Hack to provide rpc-based message passing
488 */
489 if (timeout.tv_sec == 0 && timeout.tv_usec == 0) {
490 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
491 errp->re_status = stat = RPC_TIMEDOUT;
492 goto out;
493 }
494
495 error = msleep(cr, &ct->ct_lock, ct->ct_waitflag, ct->ct_waitchan,
496 tvtohz(&timeout));
497
498 TAILQ_REMOVE(&ct->ct_pending, cr, cr_link);
499
500 if (error) {
501 /*
502 * The sleep returned an error so our request is still
503 * on the list. Turn the error code into an
504 * appropriate client status.
505 */
506 errp->re_errno = error;
507 switch (error) {
508 case EINTR:
509 case ERESTART:
510 stat = RPC_INTR;
511 break;
512 case EWOULDBLOCK:
513 stat = RPC_TIMEDOUT;
514 break;
515 default:
516 stat = RPC_CANTRECV;
517 }
518 errp->re_status = stat;
519 goto out;
520 } else {
521 /*
522 * We were woken up by the upcall. If the
523 * upcall had a receive error, report that,
524 * otherwise we have a reply.
525 */
526 if (cr->cr_error) {
527 errp->re_errno = cr->cr_error;
528 errp->re_status = stat = RPC_CANTRECV;
529 goto out;
530 }
531 }
532
533 got_reply:
534 /*
535 * Now decode and validate the response. We need to drop the
536 * lock since xdr_replymsg may end up sleeping in malloc.
537 */
538 mtx_unlock(&ct->ct_lock);
539
540 if (ext && ext->rc_feedback)
541 ext->rc_feedback(FEEDBACK_OK, proc, ext->rc_feedback_arg);
542
543 xdrmbuf_create(&xdrs, cr->cr_mrep, XDR_DECODE);
544 ok = xdr_replymsg(&xdrs, &reply_msg);
545 cr->cr_mrep = NULL;
546
547 if (ok) {
548 if ((reply_msg.rm_reply.rp_stat == MSG_ACCEPTED) &&
549 (reply_msg.acpted_rply.ar_stat == SUCCESS))
550 errp->re_status = stat = RPC_SUCCESS;
551 else
552 stat = _seterr_reply(&reply_msg, errp);
553
554 if (stat == RPC_SUCCESS) {
555 results = xdrmbuf_getall(&xdrs);
556 if (!AUTH_VALIDATE(auth, xid,
557 &reply_msg.acpted_rply.ar_verf,
558 &results)) {
559 errp->re_status = stat = RPC_AUTHERROR;
560 errp->re_why = AUTH_INVALIDRESP;
561 } else {
562 KASSERT(results,
563 ("auth validated but no result"));
564 *resultsp = results;
565 }
566 } /* end successful completion */
567 /*
568 * If unsuccesful AND error is an authentication error
569 * then refresh credentials and try again, else break
570 */
571 else if (stat == RPC_AUTHERROR)
572 /* maybe our credentials need to be refreshed ... */
573 if (nrefreshes > 0 &&
574 AUTH_REFRESH(auth, &reply_msg)) {
575 nrefreshes--;
576 XDR_DESTROY(&xdrs);
577 mtx_lock(&ct->ct_lock);
578 goto call_again;
579 }
580 /* end of unsuccessful completion */
581 } /* end of valid reply message */
582 else {
583 errp->re_status = stat = RPC_CANTDECODERES;
584 }
585 XDR_DESTROY(&xdrs);
586 mtx_lock(&ct->ct_lock);
587 out:
588 mtx_assert(&ct->ct_lock, MA_OWNED);
589
590 KASSERT(stat != RPC_SUCCESS || *resultsp,
591 ("RPC_SUCCESS without reply"));
592
593 if (mreq)
594 m_freem(mreq);
595 if (cr->cr_mrep)
596 m_freem(cr->cr_mrep);
597
598 ct->ct_threads--;
599 if (ct->ct_closing)
600 wakeup(ct);
601
602 mtx_unlock(&ct->ct_lock);
603
604 if (auth && stat != RPC_SUCCESS)
605 AUTH_VALIDATE(auth, xid, NULL, NULL);
606
607 free(cr, M_RPC);
608
609 return (stat);
610 }
611
612 static void
613 clnt_vc_geterr(CLIENT *cl, struct rpc_err *errp)
614 {
615 struct ct_data *ct = (struct ct_data *) cl->cl_private;
616
617 *errp = ct->ct_error;
618 }
619
620 static bool_t
621 clnt_vc_freeres(CLIENT *cl, xdrproc_t xdr_res, void *res_ptr)
622 {
623 XDR xdrs;
624 bool_t dummy;
625
626 xdrs.x_op = XDR_FREE;
627 dummy = (*xdr_res)(&xdrs, res_ptr);
628
629 return (dummy);
630 }
631
632 /*ARGSUSED*/
633 static void
634 clnt_vc_abort(CLIENT *cl)
635 {
636 }
637
638 static bool_t
639 clnt_vc_control(CLIENT *cl, u_int request, void *info)
640 {
641 struct ct_data *ct = (struct ct_data *)cl->cl_private;
642 void *infop = info;
643
644 mtx_lock(&ct->ct_lock);
645
646 switch (request) {
647 case CLSET_FD_CLOSE:
648 ct->ct_closeit = TRUE;
649 mtx_unlock(&ct->ct_lock);
650 return (TRUE);
651 case CLSET_FD_NCLOSE:
652 ct->ct_closeit = FALSE;
653 mtx_unlock(&ct->ct_lock);
654 return (TRUE);
655 default:
656 break;
657 }
658
659 /* for other requests which use info */
660 if (info == NULL) {
661 mtx_unlock(&ct->ct_lock);
662 return (FALSE);
663 }
664 switch (request) {
665 case CLSET_TIMEOUT:
666 if (time_not_ok((struct timeval *)info)) {
667 mtx_unlock(&ct->ct_lock);
668 return (FALSE);
669 }
670 ct->ct_wait = *(struct timeval *)infop;
671 break;
672 case CLGET_TIMEOUT:
673 *(struct timeval *)infop = ct->ct_wait;
674 break;
675 case CLGET_SERVER_ADDR:
676 (void) memcpy(info, &ct->ct_addr, (size_t)ct->ct_addr.ss_len);
677 break;
678 case CLGET_SVC_ADDR:
679 /*
680 * Slightly different semantics to userland - we use
681 * sockaddr instead of netbuf.
682 */
683 memcpy(info, &ct->ct_addr, ct->ct_addr.ss_len);
684 break;
685 case CLSET_SVC_ADDR: /* set to new address */
686 mtx_unlock(&ct->ct_lock);
687 return (FALSE);
688 case CLGET_XID:
689 *(uint32_t *)info = ct->ct_xid;
690 break;
691 case CLSET_XID:
692 /* This will set the xid of the NEXT call */
693 /* decrement by 1 as clnt_vc_call() increments once */
694 ct->ct_xid = *(uint32_t *)info - 1;
695 break;
696 case CLGET_VERS:
697 /*
698 * This RELIES on the information that, in the call body,
699 * the version number field is the fifth field from the
700 * begining of the RPC header. MUST be changed if the
701 * call_struct is changed
702 */
703 *(uint32_t *)info =
704 ntohl(*(uint32_t *)(void *)(ct->ct_mcallc +
705 4 * BYTES_PER_XDR_UNIT));
706 break;
707
708 case CLSET_VERS:
709 *(uint32_t *)(void *)(ct->ct_mcallc +
710 4 * BYTES_PER_XDR_UNIT) =
711 htonl(*(uint32_t *)info);
712 break;
713
714 case CLGET_PROG:
715 /*
716 * This RELIES on the information that, in the call body,
717 * the program number field is the fourth field from the
718 * begining of the RPC header. MUST be changed if the
719 * call_struct is changed
720 */
721 *(uint32_t *)info =
722 ntohl(*(uint32_t *)(void *)(ct->ct_mcallc +
723 3 * BYTES_PER_XDR_UNIT));
724 break;
725
726 case CLSET_PROG:
727 *(uint32_t *)(void *)(ct->ct_mcallc +
728 3 * BYTES_PER_XDR_UNIT) =
729 htonl(*(uint32_t *)info);
730 break;
731
732 case CLSET_WAITCHAN:
733 ct->ct_waitchan = (const char *)info;
734 break;
735
736 case CLGET_WAITCHAN:
737 *(const char **) info = ct->ct_waitchan;
738 break;
739
740 case CLSET_INTERRUPTIBLE:
741 if (*(int *) info)
742 ct->ct_waitflag = PCATCH | PBDRY;
743 else
744 ct->ct_waitflag = 0;
745 break;
746
747 case CLGET_INTERRUPTIBLE:
748 if (ct->ct_waitflag)
749 *(int *) info = TRUE;
750 else
751 *(int *) info = FALSE;
752 break;
753
754 default:
755 mtx_unlock(&ct->ct_lock);
756 return (FALSE);
757 }
758
759 mtx_unlock(&ct->ct_lock);
760 return (TRUE);
761 }
762
763 static void
764 clnt_vc_close(CLIENT *cl)
765 {
766 struct ct_data *ct = (struct ct_data *) cl->cl_private;
767 struct ct_request *cr;
768
769 mtx_lock(&ct->ct_lock);
770
771 if (ct->ct_closed) {
772 mtx_unlock(&ct->ct_lock);
773 return;
774 }
775
776 if (ct->ct_closing) {
777 while (ct->ct_closing)
778 msleep(ct, &ct->ct_lock, 0, "rpcclose", 0);
779 KASSERT(ct->ct_closed, ("client should be closed"));
780 mtx_unlock(&ct->ct_lock);
781 return;
782 }
783
784 if (ct->ct_socket) {
785 ct->ct_closing = TRUE;
786 mtx_unlock(&ct->ct_lock);
787
788 SOCKBUF_LOCK(&ct->ct_socket->so_rcv);
789 soupcall_clear(ct->ct_socket, SO_RCV);
790 clnt_vc_upcallsdone(ct);
791 SOCKBUF_UNLOCK(&ct->ct_socket->so_rcv);
792
793 /*
794 * Abort any pending requests and wait until everyone
795 * has finished with clnt_vc_call.
796 */
797 mtx_lock(&ct->ct_lock);
798 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) {
799 cr->cr_xid = 0;
800 cr->cr_error = ESHUTDOWN;
801 wakeup(cr);
802 }
803
804 while (ct->ct_threads)
805 msleep(ct, &ct->ct_lock, 0, "rpcclose", 0);
806 }
807
808 ct->ct_closing = FALSE;
809 ct->ct_closed = TRUE;
810 mtx_unlock(&ct->ct_lock);
811 wakeup(ct);
812 }
813
814 static void
815 clnt_vc_destroy(CLIENT *cl)
816 {
817 struct ct_data *ct = (struct ct_data *) cl->cl_private;
818 struct socket *so = NULL;
819
820 clnt_vc_close(cl);
821
822 mtx_lock(&ct->ct_lock);
823
824 if (ct->ct_socket) {
825 if (ct->ct_closeit) {
826 so = ct->ct_socket;
827 }
828 }
829
830 mtx_unlock(&ct->ct_lock);
831
832 mtx_destroy(&ct->ct_lock);
833 if (so) {
834 soshutdown(so, SHUT_WR);
835 soclose(so);
836 }
837 mem_free(ct, sizeof(struct ct_data));
838 mem_free(cl, sizeof(CLIENT));
839 }
840
841 /*
842 * Make sure that the time is not garbage. -1 value is disallowed.
843 * Note this is different from time_not_ok in clnt_dg.c
844 */
845 static bool_t
846 time_not_ok(struct timeval *t)
847 {
848 return (t->tv_sec <= -1 || t->tv_sec > 100000000 ||
849 t->tv_usec <= -1 || t->tv_usec > 1000000);
850 }
851
852 int
853 clnt_vc_soupcall(struct socket *so, void *arg, int waitflag)
854 {
855 struct ct_data *ct = (struct ct_data *) arg;
856 struct uio uio;
857 struct mbuf *m;
858 struct ct_request *cr;
859 int error, rcvflag, foundreq;
860 uint32_t xid, header;
861 bool_t do_read;
862
863 ct->ct_upcallrefs++;
864 uio.uio_td = curthread;
865 do {
866 /*
867 * If ct_record_resid is zero, we are waiting for a
868 * record mark.
869 */
870 if (ct->ct_record_resid == 0) {
871
872 /*
873 * Make sure there is either a whole record
874 * mark in the buffer or there is some other
875 * error condition
876 */
877 do_read = FALSE;
878 if (so->so_rcv.sb_cc >= sizeof(uint32_t)
879 || (so->so_rcv.sb_state & SBS_CANTRCVMORE)
880 || so->so_error)
881 do_read = TRUE;
882
883 if (!do_read)
884 break;
885
886 SOCKBUF_UNLOCK(&so->so_rcv);
887 uio.uio_resid = sizeof(uint32_t);
888 m = NULL;
889 rcvflag = MSG_DONTWAIT | MSG_SOCALLBCK;
890 error = soreceive(so, NULL, &uio, &m, NULL, &rcvflag);
891 SOCKBUF_LOCK(&so->so_rcv);
892
893 if (error == EWOULDBLOCK)
894 break;
895
896 /*
897 * If there was an error, wake up all pending
898 * requests.
899 */
900 if (error || uio.uio_resid > 0) {
901 wakeup_all:
902 mtx_lock(&ct->ct_lock);
903 if (!error) {
904 /*
905 * We must have got EOF trying
906 * to read from the stream.
907 */
908 error = ECONNRESET;
909 }
910 ct->ct_error.re_status = RPC_CANTRECV;
911 ct->ct_error.re_errno = error;
912 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) {
913 cr->cr_error = error;
914 wakeup(cr);
915 }
916 mtx_unlock(&ct->ct_lock);
917 break;
918 }
919 m_copydata(m, 0, sizeof(uint32_t), (char *)&header);
920 header = ntohl(header);
921 ct->ct_record = NULL;
922 ct->ct_record_resid = header & 0x7fffffff;
923 ct->ct_record_eor = ((header & 0x80000000) != 0);
924 m_freem(m);
925 } else {
926 /*
927 * Wait until the socket has the whole record
928 * buffered.
929 */
930 do_read = FALSE;
931 if (so->so_rcv.sb_cc >= ct->ct_record_resid
932 || (so->so_rcv.sb_state & SBS_CANTRCVMORE)
933 || so->so_error)
934 do_read = TRUE;
935
936 if (!do_read)
937 break;
938
939 /*
940 * We have the record mark. Read as much as
941 * the socket has buffered up to the end of
942 * this record.
943 */
944 SOCKBUF_UNLOCK(&so->so_rcv);
945 uio.uio_resid = ct->ct_record_resid;
946 m = NULL;
947 rcvflag = MSG_DONTWAIT | MSG_SOCALLBCK;
948 error = soreceive(so, NULL, &uio, &m, NULL, &rcvflag);
949 SOCKBUF_LOCK(&so->so_rcv);
950
951 if (error == EWOULDBLOCK)
952 break;
953
954 if (error || uio.uio_resid == ct->ct_record_resid)
955 goto wakeup_all;
956
957 /*
958 * If we have part of the record already,
959 * chain this bit onto the end.
960 */
961 if (ct->ct_record)
962 m_last(ct->ct_record)->m_next = m;
963 else
964 ct->ct_record = m;
965
966 ct->ct_record_resid = uio.uio_resid;
967
968 /*
969 * If we have the entire record, see if we can
970 * match it to a request.
971 */
972 if (ct->ct_record_resid == 0
973 && ct->ct_record_eor) {
974 /*
975 * The XID is in the first uint32_t of
976 * the reply.
977 */
978 if (ct->ct_record->m_len < sizeof(xid) &&
979 m_length(ct->ct_record, NULL) < sizeof(xid))
980 break;
981 m_copydata(ct->ct_record, 0, sizeof(xid),
982 (char *)&xid);
983 xid = ntohl(xid);
984
985 mtx_lock(&ct->ct_lock);
986 foundreq = 0;
987 TAILQ_FOREACH(cr, &ct->ct_pending, cr_link) {
988 if (cr->cr_xid == xid) {
989 /*
990 * This one
991 * matches. We leave
992 * the reply mbuf in
993 * cr->cr_mrep. Set
994 * the XID to zero so
995 * that we will ignore
996 * any duplicaed
997 * replies.
998 */
999 cr->cr_xid = 0;
1000 cr->cr_mrep = ct->ct_record;
1001 cr->cr_error = 0;
1002 foundreq = 1;
1003 wakeup(cr);
1004 break;
1005 }
1006 }
1007 mtx_unlock(&ct->ct_lock);
1008
1009 if (!foundreq)
1010 m_freem(ct->ct_record);
1011 ct->ct_record = NULL;
1012 }
1013 }
1014 } while (m);
1015 ct->ct_upcallrefs--;
1016 if (ct->ct_upcallrefs < 0)
1017 panic("rpcvc upcall refcnt");
1018 if (ct->ct_upcallrefs == 0)
1019 wakeup(&ct->ct_upcallrefs);
1020 return (SU_OK);
1021 }
1022
1023 /*
1024 * Wait for all upcalls in progress to complete.
1025 */
1026 static void
1027 clnt_vc_upcallsdone(struct ct_data *ct)
1028 {
1029
1030 SOCKBUF_LOCK_ASSERT(&ct->ct_socket->so_rcv);
1031
1032 while (ct->ct_upcallrefs > 0)
1033 (void) msleep(&ct->ct_upcallrefs,
1034 SOCKBUF_MTX(&ct->ct_socket->so_rcv), 0, "rpcvcup", 0);
1035 }
Cache object: 3fea1418706856e2c5a9352b70e152ab
|