mirror of
https://git.hardenedbsd.org/hardenedbsd/HardenedBSD.git
synced 2024-12-25 03:54:17 +01:00
ce81d24b05
- Use rpcgen to generate the unmodified boilerplate code rather than having it in the repository. - Eliminate the conflicting function names by changing them to their "natural" rpcgen generated names
589 lines
18 KiB
C
589 lines
18 KiB
C
/*
|
|
* Copyright (c) 1995
|
|
* A.R. Gordon (andrew.gordon@net-tel.co.uk). All rights reserved.
|
|
*
|
|
* Redistribution and use in source and binary forms, with or without
|
|
* modification, are permitted provided that the following conditions
|
|
* are met:
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
* notice, this list of conditions and the following disclaimer.
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
* documentation and/or other materials provided with the distribution.
|
|
* 3. All advertising materials mentioning features or use of this software
|
|
* must display the following acknowledgement:
|
|
* This product includes software developed for the FreeBSD project
|
|
* 4. Neither the name of the author nor the names of any co-contributors
|
|
* may be used to endorse or promote products derived from this software
|
|
* without specific prior written permission.
|
|
*
|
|
* THIS SOFTWARE IS PROVIDED BY ANDREW GORDON AND CONTRIBUTORS ``AS IS'' AND
|
|
* ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
|
|
* IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
|
|
* ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
|
|
* DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
|
|
* OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
|
|
* HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
|
|
* LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
|
|
* OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
|
|
* SUCH DAMAGE.
|
|
*
|
|
*/
|
|
|
|
|
|
|
|
#include "lockd.h"
|
|
|
|
#include <sys/param.h> /* for MAXHOSTNAMELEN */
|
|
#include <sys/socket.h>
|
|
#include <netinet/in.h>
|
|
#include <arpa/inet.h>
|
|
#include <netdb.h>
|
|
|
|
|
|
#define CLIENT_CACHE_SIZE 64 /* No. of client sockets cached */
|
|
#define CLIENT_CACHE_LIFETIME 120 /* In seconds */
|
|
|
|
|
|
/* log_from_addr ----------------------------------------------------------- */
|
|
/*
|
|
Purpose: Log name of function called and source address
|
|
Returns: Nothing
|
|
Notes: Extracts the source address from the transport handle
|
|
passed in as part of the called procedure specification
|
|
*/
|
|
|
|
static void log_from_addr(char *fun_name, struct svc_req *req)
|
|
{
|
|
struct sockaddr_in *addr;
|
|
struct hostent *host;
|
|
char hostname_buf[40];
|
|
|
|
addr = svc_getcaller(req->rq_xprt);
|
|
host = gethostbyaddr((char *)&(addr->sin_addr), addr->sin_len, AF_INET);
|
|
if (host)
|
|
{
|
|
strncpy(hostname_buf, host->h_name, sizeof(hostname_buf));
|
|
hostname_buf[sizeof(hostname_buf) -1] = '\0';
|
|
}
|
|
else /* No hostname available - print raw address */
|
|
{
|
|
strcpy(hostname_buf, inet_ntoa(addr->sin_addr));
|
|
}
|
|
|
|
syslog(LOG_DEBUG, "%s from %s", fun_name, hostname_buf);
|
|
}
|
|
|
|
|
|
/* get_client -------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Get a CLIENT* for making RPC calls to lockd on given host
|
|
Returns: CLIENT* pointer, from clnt_udp_create, or NULL if error
|
|
Notes: Creating a CLIENT* is quite expensive, involving a
|
|
conversation with the remote portmapper to get the
|
|
port number. Since a given client is quite likely
|
|
to make several locking requests in succession, it is
|
|
desirable to cache the created CLIENT*.
|
|
|
|
Since we are using UDP rather than TCP, there is no cost
|
|
to the remote system in keeping these cached indefinitely.
|
|
Unfortunately there is a snag: if the remote system
|
|
reboots, the cached portmapper results will be invalid,
|
|
and we will never detect this since all of the xxx_msg()
|
|
calls return no result - we just fire off a udp packet
|
|
and hope for the best.
|
|
|
|
We solve this by discarding cached values after two
|
|
minutes, regardless of whether they have been used
|
|
in the meanwhile (since a bad one might have been used
|
|
plenty of times, as the host keeps retrying the request
|
|
and we keep sending the reply back to the wrong port).
|
|
|
|
Given that the entries will always expire in the order
|
|
that they were created, there is no point in a LRU
|
|
algorithm for when the cache gets full - entries are
|
|
always re-used in sequence.
|
|
*/
|
|
|
|
static CLIENT *clnt_cache_ptr[CLIENT_CACHE_SIZE];
|
|
static long clnt_cache_time[CLIENT_CACHE_SIZE]; /* time entry created */
|
|
static struct in_addr clnt_cache_addr[CLIENT_CACHE_SIZE];
|
|
static int clnt_cache_next_to_use = 0;
|
|
|
|
static CLIENT *get_client(struct sockaddr_in *host_addr)
|
|
{
|
|
CLIENT *client;
|
|
int sock_no;
|
|
struct timeval retry_time, time_now;
|
|
int i;
|
|
|
|
gettimeofday(&time_now, NULL);
|
|
|
|
/* Search for the given client in the cache, zapping any expired */
|
|
/* entries that we happen to notice in passing. */
|
|
for (i = 0; i < CLIENT_CACHE_SIZE; i++)
|
|
{
|
|
client = clnt_cache_ptr[i];
|
|
if (client &&
|
|
((clnt_cache_time[i] + CLIENT_CACHE_LIFETIME) < time_now.tv_sec))
|
|
{
|
|
/* Cache entry has expired. */
|
|
if (debug_level > 3) syslog(LOG_DEBUG, "Expired CLIENT* in cache");
|
|
clnt_cache_time[i] = 0L;
|
|
clnt_destroy(client);
|
|
clnt_cache_ptr[i] = NULL;
|
|
client = NULL;
|
|
}
|
|
|
|
if (client && !memcmp(&clnt_cache_addr[i], &host_addr->sin_addr,
|
|
sizeof(struct in_addr)))
|
|
{
|
|
/* Found it! */
|
|
if (debug_level > 3) syslog(LOG_DEBUG, "Found CLIENT* in cache");
|
|
return (client);
|
|
}
|
|
}
|
|
|
|
/* Not found in cache. Free the next entry if it is in use */
|
|
if (clnt_cache_ptr[clnt_cache_next_to_use])
|
|
{
|
|
clnt_destroy(clnt_cache_ptr[clnt_cache_next_to_use]);
|
|
clnt_cache_ptr[clnt_cache_next_to_use] = NULL;
|
|
}
|
|
|
|
/* Create the new client handle */
|
|
|
|
sock_no = RPC_ANYSOCK;
|
|
retry_time.tv_sec = 5;
|
|
retry_time.tv_usec = 0;
|
|
host_addr->sin_port = 0; /* Force consultation with portmapper */
|
|
client = clntudp_create(host_addr, NLM_PROG, NLM_VERS, retry_time, &sock_no);
|
|
if (!client)
|
|
{
|
|
syslog(LOG_ERR, clnt_spcreateerror("clntudp_create"));
|
|
syslog(LOG_ERR, "Unable to return result to %s",
|
|
inet_ntoa(host_addr->sin_addr));
|
|
return NULL;
|
|
}
|
|
|
|
/* Success - update the cache entry */
|
|
clnt_cache_ptr[clnt_cache_next_to_use] = client;
|
|
clnt_cache_addr[clnt_cache_next_to_use] = host_addr->sin_addr;
|
|
clnt_cache_time[clnt_cache_next_to_use] = time_now.tv_sec;
|
|
if (++clnt_cache_next_to_use > CLIENT_CACHE_SIZE)
|
|
clnt_cache_next_to_use = 0;
|
|
|
|
/* Disable the default timeout, so we can specify our own in calls */
|
|
/* to clnt_call(). [note that the timeout is a different concept */
|
|
/* from the retry period set in clnt_udp_create() above.] */
|
|
retry_time.tv_sec = -1;
|
|
retry_time.tv_usec = -1;
|
|
clnt_control(client, CLSET_TIMEOUT, &retry_time);
|
|
|
|
if (debug_level > 3) syslog(LOG_DEBUG, "Created CLIENT* for %s",
|
|
inet_ntoa(host_addr->sin_addr));
|
|
return client;
|
|
}
|
|
|
|
|
|
/* transmit_result --------------------------------------------------------- */
|
|
/*
|
|
Purpose: Transmit result for nlm_xxx_msg pseudo-RPCs
|
|
Returns: Nothing - we have no idea if the datagram got there
|
|
Notes: clnt_call() will always fail (with timeout) as we are
|
|
calling it with timeout 0 as a hack to just issue a datagram
|
|
without expecting a result
|
|
*/
|
|
|
|
static void transmit_result(int opcode, nlm_res *result, struct svc_req *req)
|
|
{
|
|
static char dummy;
|
|
struct sockaddr_in *addr;
|
|
CLIENT *cli;
|
|
int success;
|
|
struct timeval timeo;
|
|
|
|
addr = svc_getcaller(req->rq_xprt);
|
|
if (cli = get_client(addr))
|
|
{
|
|
timeo.tv_sec = 0; /* No timeout - not expecting response */
|
|
timeo.tv_usec = 0;
|
|
|
|
success = clnt_call(cli, opcode, xdr_nlm_res, result, xdr_void,
|
|
&dummy, timeo);
|
|
|
|
if (debug_level > 2) syslog(LOG_DEBUG, "clnt_call returns %d\n", success);
|
|
}
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
/*
|
|
Functions for Unix<->Unix locking (ie. monitored locking, with rpc.statd
|
|
involved to ensure reclaim of locks after a crash of the "stateless"
|
|
server.
|
|
|
|
These all come in two flavours - nlm_xxx() and nlm_xxx_msg().
|
|
The first are standard RPCs with argument and result.
|
|
The nlm_xxx_msg() calls implement exactly the same functions, but
|
|
use two pseudo-RPCs (one in each direction). These calls are NOT
|
|
standard use of the RPC protocol in that they do not return a result
|
|
at all (NB. this is quite different from returning a void result).
|
|
The effect of this is to make the nlm_xxx_msg() calls simple unacknowledged
|
|
datagrams, requiring higher-level code to perform retries.
|
|
|
|
Despite the disadvantages of the nlm_xxx_msg() approach (some of which
|
|
are documented in the comments to get_client() above), this is the
|
|
interface used by all current commercial NFS implementations
|
|
[Solaris, SCO, AIX etc.]. This is presumed to be because these allow
|
|
implementations to continue using the standard RPC libraries, while
|
|
avoiding the block-until-result nature of the library interface.
|
|
|
|
No client implementations have been identified so far that make use
|
|
of the true RPC version (early SunOS releases would be a likely candidate
|
|
for testing).
|
|
*/
|
|
|
|
|
|
/* nlm_test ---------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Test whether a specified lock would be granted if requested
|
|
Returns: nlm_granted (or error code)
|
|
Notes:
|
|
*/
|
|
|
|
nlm_testres *nlm_test_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_testres res;
|
|
if (debug_level) log_from_addr("nlm_test", rqstp);
|
|
|
|
/* Copy the cookie from the argument into the result. Note that this */
|
|
/* is slightly hazardous, as the structure contains a pointer to a */
|
|
/* malloc()ed buffer that will get freed by the caller. However, the */
|
|
/* main function transmits the result before freeing the argument */
|
|
/* so it is in fact safe. */
|
|
res.cookie = arg->cookie;
|
|
res.stat.stat = nlm_granted;
|
|
return (&res);
|
|
}
|
|
|
|
void *nlm_test_msg_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
|
|
{
|
|
nlm_testres res;
|
|
static char dummy;
|
|
struct sockaddr_in *addr;
|
|
CLIENT *cli;
|
|
int success;
|
|
struct timeval timeo;
|
|
|
|
if (debug_level) log_from_addr("nlm_test_msg", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
res.stat.stat = nlm_granted;
|
|
|
|
/* nlm_test has different result type to the other operations, so */
|
|
/* can't use transmit_result() in this case */
|
|
addr = svc_getcaller(rqstp->rq_xprt);
|
|
if (cli = get_client(addr))
|
|
{
|
|
timeo.tv_sec = 0; /* No timeout - not expecting response */
|
|
timeo.tv_usec = 0;
|
|
|
|
success = clnt_call(cli, NLM_TEST_RES, xdr_nlm_testres, &res, xdr_void,
|
|
&dummy, timeo);
|
|
|
|
if (debug_level > 2) syslog(LOG_DEBUG, "clnt_call returns %d\n", success);
|
|
}
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_lock ---------------------------------------------------------------- */
|
|
/*
|
|
Purposes: Establish a lock
|
|
Returns: granted, denied or blocked
|
|
Notes: *** grace period support missing
|
|
*/
|
|
|
|
nlm_res *nlm_lock_1_svc(nlm_lockargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_lock", rqstp);
|
|
|
|
/* copy cookie from arg to result. See comment in nlm_test_1() */
|
|
res.cookie = arg->cookie;
|
|
|
|
res.stat.stat = nlm_granted;
|
|
return (&res);
|
|
}
|
|
|
|
void *nlm_lock_msg_1_svc(nlm_lockargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
|
|
if (debug_level) log_from_addr("nlm_lock_msg", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
res.stat.stat = nlm_granted;
|
|
transmit_result(NLM_LOCK_RES, &res, rqstp);
|
|
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_cancel -------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Cancel a blocked lock request
|
|
Returns: granted or denied
|
|
Notes:
|
|
*/
|
|
|
|
nlm_res *nlm_cancel_1_svc(nlm_cancargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_cancel", rqstp);
|
|
|
|
/* copy cookie from arg to result. See comment in nlm_test_1() */
|
|
res.cookie = arg->cookie;
|
|
|
|
/* Since at present we never return 'nlm_blocked', there can never be */
|
|
/* a lock to cancel, so this call always fails. */
|
|
res.stat.stat = nlm_denied;
|
|
return (&res);
|
|
}
|
|
|
|
void *nlm_cancel_msg_1_svc(nlm_cancargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_cancel_msg", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
/* Since at present we never return 'nlm_blocked', there can never be */
|
|
/* a lock to cancel, so this call always fails. */
|
|
res.stat.stat = nlm_denied;
|
|
transmit_result(NLM_CANCEL_RES, &res, rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_unlock -------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Release an existing lock
|
|
Returns: Always granted, unless during grace period
|
|
Notes: "no such lock" error condition is ignored, as the
|
|
protocol uses unreliable UDP datagrams, and may well
|
|
re-try an unlock that has already succeeded.
|
|
*/
|
|
|
|
nlm_res *nlm_unlock_1_svc(nlm_unlockargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_unlock", rqstp);
|
|
|
|
res.stat.stat= nlm_granted;
|
|
res.cookie = arg->cookie;
|
|
|
|
return (&res);
|
|
}
|
|
|
|
void *nlm_unlock_msg_1_svc(nlm_unlockargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_unlock_msg", rqstp);
|
|
|
|
res.stat.stat = nlm_granted;
|
|
res.cookie = arg->cookie;
|
|
|
|
transmit_result(NLM_UNLOCK_RES, &res, rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
/*
|
|
Client-side pseudo-RPCs for results. Note that for the client there
|
|
are only nlm_xxx_msg() versions of each call, since the 'real RPC'
|
|
version returns the results in the RPC result, and so the client
|
|
does not normally receive incoming RPCs.
|
|
|
|
The exception to this is nlm_granted(), which is genuinely an RPC
|
|
call from the server to the client - a 'call-back' in normal procedure
|
|
call terms.
|
|
*/
|
|
|
|
/* nlm_granted ------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Receive notification that formerly blocked lock now granted
|
|
Returns: always success ('granted')
|
|
Notes:
|
|
*/
|
|
|
|
nlm_res *nlm_granted_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_granted", rqstp);
|
|
|
|
/* copy cookie from arg to result. See comment in nlm_test_1() */
|
|
res.cookie = arg->cookie;
|
|
|
|
res.stat.stat = nlm_granted;
|
|
return (&res);
|
|
}
|
|
|
|
void *nlm_granted_msg_1_svc(nlm_testargs *arg, struct svc_req *rqstp)
|
|
{
|
|
nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_granted_msg", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
res.stat.stat = nlm_granted;
|
|
transmit_result(NLM_GRANTED_RES, &res, rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_test_res ------------------------------------------------------------ */
|
|
/*
|
|
Purpose: Accept result from earlier nlm_test_msg() call
|
|
Returns: Nothing
|
|
*/
|
|
|
|
void *nlm_test_res_1_svc(nlm_testres *arg, struct svc_req *rqstp)
|
|
{
|
|
if (debug_level) log_from_addr("nlm_test_res", rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_lock_res ------------------------------------------------------------ */
|
|
/*
|
|
Purpose: Accept result from earlier nlm_lock_msg() call
|
|
Returns: Nothing
|
|
*/
|
|
|
|
void *nlm_lock_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
|
|
{
|
|
if (debug_level) log_from_addr("nlm_lock_res", rqstp);
|
|
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_cancel_res ---------------------------------------------------------- */
|
|
/*
|
|
Purpose: Accept result from earlier nlm_cancel_msg() call
|
|
Returns: Nothing
|
|
*/
|
|
|
|
void *nlm_cancel_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
|
|
{
|
|
if (debug_level) log_from_addr("nlm_cancel_res", rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_unlock_res ---------------------------------------------------------- */
|
|
/*
|
|
Purpose: Accept result from earlier nlm_unlock_msg() call
|
|
Returns: Nothing
|
|
*/
|
|
|
|
void *nlm_unlock_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
|
|
{
|
|
if (debug_level) log_from_addr("nlm_unlock_res", rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* nlm_granted_res --------------------------------------------------------- */
|
|
/*
|
|
Purpose: Accept result from earlier nlm_granted_msg() call
|
|
Returns: Nothing
|
|
*/
|
|
|
|
void *nlm_granted_res_1_svc(nlm_res *arg, struct svc_req *rqstp)
|
|
{
|
|
if (debug_level) log_from_addr("nlm_granted_res", rqstp);
|
|
return (NULL);
|
|
}
|
|
|
|
/* ------------------------------------------------------------------------- */
|
|
/*
|
|
Calls for PCNFS locking (aka non-monitored locking, no involvement
|
|
of rpc.statd).
|
|
|
|
These are all genuine RPCs - no nlm_xxx_msg() nonsense here.
|
|
*/
|
|
|
|
|
|
/* nlm_share --------------------------------------------------------------- */
|
|
/*
|
|
Purpose: Establish a DOS-style lock
|
|
Returns: success or failure
|
|
Notes: Blocking locks are not supported - client is expected
|
|
to retry if required.
|
|
*/
|
|
|
|
nlm_shareres *nlm_share_3_svc(nlm_shareargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_shareres res;
|
|
if (debug_level) log_from_addr("nlm_share", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
res.stat = nlm_granted;
|
|
res.sequence = 1234356; /* X/Open says this field is ignored? */
|
|
return (&res);
|
|
}
|
|
|
|
/* nlm_unshare ------------------------------------------------------------ */
|
|
/*
|
|
Purpose: Release a DOS-style lock
|
|
Returns: nlm_granted, unless in grace period
|
|
Notes:
|
|
*/
|
|
|
|
nlm_shareres *nlm_unshare_3_svc(nlm_shareargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_shareres res;
|
|
if (debug_level) log_from_addr("nlm_unshare", rqstp);
|
|
|
|
res.cookie = arg->cookie;
|
|
res.stat = nlm_granted;
|
|
res.sequence = 1234356; /* X/Open says this field is ignored? */
|
|
return (&res);
|
|
}
|
|
|
|
/* nlm_nm_lock ------------------------------------------------------------ */
|
|
/*
|
|
Purpose: non-monitored version of nlm_lock()
|
|
Returns: as for nlm_lock()
|
|
Notes: These locks are in the same style as the standard nlm_lock,
|
|
but the rpc.statd should not be called to establish a
|
|
monitor for the client machine, since that machine is
|
|
declared not to be running a rpc.statd, and so would not
|
|
respond to the statd protocol.
|
|
*/
|
|
|
|
nlm_res *nlm_nm_lock_3_svc(nlm_lockargs *arg, struct svc_req *rqstp)
|
|
{
|
|
static nlm_res res;
|
|
if (debug_level) log_from_addr("nlm_nm_lock", rqstp);
|
|
|
|
/* copy cookie from arg to result. See comment in nlm_test_1() */
|
|
res.cookie = arg->cookie;
|
|
res.stat.stat = nlm_granted;
|
|
return (&res);
|
|
}
|
|
|
|
/* nlm_free_all ------------------------------------------------------------ */
|
|
/*
|
|
Purpose: Release all locks held by a named client
|
|
Returns: Nothing
|
|
Notes: Potential denial of service security problem here - the
|
|
locks to be released are specified by a host name, independent
|
|
of the address from which the request has arrived.
|
|
Should probably be rejected if the named host has been
|
|
using monitored locks.
|
|
*/
|
|
|
|
void *nlm_free_all_3_svc(nlm_notify *arg, struct svc_req *rqstp)
|
|
{
|
|
static char dummy;
|
|
|
|
if (debug_level) log_from_addr("nlm_free_all", rqstp);
|
|
return (&dummy);
|
|
}
|
|
|
|
|