[NET] SUNRPC: Fix whitespace errors.
Signed-off-by: YOSHIFUJI Hideaki <yoshfuji@linux-ipv6.org> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
d808ad9ab8
commit
cca5172a7e
|
@ -213,7 +213,7 @@ retry:
|
|||
rpcauth_gc_credcache(auth, &free);
|
||||
hlist_for_each_safe(pos, next, &cache->hashtable[nr]) {
|
||||
struct rpc_cred *entry;
|
||||
entry = hlist_entry(pos, struct rpc_cred, cr_hash);
|
||||
entry = hlist_entry(pos, struct rpc_cred, cr_hash);
|
||||
if (entry->cr_ops->crmatch(acred, entry, flags)) {
|
||||
hlist_del(&entry->cr_hash);
|
||||
cred = entry;
|
||||
|
|
|
@ -869,7 +869,7 @@ gss_validate(struct rpc_task *task, __be32 *p)
|
|||
|
||||
flav = ntohl(*p++);
|
||||
if ((len = ntohl(*p++)) > RPC_MAX_AUTH_SIZE)
|
||||
goto out_bad;
|
||||
goto out_bad;
|
||||
if (flav != RPC_AUTH_GSS)
|
||||
goto out_bad;
|
||||
seq = htonl(task->tk_rqstp->rq_seqno);
|
||||
|
@ -1030,7 +1030,7 @@ gss_wrap_req_priv(struct rpc_cred *cred, struct gss_cl_ctx *ctx,
|
|||
maj_stat = gss_wrap(ctx->gc_gss_ctx, offset, snd_buf, inpages);
|
||||
/* RPC_SLACK_SPACE should prevent this ever happening: */
|
||||
BUG_ON(snd_buf->len > snd_buf->buflen);
|
||||
status = -EIO;
|
||||
status = -EIO;
|
||||
/* We're assuming that when GSS_S_CONTEXT_EXPIRED, the encryption was
|
||||
* done anyway, so it's safe to put the request on the wire: */
|
||||
if (maj_stat == GSS_S_CONTEXT_EXPIRED)
|
||||
|
@ -1079,7 +1079,7 @@ gss_wrap_req(struct rpc_task *task,
|
|||
status = gss_wrap_req_integ(cred, ctx, encode,
|
||||
rqstp, p, obj);
|
||||
break;
|
||||
case RPC_GSS_SVC_PRIVACY:
|
||||
case RPC_GSS_SVC_PRIVACY:
|
||||
status = gss_wrap_req_priv(cred, ctx, encode,
|
||||
rqstp, p, obj);
|
||||
break;
|
||||
|
@ -1179,7 +1179,7 @@ gss_unwrap_resp(struct rpc_task *task,
|
|||
if (status)
|
||||
goto out;
|
||||
break;
|
||||
case RPC_GSS_SVC_PRIVACY:
|
||||
case RPC_GSS_SVC_PRIVACY:
|
||||
status = gss_unwrap_resp_priv(cred, ctx, rqstp, &p);
|
||||
if (status)
|
||||
goto out;
|
||||
|
|
|
@ -58,7 +58,7 @@ krb5_encrypt(
|
|||
int length)
|
||||
{
|
||||
u32 ret = -EINVAL;
|
||||
struct scatterlist sg[1];
|
||||
struct scatterlist sg[1];
|
||||
u8 local_iv[16] = {0};
|
||||
struct blkcipher_desc desc = { .tfm = tfm, .info = local_iv };
|
||||
|
||||
|
@ -67,7 +67,7 @@ krb5_encrypt(
|
|||
|
||||
if (crypto_blkcipher_ivsize(tfm) > 16) {
|
||||
dprintk("RPC: gss_k5encrypt: tfm iv size to large %d\n",
|
||||
crypto_blkcipher_ivsize(tfm));
|
||||
crypto_blkcipher_ivsize(tfm));
|
||||
goto out;
|
||||
}
|
||||
|
||||
|
|
|
@ -255,7 +255,7 @@ spkm3_verify_mic_token(unsigned char **tokp, int *mic_hdrlen, unsigned char **ck
|
|||
goto out;
|
||||
}
|
||||
/* checksum */
|
||||
*cksum = (&ptr[10] + ctxelen); /* ctxelen includes ptr[10] */
|
||||
*cksum = (&ptr[10] + ctxelen); /* ctxelen includes ptr[10] */
|
||||
|
||||
ret = GSS_S_COMPLETE;
|
||||
out:
|
||||
|
|
|
@ -172,8 +172,8 @@ static struct cache_head *rsi_alloc(void)
|
|||
}
|
||||
|
||||
static void rsi_request(struct cache_detail *cd,
|
||||
struct cache_head *h,
|
||||
char **bpp, int *blen)
|
||||
struct cache_head *h,
|
||||
char **bpp, int *blen)
|
||||
{
|
||||
struct rsi *rsii = container_of(h, struct rsi, h);
|
||||
|
||||
|
@ -184,7 +184,7 @@ static void rsi_request(struct cache_detail *cd,
|
|||
|
||||
|
||||
static int rsi_parse(struct cache_detail *cd,
|
||||
char *mesg, int mlen)
|
||||
char *mesg, int mlen)
|
||||
{
|
||||
/* context token expiry major minor context token */
|
||||
char *buf = mesg;
|
||||
|
|
|
@ -297,14 +297,14 @@ void cache_register(struct cache_detail *cd)
|
|||
cd->proc_ent->owner = cd->owner;
|
||||
cd->channel_ent = cd->content_ent = NULL;
|
||||
|
||||
p = create_proc_entry("flush", S_IFREG|S_IRUSR|S_IWUSR,
|
||||
cd->proc_ent);
|
||||
p = create_proc_entry("flush", S_IFREG|S_IRUSR|S_IWUSR,
|
||||
cd->proc_ent);
|
||||
cd->flush_ent = p;
|
||||
if (p) {
|
||||
p->proc_fops = &cache_flush_operations;
|
||||
p->owner = cd->owner;
|
||||
p->data = cd;
|
||||
}
|
||||
if (p) {
|
||||
p->proc_fops = &cache_flush_operations;
|
||||
p->owner = cd->owner;
|
||||
p->data = cd;
|
||||
}
|
||||
|
||||
if (cd->cache_request || cd->cache_parse) {
|
||||
p = create_proc_entry("channel", S_IFREG|S_IRUSR|S_IWUSR,
|
||||
|
@ -316,16 +316,16 @@ void cache_register(struct cache_detail *cd)
|
|||
p->data = cd;
|
||||
}
|
||||
}
|
||||
if (cd->cache_show) {
|
||||
p = create_proc_entry("content", S_IFREG|S_IRUSR|S_IWUSR,
|
||||
cd->proc_ent);
|
||||
if (cd->cache_show) {
|
||||
p = create_proc_entry("content", S_IFREG|S_IRUSR|S_IWUSR,
|
||||
cd->proc_ent);
|
||||
cd->content_ent = p;
|
||||
if (p) {
|
||||
p->proc_fops = &content_file_operations;
|
||||
p->owner = cd->owner;
|
||||
p->data = cd;
|
||||
}
|
||||
}
|
||||
if (p) {
|
||||
p->proc_fops = &content_file_operations;
|
||||
p->owner = cd->owner;
|
||||
p->data = cd;
|
||||
}
|
||||
}
|
||||
}
|
||||
rwlock_init(&cd->hash_lock);
|
||||
INIT_LIST_HEAD(&cd->queue);
|
||||
|
|
|
@ -514,7 +514,7 @@ choose_pool(struct svc_serv *serv, struct svc_pool *pool, unsigned int *state)
|
|||
if (pool != NULL)
|
||||
return pool;
|
||||
|
||||
return &serv->sv_pools[(*state)++ % serv->sv_nrpools];
|
||||
return &serv->sv_pools[(*state)++ % serv->sv_nrpools];
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -530,13 +530,13 @@ choose_victim(struct svc_serv *serv, struct svc_pool *pool, unsigned int *state)
|
|||
spin_lock_bh(&pool->sp_lock);
|
||||
} else {
|
||||
/* choose a pool in round-robin fashion */
|
||||
for (i = 0; i < serv->sv_nrpools; i++) {
|
||||
pool = &serv->sv_pools[--(*state) % serv->sv_nrpools];
|
||||
for (i = 0; i < serv->sv_nrpools; i++) {
|
||||
pool = &serv->sv_pools[--(*state) % serv->sv_nrpools];
|
||||
spin_lock_bh(&pool->sp_lock);
|
||||
if (!list_empty(&pool->sp_all_threads))
|
||||
goto found_pool;
|
||||
if (!list_empty(&pool->sp_all_threads))
|
||||
goto found_pool;
|
||||
spin_unlock_bh(&pool->sp_lock);
|
||||
}
|
||||
}
|
||||
return NULL;
|
||||
}
|
||||
|
||||
|
@ -551,7 +551,7 @@ found_pool:
|
|||
rqstp = list_entry(pool->sp_all_threads.next, struct svc_rqst, rq_all);
|
||||
list_del_init(&rqstp->rq_all);
|
||||
task = rqstp->rq_task;
|
||||
}
|
||||
}
|
||||
spin_unlock_bh(&pool->sp_lock);
|
||||
|
||||
return task;
|
||||
|
|
|
@ -484,7 +484,7 @@ svc_sendto(struct svc_rqst *rqstp, struct xdr_buf *xdr)
|
|||
if (xdr->tail[0].iov_len) {
|
||||
result = kernel_sendpage(sock, rqstp->rq_respages[0],
|
||||
((unsigned long)xdr->tail[0].iov_base)
|
||||
& (PAGE_SIZE-1),
|
||||
& (PAGE_SIZE-1),
|
||||
xdr->tail[0].iov_len, 0);
|
||||
|
||||
if (result > 0)
|
||||
|
@ -1063,7 +1063,7 @@ svc_tcp_recvfrom(struct svc_rqst *rqstp)
|
|||
|
||||
if (len < want) {
|
||||
dprintk("svc: short recvfrom while reading record length (%d of %lu)\n",
|
||||
len, want);
|
||||
len, want);
|
||||
svc_sock_received(svsk);
|
||||
return -EAGAIN; /* record header not complete */
|
||||
}
|
||||
|
@ -1651,7 +1651,7 @@ svc_delete_socket(struct svc_sock *svsk)
|
|||
|
||||
if (!test_and_set_bit(SK_DETACHED, &svsk->sk_flags))
|
||||
list_del_init(&svsk->sk_list);
|
||||
/*
|
||||
/*
|
||||
* We used to delete the svc_sock from whichever list
|
||||
* it's sk_ready node was on, but we don't actually
|
||||
* need to. This is because the only time we're called
|
||||
|
|
|
@ -1024,7 +1024,7 @@ xdr_encode_array2(struct xdr_buf *buf, unsigned int base,
|
|||
|
||||
int
|
||||
xdr_process_buf(struct xdr_buf *buf, unsigned int offset, unsigned int len,
|
||||
int (*actor)(struct scatterlist *, void *), void *data)
|
||||
int (*actor)(struct scatterlist *, void *), void *data)
|
||||
{
|
||||
int i, ret = 0;
|
||||
unsigned page_len, thislen, page_offset;
|
||||
|
|
Loading…
Reference in New Issue