Commit 8d78f0f2 authored by James Simmons's avatar James Simmons Committed by Greg Kroah-Hartman

staging: lustre: lnet: cleanup some of the > 80 line issues

Cleanup some of the checkpatch over 80 character lines
reported.
Signed-off-by: default avatarJames Simmons <jsimmons@infradead.org>
Signed-off-by: default avatarGreg Kroah-Hartman <gregkh@linuxfoundation.org>
parent 4c93630f
...@@ -96,7 +96,8 @@ ksocknal_destroy_route(struct ksock_route *route) ...@@ -96,7 +96,8 @@ ksocknal_destroy_route(struct ksock_route *route)
} }
static int static int
ksocknal_create_peer(struct ksock_peer **peerp, lnet_ni_t *ni, lnet_process_id_t id) ksocknal_create_peer(struct ksock_peer **peerp, lnet_ni_t *ni,
lnet_process_id_t id)
{ {
int cpt = lnet_cpt_of_nid(id.nid); int cpt = lnet_cpt_of_nid(id.nid);
struct ksock_net *net = ni->ni_data; struct ksock_net *net = ni->ni_data;
...@@ -319,7 +320,8 @@ ksocknal_get_peer_info(lnet_ni_t *ni, int index, ...@@ -319,7 +320,8 @@ ksocknal_get_peer_info(lnet_ni_t *ni, int index,
} }
static void static void
ksocknal_associate_route_conn_locked(struct ksock_route *route, struct ksock_conn *conn) ksocknal_associate_route_conn_locked(struct ksock_route *route,
struct ksock_conn *conn)
{ {
struct ksock_peer *peer = route->ksnr_peer; struct ksock_peer *peer = route->ksnr_peer;
int type = conn->ksnc_type; int type = conn->ksnc_type;
...@@ -821,7 +823,8 @@ ksocknal_select_ips(struct ksock_peer *peer, __u32 *peerips, int n_peerips) ...@@ -821,7 +823,8 @@ ksocknal_select_ips(struct ksock_peer *peer, __u32 *peerips, int n_peerips)
if (k < peer->ksnp_n_passive_ips) /* using it already */ if (k < peer->ksnp_n_passive_ips) /* using it already */
continue; continue;
k = ksocknal_match_peerip(iface, peerips, n_peerips); k = ksocknal_match_peerip(iface, peerips,
n_peerips);
xor = ip ^ peerips[k]; xor = ip ^ peerips[k];
this_netmatch = !(xor & iface->ksni_netmask) ? 1 : 0; this_netmatch = !(xor & iface->ksni_netmask) ? 1 : 0;
...@@ -1302,8 +1305,11 @@ ksocknal_create_conn(lnet_ni_t *ni, struct ksock_route *route, ...@@ -1302,8 +1305,11 @@ ksocknal_create_conn(lnet_ni_t *ni, struct ksock_route *route,
/* Take packets blocking for this connection. */ /* Take packets blocking for this connection. */
list_for_each_entry_safe(tx, txtmp, &peer->ksnp_tx_queue, tx_list) { list_for_each_entry_safe(tx, txtmp, &peer->ksnp_tx_queue, tx_list) {
if (conn->ksnc_proto->pro_match_tx(conn, tx, tx->tx_nonblk) == SOCKNAL_MATCH_NO) int match = conn->ksnc_proto->pro_match_tx(conn, tx,
continue; tx->tx_nonblk);
if (match == SOCKNAL_MATCH_NO)
continue;
list_del(&tx->tx_list); list_del(&tx->tx_list);
ksocknal_queue_tx_locked(tx, conn); ksocknal_queue_tx_locked(tx, conn);
...@@ -1493,8 +1499,8 @@ ksocknal_close_conn_locked(struct ksock_conn *conn, int error) ...@@ -1493,8 +1499,8 @@ ksocknal_close_conn_locked(struct ksock_conn *conn, int error)
spin_unlock_bh(&conn->ksnc_scheduler->kss_lock); spin_unlock_bh(&conn->ksnc_scheduler->kss_lock);
} }
peer->ksnp_proto = NULL; /* renegotiate protocol version */ peer->ksnp_proto = NULL; /* renegotiate protocol version */
peer->ksnp_error = error; /* stash last conn close reason */ peer->ksnp_error = error; /* stash last conn close reason */
if (list_empty(&peer->ksnp_routes)) { if (list_empty(&peer->ksnp_routes)) {
/* /*
...@@ -1786,7 +1792,8 @@ ksocknal_close_matching_conns(lnet_process_id_t id, __u32 ipaddr) ...@@ -1786,7 +1792,8 @@ ksocknal_close_matching_conns(lnet_process_id_t id, __u32 ipaddr)
(id.pid == LNET_PID_ANY || id.pid == peer->ksnp_id.pid))) (id.pid == LNET_PID_ANY || id.pid == peer->ksnp_id.pid)))
continue; continue;
count += ksocknal_close_peer_conns_locked(peer, ipaddr, 0); count += ksocknal_close_peer_conns_locked(peer, ipaddr,
0);
} }
} }
...@@ -2026,7 +2033,10 @@ ksocknal_add_interface(lnet_ni_t *ni, __u32 ipaddress, __u32 netmask) ...@@ -2026,7 +2033,10 @@ ksocknal_add_interface(lnet_ni_t *ni, __u32 ipaddress, __u32 netmask)
} }
rc = 0; rc = 0;
/* NB only new connections will pay attention to the new interface! */ /*
* NB only new connections will pay attention to the
* new interface!
*/
} }
write_unlock_bh(&ksocknal_data.ksnd_global_lock); write_unlock_bh(&ksocknal_data.ksnd_global_lock);
...@@ -2200,8 +2210,9 @@ ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg) ...@@ -2200,8 +2210,9 @@ ksocknal_ctl(lnet_ni_t *ni, unsigned int cmd, void *arg)
int txmem; int txmem;
int rxmem; int rxmem;
int nagle; int nagle;
struct ksock_conn *conn = ksocknal_get_conn_by_idx(ni, data->ioc_count); struct ksock_conn *conn;
conn = ksocknal_get_conn_by_idx(ni, data->ioc_count);
if (!conn) if (!conn)
return -ENOENT; return -ENOENT;
......
...@@ -620,7 +620,8 @@ ksocknal_launch_all_connections_locked(struct ksock_peer *peer) ...@@ -620,7 +620,8 @@ ksocknal_launch_all_connections_locked(struct ksock_peer *peer)
} }
struct ksock_conn * struct ksock_conn *
ksocknal_find_conn_locked(struct ksock_peer *peer, struct ksock_tx *tx, int nonblk) ksocknal_find_conn_locked(struct ksock_peer *peer, struct ksock_tx *tx,
int nonblk)
{ {
struct list_head *tmp; struct list_head *tmp;
struct ksock_conn *conn; struct ksock_conn *conn;
...@@ -630,10 +631,12 @@ ksocknal_find_conn_locked(struct ksock_peer *peer, struct ksock_tx *tx, int nonb ...@@ -630,10 +631,12 @@ ksocknal_find_conn_locked(struct ksock_peer *peer, struct ksock_tx *tx, int nonb
int fnob = 0; int fnob = 0;
list_for_each(tmp, &peer->ksnp_conns) { list_for_each(tmp, &peer->ksnp_conns) {
struct ksock_conn *c = list_entry(tmp, struct ksock_conn, ksnc_list); struct ksock_conn *c;
int nob = atomic_read(&c->ksnc_tx_nob) + int nob, rc;
c->ksnc_sock->sk->sk_wmem_queued;
int rc; c = list_entry(tmp, struct ksock_conn, ksnc_list);
nob = atomic_read(&c->ksnc_tx_nob) +
c->ksnc_sock->sk->sk_wmem_queued;
LASSERT(!c->ksnc_closing); LASSERT(!c->ksnc_closing);
LASSERT(c->ksnc_proto && LASSERT(c->ksnc_proto &&
...@@ -752,9 +755,9 @@ ksocknal_queue_tx_locked(struct ksock_tx *tx, struct ksock_conn *conn) ...@@ -752,9 +755,9 @@ ksocknal_queue_tx_locked(struct ksock_tx *tx, struct ksock_conn *conn)
LASSERT(msg->ksm_zc_cookies[1]); LASSERT(msg->ksm_zc_cookies[1]);
LASSERT(conn->ksnc_proto->pro_queue_tx_zcack); LASSERT(conn->ksnc_proto->pro_queue_tx_zcack);
/* ZC ACK piggybacked on ztx release tx later */
if (conn->ksnc_proto->pro_queue_tx_zcack(conn, tx, 0)) if (conn->ksnc_proto->pro_queue_tx_zcack(conn, tx, 0))
ztx = tx; /* ZC ACK piggybacked on ztx release tx later */ ztx = tx;
} else { } else {
/* /*
* It's a normal packet - can it piggback a noop zc-ack that * It's a normal packet - can it piggback a noop zc-ack that
...@@ -796,7 +799,8 @@ ksocknal_find_connectable_route_locked(struct ksock_peer *peer) ...@@ -796,7 +799,8 @@ ksocknal_find_connectable_route_locked(struct ksock_peer *peer)
LASSERT(!route->ksnr_connecting || route->ksnr_scheduled); LASSERT(!route->ksnr_connecting || route->ksnr_scheduled);
if (route->ksnr_scheduled) /* connections being established */ /* connections being established */
if (route->ksnr_scheduled)
continue; continue;
/* all route types connected ? */ /* all route types connected ? */
...@@ -1514,7 +1518,10 @@ int ksocknal_scheduler(void *arg) ...@@ -1514,7 +1518,10 @@ int ksocknal_scheduler(void *arg)
rc = ksocknal_process_transmit(conn, tx); rc = ksocknal_process_transmit(conn, tx);
if (rc == -ENOMEM || rc == -EAGAIN) { if (rc == -ENOMEM || rc == -EAGAIN) {
/* Incomplete send: replace tx on HEAD of tx_queue */ /*
* Incomplete send: replace tx on HEAD of
* tx_queue
*/
spin_lock_bh(&sched->kss_lock); spin_lock_bh(&sched->kss_lock);
list_add(&tx->tx_list, &conn->ksnc_tx_queue); list_add(&tx->tx_list, &conn->ksnc_tx_queue);
} else { } else {
...@@ -1724,7 +1731,8 @@ ksocknal_recv_hello(lnet_ni_t *ni, struct ksock_conn *conn, ...@@ -1724,7 +1731,8 @@ ksocknal_recv_hello(lnet_ni_t *ni, struct ksock_conn *conn,
timeout = active ? *ksocknal_tunables.ksnd_timeout : timeout = active ? *ksocknal_tunables.ksnd_timeout :
lnet_acceptor_timeout(); lnet_acceptor_timeout();
rc = lnet_sock_read(sock, &hello->kshm_magic, sizeof(hello->kshm_magic), timeout); rc = lnet_sock_read(sock, &hello->kshm_magic,
sizeof(hello->kshm_magic), timeout);
if (rc) { if (rc) {
CERROR("Error %d reading HELLO from %pI4h\n", CERROR("Error %d reading HELLO from %pI4h\n",
rc, &conn->ksnc_ipaddr); rc, &conn->ksnc_ipaddr);
...@@ -1798,7 +1806,8 @@ ksocknal_recv_hello(lnet_ni_t *ni, struct ksock_conn *conn, ...@@ -1798,7 +1806,8 @@ ksocknal_recv_hello(lnet_ni_t *ni, struct ksock_conn *conn,
conn->ksnc_port > LNET_ACCEPTOR_MAX_RESERVED_PORT) { conn->ksnc_port > LNET_ACCEPTOR_MAX_RESERVED_PORT) {
/* Userspace NAL assigns peer process ID from socket */ /* Userspace NAL assigns peer process ID from socket */
recv_id.pid = conn->ksnc_port | LNET_PID_USERFLAG; recv_id.pid = conn->ksnc_port | LNET_PID_USERFLAG;
recv_id.nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid), conn->ksnc_ipaddr); recv_id.nid = LNET_MKNID(LNET_NIDNET(ni->ni_nid),
conn->ksnc_ipaddr);
} else { } else {
recv_id.nid = hello->kshm_src_nid; recv_id.nid = hello->kshm_src_nid;
recv_id.pid = hello->kshm_src_pid; recv_id.pid = hello->kshm_src_pid;
...@@ -1882,7 +1891,8 @@ ksocknal_connect(struct ksock_route *route) ...@@ -1882,7 +1891,8 @@ ksocknal_connect(struct ksock_route *route)
if (peer->ksnp_accepting > 0) { if (peer->ksnp_accepting > 0) {
CDEBUG(D_NET, CDEBUG(D_NET,
"peer %s(%d) already connecting to me, retry later.\n", "peer %s(%d) already connecting to me, retry later.\n",
libcfs_nid2str(peer->ksnp_id.nid), peer->ksnp_accepting); libcfs_nid2str(peer->ksnp_id.nid),
peer->ksnp_accepting);
retry_later = 1; retry_later = 1;
} }
...@@ -2241,7 +2251,8 @@ ksocknal_connd(void *arg) ...@@ -2241,7 +2251,8 @@ ksocknal_connd(void *arg)
/* Nothing to do for 'timeout' */ /* Nothing to do for 'timeout' */
set_current_state(TASK_INTERRUPTIBLE); set_current_state(TASK_INTERRUPTIBLE);
add_wait_queue_exclusive(&ksocknal_data.ksnd_connd_waitq, &wait); add_wait_queue_exclusive(&ksocknal_data.ksnd_connd_waitq,
&wait);
spin_unlock_bh(connd_lock); spin_unlock_bh(connd_lock);
nloops = 0; nloops = 0;
...@@ -2371,7 +2382,8 @@ ksocknal_send_keepalive_locked(struct ksock_peer *peer) ...@@ -2371,7 +2382,8 @@ ksocknal_send_keepalive_locked(struct ksock_peer *peer)
struct ksock_conn *conn; struct ksock_conn *conn;
struct ksock_tx *tx; struct ksock_tx *tx;
if (list_empty(&peer->ksnp_conns)) /* last_alive will be updated by create_conn */ /* last_alive will be updated by create_conn */
if (list_empty(&peer->ksnp_conns))
return 0; return 0;
if (peer->ksnp_proto != &ksocknal_protocol_v3x) if (peer->ksnp_proto != &ksocknal_protocol_v3x)
......
...@@ -202,7 +202,8 @@ ksocknal_lib_recv_iov(struct ksock_conn *conn) ...@@ -202,7 +202,8 @@ ksocknal_lib_recv_iov(struct ksock_conn *conn)
fragnob = sum; fragnob = sum;
conn->ksnc_rx_csum = ksocknal_csum(conn->ksnc_rx_csum, conn->ksnc_rx_csum = ksocknal_csum(conn->ksnc_rx_csum,
iov[i].iov_base, fragnob); iov[i].iov_base,
fragnob);
} }
conn->ksnc_msg.ksm_csum = saved_csum; conn->ksnc_msg.ksm_csum = saved_csum;
} }
...@@ -291,7 +292,8 @@ ksocknal_lib_csum_tx(struct ksock_tx *tx) ...@@ -291,7 +292,8 @@ ksocknal_lib_csum_tx(struct ksock_tx *tx)
} }
int int
ksocknal_lib_get_conn_tunables(struct ksock_conn *conn, int *txmem, int *rxmem, int *nagle) ksocknal_lib_get_conn_tunables(struct ksock_conn *conn, int *txmem,
int *rxmem, int *nagle)
{ {
struct socket *sock = conn->ksnc_sock; struct socket *sock = conn->ksnc_sock;
int len; int len;
......
...@@ -194,7 +194,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn, ...@@ -194,7 +194,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn,
} }
if (!tx->tx_msg.ksm_zc_cookies[0]) { if (!tx->tx_msg.ksm_zc_cookies[0]) {
/* NOOP tx has only one ZC-ACK cookie, can carry at least one more */ /*
* NOOP tx has only one ZC-ACK cookie,
* can carry at least one more
*/
if (tx->tx_msg.ksm_zc_cookies[1] > cookie) { if (tx->tx_msg.ksm_zc_cookies[1] > cookie) {
tx->tx_msg.ksm_zc_cookies[0] = tx->tx_msg.ksm_zc_cookies[1]; tx->tx_msg.ksm_zc_cookies[0] = tx->tx_msg.ksm_zc_cookies[1];
tx->tx_msg.ksm_zc_cookies[1] = cookie; tx->tx_msg.ksm_zc_cookies[1] = cookie;
...@@ -203,7 +206,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn, ...@@ -203,7 +206,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn,
} }
if (tx->tx_msg.ksm_zc_cookies[0] - tx->tx_msg.ksm_zc_cookies[1] > 2) { if (tx->tx_msg.ksm_zc_cookies[0] - tx->tx_msg.ksm_zc_cookies[1] > 2) {
/* not likely to carry more ACKs, skip it to simplify logic */ /*
* not likely to carry more ACKs, skip it
* to simplify logic
*/
ksocknal_next_tx_carrier(conn); ksocknal_next_tx_carrier(conn);
} }
...@@ -237,7 +243,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn, ...@@ -237,7 +243,10 @@ ksocknal_queue_tx_zcack_v3(struct ksock_conn *conn,
} }
} else { } else {
/* ksm_zc_cookies[0] < ksm_zc_cookies[1], it is range of cookies */ /*
* ksm_zc_cookies[0] < ksm_zc_cookies[1],
* it is range of cookies
*/
if (cookie >= tx->tx_msg.ksm_zc_cookies[0] && if (cookie >= tx->tx_msg.ksm_zc_cookies[0] &&
cookie <= tx->tx_msg.ksm_zc_cookies[1]) { cookie <= tx->tx_msg.ksm_zc_cookies[1]) {
CWARN("%s: duplicated ZC cookie: %llu\n", CWARN("%s: duplicated ZC cookie: %llu\n",
...@@ -425,7 +434,8 @@ ksocknal_handle_zcack(struct ksock_conn *conn, __u64 cookie1, __u64 cookie2) ...@@ -425,7 +434,8 @@ ksocknal_handle_zcack(struct ksock_conn *conn, __u64 cookie1, __u64 cookie2)
tx_zc_list) { tx_zc_list) {
__u64 c = tx->tx_msg.ksm_zc_cookies[0]; __u64 c = tx->tx_msg.ksm_zc_cookies[0];
if (c == cookie1 || c == cookie2 || (cookie1 < c && c < cookie2)) { if (c == cookie1 || c == cookie2 ||
(cookie1 < c && c < cookie2)) {
tx->tx_msg.ksm_zc_cookies[0] = 0; tx->tx_msg.ksm_zc_cookies[0] = 0;
list_del(&tx->tx_zc_list); list_del(&tx->tx_zc_list);
list_add(&tx->tx_zc_list, &zlist); list_add(&tx->tx_zc_list, &zlist);
...@@ -639,7 +649,8 @@ ksocknal_recv_hello_v1(struct ksock_conn *conn, ksock_hello_msg_t *hello, ...@@ -639,7 +649,8 @@ ksocknal_recv_hello_v1(struct ksock_conn *conn, ksock_hello_msg_t *hello,
} }
static int static int
ksocknal_recv_hello_v2(struct ksock_conn *conn, ksock_hello_msg_t *hello, int timeout) ksocknal_recv_hello_v2(struct ksock_conn *conn, ksock_hello_msg_t *hello,
int timeout)
{ {
struct socket *sock = conn->ksnc_sock; struct socket *sock = conn->ksnc_sock;
int rc; int rc;
...@@ -737,7 +748,10 @@ ksocknal_pack_msg_v2(struct ksock_tx *tx) ...@@ -737,7 +748,10 @@ ksocknal_pack_msg_v2(struct ksock_tx *tx)
tx->tx_nob = offsetof(ksock_msg_t, ksm_u.lnetmsg.ksnm_hdr); tx->tx_nob = offsetof(ksock_msg_t, ksm_u.lnetmsg.ksnm_hdr);
tx->tx_resid = offsetof(ksock_msg_t, ksm_u.lnetmsg.ksnm_hdr); tx->tx_resid = offsetof(ksock_msg_t, ksm_u.lnetmsg.ksnm_hdr);
} }
/* Don't checksum before start sending, because packet can be piggybacked with ACK */ /*
* Don't checksum before start sending, because packet can be
* piggybacked with ACK
*/
} }
static void static void
......
...@@ -87,7 +87,8 @@ lstcon_rpc_done(struct srpc_client_rpc *rpc) ...@@ -87,7 +87,8 @@ lstcon_rpc_done(struct srpc_client_rpc *rpc)
static int static int
lstcon_rpc_init(struct lstcon_node *nd, int service, unsigned int feats, lstcon_rpc_init(struct lstcon_node *nd, int service, unsigned int feats,
int bulk_npg, int bulk_len, int embedded, struct lstcon_rpc *crpc) int bulk_npg, int bulk_len, int embedded,
struct lstcon_rpc *crpc)
{ {
crpc->crp_rpc = sfw_create_rpc(nd->nd_id, service, crpc->crp_rpc = sfw_create_rpc(nd->nd_id, service,
feats, bulk_npg, bulk_len, feats, bulk_npg, bulk_len,
...@@ -292,8 +293,8 @@ lstcon_rpc_trans_abort(struct lstcon_rpc_trans *trans, int error) ...@@ -292,8 +293,8 @@ lstcon_rpc_trans_abort(struct lstcon_rpc_trans *trans, int error)
spin_lock(&rpc->crpc_lock); spin_lock(&rpc->crpc_lock);
if (!crpc->crp_posted || /* not posted */ if (!crpc->crp_posted || /* not posted */
crpc->crp_stamp) { /* rpc done or aborted already */ crpc->crp_stamp) { /* rpc done or aborted already */
if (!crpc->crp_stamp) { if (!crpc->crp_stamp) {
crpc->crp_stamp = cfs_time_current(); crpc->crp_stamp = cfs_time_current();
crpc->crp_status = -EINTR; crpc->crp_status = -EINTR;
...@@ -778,7 +779,8 @@ lstcon_pingrpc_prep(lst_test_ping_param_t *param, struct srpc_test_reqst *req) ...@@ -778,7 +779,8 @@ lstcon_pingrpc_prep(lst_test_ping_param_t *param, struct srpc_test_reqst *req)
} }
static int static int
lstcon_bulkrpc_v0_prep(lst_test_bulk_param_t *param, struct srpc_test_reqst *req) lstcon_bulkrpc_v0_prep(lst_test_bulk_param_t *param,
struct srpc_test_reqst *req)
{ {
struct test_bulk_req *brq = &req->tsr_u.bulk_v0; struct test_bulk_req *brq = &req->tsr_u.bulk_v0;
......
...@@ -131,7 +131,8 @@ int lstcon_rpc_trans_interpreter(struct lstcon_rpc_trans *trans, ...@@ -131,7 +131,8 @@ int lstcon_rpc_trans_interpreter(struct lstcon_rpc_trans *trans,
lstcon_rpc_readent_func_t readent); lstcon_rpc_readent_func_t readent);
void lstcon_rpc_trans_abort(struct lstcon_rpc_trans *trans, int error); void lstcon_rpc_trans_abort(struct lstcon_rpc_trans *trans, int error);
void lstcon_rpc_trans_destroy(struct lstcon_rpc_trans *trans); void lstcon_rpc_trans_destroy(struct lstcon_rpc_trans *trans);
void lstcon_rpc_trans_addreq(struct lstcon_rpc_trans *trans, struct lstcon_rpc *req); void lstcon_rpc_trans_addreq(struct lstcon_rpc_trans *trans,
struct lstcon_rpc *req);
int lstcon_rpc_trans_postwait(struct lstcon_rpc_trans *trans, int timeout); int lstcon_rpc_trans_postwait(struct lstcon_rpc_trans *trans, int timeout);
int lstcon_rpc_pinger_start(void); int lstcon_rpc_pinger_start(void);
void lstcon_rpc_pinger_stop(void); void lstcon_rpc_pinger_stop(void);
......
...@@ -86,7 +86,7 @@ lstcon_node_find(lnet_process_id_t id, struct lstcon_node **ndpp, int create) ...@@ -86,7 +86,7 @@ lstcon_node_find(lnet_process_id_t id, struct lstcon_node **ndpp, int create)
if (!create) if (!create)
return -ENOENT; return -ENOENT;
LIBCFS_ALLOC(*ndpp, sizeof(struct lstcon_node) + sizeof(struct lstcon_ndlink)); LIBCFS_ALLOC(*ndpp, sizeof(*ndpp) + sizeof(*ndl));
if (!*ndpp) if (!*ndpp)
return -ENOMEM; return -ENOMEM;
...@@ -131,12 +131,12 @@ lstcon_node_put(struct lstcon_node *nd) ...@@ -131,12 +131,12 @@ lstcon_node_put(struct lstcon_node *nd)
list_del(&ndl->ndl_link); list_del(&ndl->ndl_link);
list_del(&ndl->ndl_hlink); list_del(&ndl->ndl_hlink);
LIBCFS_FREE(nd, sizeof(struct lstcon_node) + sizeof(struct lstcon_ndlink)); LIBCFS_FREE(nd, sizeof(*nd) + sizeof(*ndl));
} }
static int static int
lstcon_ndlink_find(struct list_head *hash, lstcon_ndlink_find(struct list_head *hash, lnet_process_id_t id,
lnet_process_id_t id, struct lstcon_ndlink **ndlpp, int create) struct lstcon_ndlink **ndlpp, int create)
{ {
unsigned int idx = LNET_NIDADDR(id.nid) % LST_NODE_HASHSIZE; unsigned int idx = LNET_NIDADDR(id.nid) % LST_NODE_HASHSIZE;
struct lstcon_ndlink *ndl; struct lstcon_ndlink *ndl;
...@@ -230,7 +230,8 @@ lstcon_group_addref(struct lstcon_group *grp) ...@@ -230,7 +230,8 @@ lstcon_group_addref(struct lstcon_group *grp)
grp->grp_ref++; grp->grp_ref++;
} }
static void lstcon_group_ndlink_release(struct lstcon_group *, struct lstcon_ndlink *); static void lstcon_group_ndlink_release(struct lstcon_group *,
struct lstcon_ndlink *);
static void static void
lstcon_group_drain(struct lstcon_group *grp, int keep) lstcon_group_drain(struct lstcon_group *grp, int keep)
...@@ -1183,7 +1184,8 @@ lstcon_testrpc_condition(int transop, struct lstcon_node *nd, void *arg) ...@@ -1183,7 +1184,8 @@ lstcon_testrpc_condition(int transop, struct lstcon_node *nd, void *arg)
} }
static int static int
lstcon_test_nodes_add(struct lstcon_test *test, struct list_head __user *result_up) lstcon_test_nodes_add(struct lstcon_test *test,
struct list_head __user *result_up)
{ {
struct lstcon_rpc_trans *trans; struct lstcon_rpc_trans *trans;
struct lstcon_group *grp; struct lstcon_group *grp;
...@@ -1366,7 +1368,8 @@ lstcon_test_add(char *batch_name, int type, int loop, ...@@ -1366,7 +1368,8 @@ lstcon_test_add(char *batch_name, int type, int loop,
} }
static int static int
lstcon_test_find(struct lstcon_batch *batch, int idx, struct lstcon_test **testpp) lstcon_test_find(struct lstcon_batch *batch, int idx,
struct lstcon_test **testpp)
{ {
struct lstcon_test *test; struct lstcon_test *test;
......
...@@ -131,7 +131,8 @@ sfw_find_test_case(int id) ...@@ -131,7 +131,8 @@ sfw_find_test_case(int id)
} }
static int static int
sfw_register_test(struct srpc_service *service, struct sfw_test_client_ops *cliops) sfw_register_test(struct srpc_service *service,
struct sfw_test_client_ops *cliops)
{ {
struct sfw_test_case *tsc; struct sfw_test_case *tsc;
...@@ -469,7 +470,8 @@ sfw_make_session(struct srpc_mksn_reqst *request, struct srpc_mksn_reply *reply) ...@@ -469,7 +470,8 @@ sfw_make_session(struct srpc_mksn_reqst *request, struct srpc_mksn_reply *reply)
} }
static int static int
sfw_remove_session(struct srpc_rmsn_reqst *request, struct srpc_rmsn_reply *reply) sfw_remove_session(struct srpc_rmsn_reqst *request,
struct srpc_rmsn_reply *reply)
{ {
struct sfw_session *sn = sfw_data.fw_session; struct sfw_session *sn = sfw_data.fw_session;
...@@ -501,7 +503,8 @@ sfw_remove_session(struct srpc_rmsn_reqst *request, struct srpc_rmsn_reply *repl ...@@ -501,7 +503,8 @@ sfw_remove_session(struct srpc_rmsn_reqst *request, struct srpc_rmsn_reply *repl
} }
static int static int
sfw_debug_session(struct srpc_debug_reqst *request, struct srpc_debug_reply *reply) sfw_debug_session(struct srpc_debug_reqst *request,
struct srpc_debug_reply *reply)
{ {
struct sfw_session *sn = sfw_data.fw_session; struct sfw_session *sn = sfw_data.fw_session;
...@@ -1064,7 +1067,8 @@ sfw_stop_batch(struct sfw_batch *tsb, int force) ...@@ -1064,7 +1067,8 @@ sfw_stop_batch(struct sfw_batch *tsb, int force)
} }
static int static int
sfw_query_batch(struct sfw_batch *tsb, int testidx, struct srpc_batch_reply *reply) sfw_query_batch(struct sfw_batch *tsb, int testidx,
struct srpc_batch_reply *reply)
{ {
struct sfw_test_instance *tsi; struct sfw_test_instance *tsi;
...@@ -1179,7 +1183,8 @@ sfw_add_test(struct srpc_server_rpc *rpc) ...@@ -1179,7 +1183,8 @@ sfw_add_test(struct srpc_server_rpc *rpc)
} }
static int static int
sfw_control_batch(struct srpc_batch_reqst *request, struct srpc_batch_reply *reply) sfw_control_batch(struct srpc_batch_reqst *request,
struct srpc_batch_reply *reply)
{ {
struct sfw_session *sn = sfw_data.fw_session; struct sfw_session *sn = sfw_data.fw_session;
int rc = 0; int rc = 0;
......
...@@ -696,7 +696,8 @@ srpc_finish_service(struct srpc_service *sv) ...@@ -696,7 +696,8 @@ srpc_finish_service(struct srpc_service *sv)
/* called with sv->sv_lock held */ /* called with sv->sv_lock held */
static void static void
srpc_service_recycle_buffer(struct srpc_service_cd *scd, struct srpc_buffer *buf) srpc_service_recycle_buffer(struct srpc_service_cd *scd,
struct srpc_buffer *buf)
__must_hold(&scd->scd_lock) __must_hold(&scd->scd_lock)
{ {
if (!scd->scd_svc->sv_shuttingdown && scd->scd_buf_adjust >= 0) { if (!scd->scd_svc->sv_shuttingdown && scd->scd_buf_adjust >= 0) {
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment