forked from luck/tmp_suning_uos_patched
f94b533d09
Add a driver for the Ammasso 1100 gigabit ethernet RNIC. Signed-off-by: Tom Tucker <tom@opengridcomputing.com> Signed-off-by: Steve Wise <swise@opengridcomputing.com> Signed-off-by: Roland Dreier <rolandd@cisco.com>
376 lines
8.7 KiB
C
376 lines
8.7 KiB
C
/*
|
|
* Copyright (c) 2005 Ammasso, Inc. All rights reserved.
|
|
* Copyright (c) 2005 Open Grid Computing, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
#include "c2.h"
|
|
#include "c2_vq.h"
|
|
|
|
#define PBL_VIRT 1
|
|
#define PBL_PHYS 2
|
|
|
|
/*
|
|
* Send all the PBL messages to convey the remainder of the PBL
|
|
* Wait for the adapter's reply on the last one.
|
|
* This is indicated by setting the MEM_PBL_COMPLETE in the flags.
|
|
*
|
|
* NOTE: vq_req is _not_ freed by this function. The VQ Host
|
|
* Reply buffer _is_ freed by this function.
|
|
*/
|
|
static int
|
|
send_pbl_messages(struct c2_dev *c2dev, u32 stag_index,
|
|
unsigned long va, u32 pbl_depth,
|
|
struct c2_vq_req *vq_req, int pbl_type)
|
|
{
|
|
u32 pbe_count; /* amt that fits in a PBL msg */
|
|
u32 count; /* amt in this PBL MSG. */
|
|
struct c2wr_nsmr_pbl_req *wr; /* PBL WR ptr */
|
|
struct c2wr_nsmr_pbl_rep *reply; /* reply ptr */
|
|
int err, pbl_virt, pbl_index, i;
|
|
|
|
switch (pbl_type) {
|
|
case PBL_VIRT:
|
|
pbl_virt = 1;
|
|
break;
|
|
case PBL_PHYS:
|
|
pbl_virt = 0;
|
|
break;
|
|
default:
|
|
return -EINVAL;
|
|
break;
|
|
}
|
|
|
|
pbe_count = (c2dev->req_vq.msg_size -
|
|
sizeof(struct c2wr_nsmr_pbl_req)) / sizeof(u64);
|
|
wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
|
|
if (!wr) {
|
|
return -ENOMEM;
|
|
}
|
|
c2_wr_set_id(wr, CCWR_NSMR_PBL);
|
|
|
|
/*
|
|
* Only the last PBL message will generate a reply from the verbs,
|
|
* so we set the context to 0 indicating there is no kernel verbs
|
|
* handler blocked awaiting this reply.
|
|
*/
|
|
wr->hdr.context = 0;
|
|
wr->rnic_handle = c2dev->adapter_handle;
|
|
wr->stag_index = stag_index; /* already swapped */
|
|
wr->flags = 0;
|
|
pbl_index = 0;
|
|
while (pbl_depth) {
|
|
count = min(pbe_count, pbl_depth);
|
|
wr->addrs_length = cpu_to_be32(count);
|
|
|
|
/*
|
|
* If this is the last message, then reference the
|
|
* vq request struct cuz we're gonna wait for a reply.
|
|
* also make this PBL msg as the last one.
|
|
*/
|
|
if (count == pbl_depth) {
|
|
/*
|
|
* reference the request struct. dereferenced in the
|
|
* int handler.
|
|
*/
|
|
vq_req_get(c2dev, vq_req);
|
|
wr->flags = cpu_to_be32(MEM_PBL_COMPLETE);
|
|
|
|
/*
|
|
* This is the last PBL message.
|
|
* Set the context to our VQ Request Object so we can
|
|
* wait for the reply.
|
|
*/
|
|
wr->hdr.context = (unsigned long) vq_req;
|
|
}
|
|
|
|
/*
|
|
* If pbl_virt is set then va is a virtual address
|
|
* that describes a virtually contiguous memory
|
|
* allocation. The wr needs the start of each virtual page
|
|
* to be converted to the corresponding physical address
|
|
* of the page. If pbl_virt is not set then va is an array
|
|
* of physical addresses and there is no conversion to do.
|
|
* Just fill in the wr with what is in the array.
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
if (pbl_virt) {
|
|
va += PAGE_SIZE;
|
|
} else {
|
|
wr->paddrs[i] =
|
|
cpu_to_be64(((u64 *)va)[pbl_index + i]);
|
|
}
|
|
}
|
|
|
|
/*
|
|
* Send WR to adapter
|
|
*/
|
|
err = vq_send_wr(c2dev, (union c2wr *) wr);
|
|
if (err) {
|
|
if (count <= pbe_count) {
|
|
vq_req_put(c2dev, vq_req);
|
|
}
|
|
goto bail0;
|
|
}
|
|
pbl_depth -= count;
|
|
pbl_index += count;
|
|
}
|
|
|
|
/*
|
|
* Now wait for the reply...
|
|
*/
|
|
err = vq_wait_for_reply(c2dev, vq_req);
|
|
if (err) {
|
|
goto bail0;
|
|
}
|
|
|
|
/*
|
|
* Process reply
|
|
*/
|
|
reply = (struct c2wr_nsmr_pbl_rep *) (unsigned long) vq_req->reply_msg;
|
|
if (!reply) {
|
|
err = -ENOMEM;
|
|
goto bail0;
|
|
}
|
|
|
|
err = c2_errno(reply);
|
|
|
|
vq_repbuf_free(c2dev, reply);
|
|
bail0:
|
|
kfree(wr);
|
|
return err;
|
|
}
|
|
|
|
#define C2_PBL_MAX_DEPTH 131072
|
|
int
|
|
c2_nsmr_register_phys_kern(struct c2_dev *c2dev, u64 *addr_list,
|
|
int page_size, int pbl_depth, u32 length,
|
|
u32 offset, u64 *va, enum c2_acf acf,
|
|
struct c2_mr *mr)
|
|
{
|
|
struct c2_vq_req *vq_req;
|
|
struct c2wr_nsmr_register_req *wr;
|
|
struct c2wr_nsmr_register_rep *reply;
|
|
u16 flags;
|
|
int i, pbe_count, count;
|
|
int err;
|
|
|
|
if (!va || !length || !addr_list || !pbl_depth)
|
|
return -EINTR;
|
|
|
|
/*
|
|
* Verify PBL depth is within rnic max
|
|
*/
|
|
if (pbl_depth > C2_PBL_MAX_DEPTH) {
|
|
return -EINTR;
|
|
}
|
|
|
|
/*
|
|
* allocate verbs request object
|
|
*/
|
|
vq_req = vq_req_alloc(c2dev);
|
|
if (!vq_req)
|
|
return -ENOMEM;
|
|
|
|
wr = kmalloc(c2dev->req_vq.msg_size, GFP_KERNEL);
|
|
if (!wr) {
|
|
err = -ENOMEM;
|
|
goto bail0;
|
|
}
|
|
|
|
/*
|
|
* build the WR
|
|
*/
|
|
c2_wr_set_id(wr, CCWR_NSMR_REGISTER);
|
|
wr->hdr.context = (unsigned long) vq_req;
|
|
wr->rnic_handle = c2dev->adapter_handle;
|
|
|
|
flags = (acf | MEM_VA_BASED | MEM_REMOTE);
|
|
|
|
/*
|
|
* compute how many pbes can fit in the message
|
|
*/
|
|
pbe_count = (c2dev->req_vq.msg_size -
|
|
sizeof(struct c2wr_nsmr_register_req)) / sizeof(u64);
|
|
|
|
if (pbl_depth <= pbe_count) {
|
|
flags |= MEM_PBL_COMPLETE;
|
|
}
|
|
wr->flags = cpu_to_be16(flags);
|
|
wr->stag_key = 0; //stag_key;
|
|
wr->va = cpu_to_be64(*va);
|
|
wr->pd_id = mr->pd->pd_id;
|
|
wr->pbe_size = cpu_to_be32(page_size);
|
|
wr->length = cpu_to_be32(length);
|
|
wr->pbl_depth = cpu_to_be32(pbl_depth);
|
|
wr->fbo = cpu_to_be32(offset);
|
|
count = min(pbl_depth, pbe_count);
|
|
wr->addrs_length = cpu_to_be32(count);
|
|
|
|
/*
|
|
* fill out the PBL for this message
|
|
*/
|
|
for (i = 0; i < count; i++) {
|
|
wr->paddrs[i] = cpu_to_be64(addr_list[i]);
|
|
}
|
|
|
|
/*
|
|
* regerence the request struct
|
|
*/
|
|
vq_req_get(c2dev, vq_req);
|
|
|
|
/*
|
|
* send the WR to the adapter
|
|
*/
|
|
err = vq_send_wr(c2dev, (union c2wr *) wr);
|
|
if (err) {
|
|
vq_req_put(c2dev, vq_req);
|
|
goto bail1;
|
|
}
|
|
|
|
/*
|
|
* wait for reply from adapter
|
|
*/
|
|
err = vq_wait_for_reply(c2dev, vq_req);
|
|
if (err) {
|
|
goto bail1;
|
|
}
|
|
|
|
/*
|
|
* process reply
|
|
*/
|
|
reply =
|
|
(struct c2wr_nsmr_register_rep *) (unsigned long) (vq_req->reply_msg);
|
|
if (!reply) {
|
|
err = -ENOMEM;
|
|
goto bail1;
|
|
}
|
|
if ((err = c2_errno(reply))) {
|
|
goto bail2;
|
|
}
|
|
//*p_pb_entries = be32_to_cpu(reply->pbl_depth);
|
|
mr->ibmr.lkey = mr->ibmr.rkey = be32_to_cpu(reply->stag_index);
|
|
vq_repbuf_free(c2dev, reply);
|
|
|
|
/*
|
|
* if there are still more PBEs we need to send them to
|
|
* the adapter and wait for a reply on the final one.
|
|
* reuse vq_req for this purpose.
|
|
*/
|
|
pbl_depth -= count;
|
|
if (pbl_depth) {
|
|
|
|
vq_req->reply_msg = (unsigned long) NULL;
|
|
atomic_set(&vq_req->reply_ready, 0);
|
|
err = send_pbl_messages(c2dev,
|
|
cpu_to_be32(mr->ibmr.lkey),
|
|
(unsigned long) &addr_list[i],
|
|
pbl_depth, vq_req, PBL_PHYS);
|
|
if (err) {
|
|
goto bail1;
|
|
}
|
|
}
|
|
|
|
vq_req_free(c2dev, vq_req);
|
|
kfree(wr);
|
|
|
|
return err;
|
|
|
|
bail2:
|
|
vq_repbuf_free(c2dev, reply);
|
|
bail1:
|
|
kfree(wr);
|
|
bail0:
|
|
vq_req_free(c2dev, vq_req);
|
|
return err;
|
|
}
|
|
|
|
int c2_stag_dealloc(struct c2_dev *c2dev, u32 stag_index)
|
|
{
|
|
struct c2_vq_req *vq_req; /* verbs request object */
|
|
struct c2wr_stag_dealloc_req wr; /* work request */
|
|
struct c2wr_stag_dealloc_rep *reply; /* WR reply */
|
|
int err;
|
|
|
|
|
|
/*
|
|
* allocate verbs request object
|
|
*/
|
|
vq_req = vq_req_alloc(c2dev);
|
|
if (!vq_req) {
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/*
|
|
* Build the WR
|
|
*/
|
|
c2_wr_set_id(&wr, CCWR_STAG_DEALLOC);
|
|
wr.hdr.context = (u64) (unsigned long) vq_req;
|
|
wr.rnic_handle = c2dev->adapter_handle;
|
|
wr.stag_index = cpu_to_be32(stag_index);
|
|
|
|
/*
|
|
* reference the request struct. dereferenced in the int handler.
|
|
*/
|
|
vq_req_get(c2dev, vq_req);
|
|
|
|
/*
|
|
* Send WR to adapter
|
|
*/
|
|
err = vq_send_wr(c2dev, (union c2wr *) & wr);
|
|
if (err) {
|
|
vq_req_put(c2dev, vq_req);
|
|
goto bail0;
|
|
}
|
|
|
|
/*
|
|
* Wait for reply from adapter
|
|
*/
|
|
err = vq_wait_for_reply(c2dev, vq_req);
|
|
if (err) {
|
|
goto bail0;
|
|
}
|
|
|
|
/*
|
|
* Process reply
|
|
*/
|
|
reply = (struct c2wr_stag_dealloc_rep *) (unsigned long) vq_req->reply_msg;
|
|
if (!reply) {
|
|
err = -ENOMEM;
|
|
goto bail0;
|
|
}
|
|
|
|
err = c2_errno(reply);
|
|
|
|
vq_repbuf_free(c2dev, reply);
|
|
bail0:
|
|
vq_req_free(c2dev, vq_req);
|
|
return err;
|
|
}
|