/external/syslinux/gpxe/src/net/ |
infiniband.c | 75 * @ret cq New completion queue 80 struct ib_completion_queue *cq; local 86 cq = zalloc ( sizeof ( *cq ) ); 87 if ( ! cq ) 89 cq->ibdev = ibdev; 90 list_add ( &cq->list, &ibdev->cqs ); 91 cq->num_cqes = num_cqes; 92 INIT_LIST_HEAD ( &cq->work_queues ); 93 cq->op = op 815 struct ib_completion_queue *cq; local [all...] |
/external/syslinux/gpxe/src/drivers/net/ |
mtnic.c | 177 u32 size, u16 stride, u16 cq, u8 is_rx) 187 ring->cq = cq; /* CQ number associated with this ring */ 247 /* Map Tx+CQ doorbells */ 265 * mtnic alloc CQ 267 * Alloc and configure CQ. 271 mtnic_alloc_cq(struct net_device *dev, int num, struct mtnic_cq *cq, 277 cq->num = num; 278 cq->dev = dev 1316 struct mtnic_cq *cq; local 1491 struct mtnic_cq *cq; local [all...] |
mtnic.h | 170 MTNIC_IF_CAP_TX_CQ_DB_OFFSET = 0x13, /* offset in bytes for TX, CQ doorbell record */ 262 u16 cq; /* index of port CQ associated with this ring */ member in struct:mtnic_ring 283 * CQ 314 int num; /* CQ number (on attached port) */ 315 u32 size; /* number of CQEs in CQ */ 322 u16 ring; /* ring associated with this CQ */ 440 struct mtnic_cq cq[NUM_CQS]; member in struct:mtnic_port 519 u8 log_cq_p1; /* log2 cq for port1 */ 527 u8 log_cq_p2; /* log2 cq for port1 * 592 u8 cq; member in struct:mtnic_if_config_cq_in_mbox 619 u8 cq; \/* CQ associated with this ring *\/ member in struct:mtnic_if_config_rx_ring_in_mbox 653 u16 cq; \/* CQ associated with this ring *\/ member in struct:mtnic_if_config_send_ring_in_mbox [all...] |
ipoib.c | 57 struct ib_completion_queue *cq; member in struct:ipoib_device 618 ipoib->cq = ib_create_cq ( ibdev, IPOIB_NUM_CQES, &ipoib_cq_op ); 619 if ( ! ipoib->cq ) { 628 IPOIB_NUM_SEND_WQES, ipoib->cq, 629 IPOIB_NUM_RECV_WQES, ipoib->cq ); 651 ib_destroy_cq ( ibdev, ipoib->cq ); 676 ib_destroy_cq ( ibdev, ipoib->cq );
|
/external/flatbuffers/tests/ |
monster_test.grpc.fb.h | 35 std::unique_ptr< ::grpc::ClientAsyncResponseReaderInterface< flatbuffers::BufferRef<Stat>>> AsyncStore(::grpc::ClientContext* context, const flatbuffers::BufferRef<Monster>& request, ::grpc::CompletionQueue* cq) { 36 return std::unique_ptr< ::grpc::ClientAsyncResponseReaderInterface< flatbuffers::BufferRef<Stat>>>(AsyncStoreRaw(context, request, cq)); 41 std::unique_ptr< ::grpc::ClientAsyncReaderInterface< flatbuffers::BufferRef<Monster>>> AsyncRetrieve(::grpc::ClientContext* context, const flatbuffers::BufferRef<Stat>& request, ::grpc::CompletionQueue* cq, void* tag) { 42 return std::unique_ptr< ::grpc::ClientAsyncReaderInterface< flatbuffers::BufferRef<Monster>>>(AsyncRetrieveRaw(context, request, cq, tag)); 45 virtual ::grpc::ClientAsyncResponseReaderInterface< flatbuffers::BufferRef<Stat>>* AsyncStoreRaw(::grpc::ClientContext* context, const flatbuffers::BufferRef<Monster>& request, ::grpc::CompletionQueue* cq) = 0; 47 virtual ::grpc::ClientAsyncReaderInterface< flatbuffers::BufferRef<Monster>>* AsyncRetrieveRaw(::grpc::ClientContext* context, const flatbuffers::BufferRef<Stat>& request, ::grpc::CompletionQueue* cq, void* tag) = 0; 53 std::unique_ptr< ::grpc::ClientAsyncResponseReader< flatbuffers::BufferRef<Stat>>> AsyncStore(::grpc::ClientContext* context, const flatbuffers::BufferRef<Monster>& request, ::grpc::CompletionQueue* cq) { 54 return std::unique_ptr< ::grpc::ClientAsyncResponseReader< flatbuffers::BufferRef<Stat>>>(AsyncStoreRaw(context, request, cq)); 59 std::unique_ptr< ::grpc::ClientAsyncReader< flatbuffers::BufferRef<Monster>>> AsyncRetrieve(::grpc::ClientContext* context, const flatbuffers::BufferRef<Stat>& request, ::grpc::CompletionQueue* cq, void* tag) { 60 return std::unique_ptr< ::grpc::ClientAsyncReader< flatbuffers::BufferRef<Monster>>>(AsyncRetrieveRaw(context, request, cq, tag)) [all...] |
monster_test.grpc.fb.cc | 39 ::grpc::ClientAsyncResponseReader< flatbuffers::BufferRef<Stat>>* MonsterStorage::Stub::AsyncStoreRaw(::grpc::ClientContext* context, const flatbuffers::BufferRef<Monster>& request, ::grpc::CompletionQueue* cq) { 40 return new ::grpc::ClientAsyncResponseReader< flatbuffers::BufferRef<Stat>>(channel_.get(), cq, rpcmethod_Store_, context, request); 47 ::grpc::ClientAsyncReader< flatbuffers::BufferRef<Monster>>* MonsterStorage::Stub::AsyncRetrieveRaw(::grpc::ClientContext* context, const flatbuffers::BufferRef<Stat>& request, ::grpc::CompletionQueue* cq, void* tag) { 48 return new ::grpc::ClientAsyncReader< flatbuffers::BufferRef<Monster>>(channel_.get(), cq, rpcmethod_Retrieve_, context, request, tag);
|
/external/syslinux/gpxe/src/include/gpxe/ |
infiniband.h | 100 struct ib_completion_queue *cq; member in struct:ib_work_queue 235 * @v cq Completion queue 239 struct ib_completion_queue *cq ); 243 * @v cq Completion queue 246 struct ib_completion_queue *cq ); 305 * @v cq Completion queue 311 struct ib_completion_queue *cq ); 439 struct ib_completion_queue *cq ); 441 struct ib_completion_queue *cq ); 454 extern struct ib_work_queue * ib_find_wq ( struct ib_completion_queue *cq, [all...] |
ib_mi.h | 92 struct ib_completion_queue *cq; member in struct:ib_mad_interface
|
/external/autotest/contrib/ |
move_cq_hosts.sh | 3 TO_HOST=cautotest-cq 4 POOL=pool:cq
|
/external/syslinux/gpxe/src/net/infiniband/ |
ib_cmrc.c | 77 struct ib_completion_queue *cq; member in struct:ib_cmrc_connection 119 ib_destroy_cq ( cmrc->ibdev, cmrc->cq ); 401 cmrc->cq = ib_create_cq ( ibdev, IB_CMRC_NUM_CQES, 403 if ( ! cmrc->cq ) { 412 cmrc->cq, IB_CMRC_NUM_RECV_WQES, cmrc->cq ); 429 ib_destroy_cq ( ibdev, cmrc->cq );
|
ib_mi.c | 347 mi->cq = ib_create_cq ( ibdev, IB_MI_NUM_CQES, &ib_mi_completion_ops ); 348 if ( ! mi->cq ) { 354 mi->qp = ib_create_qp ( ibdev, type, IB_MI_NUM_SEND_WQES, mi->cq, 355 IB_MI_NUM_RECV_WQES, mi->cq ); 379 ib_destroy_cq ( ibdev, mi->cq ); 404 ib_destroy_cq ( ibdev, mi->cq );
|
/external/llvm/test/MC/Sparc/ |
sparc-coproc.s | 66 ! CHECK: std %cq, [%o3] ! encoding: [0xc1,0xb2,0xc0,0x00] 67 ! CHECK: std %cq, [%o3+-93] ! encoding: [0xc1,0xb2,0xff,0xa3] 68 ! CHECK: std %cq, [%o3+%l5] ! encoding: [0xc1,0xb2,0xc0,0x15] 69 std %cq, [%o3] 70 std %cq, [%o3-93] 71 std %cq, [%o3+%l5]
|
/external/syslinux/lzo/ |
Makefile | 29 $(AR) cq $@ $^
|
/external/iproute2/examples/bpf/ |
bpf_agent.c | 74 struct count_queue cq; local 77 memset(&cq, 0, sizeof(cq)); 78 ret = bpf_lookup_elem(fd, &i, &cq); 82 i, cq.total, cq.mismatch);
|
bpf_prog.c | 407 struct count_queue *cq, _cq; local 412 cq = map_lookup_elem(&map_queue, &queue); 413 if (likely(cq)) { 414 lock_xadd(&cq->total, 1); 416 lock_xadd(&cq->mismatch, 1);
|
/external/syslinux/com32/libupload/ |
Makefile | 20 $(AR) cq $@ $^
|
/external/syslinux/com32/libutil/ |
Makefile | 48 $(AR) cq $@ $(LNXLIBOBJS)
|
/external/testng/src/main/java/org/testng/reporters/ |
EmailableReporter.java | 138 int cq = 0; local 151 cq += 1; 153 + (cq % 2 == 0 ? "even" : "odd") + "\">" + "<td"); 176 buff.append("<tr class=\"" + style + (cq % 2 == 0 ? "odd" : "even") 194 cq += 1; 195 m_out.print("<tr class=\"" + style + (cq % 2 == 0 ? "even" : "odd") 443 private void titleRow(String label, int cq) { 444 titleRow(label, cq, null); 447 private void titleRow(String label, int cq, String id) { 452 m_out.println( "><th colspan=\"" + cq + "\">" + label + "</th></tr>") [all...] |
/external/elfutils/src/ |
make-debug-archive.in | 129 (cd "$outdir" && $LS | $XARGS $AR cq "$new_archive") &&
|
/external/fio/engines/ |
rdma.c | 169 struct ibv_cq *cq; member in struct:rdmaio_data 267 while ((ret = ibv_poll_cq(rd->cq, 1, &wc)) == 1) { 272 log_err("fio: cq completion status %d(%s)\n", 383 log_err("fio: Failed to get cq event!\n"); 386 if (ev_cq != rd->cq) { 387 log_err("fio: Unknown CQ!\n"); 390 if (ibv_req_notify_cq(rd->cq, 0) != 0) { 399 ibv_ack_cq_events(rd->cq, ret); 435 rd->cq = ibv_create_cq(rd->child_cm_id->verbs, 438 rd->cq = ibv_create_cq(rd->cm_id->verbs [all...] |
/external/syslinux/com32/elflink/ldlinux/ |
Makefile | 42 $(AR) cq $@ $(LNXLIBOBJS)
|
/external/toolchain-utils/ |
generate-waterfall-reports.py | 37 ('bvt-cq', 'HWTest'), 241 out_file.write(' Build bvt- bvt-cq ' 261 cq = build_dict.get('bvt-cq', '[??/ ?? /??]') 263 cq_color = build_dict.get('bvt-cq-color', '') 279 inline, cq, 288 cq, toolchain, security)) 292 out_file.write('%25s %3s %s %s\n' % (builder, status, inline, cq))
|
/external/syslinux/gpxe/src/drivers/infiniband/ |
arbel.c | 544 * @v cq Completion queue 548 struct ib_completion_queue *cq ) { 565 cq->cqn = ( arbel->limits.reserved_cqs + cqn_offset ); 577 arbel_cq->cqe_size = ( cq->num_cqes * sizeof ( arbel_cq->cqe[0] ) ); 585 for ( i = 0 ; i < cq->num_cqes ; i++ ) { 595 cq_number, cq->cqn ); 600 cq_number, cq->cqn ); 609 log_cq_size, fls ( cq->num_cqes - 1 ) ); 613 MLX_FILL_1 ( &cqctx, 12, cqn, cq->cqn ); 618 if ( ( rc = arbel_cmd_sw2hw_cq ( arbel, cq->cqn, &cqctx ) ) != 0 ) [all...] |
hermon.c | 699 * @v cq Completion queue 703 struct ib_completion_queue *cq ) { 720 cq->cqn = ( hermon->cap.reserved_cqs + cqn_offset ); 730 hermon_cq->cqe_size = ( cq->num_cqes * sizeof ( hermon_cq->cqe[0] ) ); 738 for ( i = 0 ; i < cq->num_cqes ; i++ ) { 756 log_cq_size, fls ( cq->num_cqes - 1 ) ); 761 if ( ( rc = hermon_cmd_sw2hw_cq ( hermon, cq->cqn, &cqctx ) ) != 0 ) { 768 hermon, cq->cqn, hermon_cq->cqe, 770 ib_cq_set_drvdata ( cq, hermon_cq ); 789 * @v cq Completion queu [all...] |
/external/syslinux/dos/ |
Makefile | 65 $(AR) cq $@ $^
|