Home
last modified time | relevance | path

Searched refs:queue_type (Results 1 – 25 of 54) sorted by relevance

123

/linux-6.14.4/drivers/infiniband/sw/rxe/
Drxe_queue.h53 enum queue_type { enum
70 enum queue_type type;
86 unsigned int elem_size, enum queue_type type);
101 enum queue_type type) in queue_get_producer()
128 enum queue_type type) in queue_get_consumer()
154 static inline int queue_empty(struct rxe_queue *q, enum queue_type type) in queue_empty()
162 static inline int queue_full(struct rxe_queue *q, enum queue_type type) in queue_full()
171 enum queue_type type) in queue_count()
180 enum queue_type type) in queue_advance_producer()
216 enum queue_type type) in queue_advance_consumer()
[all …]
Drxe_queue.c56 unsigned int elem_size, enum queue_type type) in rxe_queue_init()
113 enum queue_type type = q->type; in resize_finish()
Drxe_cq.c47 enum queue_type type; in rxe_cq_from_init()
/linux-6.14.4/drivers/gpu/drm/amd/amdgpu/
Damdgpu_mes.h188 int queue_type; member
194 int queue_type; member
232 uint32_t queue_type; member
255 uint32_t queue_type; member
264 uint32_t queue_type; member
274 uint32_t queue_type; member
295 uint32_t queue_type; member
424 int amdgpu_mes_reset_hw_queue_mmio(struct amdgpu_device *adev, int queue_type,
455 int queue_type, int idx,
Damdgpu_mes.c538 struct amdgpu_mqd *mqd_mgr = &adev->mqds[p->queue_type]; in amdgpu_mes_queue_alloc_mqd()
569 struct amdgpu_mqd *mqd_mgr = &adev->mqds[p->queue_type]; in amdgpu_mes_queue_init_mqd()
584 if (p->queue_type == AMDGPU_RING_TYPE_GFX || in amdgpu_mes_queue_init_mqd()
585 p->queue_type == AMDGPU_RING_TYPE_COMPUTE) { in amdgpu_mes_queue_init_mqd()
592 if (p->queue_type == AMDGPU_RING_TYPE_GFX || in amdgpu_mes_queue_init_mqd()
593 p->queue_type == AMDGPU_RING_TYPE_COMPUTE) { in amdgpu_mes_queue_init_mqd()
651 qprops->queue_type, in amdgpu_mes_add_hw_queue()
679 queue_input.queue_type = qprops->queue_type; in amdgpu_mes_add_hw_queue()
692 gang->process->pasid, gang_id, qprops->queue_type, in amdgpu_mes_add_hw_queue()
698 queue->queue_type = qprops->queue_type; in amdgpu_mes_add_hw_queue()
[all …]
Dmes_v11_0.c274 static int convert_to_mes_queue_type(int queue_type) in convert_to_mes_queue_type() argument
276 if (queue_type == AMDGPU_RING_TYPE_GFX) in convert_to_mes_queue_type()
278 else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) in convert_to_mes_queue_type()
280 else if (queue_type == AMDGPU_RING_TYPE_SDMA) in convert_to_mes_queue_type()
322 mes_add_queue_pkt.queue_type = in mes_v11_0_add_hw_queue()
323 convert_to_mes_queue_type(input->queue_type); in mes_v11_0_add_hw_queue()
364 static int mes_v11_0_reset_queue_mmio(struct amdgpu_mes *mes, uint32_t queue_type, in mes_v11_0_reset_queue_mmio() argument
374 if (queue_type == AMDGPU_RING_TYPE_GFX) { in mes_v11_0_reset_queue_mmio()
407 } else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) { in mes_v11_0_reset_queue_mmio()
427 } else if (queue_type == AMDGPU_RING_TYPE_SDMA) { in mes_v11_0_reset_queue_mmio()
[all …]
Dmes_v12_0.c262 static int convert_to_mes_queue_type(int queue_type) in convert_to_mes_queue_type() argument
264 if (queue_type == AMDGPU_RING_TYPE_GFX) in convert_to_mes_queue_type()
266 else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) in convert_to_mes_queue_type()
268 else if (queue_type == AMDGPU_RING_TYPE_SDMA) in convert_to_mes_queue_type()
270 else if (queue_type == AMDGPU_RING_TYPE_MES) in convert_to_mes_queue_type()
308 mes_add_queue_pkt.queue_type = in mes_v12_0_add_hw_queue()
309 convert_to_mes_queue_type(input->queue_type); in mes_v12_0_add_hw_queue()
386 static int mes_v12_0_reset_queue_mmio(struct amdgpu_mes *mes, uint32_t queue_type, in mes_v12_0_reset_queue_mmio() argument
396 if (queue_type == AMDGPU_RING_TYPE_GFX) { in mes_v12_0_reset_queue_mmio()
429 } else if (queue_type == AMDGPU_RING_TYPE_COMPUTE) { in mes_v12_0_reset_queue_mmio()
[all …]
/linux-6.14.4/drivers/accel/habanalabs/common/
Dhw_queue.c53 if (!hdev->asic_prop.max_queues || q->queue_type == QUEUE_TYPE_HW) in hl_hw_queue_update_ci()
62 if (!cs_needs_completion(cs) || q->queue_type == QUEUE_TYPE_INT) in hl_hw_queue_update_ci()
257 if (q->queue_type != QUEUE_TYPE_HW) { in hl_hw_queue_send_cb_no_cmpl()
666 switch (q->queue_type) { in hl_hw_queue_schedule_cs()
683 q->queue_type); in hl_hw_queue_schedule_cs()
695 if (q->queue_type == QUEUE_TYPE_EXT) in hl_hw_queue_schedule_cs()
773 switch (job->queue_type) { in hl_hw_queue_schedule_cs()
796 if ((q->queue_type == QUEUE_TYPE_EXT) && in hl_hw_queue_schedule_cs()
1003 switch (q->queue_type) { in queue_init()
1021 q->queue_type); in queue_init()
[all …]
Dcommand_submission.c302 return (job->queue_type == QUEUE_TYPE_EXT); in is_cb_patched()
331 parser.queue_type = job->queue_type; in cs_parser()
384 (job->queue_type == QUEUE_TYPE_HW || job->queue_type == QUEUE_TYPE_INT)) { in hl_complete_job()
412 (job->queue_type == QUEUE_TYPE_EXT || job->queue_type == QUEUE_TYPE_HW)) { in hl_complete_job()
1205 enum hl_queue_type *queue_type, in validate_queue_index() argument
1270 *queue_type = hw_queue_prop->type; in validate_queue_index()
1301 enum hl_queue_type queue_type, bool is_kernel_allocated_cb) in hl_cs_allocate_job() argument
1313 job->queue_type = queue_type; in hl_cs_allocate_job()
1319 if (job->queue_type == QUEUE_TYPE_EXT) in hl_cs_allocate_job()
1545 enum hl_queue_type queue_type; in cs_ioctl_default() local
[all …]
/linux-6.14.4/drivers/staging/octeon/
Dethernet-tx.c134 enum {QUEUE_CORE, QUEUE_HW, QUEUE_DROP} queue_type; in cvm_oct_xmit() local
191 queue_type = QUEUE_DROP; in cvm_oct_xmit()
398 queue_type = QUEUE_CORE; in cvm_oct_xmit()
401 queue_type = QUEUE_HW; in cvm_oct_xmit()
421 queue_type = QUEUE_DROP; in cvm_oct_xmit()
436 queue_type = QUEUE_DROP; in cvm_oct_xmit()
441 switch (queue_type) { in cvm_oct_xmit()
/linux-6.14.4/drivers/gpu/drm/amd/amdkfd/
Dkfd_packet_manager_vi.c125 packet->bitfields2.queue_type = in pm_set_resources_vi()
159 packet->bitfields2.queue_type = in pm_map_queues_vi()
165 packet->bitfields2.queue_type = in pm_map_queues_vi()
169 packet->bitfields2.queue_type = in pm_map_queues_vi()
Dkfd_packet_manager_v9.c201 packet->bitfields2.queue_type = in pm_set_resources_v9()
243 packet->bitfields2.queue_type = in pm_map_queues_v9()
249 packet->bitfields2.queue_type = in pm_map_queues_v9()
253 packet->bitfields2.queue_type = in pm_map_queues_v9()
Dkfd_pm4_headers_vi.h67 enum mes_set_resources_queue_type_enum queue_type:3; member
238 enum mes_map_queues_queue_type_vi_enum queue_type:3; member
Dkfd_pm4_headers_ai.h67 enum mes_set_resources_queue_type_enum queue_type:3; member
288 enum mes_map_queues_queue_type_enum queue_type:3; member
/linux-6.14.4/drivers/net/wireless/ath/ath5k/
Dqcu.c203 ath5k_hw_setup_tx_queue(struct ath5k_hw *ah, enum ath5k_tx_queue queue_type, in ath5k_hw_setup_tx_queue() argument
214 switch (queue_type) { in ath5k_hw_setup_tx_queue()
226 switch (queue_type) { in ath5k_hw_setup_tx_queue()
248 ah->ah_txq[queue].tqi_type = queue_type; in ath5k_hw_setup_tx_queue()
251 queue_info->tqi_type = queue_type; in ath5k_hw_setup_tx_queue()
/linux-6.14.4/drivers/net/
Dvirtio_net.c4405 u32 queue_type; in virtnet_stats_ctx_init() local
4414 queue_type = VIRTNET_Q_TYPE_RX; in virtnet_stats_ctx_init()
4417 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_BASIC; in virtnet_stats_ctx_init()
4418 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_basic_desc_qstat); in virtnet_stats_ctx_init()
4419 ctx->size[queue_type] += sizeof(struct virtio_net_stats_rx_basic); in virtnet_stats_ctx_init()
4423 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_CSUM; in virtnet_stats_ctx_init()
4424 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_csum_desc_qstat); in virtnet_stats_ctx_init()
4425 ctx->size[queue_type] += sizeof(struct virtio_net_stats_rx_csum); in virtnet_stats_ctx_init()
4429 ctx->bitmap[queue_type] |= VIRTIO_NET_STATS_TYPE_RX_GSO; in virtnet_stats_ctx_init()
4430 ctx->desc_num[queue_type] += ARRAY_SIZE(virtnet_stats_rx_gso_desc_qstat); in virtnet_stats_ctx_init()
[all …]
/linux-6.14.4/drivers/scsi/qla2xxx/
Dqla_tmpl.h135 uint8_t queue_type; member
196 uint8_t queue_type; member
Dqla_tmpl.c329 uint type = ent->t263.queue_type; in qla27xx_fwdt_entry_t263()
365 ent->t263.queue_type == T263_QUEUE_TYPE_ATIO) { in qla27xx_fwdt_entry_t263()
633 ulong type = ent->t274.queue_type; in qla27xx_fwdt_entry_t274()
664 ent->t274.queue_type == T274_QUEUE_TYPE_ATIO_SHAD) { in qla27xx_fwdt_entry_t274()
/linux-6.14.4/drivers/gpu/drm/amd/include/
Dmes_v12_api_def.h305 enum MES_QUEUE_TYPE queue_type; member
370 enum MES_QUEUE_TYPE queue_type; member
515 enum MES_QUEUE_TYPE queue_type; member
627 enum MES_QUEUE_TYPE queue_type; member
Dmes_v11_api_def.h293 enum MES_QUEUE_TYPE queue_type; member
356 enum MES_QUEUE_TYPE queue_type; member
472 enum MES_QUEUE_TYPE queue_type; member
/linux-6.14.4/drivers/net/ethernet/ibm/ehea/
Dehea_phyp.c416 const u8 queue_type, const u64 resource_handle, in ehea_h_register_rpage() argument
422 | EHEA_BMASK_SET(H_REG_RPAGE_QT, queue_type); in ehea_h_register_rpage()
500 const u8 pagesize, const u8 queue_type, in ehea_h_register_rpage_mr() argument
509 queue_type, mr_handle, in ehea_h_register_rpage_mr()
Dehea_phyp.h376 const u8 queue_type,
397 const u8 pagesize, const u8 queue_type,
/linux-6.14.4/drivers/infiniband/hw/irdma/
Dprotos.h75 u8 irdma_get_encoded_wqe_size(u32 wqsize, enum irdma_queue_type queue_type);
/linux-6.14.4/drivers/net/ethernet/huawei/hinic/
Dhinic_hw_qp_ctxt.h150 u16 queue_type; member
/linux-6.14.4/drivers/s390/cio/
Dqdio.h263 #define queue_type(q) q->irq_ptr->qib.qfmt macro

123