/drivers/gpu/drm/nouveau/nvkm/subdev/fault/ |
D | base.c | 30 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_fini() local 31 fault->func->buffer.intr(fault->buffer[index], false); in nvkm_fault_ntfy_fini() 37 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_init() local 38 fault->func->buffer.intr(fault->buffer[index], true); in nvkm_fault_ntfy_init() 65 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_intr() local 66 return fault->func->intr(fault); in nvkm_fault_intr() 72 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_fini() local 73 if (fault->func->fini) in nvkm_fault_fini() 74 fault->func->fini(fault); in nvkm_fault_fini() 81 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_init() local [all …]
|
D | gv100.c | 33 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_process() 42 const u32 base = get * buffer->fault->func->buffer.entry_size; in gv100_fault_buffer_process() 76 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_intr() 87 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_fini() 95 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_init() 107 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_info() 120 struct nvkm_fault *fault = container_of(notify, typeof(*fault), nrpfb); in gv100_fault_ntfy_nrpfb() local 121 gv100_fault_buffer_process(fault->buffer[0]); in gv100_fault_ntfy_nrpfb() 126 gv100_fault_intr_fault(struct nvkm_fault *fault) in gv100_fault_intr_fault() argument 128 struct nvkm_subdev *subdev = &fault->subdev; in gv100_fault_intr_fault() [all …]
|
D | tu102.c | 38 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_intr() 46 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_fini() 59 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_init() 75 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_info() 86 tu102_fault_intr_fault(struct nvkm_fault *fault) in tu102_fault_intr_fault() argument 88 struct nvkm_subdev *subdev = &fault->subdev; in tu102_fault_intr_fault() 112 tu102_fault_intr(struct nvkm_fault *fault) in tu102_fault_intr() argument 114 struct nvkm_subdev *subdev = &fault->subdev; in tu102_fault_intr() 119 tu102_fault_intr_fault(fault); in tu102_fault_intr() 128 if (fault->buffer[0]) { in tu102_fault_intr() [all …]
|
D | user.c | 35 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map() 48 *pevent = &buffer->fault->event; in nvkm_ufault_ntfy() 58 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini() 66 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init() 92 struct nvkm_fault *fault = device->fault; in nvkm_ufault_new() local 93 struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; in nvkm_ufault_new()
|
D | Kbuild | 2 nvkm-y += nvkm/subdev/fault/base.o 3 nvkm-y += nvkm/subdev/fault/user.o 4 nvkm-y += nvkm/subdev/fault/gp100.o 5 nvkm-y += nvkm/subdev/fault/gp10b.o 6 nvkm-y += nvkm/subdev/fault/gv100.o 7 nvkm-y += nvkm/subdev/fault/tu102.o
|
D | gp100.c | 32 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_intr() 39 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_fini() 46 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_init() 60 buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78); in gp100_fault_buffer_info() 66 gp100_fault_intr(struct nvkm_fault *fault) in gp100_fault_intr() argument 68 nvkm_event_send(&fault->event, 1, 0, NULL, 0); in gp100_fault_intr()
|
/drivers/infiniband/hw/hfi1/ |
D | fault.c | 70 if (!ibd->fault->n_rxfaults[i] && !ibd->fault->n_txfaults[i]) in _fault_stats_seq_show() 75 (unsigned long long)ibd->fault->n_rxfaults[i], in _fault_stats_seq_show() 76 (unsigned long long)ibd->fault->n_txfaults[i]); in _fault_stats_seq_show() 97 struct fault *fault = file->private_data; in fault_opcodes_write() local 139 bitmap_zero(fault->opcodes, sizeof(fault->opcodes) * in fault_opcodes_write() 149 clear_bit(i, fault->opcodes); in fault_opcodes_write() 151 set_bit(i, fault->opcodes); in fault_opcodes_write() 171 struct fault *fault = file->private_data; in fault_opcodes_read() local 172 size_t bitsize = sizeof(fault->opcodes) * BITS_PER_BYTE; in fault_opcodes_read() 180 bit = find_first_bit(fault->opcodes, bitsize); in fault_opcodes_read() [all …]
|
/drivers/iommu/amd/ |
D | iommu_v2.c | 67 struct fault { struct 427 static void handle_fault_error(struct fault *fault) in handle_fault_error() argument 431 if (!fault->dev_state->inv_ppr_cb) { in handle_fault_error() 432 set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); in handle_fault_error() 436 status = fault->dev_state->inv_ppr_cb(fault->dev_state->pdev, in handle_fault_error() 437 fault->pasid, in handle_fault_error() 438 fault->address, in handle_fault_error() 439 fault->flags); in handle_fault_error() 442 set_pri_tag_status(fault->state, fault->tag, PPR_SUCCESS); in handle_fault_error() 445 set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); in handle_fault_error() [all …]
|
/drivers/gpu/drm/nouveau/ |
D | nouveau_svm.c | 64 u8 fault; member 66 } **fault; member 413 struct nouveau_svm_fault *fault) in nouveau_svm_fault_cancel_fault() argument 415 nouveau_svm_fault_cancel(svm, fault->inst, in nouveau_svm_fault_cancel_fault() 416 fault->hub, in nouveau_svm_fault_cancel_fault() 417 fault->gpc, in nouveau_svm_fault_cancel_fault() 418 fault->client); in nouveau_svm_fault_cancel_fault() 422 nouveau_svm_fault_priority(u8 fault) in nouveau_svm_fault_priority() argument 424 switch (fault) { in nouveau_svm_fault_priority() 470 struct nouveau_svm_fault *fault; in nouveau_svm_fault_cache() local [all …]
|
/drivers/iommu/ |
D | io-pgfault.c | 44 struct iommu_fault fault; member 60 .pasid = iopf->fault.prm.pasid, in iopf_complete_group() 61 .grpid = iopf->fault.prm.grpid, in iopf_complete_group() 65 if ((iopf->fault.prm.flags & IOMMU_FAULT_PAGE_REQUEST_PASID_VALID) && in iopf_complete_group() 66 (iopf->fault.prm.flags & IOMMU_FAULT_PAGE_RESPONSE_NEEDS_PASID)) in iopf_complete_group() 80 struct iommu_fault_page_request *prm = &iopf->fault.prm; in iopf_handle_single() 144 if (!(iopf->fault.prm.flags & in iopf_handle_group() 186 int iommu_queue_iopf(struct iommu_fault *fault, void *cookie) in iommu_queue_iopf() argument 198 if (fault->type != IOMMU_FAULT_PAGE_REQ) in iommu_queue_iopf() 210 if (!(fault->prm.flags & IOMMU_FAULT_PAGE_REQUEST_LAST_PAGE)) { in iommu_queue_iopf() [all …]
|
/drivers/gpu/drm/nouveau/nvkm/engine/fifo/ |
D | gk20a.c | 29 .intr.fault = gf100_fifo_intr_fault, 31 .fault.access = gk104_fifo_fault_access, 32 .fault.engine = gk104_fifo_fault_engine, 33 .fault.reason = gk104_fifo_fault_reason, 34 .fault.hubclient = gk104_fifo_fault_hubclient, 35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gm20b.c | 29 .intr.fault = gm107_fifo_intr_fault, 31 .fault.access = gk104_fifo_fault_access, 32 .fault.engine = gm107_fifo_fault_engine, 33 .fault.reason = gk104_fifo_fault_reason, 34 .fault.hubclient = gk104_fifo_fault_hubclient, 35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gp10b.c | 29 .intr.fault = gp100_fifo_intr_fault, 31 .fault.access = gk104_fifo_fault_access, 32 .fault.engine = gp100_fifo_fault_engine, 33 .fault.reason = gk104_fifo_fault_reason, 34 .fault.hubclient = gk104_fifo_fault_hubclient, 35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gm200.c | 45 .intr.fault = gm107_fifo_intr_fault, 47 .fault.access = gk104_fifo_fault_access, 48 .fault.engine = gm107_fifo_fault_engine, 49 .fault.reason = gk104_fifo_fault_reason, 50 .fault.hubclient = gk104_fifo_fault_hubclient, 51 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gk110.c | 51 .intr.fault = gf100_fifo_intr_fault, 53 .fault.access = gk104_fifo_fault_access, 54 .fault.engine = gk104_fifo_fault_engine, 55 .fault.reason = gk104_fifo_fault_reason, 56 .fault.hubclient = gk104_fifo_fault_hubclient, 57 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gk208.c | 48 .intr.fault = gf100_fifo_intr_fault, 50 .fault.access = gk104_fifo_fault_access, 51 .fault.engine = gk104_fifo_fault_engine, 52 .fault.reason = gk104_fifo_fault_reason, 53 .fault.hubclient = gk104_fifo_fault_hubclient, 54 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gp100.c | 81 .intr.fault = gp100_fifo_intr_fault, 83 .fault.access = gk104_fifo_fault_access, 84 .fault.engine = gp100_fifo_fault_engine, 85 .fault.reason = gk104_fifo_fault_reason, 86 .fault.hubclient = gk104_fifo_fault_hubclient, 87 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gm107.c | 97 .intr.fault = gm107_fifo_intr_fault, 99 .fault.access = gk104_fifo_fault_access, 100 .fault.engine = gm107_fifo_fault_engine, 101 .fault.reason = gk104_fifo_fault_reason, 102 .fault.hubclient = gk104_fifo_fault_hubclient, 103 .fault.gpcclient = gk104_fifo_fault_gpcclient,
|
D | gv100.c | 292 .fault.access = gv100_fifo_fault_access, 293 .fault.engine = gv100_fifo_fault_engine, 294 .fault.reason = gv100_fifo_fault_reason, 295 .fault.hubclient = gv100_fifo_fault_hubclient, 296 .fault.gpcclient = gv100_fifo_fault_gpcclient,
|
/drivers/media/i2c/ |
D | adp1653.c | 80 int fault; in adp1653_get_fault() local 83 fault = i2c_smbus_read_byte_data(client, ADP1653_REG_FAULT); in adp1653_get_fault() 84 if (fault < 0) in adp1653_get_fault() 85 return fault; in adp1653_get_fault() 87 flash->fault |= fault; in adp1653_get_fault() 89 if (!flash->fault) in adp1653_get_fault() 103 return flash->fault; in adp1653_get_fault() 152 if (flash->fault & ADP1653_REG_FAULT_FLT_SCP) in adp1653_get_ctrl() 154 if (flash->fault & ADP1653_REG_FAULT_FLT_OT) in adp1653_get_ctrl() 156 if (flash->fault & ADP1653_REG_FAULT_FLT_TMR) in adp1653_get_ctrl() [all …]
|
/drivers/gpu/drm/amd/amdgpu/ |
D | amdgpu_gmc.c | 372 struct amdgpu_gmc_fault *fault; in amdgpu_gmc_filter_faults() local 383 fault = &gmc->fault_ring[gmc->fault_hash[hash].idx]; in amdgpu_gmc_filter_faults() 384 while (fault->timestamp >= stamp) { in amdgpu_gmc_filter_faults() 387 if (atomic64_read(&fault->key) == key) in amdgpu_gmc_filter_faults() 390 tmp = fault->timestamp; in amdgpu_gmc_filter_faults() 391 fault = &gmc->fault_ring[fault->next]; in amdgpu_gmc_filter_faults() 394 if (fault->timestamp >= tmp) in amdgpu_gmc_filter_faults() 399 fault = &gmc->fault_ring[gmc->last_fault]; in amdgpu_gmc_filter_faults() 400 atomic64_set(&fault->key, key); in amdgpu_gmc_filter_faults() 401 fault->timestamp = timestamp; in amdgpu_gmc_filter_faults() [all …]
|
/drivers/power/supply/ |
D | max8903_charger.c | 35 bool fault; member 72 if (data->fault) in max8903_get_property() 201 bool fault; in max8903_fault() local 210 fault = gpiod_get_value(data->flt); in max8903_fault() 212 if (fault == data->fault) in max8903_fault() 215 data->fault = fault; in max8903_fault() 217 if (fault) in max8903_fault() 320 data->fault = false; in max8903_setup_gpios()
|
/drivers/iio/dac/ |
D | ad5421.c | 207 unsigned int fault; in ad5421_fault_handler() local 211 fault = ad5421_read(indio_dev, AD5421_REG_FAULT); in ad5421_fault_handler() 212 if (!fault) in ad5421_fault_handler() 234 if (fault == 0xffff) in ad5421_fault_handler() 235 fault = 0; in ad5421_fault_handler() 238 events = (old_fault ^ fault) & fault; in ad5421_fault_handler() 268 old_fault = fault; in ad5421_fault_handler() 269 fault = ad5421_read(indio_dev, AD5421_REG_FAULT); in ad5421_fault_handler() 272 if (fault & AD5421_FAULT_TRIGGER_IRQ) in ad5421_fault_handler() 275 } while (fault & AD5421_FAULT_TRIGGER_IRQ); in ad5421_fault_handler()
|
/drivers/gpu/drm/nouveau/nvkm/subdev/top/ |
D | base.c | 34 info->fault = -1; in nvkm_top_device_new() 99 if (info->type == type && info->inst == inst && info->fault >= 0) in nvkm_top_fault_id() 100 return info->fault; in nvkm_top_fault_id() 107 nvkm_top_fault(struct nvkm_device *device, int fault) in nvkm_top_fault() argument 113 if (info->fault == fault) in nvkm_top_fault()
|
/drivers/misc/ocxl/ |
D | link.c | 141 struct xsl_fault *fault = container_of(fault_work, struct xsl_fault, in xsl_fault_handler_bh() local 143 struct spa *spa = container_of(fault, struct spa, xsl_fault); in xsl_fault_handler_bh() 151 rc = copro_handle_mm_fault(fault->pe_data.mm, fault->dar, fault->dsisr, in xsl_fault_handler_bh() 155 if (fault->pe_data.xsl_err_cb) { in xsl_fault_handler_bh() 156 fault->pe_data.xsl_err_cb( in xsl_fault_handler_bh() 157 fault->pe_data.xsl_err_data, in xsl_fault_handler_bh() 158 fault->dar, fault->dsisr); in xsl_fault_handler_bh() 171 if (fault->dsisr & SPA_XSL_S) in xsl_fault_handler_bh() 174 if (get_region_id(fault->dar) != USER_REGION_ID) in xsl_fault_handler_bh() 178 hash_page_mm(fault->pe_data.mm, fault->dar, access, 0x300, in xsl_fault_handler_bh() [all …]
|