Home
last modified time | relevance | path

Searched refs:fault (Results 1 – 25 of 133) sorted by relevance

123456

/drivers/gpu/drm/nouveau/nvkm/subdev/fault/
Dbase.c30 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_fini() local
31 fault->func->buffer.intr(fault->buffer[index], false); in nvkm_fault_ntfy_fini()
37 struct nvkm_fault *fault = container_of(event, typeof(*fault), event); in nvkm_fault_ntfy_init() local
38 fault->func->buffer.intr(fault->buffer[index], true); in nvkm_fault_ntfy_init()
65 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_intr() local
66 return fault->func->intr(fault); in nvkm_fault_intr()
72 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_fini() local
73 if (fault->func->fini) in nvkm_fault_fini()
74 fault->func->fini(fault); in nvkm_fault_fini()
81 struct nvkm_fault *fault = nvkm_fault(subdev); in nvkm_fault_init() local
[all …]
Dgv100.c33 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_process()
42 const u32 base = get * buffer->fault->func->buffer.entry_size; in gv100_fault_buffer_process()
76 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_intr()
87 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_fini()
95 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_init()
107 struct nvkm_device *device = buffer->fault->subdev.device; in gv100_fault_buffer_info()
120 struct nvkm_fault *fault = container_of(notify, typeof(*fault), nrpfb); in gv100_fault_ntfy_nrpfb() local
121 gv100_fault_buffer_process(fault->buffer[0]); in gv100_fault_ntfy_nrpfb()
126 gv100_fault_intr_fault(struct nvkm_fault *fault) in gv100_fault_intr_fault() argument
128 struct nvkm_subdev *subdev = &fault->subdev; in gv100_fault_intr_fault()
[all …]
Dtu102.c42 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_fini()
50 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_init()
62 struct nvkm_device *device = buffer->fault->subdev.device; in tu102_fault_buffer_info()
73 tu102_fault_intr_fault(struct nvkm_fault *fault) in tu102_fault_intr_fault() argument
75 struct nvkm_subdev *subdev = &fault->subdev; in tu102_fault_intr_fault()
99 tu102_fault_intr(struct nvkm_fault *fault) in tu102_fault_intr() argument
101 struct nvkm_subdev *subdev = &fault->subdev; in tu102_fault_intr()
106 tu102_fault_intr_fault(fault); in tu102_fault_intr()
112 if (fault->buffer[0]) { in tu102_fault_intr()
113 nvkm_event_send(&fault->event, 1, 0, NULL, 0); in tu102_fault_intr()
[all …]
Duser.c35 struct nvkm_device *device = buffer->fault->subdev.device; in nvkm_ufault_map()
48 *pevent = &buffer->fault->event; in nvkm_ufault_ntfy()
58 buffer->fault->func->buffer.fini(buffer); in nvkm_ufault_fini()
66 buffer->fault->func->buffer.init(buffer); in nvkm_ufault_init()
92 struct nvkm_fault *fault = device->fault; in nvkm_ufault_new() local
93 struct nvkm_fault_buffer *buffer = fault->buffer[fault->func->user.rp]; in nvkm_ufault_new()
Dgp100.c31 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_intr()
38 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_fini()
45 struct nvkm_device *device = buffer->fault->subdev.device; in gp100_fault_buffer_init()
54 buffer->entries = nvkm_rd32(buffer->fault->subdev.device, 0x002a78); in gp100_fault_buffer_info()
60 gp100_fault_intr(struct nvkm_fault *fault) in gp100_fault_intr() argument
62 nvkm_event_send(&fault->event, 1, 0, NULL, 0); in gp100_fault_intr()
DKbuild2 nvkm-y += nvkm/subdev/fault/base.o
3 nvkm-y += nvkm/subdev/fault/user.o
4 nvkm-y += nvkm/subdev/fault/gp100.o
5 nvkm-y += nvkm/subdev/fault/gv100.o
6 nvkm-y += nvkm/subdev/fault/tu102.o
/drivers/infiniband/hw/hfi1/
Dfault.c111 if (!ibd->fault->n_rxfaults[i] && !ibd->fault->n_txfaults[i]) in _fault_stats_seq_show()
116 (unsigned long long)ibd->fault->n_rxfaults[i], in _fault_stats_seq_show()
117 (unsigned long long)ibd->fault->n_txfaults[i]); in _fault_stats_seq_show()
138 struct fault *fault = file->private_data; in fault_opcodes_write() local
180 bitmap_zero(fault->opcodes, sizeof(fault->opcodes) * in fault_opcodes_write()
190 clear_bit(i, fault->opcodes); in fault_opcodes_write()
192 set_bit(i, fault->opcodes); in fault_opcodes_write()
212 struct fault *fault = file->private_data; in fault_opcodes_read() local
213 size_t bitsize = sizeof(fault->opcodes) * BITS_PER_BYTE; in fault_opcodes_read()
221 bit = find_first_bit(fault->opcodes, bitsize); in fault_opcodes_read()
[all …]
/drivers/gpu/drm/nouveau/
Dnouveau_svm.c63 u8 fault; member
65 } **fault; member
406 struct nouveau_svm_fault *fault) in nouveau_svm_fault_cancel_fault() argument
408 nouveau_svm_fault_cancel(svm, fault->inst, in nouveau_svm_fault_cancel_fault()
409 fault->hub, in nouveau_svm_fault_cancel_fault()
410 fault->gpc, in nouveau_svm_fault_cancel_fault()
411 fault->client); in nouveau_svm_fault_cancel_fault()
446 struct nouveau_svm_fault *fault; in nouveau_svm_fault_cache() local
454 if (!buffer->fault[buffer->fault_nr]) { in nouveau_svm_fault_cache()
455 fault = kmalloc(sizeof(*fault), GFP_KERNEL); in nouveau_svm_fault_cache()
[all …]
/drivers/iommu/
Damd_iommu_v2.c68 struct fault { struct
428 static void handle_fault_error(struct fault *fault) in handle_fault_error() argument
432 if (!fault->dev_state->inv_ppr_cb) { in handle_fault_error()
433 set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); in handle_fault_error()
437 status = fault->dev_state->inv_ppr_cb(fault->dev_state->pdev, in handle_fault_error()
438 fault->pasid, in handle_fault_error()
439 fault->address, in handle_fault_error()
440 fault->flags); in handle_fault_error()
443 set_pri_tag_status(fault->state, fault->tag, PPR_SUCCESS); in handle_fault_error()
446 set_pri_tag_status(fault->state, fault->tag, PPR_INVALID); in handle_fault_error()
[all …]
/drivers/gpu/drm/i915/gt/
Dintel_gt.c101 u32 fault; in gen6_check_faults() local
104 fault = GEN6_RING_FAULT_REG_READ(engine); in gen6_check_faults()
105 if (fault & RING_FAULT_VALID) { in gen6_check_faults()
111 fault & PAGE_MASK, in gen6_check_faults()
112 fault & RING_FAULT_GTTSEL_MASK ? in gen6_check_faults()
114 RING_FAULT_SRCID(fault), in gen6_check_faults()
115 RING_FAULT_FAULT_TYPE(fault)); in gen6_check_faults()
124 u32 fault; in gen8_check_faults() local
136 fault = intel_uncore_read(uncore, fault_reg); in gen8_check_faults()
137 if (fault & RING_FAULT_VALID) { in gen8_check_faults()
[all …]
/drivers/gpu/drm/nouveau/nvkm/engine/fifo/
Dgk20a.c29 .intr.fault = gf100_fifo_intr_fault,
31 .fault.access = gk104_fifo_fault_access,
32 .fault.engine = gk104_fifo_fault_engine,
33 .fault.reason = gk104_fifo_fault_reason,
34 .fault.hubclient = gk104_fifo_fault_hubclient,
35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgm20b.c29 .intr.fault = gm107_fifo_intr_fault,
31 .fault.access = gk104_fifo_fault_access,
32 .fault.engine = gm107_fifo_fault_engine,
33 .fault.reason = gk104_fifo_fault_reason,
34 .fault.hubclient = gk104_fifo_fault_hubclient,
35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgp10b.c29 .intr.fault = gp100_fifo_intr_fault,
31 .fault.access = gk104_fifo_fault_access,
32 .fault.engine = gp100_fifo_fault_engine,
33 .fault.reason = gk104_fifo_fault_reason,
34 .fault.hubclient = gk104_fifo_fault_hubclient,
35 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgm200.c45 .intr.fault = gm107_fifo_intr_fault,
47 .fault.access = gk104_fifo_fault_access,
48 .fault.engine = gm107_fifo_fault_engine,
49 .fault.reason = gk104_fifo_fault_reason,
50 .fault.hubclient = gk104_fifo_fault_hubclient,
51 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgk110.c51 .intr.fault = gf100_fifo_intr_fault,
53 .fault.access = gk104_fifo_fault_access,
54 .fault.engine = gk104_fifo_fault_engine,
55 .fault.reason = gk104_fifo_fault_reason,
56 .fault.hubclient = gk104_fifo_fault_hubclient,
57 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgk208.c48 .intr.fault = gf100_fifo_intr_fault,
50 .fault.access = gk104_fifo_fault_access,
51 .fault.engine = gk104_fifo_fault_engine,
52 .fault.reason = gk104_fifo_fault_reason,
53 .fault.hubclient = gk104_fifo_fault_hubclient,
54 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgp100.c81 .intr.fault = gp100_fifo_intr_fault,
83 .fault.access = gk104_fifo_fault_access,
84 .fault.engine = gp100_fifo_fault_engine,
85 .fault.reason = gk104_fifo_fault_reason,
86 .fault.hubclient = gk104_fifo_fault_hubclient,
87 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dgm107.c97 .intr.fault = gm107_fifo_intr_fault,
99 .fault.access = gk104_fifo_fault_access,
100 .fault.engine = gm107_fifo_fault_engine,
101 .fault.reason = gk104_fifo_fault_reason,
102 .fault.hubclient = gk104_fifo_fault_hubclient,
103 .fault.gpcclient = gk104_fifo_fault_gpcclient,
Dtu102.c101 .fault.access = gv100_fifo_fault_access,
102 .fault.engine = tu102_fifo_fault_engine,
103 .fault.reason = gv100_fifo_fault_reason,
104 .fault.hubclient = gv100_fifo_fault_hubclient,
105 .fault.gpcclient = gv100_fifo_fault_gpcclient,
Dgv100.c292 .fault.access = gv100_fifo_fault_access,
293 .fault.engine = gv100_fifo_fault_engine,
294 .fault.reason = gv100_fifo_fault_reason,
295 .fault.hubclient = gv100_fifo_fault_hubclient,
296 .fault.gpcclient = gv100_fifo_fault_gpcclient,
/drivers/gpu/drm/amd/amdgpu/
Damdgpu_gmc.c272 struct amdgpu_gmc_fault *fault; in amdgpu_gmc_filter_faults() local
283 fault = &gmc->fault_ring[gmc->fault_hash[hash].idx]; in amdgpu_gmc_filter_faults()
284 while (fault->timestamp >= stamp) { in amdgpu_gmc_filter_faults()
287 if (fault->key == key) in amdgpu_gmc_filter_faults()
290 tmp = fault->timestamp; in amdgpu_gmc_filter_faults()
291 fault = &gmc->fault_ring[fault->next]; in amdgpu_gmc_filter_faults()
294 if (fault->timestamp >= tmp) in amdgpu_gmc_filter_faults()
299 fault = &gmc->fault_ring[gmc->last_fault]; in amdgpu_gmc_filter_faults()
300 fault->key = key; in amdgpu_gmc_filter_faults()
301 fault->timestamp = timestamp; in amdgpu_gmc_filter_faults()
[all …]
/drivers/media/i2c/
Dadp1653.c80 int fault; in adp1653_get_fault() local
83 fault = i2c_smbus_read_byte_data(client, ADP1653_REG_FAULT); in adp1653_get_fault()
84 if (fault < 0) in adp1653_get_fault()
85 return fault; in adp1653_get_fault()
87 flash->fault |= fault; in adp1653_get_fault()
89 if (!flash->fault) in adp1653_get_fault()
103 return flash->fault; in adp1653_get_fault()
152 if (flash->fault & ADP1653_REG_FAULT_FLT_SCP) in adp1653_get_ctrl()
154 if (flash->fault & ADP1653_REG_FAULT_FLT_OT) in adp1653_get_ctrl()
156 if (flash->fault & ADP1653_REG_FAULT_FLT_TMR) in adp1653_get_ctrl()
[all …]
/drivers/iio/dac/
Dad5421.c204 unsigned int fault; in ad5421_fault_handler() local
208 fault = ad5421_read(indio_dev, AD5421_REG_FAULT); in ad5421_fault_handler()
209 if (!fault) in ad5421_fault_handler()
231 if (fault == 0xffff) in ad5421_fault_handler()
232 fault = 0; in ad5421_fault_handler()
235 events = (old_fault ^ fault) & fault; in ad5421_fault_handler()
265 old_fault = fault; in ad5421_fault_handler()
266 fault = ad5421_read(indio_dev, AD5421_REG_FAULT); in ad5421_fault_handler()
269 if (fault & AD5421_FAULT_TRIGGER_IRQ) in ad5421_fault_handler()
272 } while (fault & AD5421_FAULT_TRIGGER_IRQ); in ad5421_fault_handler()
/drivers/misc/ocxl/
Dlink.c133 struct xsl_fault *fault = container_of(fault_work, struct xsl_fault, in xsl_fault_handler_bh() local
135 struct spa *spa = container_of(fault, struct spa, xsl_fault); in xsl_fault_handler_bh()
143 rc = copro_handle_mm_fault(fault->pe_data.mm, fault->dar, fault->dsisr, in xsl_fault_handler_bh()
147 if (fault->pe_data.xsl_err_cb) { in xsl_fault_handler_bh()
148 fault->pe_data.xsl_err_cb( in xsl_fault_handler_bh()
149 fault->pe_data.xsl_err_data, in xsl_fault_handler_bh()
150 fault->dar, fault->dsisr); in xsl_fault_handler_bh()
163 if (fault->dsisr & SPA_XSL_S) in xsl_fault_handler_bh()
166 if (get_region_id(fault->dar) != USER_REGION_ID) in xsl_fault_handler_bh()
170 hash_page_mm(fault->pe_data.mm, fault->dar, access, 0x300, in xsl_fault_handler_bh()
[all …]
/drivers/gpu/drm/nouveau/nvkm/subdev/top/
Dbase.c33 info->fault = -1; in nvkm_top_device_new()
121 if (info->index == devidx && info->fault >= 0) in nvkm_top_fault_id()
122 return info->fault; in nvkm_top_fault_id()
129 nvkm_top_fault(struct nvkm_device *device, int fault) in nvkm_top_fault() argument
135 if (info->fault == fault) in nvkm_top_fault()

123456