Home
last modified time | relevance | path

Searched refs:entry (Results 1 – 25 of 963) sorted by relevance

12345678910>>...39

/drivers/acpi/
Dnvs.c96 struct nvs_page *entry, *next; in suspend_nvs_register() local
104 entry = kzalloc(sizeof(struct nvs_page), GFP_KERNEL); in suspend_nvs_register()
105 if (!entry) in suspend_nvs_register()
108 list_add_tail(&entry->node, &nvs_list); in suspend_nvs_register()
109 entry->phys_start = start; in suspend_nvs_register()
111 entry->size = (size < nr_bytes) ? size : nr_bytes; in suspend_nvs_register()
113 start += entry->size; in suspend_nvs_register()
114 size -= entry->size; in suspend_nvs_register()
119 list_for_each_entry_safe(entry, next, &nvs_list, node) { in suspend_nvs_register()
120 list_del(&entry->node); in suspend_nvs_register()
[all …]
Dpci_irq.c125 static void do_prt_fixups(struct acpi_prt_entry *entry, in do_prt_fixups() argument
136 entry->id.segment == quirk->segment && in do_prt_fixups()
137 entry->id.bus == quirk->bus && in do_prt_fixups()
138 entry->id.device == quirk->device && in do_prt_fixups()
139 entry->pin == quirk->pin && in do_prt_fixups()
145 entry->id.segment, entry->id.bus, in do_prt_fixups()
146 entry->id.device, pin_name(entry->pin), in do_prt_fixups()
160 struct acpi_prt_entry *entry; in acpi_pci_irq_check_entry() local
166 entry = kzalloc(sizeof(struct acpi_prt_entry), GFP_KERNEL); in acpi_pci_irq_check_entry()
167 if (!entry) in acpi_pci_irq_check_entry()
[all …]
/drivers/firmware/
Dmemmap.c52 static ssize_t start_show(struct firmware_map_entry *entry, char *buf);
53 static ssize_t end_show(struct firmware_map_entry *entry, char *buf);
54 static ssize_t type_show(struct firmware_map_entry *entry, char *buf);
65 ssize_t (*show)(struct firmware_map_entry *entry, char *buf);
108 struct firmware_map_entry *entry = to_memmap_entry(kobj); in release_firmware_map_entry() local
110 if (PageReserved(virt_to_page(entry))) { in release_firmware_map_entry()
118 list_add(&entry->list, &map_entries_bootmem); in release_firmware_map_entry()
124 kfree(entry); in release_firmware_map_entry()
152 struct firmware_map_entry *entry) in firmware_map_add_entry() argument
156 entry->start = start; in firmware_map_add_entry()
[all …]
Ddmi-sysfs.c53 ssize_t (*show)(struct dmi_sysfs_entry *entry, char *buf);
68 ssize_t (*show)(struct dmi_sysfs_entry *entry,
100 struct dmi_sysfs_entry *entry = to_entry(kobj); in dmi_sysfs_attr_show() local
107 return attr->show(entry, buf); in dmi_sysfs_attr_show()
118 struct dmi_sysfs_entry *entry; member
129 struct dmi_sysfs_entry *entry = data->entry; in find_dmi_entry_helper() local
132 if (dh->type != entry->dh.type) in find_dmi_entry_helper()
148 data->ret = data->callback(entry, dh, data->private); in find_dmi_entry_helper()
158 static ssize_t find_dmi_entry(struct dmi_sysfs_entry *entry, in find_dmi_entry() argument
162 .entry = entry, in find_dmi_entry()
[all …]
/drivers/staging/media/atomisp/pci/atomisp2/css2400/base/refcount/src/
Drefcount.c97 struct ia_css_refcount_entry *entry; in ia_css_refcount_uninit() local
106 entry = myrefcount.items + i; in ia_css_refcount_uninit()
107 if (entry->data != mmgr_NULL) { in ia_css_refcount_uninit()
111 hmm_free(entry->data); in ia_css_refcount_uninit()
112 entry->data = mmgr_NULL; in ia_css_refcount_uninit()
113 entry->count = 0; in ia_css_refcount_uninit()
114 entry->id = 0; in ia_css_refcount_uninit()
126 struct ia_css_refcount_entry *entry; in ia_css_refcount_increment() local
131 entry = refcount_find_entry(ptr, false); in ia_css_refcount_increment()
136 if (!entry) { in ia_css_refcount_increment()
[all …]
/drivers/gpu/drm/
Ddrm_scatter.c50 static void drm_sg_cleanup(struct drm_sg_mem * entry) in drm_sg_cleanup() argument
55 for (i = 0; i < entry->pages; i++) { in drm_sg_cleanup()
56 page = entry->pagelist[i]; in drm_sg_cleanup()
61 vfree(entry->virtual); in drm_sg_cleanup()
63 kfree(entry->busaddr); in drm_sg_cleanup()
64 kfree(entry->pagelist); in drm_sg_cleanup()
65 kfree(entry); in drm_sg_cleanup()
86 struct drm_sg_mem *entry; in drm_legacy_sg_alloc() local
100 entry = kzalloc(sizeof(*entry), GFP_KERNEL); in drm_legacy_sg_alloc()
101 if (!entry) in drm_legacy_sg_alloc()
[all …]
Ddrm_agpsupport.c199 struct drm_agp_mem *entry; in drm_agp_alloc() local
206 if (!(entry = kzalloc(sizeof(*entry), GFP_KERNEL))) in drm_agp_alloc()
212 kfree(entry); in drm_agp_alloc()
216 entry->handle = (unsigned long)memory->key + 1; in drm_agp_alloc()
217 entry->memory = memory; in drm_agp_alloc()
218 entry->bound = 0; in drm_agp_alloc()
219 entry->pages = pages; in drm_agp_alloc()
220 list_add(&entry->head, &dev->agp->memory); in drm_agp_alloc()
222 request->handle = entry->handle; in drm_agp_alloc()
250 struct drm_agp_mem *entry; in drm_agp_lookup_entry() local
[all …]
/drivers/net/dsa/mv88e6xxx/
Dglobal1_vtu.c20 struct mv88e6xxx_vtu_entry *entry) in mv88e6xxx_g1_vtu_fid_read() argument
29 entry->fid = val & MV88E6352_G1_VTU_FID_MASK; in mv88e6xxx_g1_vtu_fid_read()
35 struct mv88e6xxx_vtu_entry *entry) in mv88e6xxx_g1_vtu_fid_write() argument
37 u16 val = entry->fid & MV88E6352_G1_VTU_FID_MASK; in mv88e6xxx_g1_vtu_fid_write()
45 struct mv88e6xxx_vtu_entry *entry) in mv88e6xxx_g1_vtu_sid_read() argument
54 entry->sid = val & MV88E6352_G1_VTU_SID_MASK; in mv88e6xxx_g1_vtu_sid_read()
60 struct mv88e6xxx_vtu_entry *entry) in mv88e6xxx_g1_vtu_sid_write() argument
62 u16 val = entry->sid & MV88E6352_G1_VTU_SID_MASK; in mv88e6xxx_g1_vtu_sid_write()
90 struct mv88e6xxx_vtu_entry *entry) in mv88e6xxx_g1_vtu_vid_read() argument
99 entry->vid = val & 0xfff; in mv88e6xxx_g1_vtu_vid_read()
[all …]
/drivers/staging/irda/net/
Dirqueue.c445 void hashbin_insert(hashbin_t* hashbin, irda_queue_t* entry, long hashv, in hashbin_insert() argument
469 entry->q_hash = hashv; in hashbin_insert()
471 strlcpy( entry->q_name, name, sizeof(entry->q_name)); in hashbin_insert()
477 entry); in hashbin_insert()
499 irda_queue_t *entry = NULL; in hashbin_remove_first() local
506 entry = hashbin_get_first( hashbin); in hashbin_remove_first()
507 if ( entry != NULL) { in hashbin_remove_first()
513 hashv = entry->q_hash; in hashbin_remove_first()
520 entry); in hashbin_remove_first()
522 entry->q_next = NULL; in hashbin_remove_first()
[all …]
/drivers/gpu/drm/vmwgfx/
Dvmwgfx_cmdbuf_res.c106 struct vmw_cmdbuf_res *entry) in vmw_cmdbuf_res_free() argument
108 list_del(&entry->head); in vmw_cmdbuf_res_free()
109 WARN_ON(drm_ht_remove_item(&man->resources, &entry->hash)); in vmw_cmdbuf_res_free()
110 vmw_resource_unreference(&entry->res); in vmw_cmdbuf_res_free()
111 kfree(entry); in vmw_cmdbuf_res_free()
126 struct vmw_cmdbuf_res *entry, *next; in vmw_cmdbuf_res_commit() local
128 list_for_each_entry_safe(entry, next, list, head) { in vmw_cmdbuf_res_commit()
129 list_del(&entry->head); in vmw_cmdbuf_res_commit()
130 if (entry->res->func->commit_notify) in vmw_cmdbuf_res_commit()
131 entry->res->func->commit_notify(entry->res, in vmw_cmdbuf_res_commit()
[all …]
/drivers/net/ethernet/rocker/
Drocker_ofdpa.c96 struct hlist_node entry; member
106 struct hlist_node entry; member
132 struct hlist_node entry; member
144 struct hlist_node entry; member
151 struct hlist_node entry; member
309 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_ig_port() argument
312 entry->key.ig_port.in_pport)) in ofdpa_cmd_flow_tbl_add_ig_port()
315 entry->key.ig_port.in_pport_mask)) in ofdpa_cmd_flow_tbl_add_ig_port()
318 entry->key.ig_port.goto_tbl)) in ofdpa_cmd_flow_tbl_add_ig_port()
326 const struct ofdpa_flow_tbl_entry *entry) in ofdpa_cmd_flow_tbl_add_vlan() argument
[all …]
/drivers/misc/vmw_vmci/
Dvmci_queue_pair.c900 struct qp_entry *entry; in qp_list_find() local
905 list_for_each_entry(entry, &qp_list->head, list_item) { in qp_list_find()
906 if (vmci_handle_is_equal(entry->handle, handle)) in qp_list_find()
907 return entry; in qp_list_find()
919 struct qp_guest_endpoint *entry; in qp_guest_handle_to_entry() local
922 entry = qp ? container_of( in qp_guest_handle_to_entry()
924 return entry; in qp_guest_handle_to_entry()
933 struct qp_broker_entry *entry; in qp_broker_handle_to_entry() local
936 entry = qp ? container_of( in qp_broker_handle_to_entry()
938 return entry; in qp_broker_handle_to_entry()
[all …]
Dvmci_doorbell.c99 struct dbell_entry *entry; in vmci_dbell_get_priv_flags() local
107 entry = container_of(resource, struct dbell_entry, resource); in vmci_dbell_get_priv_flags()
108 *priv_flags = entry->priv_flags; in vmci_dbell_get_priv_flags()
145 static void dbell_index_table_add(struct dbell_entry *entry) in dbell_index_table_add() argument
150 vmci_resource_get(&entry->resource); in dbell_index_table_add()
194 entry->idx = new_notify_idx; in dbell_index_table_add()
195 bucket = VMCI_DOORBELL_HASH(entry->idx); in dbell_index_table_add()
196 hlist_add_head(&entry->node, &vmci_doorbell_it.entries[bucket]); in dbell_index_table_add()
205 static void dbell_index_table_remove(struct dbell_entry *entry) in dbell_index_table_remove() argument
209 hlist_del_init(&entry->node); in dbell_index_table_remove()
[all …]
/drivers/hid/
Dhid-lg4ff.c310 struct lg4ff_device_entry *entry = drv_data->device_props; in lg4ff_adjust_input_event() local
313 if (!entry) { in lg4ff_adjust_input_event()
318 switch (entry->wdata.product_id) { in lg4ff_adjust_input_event()
322 new_value = lg4ff_adjust_dfp_x_axis(value, entry->wdata.range); in lg4ff_adjust_input_event()
337 struct lg4ff_device_entry *entry = drv_data->device_props; in lg4ff_raw_event() local
339 if (!entry) in lg4ff_raw_event()
343 if (entry->wdata.combine) { in lg4ff_raw_event()
344 switch (entry->wdata.product_id) { in lg4ff_raw_event()
415 struct lg4ff_device_entry *entry; in lg4ff_play() local
427 entry = drv_data->device_props; in lg4ff_play()
[all …]
/drivers/isdn/mISDN/
Ddsp_pipeline.c82 struct dsp_element_entry *entry = in mISDN_dsp_dev_release() local
84 list_del(&entry->list); in mISDN_dsp_dev_release()
85 kfree(entry); in mISDN_dsp_dev_release()
90 struct dsp_element_entry *entry; in mISDN_dsp_element_register() local
96 entry = kzalloc(sizeof(struct dsp_element_entry), GFP_ATOMIC); in mISDN_dsp_element_register()
97 if (!entry) in mISDN_dsp_element_register()
100 entry->elem = elem; in mISDN_dsp_element_register()
102 entry->dev.class = elements_class; in mISDN_dsp_element_register()
103 entry->dev.release = mISDN_dsp_dev_release; in mISDN_dsp_element_register()
104 dev_set_drvdata(&entry->dev, elem); in mISDN_dsp_element_register()
[all …]
/drivers/soc/qcom/
Dsmp2p.c171 struct smp2p_entry *entry; in qcom_smp2p_intr() local
198 list_for_each_entry(entry, &smp2p->inbound, node) { in qcom_smp2p_intr()
200 if (!strcmp(buf, entry->name)) { in qcom_smp2p_intr()
201 entry->value = &in->entries[i].value; in qcom_smp2p_intr()
209 list_for_each_entry(entry, &smp2p->inbound, node) { in qcom_smp2p_intr()
211 if (!entry->value) in qcom_smp2p_intr()
214 val = readl(entry->value); in qcom_smp2p_intr()
216 status = val ^ entry->last_value; in qcom_smp2p_intr()
217 entry->last_value = val; in qcom_smp2p_intr()
223 for_each_set_bit(i, entry->irq_enabled, 32) { in qcom_smp2p_intr()
[all …]
/drivers/sh/intc/
Dvirq.c27 #define for_each_virq(entry, head) \ argument
28 for (entry = head; entry; entry = entry->next)
86 struct intc_virq_list *entry; in add_virq_to_pirq() local
90 for_each_virq(entry, irq_get_handler_data(irq)) { in add_virq_to_pirq()
91 if (entry->irq == virq) in add_virq_to_pirq()
93 last = &entry->next; in add_virq_to_pirq()
96 entry = kzalloc(sizeof(struct intc_virq_list), GFP_ATOMIC); in add_virq_to_pirq()
97 if (!entry) in add_virq_to_pirq()
100 entry->irq = virq; in add_virq_to_pirq()
103 *last = entry; in add_virq_to_pirq()
[all …]
/drivers/infiniband/hw/cxgb4/
Dresource.c84 u32 entry; in c4iw_get_resource() local
85 entry = c4iw_id_alloc(id_table); in c4iw_get_resource()
86 if (entry == (u32)(-1)) in c4iw_get_resource()
88 return entry; in c4iw_get_resource()
91 void c4iw_put_resource(struct c4iw_id_table *id_table, u32 entry) in c4iw_put_resource() argument
93 pr_debug("%s entry 0x%x\n", __func__, entry); in c4iw_put_resource()
94 c4iw_id_free(id_table, entry); in c4iw_put_resource()
99 struct c4iw_qid_list *entry; in c4iw_get_cqid() local
105 entry = list_entry(uctx->cqids.next, struct c4iw_qid_list, in c4iw_get_cqid()
106 entry); in c4iw_get_cqid()
[all …]
/drivers/pci/
Dmsi.c93 struct msi_desc *entry; in arch_setup_msi_irqs() local
105 for_each_pci_msi_entry(entry, dev) { in arch_setup_msi_irqs()
106 ret = arch_setup_msi_irq(dev, entry); in arch_setup_msi_irqs()
123 struct msi_desc *entry; in default_teardown_msi_irqs() local
125 for_each_pci_msi_entry(entry, dev) in default_teardown_msi_irqs()
126 if (entry->irq) in default_teardown_msi_irqs()
127 for (i = 0; i < entry->nvec_used; i++) in default_teardown_msi_irqs()
128 arch_teardown_msi_irq(entry->irq + i); in default_teardown_msi_irqs()
138 struct msi_desc *entry; in default_restore_msi_irq() local
140 entry = NULL; in default_restore_msi_irq()
[all …]
/drivers/gpu/drm/amd/powerplay/smumgr/
Dvega10_smumgr.c231 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_from_smc()
233 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_from_smc()
237 priv->smu_tables.entry[table_id].table_addr_high) == 0, in vega10_copy_table_from_smc()
241 priv->smu_tables.entry[table_id].table_addr_low) == 0, in vega10_copy_table_from_smc()
246 priv->smu_tables.entry[table_id].table_id) == 0, in vega10_copy_table_from_smc()
250 memcpy(table, priv->smu_tables.entry[table_id].table, in vega10_copy_table_from_smc()
251 priv->smu_tables.entry[table_id].size); in vega10_copy_table_from_smc()
269 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].version != 0, in vega10_copy_table_to_smc()
271 PP_ASSERT_WITH_CODE(priv->smu_tables.entry[table_id].size != 0, in vega10_copy_table_to_smc()
274 memcpy(priv->smu_tables.entry[table_id].table, table, in vega10_copy_table_to_smc()
[all …]
/drivers/parisc/
Dpdc_stable.c111 ssize_t (*show)(struct pdcspath_entry *entry, char *buf);
112 ssize_t (*store)(struct pdcspath_entry *entry, const char *buf, size_t count);
152 pdcspath_fetch(struct pdcspath_entry *entry) in pdcspath_fetch() argument
156 if (!entry) in pdcspath_fetch()
159 devpath = &entry->devpath; in pdcspath_fetch()
162 entry, devpath, entry->addr); in pdcspath_fetch()
165 if (pdc_stable_read(entry->addr, devpath, sizeof(*devpath)) != PDC_OK) in pdcspath_fetch()
171 entry->dev = hwpath_to_device((struct hardware_path *)devpath); in pdcspath_fetch()
173 entry->ready = 1; in pdcspath_fetch()
175 DPRINTK("%s: device: 0x%p\n", __func__, entry->dev); in pdcspath_fetch()
[all …]
/drivers/oprofile/
Dcpu_buffer.c144 *op_cpu_buffer_write_reserve(struct op_entry *entry, unsigned long size) in op_cpu_buffer_write_reserve() argument
146 entry->event = ring_buffer_lock_reserve in op_cpu_buffer_write_reserve()
148 size * sizeof(entry->sample->data[0])); in op_cpu_buffer_write_reserve()
149 if (!entry->event) in op_cpu_buffer_write_reserve()
151 entry->sample = ring_buffer_event_data(entry->event); in op_cpu_buffer_write_reserve()
152 entry->size = size; in op_cpu_buffer_write_reserve()
153 entry->data = entry->sample->data; in op_cpu_buffer_write_reserve()
155 return entry->sample; in op_cpu_buffer_write_reserve()
158 int op_cpu_buffer_write_commit(struct op_entry *entry) in op_cpu_buffer_write_commit() argument
160 return ring_buffer_unlock_commit(op_ring_buffer, entry->event); in op_cpu_buffer_write_commit()
[all …]
/drivers/gpu/drm/ttm/
Dttm_execbuf_util.c36 struct ttm_validate_buffer *entry) in ttm_eu_backoff_reservation_reverse() argument
38 list_for_each_entry_continue_reverse(entry, list, head) { in ttm_eu_backoff_reservation_reverse()
39 struct ttm_buffer_object *bo = entry->bo; in ttm_eu_backoff_reservation_reverse()
47 struct ttm_validate_buffer *entry; in ttm_eu_del_from_lru_locked() local
49 list_for_each_entry(entry, list, head) { in ttm_eu_del_from_lru_locked()
50 struct ttm_buffer_object *bo = entry->bo; in ttm_eu_del_from_lru_locked()
58 struct ttm_validate_buffer *entry; in ttm_eu_backoff_reservation() local
64 entry = list_first_entry(list, struct ttm_validate_buffer, head); in ttm_eu_backoff_reservation()
65 glob = entry->bo->glob; in ttm_eu_backoff_reservation()
68 list_for_each_entry(entry, list, head) { in ttm_eu_backoff_reservation()
[all …]
/drivers/ntb/
Dntb_transport.c103 struct list_head entry; member
122 unsigned int entry; member
204 struct list_head entry; member
210 struct list_head entry; member
263 struct ntb_queue_entry *entry);
264 static void ntb_memcpy_tx(struct ntb_queue_entry *entry, void __iomem *offset);
265 static int ntb_async_rx_submit(struct ntb_queue_entry *entry, void *offset);
266 static void ntb_memcpy_rx(struct ntb_queue_entry *entry, void *offset);
313 list_add_tail(&nt->entry, &ntb_transport_list); in ntb_bus_init()
321 list_for_each_entry_safe(client_dev, cd, &nt->client_devs, entry) { in ntb_bus_remove()
[all …]
/drivers/firmware/efi/
Druntime-map.c26 ssize_t (*show)(struct efi_runtime_map_entry *entry, char *buf);
34 static ssize_t type_show(struct efi_runtime_map_entry *entry, char *buf) in type_show() argument
36 return snprintf(buf, PAGE_SIZE, "0x%x\n", entry->md.type); in type_show()
39 #define EFI_RUNTIME_FIELD(var) entry->md.var
42 static ssize_t name##_show(struct efi_runtime_map_entry *entry, char *buf) \
60 struct efi_runtime_map_entry *entry = to_map_entry(kobj); in map_attr_show() local
63 return map_attr->show(entry, buf); in map_attr_show()
90 struct efi_runtime_map_entry *entry; in map_release() local
92 entry = to_map_entry(kobj); in map_release()
93 kfree(entry); in map_release()
[all …]

12345678910>>...39