/arch/x86/power/ |
D | cpu.c | 226 loadsegment(ss, __KERNEL_DS); in __restore_processor_state() 227 loadsegment(ds, __USER_DS); in __restore_processor_state() 228 loadsegment(es, __USER_DS); in __restore_processor_state() 237 loadsegment(fs, __KERNEL_PERCPU); in __restore_processor_state() 238 loadsegment(gs, __KERNEL_STACK_CANARY); in __restore_processor_state() 249 loadsegment(ds, ctxt->es); in __restore_processor_state() 250 loadsegment(es, ctxt->es); in __restore_processor_state() 251 loadsegment(fs, ctxt->fs); in __restore_processor_state() 262 loadsegment(gs, ctxt->gs); in __restore_processor_state()
|
/arch/x86/kernel/ |
D | process_64.c | 214 loadsegment(fs, sel); in loadseg() 360 loadsegment(fs, __USER_DS); in start_thread_common() 364 loadsegment(fs, 0); in start_thread_common() 365 loadsegment(es, _ds); in start_thread_common() 366 loadsegment(ds, _ds); in start_thread_common() 457 loadsegment(es, next->es); in __switch_to() 461 loadsegment(ds, next->ds); in __switch_to() 517 loadsegment(ss, __KERNEL_DS); in __switch_to() 638 loadsegment(fs, 0); in do_arch_prctl_64()
|
D | tls.c | 158 loadsegment(ds, sel); in do_set_thread_area() 162 loadsegment(es, sel); in do_set_thread_area() 166 loadsegment(fs, sel); in do_set_thread_area() 176 loadsegment(gs, sel); in do_set_thread_area()
|
D | ldt.c | 43 loadsegment(ds, sel); in refresh_ldt_segments() 47 loadsegment(es, sel); in refresh_ldt_segments()
|
D | ptrace.c | 311 loadsegment(fs, task->thread.fsindex); in set_segment_reg() 321 loadsegment(ds, task->thread.ds); in set_segment_reg() 326 loadsegment(es, task->thread.es); in set_segment_reg()
|
D | signal.c | 595 loadsegment(ds, __USER_DS); in x32_setup_rt_frame() 596 loadsegment(es, __USER_DS); in x32_setup_rt_frame()
|
D | apm_32.c | 567 loadsegment(fs, saved_fs); loadsegment(gs, saved_gs)
|
/arch/x86/include/asm/ |
D | segment.h | 310 #define loadsegment(seg, value) __loadsegment_ ## seg (value) macro 324 # define set_user_gs(regs, v) loadsegment(gs, (unsigned long)(v)) 327 # define lazy_load_gs(v) loadsegment(gs, (v))
|
D | mmu_context.h | 235 loadsegment(fs, 0); \
|
/arch/x86/ia32/ |
D | ia32_signal.c | 43 #define loadsegment_fs(v) loadsegment(fs, v) 44 #define loadsegment_ds(v) loadsegment(ds, v) 45 #define loadsegment_es(v) loadsegment(es, v) 326 loadsegment(ds, __USER32_DS); in ia32_setup_frame() 327 loadsegment(es, __USER32_DS); in ia32_setup_frame() 405 loadsegment(ds, __USER32_DS); in ia32_setup_rt_frame() 406 loadsegment(es, __USER32_DS); in ia32_setup_rt_frame()
|
D | ia32_aout.c | 392 loadsegment(fs, 0); in load_aout_binary() 393 loadsegment(ds, __USER32_DS); in load_aout_binary() 394 loadsegment(es, __USER32_DS); in load_aout_binary()
|
/arch/x86/xen/ |
D | smp_pv.c | 214 loadsegment(ds, __USER_DS); in xen_pv_smp_prepare_boot_cpu() 215 loadsegment(es, __USER_DS); in xen_pv_smp_prepare_boot_cpu()
|
D | enlighten_pv.c | 554 loadsegment(fs, 0); in xen_load_tls()
|
/arch/x86/kernel/cpu/ |
D | common.c | 470 loadsegment(fs, __KERNEL_PERCPU); in load_percpu_segment() 1178 loadsegment(fs, 0); in detect_null_seg_behavior() 1713 loadsegment(fs, 0); in cpu_init()
|
/arch/x86/kvm/ |
D | svm.c | 1868 loadsegment(fs, svm->host.fs); in svm_vcpu_put() 1873 loadsegment(gs, svm->host.gs); in svm_vcpu_put() 5267 loadsegment(fs, svm->host.fs); in svm_vcpu_run() 5269 loadsegment(gs, svm->host.gs); in svm_vcpu_run()
|
D | vmx.c | 2409 loadsegment(gs, vmx->host_state.gs_sel); in __vmx_load_host_state() 2413 loadsegment(fs, vmx->host_state.fs_sel); in __vmx_load_host_state() 2416 loadsegment(ds, vmx->host_state.ds_sel); in __vmx_load_host_state() 2417 loadsegment(es, vmx->host_state.es_sel); in __vmx_load_host_state() 9966 loadsegment(ds, __USER_DS); in vmx_vcpu_run() 9967 loadsegment(es, __USER_DS); in vmx_vcpu_run()
|