gs 52 arch/x86/ia32/ia32_aout.c u32 fs, gs; gs 90 arch/x86/ia32/ia32_aout.c savesegment(gs, gs); gs 91 arch/x86/ia32/ia32_aout.c dump->regs.gs = gs; gs 218 arch/x86/ia32/ia32_signal.c unsigned int tmpflags, gs, oldgs, err = 0; gs 237 arch/x86/ia32/ia32_signal.c err |= __get_user(gs, &sc->gs); gs 238 arch/x86/ia32/ia32_signal.c gs |= 3; gs 239 arch/x86/ia32/ia32_signal.c savesegment(gs, oldgs); gs 240 arch/x86/ia32/ia32_signal.c if (gs != oldgs) gs 241 arch/x86/ia32/ia32_signal.c load_gs_index(gs); gs 345 arch/x86/ia32/ia32_signal.c savesegment(gs, tmp); gs 346 arch/x86/ia32/ia32_signal.c err |= __put_user(tmp, (unsigned int __user *)&sc->gs); gs 566 arch/x86/kernel/apm_32.c savesegment(fs, saved_fs); savesegment(gs, saved_gs) gs 568 arch/x86/kernel/apm_32.c loadsegment(fs, saved_fs); loadsegment(gs, saved_gs) gs 868 arch/x86/kernel/cpu/common.c loadsegment(gs, 0); gs 131 arch/x86/kernel/process_32.c unsigned short ss, gs; gs 137 arch/x86/kernel/process_32.c savesegment(gs, gs); gs 141 arch/x86/kernel/process_32.c savesegment(gs, gs); gs 165 arch/x86/kernel/process_32.c (u16)regs->ds, (u16)regs->es, (u16)regs->fs, gs, ss); gs 316 arch/x86/kernel/process_32.c savesegment(gs, p->thread.gs); gs 582 arch/x86/kernel/process_32.c savesegment(gs, prev->gs); gs 627 arch/x86/kernel/process_32.c if (prev->gs | next->gs) gs 628 arch/x86/kernel/process_32.c loadsegment(gs, next->gs); gs 141 arch/x86/kernel/process_64.c unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs; gs 175 arch/x86/kernel/process_64.c rdmsrl(MSR_GS_BASE, gs); gs 187 arch/x86/kernel/process_64.c fs, fsindex, gs, gsindex, shadowgs); gs 336 arch/x86/kernel/process_64.c p->thread.gs = me->thread.gs; gs 338 arch/x86/kernel/process_64.c savesegment(gs, p->thread.gsindex); gs 585 arch/x86/kernel/process_64.c savesegment(gs, gsindex); gs 620 arch/x86/kernel/process_64.c if (unlikely(gsindex | next->gsindex | prev->gs)) { gs 623 arch/x86/kernel/process_64.c prev->gs = 0; gs 625 arch/x86/kernel/process_64.c if (next->gs) gs 626 arch/x86/kernel/process_64.c wrmsrl(MSR_KERNEL_GS_BASE, next->gs); gs 777 arch/x86/kernel/process_64.c task->thread.gs = 0; gs 780 arch/x86/kernel/process_64.c task->thread.gs = addr; gs 833 arch/x86/kernel/process_64.c savesegment(gs, gsindex); gs 837 arch/x86/kernel/process_64.c base = task->thread.gs; gs 839 arch/x86/kernel/process_64.c base = task->thread.gs; gs 90 arch/x86/kernel/ptrace.c if (offset != offsetof(struct user_regs_struct, gs)) gs 93 arch/x86/kernel/ptrace.c retval = task->thread.gs; gs 95 arch/x86/kernel/ptrace.c savesegment(gs, retval); gs 128 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs): gs 129 arch/x86/kernel/ptrace.c task->thread.gs = value; gs 135 arch/x86/kernel/ptrace.c loadsegment(gs, value); gs 171 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs): gs 221 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct,gs): gs 227 arch/x86/kernel/ptrace.c task->thread.gs != 0) || gs 229 arch/x86/kernel/ptrace.c task->thread.gs == 0)) gs 321 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs): gs 356 arch/x86/kernel/ptrace.c if (child->thread.gs != value) gs 373 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs): gs 401 arch/x86/kernel/ptrace.c if (task->thread.gs != 0) gs 402 arch/x86/kernel/ptrace.c return task->thread.gs; gs 1169 arch/x86/kernel/ptrace.c SEG32(gs); gs 1233 arch/x86/kernel/ptrace.c SEG32(gs); gs 152 arch/x86/kernel/signal_32.c GET_SEG(gs); gs 263 arch/x86/kernel/signal_32.c savesegment(gs, tmp); gs 264 arch/x86/kernel/signal_32.c err |= __put_user(tmp, (unsigned int __user *)&sc->gs); gs 153 arch/x86/kernel/signal_64.c err |= __put_user(0, &sc->gs); gs 161 arch/x86/kernel/vm86_32.c loadsegment(gs, current->thread.saved_gs); gs 326 arch/x86/kernel/vm86_32.c savesegment(gs, tsk->thread.saved_gs); gs 558 arch/x86/kernel/vmi_32.c ap.gs = 0; gs 558 arch/x86/kvm/svm.c init_seg(&save->gs); gs 760 arch/x86/kvm/svm.c case VCPU_SREG_GS: return &save->gs; gs 130 arch/x86/kvm/svm.h struct vmcb_seg gs; gs 28 arch/x86/kvm/tss.h u32 gs; gs 1310 arch/x86/kvm/vmx.c fix_pmode_dataseg(VCPU_SREG_GS, &vcpu->arch.rmode.gs); gs 1382 arch/x86/kvm/vmx.c fix_rmode_seg(VCPU_SREG_GS, &vcpu->arch.rmode.gs); gs 3120 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &sregs->gs, VCPU_SREG_GS); gs 3335 arch/x86/kvm/x86.c tss->gs = get_segment_selector(vcpu, VCPU_SREG_GS); gs 3375 arch/x86/kvm/x86.c if (kvm_load_segment_descriptor(vcpu, tss->gs, 1, VCPU_SREG_GS)) gs 3631 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &sregs->gs, VCPU_SREG_GS); gs 281 arch/x86/lguest/boot.c loadsegment(gs, 0); gs 160 arch/x86/math-emu/get_address.c savesegment(gs, addr->selector); gs 40 arch/x86/power/cpu_32.c savesegment(gs, ctxt->gs); gs 121 arch/x86/power/cpu_32.c loadsegment(gs, ctxt->gs); gs 56 arch/x86/power/cpu_64.c asm volatile ("movw %%gs, %0" : "=m" (ctxt->gs)); gs 119 arch/x86/power/cpu_64.c load_gs_index(ctxt->gs); gs 377 arch/x86/xen/enlighten.c loadsegment(gs, 0); gs 26 include/asm-x86/a.out-core.h u16 gs; gs 60 include/asm-x86/a.out-core.h savesegment(gs, gs); gs 31 include/asm-x86/current.h #define GET_CURRENT(reg) movq %gs:(pda_pcurrent),reg gs 127 include/asm-x86/elf.h savesegment(gs, pr_reg[10]); \ gs 169 include/asm-x86/elf.h t->fs = t->gs = 0; gs 231 include/asm-x86/elf.h (pr_reg)[22] = current->thread.gs; \ gs 108 include/asm-x86/kvm.h struct kvm_segment cs, ds, es, fs, gs, ss; gs 280 include/asm-x86/kvm_host.h } tr, es, ds, fs, gs; gs 41 include/asm-x86/lguest.h unsigned long gs; gs 884 include/asm-x86/paravirt.h PVOP_VCALL1(pv_cpu_ops.load_gs_index, gs); gs 216 include/asm-x86/processor.h unsigned short gs, __gsh; gs 403 include/asm-x86/processor.h unsigned long gs; gs 103 include/asm-x86/sigcontext.h unsigned short gs, __gsh; gs 139 include/asm-x86/sigcontext.h unsigned short gs, __gsh; gs 210 include/asm-x86/sigcontext.h unsigned short gs; gs 252 include/asm-x86/sigcontext.h unsigned short gs; gs 51 include/asm-x86/sigcontext32.h unsigned short gs, __gsh; gs 16 include/asm-x86/suspend_32.h u16 es, fs, gs, ss; gs 27 include/asm-x86/suspend_64.h u16 ds, es, fs, gs, ss; gs 224 include/asm-x86/thread_info.h movq %gs:pda_kernelstack,reg ; \ gs 38 include/asm-x86/user32.h unsigned short fs, __fs, gs, __gs; gs 88 include/asm-x86/user_32.h unsigned long gs; gs 95 include/asm-x86/user_64.h unsigned long gs; gs 89 include/asm-x86/vm86.h unsigned short gs, __gsh; gs 152 include/asm-x86/vm86.h unsigned short gs, __gsh; gs 254 include/asm-x86/vmi.h u16 gs; gs 63 include/asm-x86/xen/interface_32.h uint16_t gs, _pad5; gs 137 include/asm-x86/xen/interface_64.h uint16_t gs, _pad6[3]; /* Non-zero => takes precedence over gs_base_usr. */ gs 20 include/video/uvesafb.h __u16 gs; gs 1105 scripts/kconfig/expr.c expr_print(e, expr_print_gstr_helper, gs, E_NONE); gs 185 scripts/kconfig/expr.h void expr_gstr_print(struct expr *e, struct gstr *gs); gs 111 scripts/kconfig/lkc.h void str_free(struct gstr *gs); gs 112 scripts/kconfig/lkc.h void str_append(struct gstr *gs, const char *s); gs 113 scripts/kconfig/lkc.h void str_printf(struct gstr *gs, const char *fmt, ...); gs 114 scripts/kconfig/lkc.h const char *str_get(struct gstr *gs); gs 78 scripts/kconfig/util.c struct gstr gs; gs 79 scripts/kconfig/util.c gs.s = malloc(sizeof(char) * 64); gs 80 scripts/kconfig/util.c gs.len = 64; gs 81 scripts/kconfig/util.c strcpy(gs.s, "\0"); gs 82 scripts/kconfig/util.c return gs; gs 88 scripts/kconfig/util.c struct gstr gs; gs 89 scripts/kconfig/util.c gs.s = strdup(s); gs 90 scripts/kconfig/util.c gs.len = strlen(s) + 1; gs 91 scripts/kconfig/util.c return gs; gs 97 scripts/kconfig/util.c if (gs->s) gs 98 scripts/kconfig/util.c free(gs->s); gs 99 scripts/kconfig/util.c gs->s = NULL; gs 100 scripts/kconfig/util.c gs->len = 0; gs 108 scripts/kconfig/util.c l = strlen(gs->s) + strlen(s) + 1; gs 109 scripts/kconfig/util.c if (l > gs->len) { gs 110 scripts/kconfig/util.c gs->s = realloc(gs->s, l); gs 111 scripts/kconfig/util.c gs->len = l; gs 113 scripts/kconfig/util.c strcat(gs->s, s); gs 124 scripts/kconfig/util.c str_append(gs, s); gs 131 scripts/kconfig/util.c return gs->s;