gs 52 arch/x86/ia32/ia32_aout.c u32 fs, gs;
gs 90 arch/x86/ia32/ia32_aout.c savesegment(gs, gs);
gs 91 arch/x86/ia32/ia32_aout.c dump->regs.gs = gs;
gs 218 arch/x86/ia32/ia32_signal.c unsigned int tmpflags, gs, oldgs, err = 0;
gs 237 arch/x86/ia32/ia32_signal.c err |= __get_user(gs, &sc->gs);
gs 238 arch/x86/ia32/ia32_signal.c gs |= 3;
gs 239 arch/x86/ia32/ia32_signal.c savesegment(gs, oldgs);
gs 240 arch/x86/ia32/ia32_signal.c if (gs != oldgs)
gs 241 arch/x86/ia32/ia32_signal.c load_gs_index(gs);
gs 345 arch/x86/ia32/ia32_signal.c savesegment(gs, tmp);
gs 346 arch/x86/ia32/ia32_signal.c err |= __put_user(tmp, (unsigned int __user *)&sc->gs);
gs 566 arch/x86/kernel/apm_32.c savesegment(fs, saved_fs); savesegment(gs, saved_gs)
gs 568 arch/x86/kernel/apm_32.c loadsegment(fs, saved_fs); loadsegment(gs, saved_gs)
gs 868 arch/x86/kernel/cpu/common.c loadsegment(gs, 0);
gs 131 arch/x86/kernel/process_32.c unsigned short ss, gs;
gs 137 arch/x86/kernel/process_32.c savesegment(gs, gs);
gs 141 arch/x86/kernel/process_32.c savesegment(gs, gs);
gs 165 arch/x86/kernel/process_32.c (u16)regs->ds, (u16)regs->es, (u16)regs->fs, gs, ss);
gs 316 arch/x86/kernel/process_32.c savesegment(gs, p->thread.gs);
gs 582 arch/x86/kernel/process_32.c savesegment(gs, prev->gs);
gs 627 arch/x86/kernel/process_32.c if (prev->gs | next->gs)
gs 628 arch/x86/kernel/process_32.c loadsegment(gs, next->gs);
gs 141 arch/x86/kernel/process_64.c unsigned long cr0 = 0L, cr2 = 0L, cr3 = 0L, cr4 = 0L, fs, gs, shadowgs;
gs 175 arch/x86/kernel/process_64.c rdmsrl(MSR_GS_BASE, gs);
gs 187 arch/x86/kernel/process_64.c fs, fsindex, gs, gsindex, shadowgs);
gs 336 arch/x86/kernel/process_64.c p->thread.gs = me->thread.gs;
gs 338 arch/x86/kernel/process_64.c savesegment(gs, p->thread.gsindex);
gs 585 arch/x86/kernel/process_64.c savesegment(gs, gsindex);
gs 620 arch/x86/kernel/process_64.c if (unlikely(gsindex | next->gsindex | prev->gs)) {
gs 623 arch/x86/kernel/process_64.c prev->gs = 0;
gs 625 arch/x86/kernel/process_64.c if (next->gs)
gs 626 arch/x86/kernel/process_64.c wrmsrl(MSR_KERNEL_GS_BASE, next->gs);
gs 777 arch/x86/kernel/process_64.c task->thread.gs = 0;
gs 780 arch/x86/kernel/process_64.c task->thread.gs = addr;
gs 833 arch/x86/kernel/process_64.c savesegment(gs, gsindex);
gs 837 arch/x86/kernel/process_64.c base = task->thread.gs;
gs 839 arch/x86/kernel/process_64.c base = task->thread.gs;
gs 90 arch/x86/kernel/ptrace.c if (offset != offsetof(struct user_regs_struct, gs))
gs 93 arch/x86/kernel/ptrace.c retval = task->thread.gs;
gs 95 arch/x86/kernel/ptrace.c savesegment(gs, retval);
gs 128 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs):
gs 129 arch/x86/kernel/ptrace.c task->thread.gs = value;
gs 135 arch/x86/kernel/ptrace.c loadsegment(gs, value);
gs 171 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs):
gs 221 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct,gs):
gs 227 arch/x86/kernel/ptrace.c task->thread.gs != 0) ||
gs 229 arch/x86/kernel/ptrace.c task->thread.gs == 0))
gs 321 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs):
gs 356 arch/x86/kernel/ptrace.c if (child->thread.gs != value)
gs 373 arch/x86/kernel/ptrace.c case offsetof(struct user_regs_struct, gs):
gs 401 arch/x86/kernel/ptrace.c if (task->thread.gs != 0)
gs 402 arch/x86/kernel/ptrace.c return task->thread.gs;
gs 1169 arch/x86/kernel/ptrace.c SEG32(gs);
gs 1233 arch/x86/kernel/ptrace.c SEG32(gs);
gs 152 arch/x86/kernel/signal_32.c GET_SEG(gs);
gs 263 arch/x86/kernel/signal_32.c savesegment(gs, tmp);
gs 264 arch/x86/kernel/signal_32.c err |= __put_user(tmp, (unsigned int __user *)&sc->gs);
gs 153 arch/x86/kernel/signal_64.c err |= __put_user(0, &sc->gs);
gs 161 arch/x86/kernel/vm86_32.c loadsegment(gs, current->thread.saved_gs);
gs 326 arch/x86/kernel/vm86_32.c savesegment(gs, tsk->thread.saved_gs);
gs 558 arch/x86/kernel/vmi_32.c ap.gs = 0;
gs 558 arch/x86/kvm/svm.c init_seg(&save->gs);
gs 760 arch/x86/kvm/svm.c case VCPU_SREG_GS: return &save->gs;
gs 130 arch/x86/kvm/svm.h struct vmcb_seg gs;
gs 28 arch/x86/kvm/tss.h u32 gs;
gs 1310 arch/x86/kvm/vmx.c fix_pmode_dataseg(VCPU_SREG_GS, &vcpu->arch.rmode.gs);
gs 1382 arch/x86/kvm/vmx.c fix_rmode_seg(VCPU_SREG_GS, &vcpu->arch.rmode.gs);
gs 3120 arch/x86/kvm/x86.c kvm_get_segment(vcpu, &sregs->gs, VCPU_SREG_GS);
gs 3335 arch/x86/kvm/x86.c tss->gs = get_segment_selector(vcpu, VCPU_SREG_GS);
gs 3375 arch/x86/kvm/x86.c if (kvm_load_segment_descriptor(vcpu, tss->gs, 1, VCPU_SREG_GS))
gs 3631 arch/x86/kvm/x86.c kvm_set_segment(vcpu, &sregs->gs, VCPU_SREG_GS);
gs 281 arch/x86/lguest/boot.c loadsegment(gs, 0);
gs 160 arch/x86/math-emu/get_address.c savesegment(gs, addr->selector);
gs 40 arch/x86/power/cpu_32.c savesegment(gs, ctxt->gs);
gs 121 arch/x86/power/cpu_32.c loadsegment(gs, ctxt->gs);
gs 56 arch/x86/power/cpu_64.c asm volatile ("movw %%gs, %0" : "=m" (ctxt->gs));
gs 119 arch/x86/power/cpu_64.c load_gs_index(ctxt->gs);
gs 377 arch/x86/xen/enlighten.c loadsegment(gs, 0);
gs 26 include/asm-x86/a.out-core.h u16 gs;
gs 60 include/asm-x86/a.out-core.h savesegment(gs, gs);
gs 31 include/asm-x86/current.h #define GET_CURRENT(reg) movq %gs:(pda_pcurrent),reg
gs 127 include/asm-x86/elf.h savesegment(gs, pr_reg[10]); \
gs 169 include/asm-x86/elf.h t->fs = t->gs = 0;
gs 231 include/asm-x86/elf.h (pr_reg)[22] = current->thread.gs; \
gs 108 include/asm-x86/kvm.h struct kvm_segment cs, ds, es, fs, gs, ss;
gs 280 include/asm-x86/kvm_host.h } tr, es, ds, fs, gs;
gs 41 include/asm-x86/lguest.h unsigned long gs;
gs 884 include/asm-x86/paravirt.h PVOP_VCALL1(pv_cpu_ops.load_gs_index, gs);
gs 216 include/asm-x86/processor.h unsigned short gs, __gsh;
gs 403 include/asm-x86/processor.h unsigned long gs;
gs 103 include/asm-x86/sigcontext.h unsigned short gs, __gsh;
gs 139 include/asm-x86/sigcontext.h unsigned short gs, __gsh;
gs 210 include/asm-x86/sigcontext.h unsigned short gs;
gs 252 include/asm-x86/sigcontext.h unsigned short gs;
gs 51 include/asm-x86/sigcontext32.h unsigned short gs, __gsh;
gs 16 include/asm-x86/suspend_32.h u16 es, fs, gs, ss;
gs 27 include/asm-x86/suspend_64.h u16 ds, es, fs, gs, ss;
gs 224 include/asm-x86/thread_info.h movq %gs:pda_kernelstack,reg ; \
gs 38 include/asm-x86/user32.h unsigned short fs, __fs, gs, __gs;
gs 88 include/asm-x86/user_32.h unsigned long gs;
gs 95 include/asm-x86/user_64.h unsigned long gs;
gs 89 include/asm-x86/vm86.h unsigned short gs, __gsh;
gs 152 include/asm-x86/vm86.h unsigned short gs, __gsh;
gs 254 include/asm-x86/vmi.h u16 gs;
gs 63 include/asm-x86/xen/interface_32.h uint16_t gs, _pad5;
gs 137 include/asm-x86/xen/interface_64.h uint16_t gs, _pad6[3]; /* Non-zero => takes precedence over gs_base_usr. */
gs 20 include/video/uvesafb.h __u16 gs;
gs 1105 scripts/kconfig/expr.c expr_print(e, expr_print_gstr_helper, gs, E_NONE);
gs 185 scripts/kconfig/expr.h void expr_gstr_print(struct expr *e, struct gstr *gs);
gs 111 scripts/kconfig/lkc.h void str_free(struct gstr *gs);
gs 112 scripts/kconfig/lkc.h void str_append(struct gstr *gs, const char *s);
gs 113 scripts/kconfig/lkc.h void str_printf(struct gstr *gs, const char *fmt, ...);
gs 114 scripts/kconfig/lkc.h const char *str_get(struct gstr *gs);
gs 78 scripts/kconfig/util.c struct gstr gs;
gs 79 scripts/kconfig/util.c gs.s = malloc(sizeof(char) * 64);
gs 80 scripts/kconfig/util.c gs.len = 64;
gs 81 scripts/kconfig/util.c strcpy(gs.s, "\0");
gs 82 scripts/kconfig/util.c return gs;
gs 88 scripts/kconfig/util.c struct gstr gs;
gs 89 scripts/kconfig/util.c gs.s = strdup(s);
gs 90 scripts/kconfig/util.c gs.len = strlen(s) + 1;
gs 91 scripts/kconfig/util.c return gs;
gs 97 scripts/kconfig/util.c if (gs->s)
gs 98 scripts/kconfig/util.c free(gs->s);
gs 99 scripts/kconfig/util.c gs->s = NULL;
gs 100 scripts/kconfig/util.c gs->len = 0;
gs 108 scripts/kconfig/util.c l = strlen(gs->s) + strlen(s) + 1;
gs 109 scripts/kconfig/util.c if (l > gs->len) {
gs 110 scripts/kconfig/util.c gs->s = realloc(gs->s, l);
gs 111 scripts/kconfig/util.c gs->len = l;
gs 113 scripts/kconfig/util.c strcat(gs->s, s);
gs 124 scripts/kconfig/util.c str_append(gs, s);
gs 131 scripts/kconfig/util.c return gs->s;