- asm ("movq %%rax,%c[rax_off](%[regs])\n\t" \
- "movq %%rbx,%c[rbx_off](%[regs])\n\t" \
- "movq %%rcx,%c[rcx_off](%[regs])\n\t" \
- "movq %%rdx,%c[rdx_off](%[regs])\n\t" \
- "movq %%rbp,%c[rbp_off](%[regs])\n\t" \
- "movq %%rsp,%c[rsp_off](%[regs])\n\t" \
- "movq %%rsi,%c[rsi_off](%[regs])\n\t" \
- "movq %%rdi,%c[rdi_off](%[regs])\n\t" \
- "movq %%r8,%c[r8_off](%[regs])\n\t" \
- "movq %%r9,%c[r9_off](%[regs])\n\t" \
- "movq %%r10,%c[r10_off](%[regs])\n\t" \
- "movq %%r11,%c[r11_off](%[regs])\n\t" \
- "movq %%r12,%c[r12_off](%[regs])\n\t" \
- "movq %%r13,%c[r13_off](%[regs])\n\t" \
- "movq %%r14,%c[r14_off](%[regs])\n\t" \
- "movq %%r15,%c[r15_off](%[regs])\n\t" \
- "movw %%cs,%c[cs_off](%[regs])\n\t" \
- "movw %%ss,%c[ss_off](%[regs])\n\t" \
+ if(ust_reg_stack_ptr == NULL) { \
+ ust_reg_stack_ptr = (long*)((long)ust_reg_stack)+500; \
+ } \
+ asm volatile ( \
+ /* save original rsp */ \
+ "pushq %%rsp\n\t" \
+ /* rax will hold the ptr to the private stack bottom */ \
+ "pushq %%rax\n\t" \
+ /* rbx will be used to temporarily hold the stack bottom addr */ \
+ "pushq %%rbx\n\t" \
+ /* rdi is the input to __tls_get_addr, and also a temp var */ \
+ "pushq %%rdi\n\t" \
+ /* Start TLS access of private reg stack */ \
+ ".byte 0x66\n\t" \
+ "leaq ust_reg_stack_ptr@tlsgd(%%rip), %%rdi\n\t" \
+ ".word 0x6666\n\t" \
+ "rex64\n\t" \
+ "call __tls_get_addr@plt\n\t" \
+ /* --- End TLS access */ \
+ /* Manually push rsp to private stack */ \
+ "addq $-8,(%%rax)\n\t" \
+ "movq 24(%%rsp), %%rdi\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rdi, (%%rbx)\n\t" \
+ /* Manually push rax to private stack */ \
+ "addq $-8,(%%rax)\n\t" \
+ "movq 16(%%rsp), %%rdi\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rdi, (%%rbx)\n\t" \
+ /* Manually push rbx to private stack */ \
+ "addq $-8,(%%rax)\n\t" \
+ "movq 8(%%rsp), %%rdi\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rdi, (%%rbx)\n\t" \
+ /* Manually push rdi to private stack */ \
+ "addq $-8,(%%rax)\n\t" \
+ "movq 0(%%rsp), %%rdi\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rdi, (%%rbx)\n\t" \
+ /* now push regs to tls */ \
+ /* -- rsp already pushed -- */ \
+ /* -- rax already pushed -- */ \
+ /* -- rbx already pushed -- */ \
+ /* -- rdi already pushed -- */ \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rcx,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rdx,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rbp,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%rsi,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r8,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r9,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r10,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r11,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r12,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r13,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r14,(%%rbx)\n\t" \
+ "addq $-8,(%%rax)\n\t" \
+ "movq (%%rax), %%rbx\n\t" \
+ "movq %%r15,(%%rbx)\n\t" \