#include <asm/sigcontext32.h>
 #include <asm/fpu32.h>
 #include <asm/proto.h>
-#include <asm/vsyscall32.h>
+#include <asm/vdso.h>
 
 #define DEBUG_SIG 0
 
                        goto give_sigsegv;
        }
 
-       /* Return stub is in 32bit vsyscall page */
-       if (current->binfmt->hasvdso)
-               restorer = VSYSCALL32_SIGRETURN;
-       else
-               restorer = (void *)&frame->retcode;
-       if (ka->sa.sa_flags & SA_RESTORER)
+       if (ka->sa.sa_flags & SA_RESTORER) {
                restorer = ka->sa.sa_restorer;
+       } else {
+               /* Return stub is in 32bit vsyscall page */
+               if (current->binfmt->hasvdso)
+                       restorer = VDSO32_SYMBOL(current->mm->context.vdso,
+                                                sigreturn);
+               else
+                       restorer = (void *)&frame->retcode;
+       }
        err |= __put_user(ptr_to_compat(restorer), &frame->pretcode);
 
        /*
 {
        struct rt_sigframe __user *frame;
        struct exec_domain *ed = current_thread_info()->exec_domain;
-       void __user *restorer = VSYSCALL32_RTSIGRETURN;
+       void __user *restorer;
        int err = 0;
 
        /* __copy_to_user optimizes that into a single 8 byte store */
 
        if (ka->sa.sa_flags & SA_RESTORER)
                restorer = ka->sa.sa_restorer;
+       else
+               restorer = VDSO32_SYMBOL(current->mm->context.vdso,
+                                        rt_sigreturn);
        err |= __put_user(ptr_to_compat(restorer), &frame->pretcode);
 
        /*
 
 #include <asm/elf.h>
 #include <asm/tlbflush.h>
 #include <asm/vdso.h>
+#include <asm/proto.h>
 
 enum {
        VDSO_DISABLED = 0,
 #define VDSO_DEFAULT   VDSO_ENABLED
 #endif
 
+#ifdef CONFIG_X86_64
+#define vdso_enabled                   sysctl_vsyscall32
+#define arch_setup_additional_pages    syscall32_setup_pages
+#endif
+
+/*
+ * This is the difference between the prelinked addresses in the vDSO images
+ * and the VDSO_HIGH_BASE address where CONFIG_COMPAT_VDSO places the vDSO
+ * in the user address space.
+ */
+#define VDSO_ADDR_ADJUST       (VDSO_HIGH_BASE - (unsigned long)VDSO32_PRELINK)
+
 /*
  * Should the kernel map a VDSO page into processes and pass its
  * address down to glibc upon exec()?
  */
 unsigned int __read_mostly vdso_enabled = VDSO_DEFAULT;
 
-EXPORT_SYMBOL_GPL(vdso_enabled);
-
 static int __init vdso_setup(char *s)
 {
        vdso_enabled = simple_strtoul(s, NULL, 0);
        return 1;
 }
 
-__setup("vdso=", vdso_setup);
+/*
+ * For consistency, the argument vdso32=[012] affects the 32-bit vDSO
+ * behavior on both 64-bit and 32-bit kernels.
+ * On 32-bit kernels, vdso=[012] means the same thing.
+ */
+__setup("vdso32=", vdso_setup);
 
-extern asmlinkage void ia32_sysenter_target(void);
+#ifdef CONFIG_X86_32
+__setup_param("vdso=", vdso32_setup, vdso_setup, 0);
+
+EXPORT_SYMBOL_GPL(vdso_enabled);
+#endif
 
 static __init void reloc_symtab(Elf32_Ehdr *ehdr,
                                unsigned offset, unsigned size)
                case STT_FUNC:
                case STT_SECTION:
                case STT_FILE:
-                       sym->st_value += VDSO_HIGH_BASE;
+                       sym->st_value += VDSO_ADDR_ADJUST;
                }
        }
 }
                case DT_VERNEED:
                case DT_ADDRRNGLO ... DT_ADDRRNGHI:
                        /* definitely pointers needing relocation */
-                       dyn->d_un.d_ptr += VDSO_HIGH_BASE;
+                       dyn->d_un.d_ptr += VDSO_ADDR_ADJUST;
                        break;
 
                case DT_ENCODING ... OLD_DT_LOOS-1:
                           they're even */
                        if (dyn->d_tag >= DT_ENCODING &&
                            (dyn->d_tag & 1) == 0)
-                               dyn->d_un.d_ptr += VDSO_HIGH_BASE;
+                               dyn->d_un.d_ptr += VDSO_ADDR_ADJUST;
                        break;
 
                case DT_VERDEFNUM:
        int i;
 
        BUG_ON(memcmp(ehdr->e_ident, ELFMAG, 4) != 0 ||
-              !elf_check_arch(ehdr) ||
+              !elf_check_arch_ia32(ehdr) ||
               ehdr->e_type != ET_DYN);
 
-       ehdr->e_entry += VDSO_HIGH_BASE;
+       ehdr->e_entry += VDSO_ADDR_ADJUST;
 
        /* rebase phdrs */
        phdr = (void *)ehdr + ehdr->e_phoff;
        for (i = 0; i < ehdr->e_phnum; i++) {
-               phdr[i].p_vaddr += VDSO_HIGH_BASE;
+               phdr[i].p_vaddr += VDSO_ADDR_ADJUST;
 
                /* relocate dynamic stuff */
                if (phdr[i].p_type == PT_DYNAMIC)
                if (!(shdr[i].sh_flags & SHF_ALLOC))
                        continue;
 
-               shdr[i].sh_addr += VDSO_HIGH_BASE;
+               shdr[i].sh_addr += VDSO_ADDR_ADJUST;
 
                if (shdr[i].sh_type == SHT_SYMTAB ||
                    shdr[i].sh_type == SHT_DYNSYM)
        }
 }
 
+/*
+ * These symbols are defined by vdso32.S to mark the bounds
+ * of the ELF DSO images included therein.
+ */
+extern const char vdso32_default_start, vdso32_default_end;
+extern const char vdso32_sysenter_start, vdso32_sysenter_end;
+static struct page *vdso32_pages[1];
+
+#ifdef CONFIG_X86_64
+
+static int use_sysenter __read_mostly = -1;
+
+#define        vdso32_sysenter()       (use_sysenter > 0)
+
+/* May not be __init: called during resume */
+void syscall32_cpu_init(void)
+{
+       if (use_sysenter < 0)
+               use_sysenter = (boot_cpu_data.x86_vendor == X86_VENDOR_INTEL);
+
+       /* Load these always in case some future AMD CPU supports
+          SYSENTER from compat mode too. */
+       checking_wrmsrl(MSR_IA32_SYSENTER_CS, (u64)__KERNEL_CS);
+       checking_wrmsrl(MSR_IA32_SYSENTER_ESP, 0ULL);
+       checking_wrmsrl(MSR_IA32_SYSENTER_EIP, (u64)ia32_sysenter_target);
+
+       wrmsrl(MSR_CSTAR, ia32_cstar_target);
+}
+
+#define compat_uses_vma                1
+
+static inline void map_compat_vdso(int map)
+{
+}
+
+#else  /* CONFIG_X86_32 */
+
+#define vdso32_sysenter()      (boot_cpu_has(X86_FEATURE_SEP))
+
 void enable_sep_cpu(void)
 {
        int cpu = get_cpu();
        return 0;
 }
 
-/*
- * These symbols are defined by vsyscall.o to mark the bounds
- * of the ELF DSO images included therein.
- */
-extern const char vsyscall_int80_start, vsyscall_int80_end;
-extern const char vsyscall_sysenter_start, vsyscall_sysenter_end;
-static struct page *syscall_pages[1];
+#define compat_uses_vma                0
 
 static void map_compat_vdso(int map)
 {
 
        vdso_mapped = map;
 
-       __set_fixmap(FIX_VDSO, page_to_pfn(syscall_pages[0]) << PAGE_SHIFT,
+       __set_fixmap(FIX_VDSO, page_to_pfn(vdso32_pages[0]) << PAGE_SHIFT,
                     map ? PAGE_READONLY_EXEC : PAGE_NONE);
 
        /* flush stray tlbs */
        flush_tlb_all();
 }
 
+#endif /* CONFIG_X86_64 */
+
 int __init sysenter_setup(void)
 {
        void *syscall_page = (void *)get_zeroed_page(GFP_ATOMIC);
        const void *vsyscall;
        size_t vsyscall_len;
 
-       syscall_pages[0] = virt_to_page(syscall_page);
+       vdso32_pages[0] = virt_to_page(syscall_page);
 
+#ifdef CONFIG_X86_32
        gate_vma_init();
 
        printk("Compat vDSO mapped to %08lx.\n", __fix_to_virt(FIX_VDSO));
+#endif
 
-       if (!boot_cpu_has(X86_FEATURE_SEP)) {
-               vsyscall = &vsyscall_int80_start;
-               vsyscall_len = &vsyscall_int80_end - &vsyscall_int80_start;
+       if (!vdso32_sysenter()) {
+               vsyscall = &vdso32_default_start;
+               vsyscall_len = &vdso32_default_end - &vdso32_default_start;
        } else {
-               vsyscall = &vsyscall_sysenter_start;
-               vsyscall_len = &vsyscall_sysenter_end - &vsyscall_sysenter_start;
+               vsyscall = &vdso32_sysenter_start;
+               vsyscall_len = &vdso32_sysenter_end - &vdso32_sysenter_start;
        }
 
        memcpy(syscall_page, vsyscall, vsyscall_len);
                        ret = addr;
                        goto up_fail;
                }
+       }
 
+       if (compat_uses_vma || !compat) {
                /*
                 * MAYWRITE to allow gdb to COW and set breakpoints
                 *
                                              VM_READ|VM_EXEC|
                                              VM_MAYREAD|VM_MAYWRITE|VM_MAYEXEC|
                                              VM_ALWAYSDUMP,
-                                             syscall_pages);
+                                             vdso32_pages);
 
                if (ret)
                        goto up_fail;
        return ret;
 }
 
+#ifdef CONFIG_X86_64
+
+__initcall(sysenter_setup);
+
+#else  /* CONFIG_X86_32 */
+
 const char *arch_vma_name(struct vm_area_struct *vma)
 {
        if (vma->vm_mm && vma->vm_start == (long)vma->vm_mm->context.vdso)
 {
        return 0;
 }
+
+#endif /* CONFIG_X86_64 */
 
 
 #ifdef __KERNEL__
 
+/*
+ * This is used to ensure we don't load something for the wrong architecture.
+ */
+#define elf_check_arch_ia32(x) \
+       (((x)->e_machine == EM_386) || ((x)->e_machine == EM_486))
+
 #ifdef CONFIG_X86_32
 #include <asm/processor.h>
 #include <asm/system.h>                /* for savesegment */
 #include <asm/desc.h>
 #include <asm/vdso.h>
 
-/*
- * This is used to ensure we don't load something for the wrong architecture.
- */
-#define elf_check_arch(x) \
-       (((x)->e_machine == EM_386) || ((x)->e_machine == EM_486))
+#define elf_check_arch(x)      elf_check_arch_ia32(x)
 
 /* SVR4/i386 ABI (pages 3-31, 3-32) says that when the program starts %edx
    contains a pointer to a function which might be registered using `atexit'.
 #define ELF_CORE_XFPREG_TYPE NT_PRXFPREG
 
 #define VDSO_HIGH_BASE         (__fix_to_virt(FIX_VDSO))
-#define VDSO_CURRENT_BASE      ((unsigned long)current->mm->context.vdso)
-
-#define VDSO_ENTRY \
-       ((unsigned long) VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall))
 
 /* update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT entries changes */
 
 
 #else /* CONFIG_X86_32 */
 
+#define VDSO_HIGH_BASE         0xffffe000U /* CONFIG_COMPAT_VDSO address */
+
 /* 1GB for 64bit, 8MB for 32bit */
 #define STACK_RND_MASK (test_thread_flag(TIF_IA32) ? 0x7ff : 0x3fffff)
 
 
 #endif /* !CONFIG_X86_32 */
 
+#define VDSO_CURRENT_BASE      ((unsigned long)current->mm->context.vdso)
+
+#define VDSO_ENTRY \
+       ((unsigned long) VDSO32_SYMBOL(VDSO_CURRENT_BASE, vsyscall))
+
 struct linux_binprm;
 
 #define ARCH_HAS_SETUP_ADDITIONAL_PAGES 1