mirror of
https://github.com/torvalds/linux.git
synced 2024-11-15 16:41:58 +00:00
16dd07bc64
More trimming of the page fault path. Permissions are passed around in a single int rather than one bit per int. The permission values are copied from libc so that they can be passed to mmap and mprotect without any further conversion. The register sets used by do_syscall_stub and copy_context_skas0 are initialized once, at boot time, rather than once per call. wait_stub_done checks whether it is getting the signals it expects by comparing the wait status to a mask containing bits for the signals of interest rather than comparing individually to the signal numbers. It also has one check for a wait failure instead of two. The caller is expected to do the initial continue of the stub. This gets rid of an argument and some logic. The fname argument is gone, as that can be had from a stack trace. user_signal() is collapsed into userspace() as it is basically one or two lines of code afterwards. The physical memory remapping stuff is gone, as it is unused. flush_tlb_page is inlined. Signed-off-by: Jeff Dike <jdike@linux.intel.com> Cc: Paolo 'Blaisorblade' Giarrusso <blaisorblade@yahoo.it> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
121 lines
3.4 KiB
C
121 lines
3.4 KiB
C
/*
|
|
* Copyright (C) 2000 - 2003 Jeff Dike (jdike@addtoit.com)
|
|
* Copyright 2003 PathScale, Inc.
|
|
* Licensed under the GPL
|
|
*/
|
|
|
|
#ifndef __UM_PAGE_H
|
|
#define __UM_PAGE_H
|
|
|
|
struct page;
|
|
|
|
#include <asm/vm-flags.h>
|
|
|
|
/* PAGE_SHIFT determines the page size */
|
|
#define PAGE_SHIFT 12
|
|
#define PAGE_SIZE (1UL << PAGE_SHIFT)
|
|
#define PAGE_MASK (~(PAGE_SIZE-1))
|
|
|
|
/*
|
|
* These are used to make use of C type-checking..
|
|
*/
|
|
|
|
#define clear_page(page) memset((void *)(page), 0, PAGE_SIZE)
|
|
#define copy_page(to,from) memcpy((void *)(to), (void *)(from), PAGE_SIZE)
|
|
|
|
#define clear_user_page(page, vaddr, pg) clear_page(page)
|
|
#define copy_user_page(to, from, vaddr, pg) copy_page(to, from)
|
|
|
|
#if defined(CONFIG_3_LEVEL_PGTABLES) && !defined(CONFIG_64BIT)
|
|
|
|
typedef struct { unsigned long pte_low, pte_high; } pte_t;
|
|
typedef struct { unsigned long long pmd; } pmd_t;
|
|
typedef struct { unsigned long pgd; } pgd_t;
|
|
#define pte_val(x) ((x).pte_low | ((unsigned long long) (x).pte_high << 32))
|
|
|
|
#define pte_get_bits(pte, bits) ((pte).pte_low & (bits))
|
|
#define pte_set_bits(pte, bits) ((pte).pte_low |= (bits))
|
|
#define pte_clear_bits(pte, bits) ((pte).pte_low &= ~(bits))
|
|
#define pte_copy(to, from) ({ (to).pte_high = (from).pte_high; \
|
|
smp_wmb(); \
|
|
(to).pte_low = (from).pte_low; })
|
|
#define pte_is_zero(pte) (!((pte).pte_low & ~_PAGE_NEWPAGE) && !(pte).pte_high)
|
|
#define pte_set_val(pte, phys, prot) \
|
|
({ (pte).pte_high = (phys) >> 32; \
|
|
(pte).pte_low = (phys) | pgprot_val(prot); })
|
|
|
|
#define pmd_val(x) ((x).pmd)
|
|
#define __pmd(x) ((pmd_t) { (x) } )
|
|
|
|
typedef unsigned long long pfn_t;
|
|
typedef unsigned long long phys_t;
|
|
|
|
#else
|
|
|
|
typedef struct { unsigned long pte; } pte_t;
|
|
typedef struct { unsigned long pgd; } pgd_t;
|
|
|
|
#ifdef CONFIG_3_LEVEL_PGTABLES
|
|
typedef struct { unsigned long pmd; } pmd_t;
|
|
#define pmd_val(x) ((x).pmd)
|
|
#define __pmd(x) ((pmd_t) { (x) } )
|
|
#endif
|
|
|
|
#define pte_val(x) ((x).pte)
|
|
|
|
|
|
#define pte_get_bits(p, bits) ((p).pte & (bits))
|
|
#define pte_set_bits(p, bits) ((p).pte |= (bits))
|
|
#define pte_clear_bits(p, bits) ((p).pte &= ~(bits))
|
|
#define pte_copy(to, from) ((to).pte = (from).pte)
|
|
#define pte_is_zero(p) (!((p).pte & ~_PAGE_NEWPAGE))
|
|
#define pte_set_val(p, phys, prot) (p).pte = (phys | pgprot_val(prot))
|
|
|
|
typedef unsigned long pfn_t;
|
|
typedef unsigned long phys_t;
|
|
|
|
#endif
|
|
|
|
typedef struct { unsigned long pgprot; } pgprot_t;
|
|
|
|
#define pgd_val(x) ((x).pgd)
|
|
#define pgprot_val(x) ((x).pgprot)
|
|
|
|
#define __pte(x) ((pte_t) { (x) } )
|
|
#define __pgd(x) ((pgd_t) { (x) } )
|
|
#define __pgprot(x) ((pgprot_t) { (x) } )
|
|
|
|
/* to align the pointer to the (next) page boundary */
|
|
#define PAGE_ALIGN(addr) (((addr)+PAGE_SIZE-1)&PAGE_MASK)
|
|
|
|
extern unsigned long uml_physmem;
|
|
|
|
#define PAGE_OFFSET (uml_physmem)
|
|
#define KERNELBASE PAGE_OFFSET
|
|
|
|
#define __va_space (8*1024*1024)
|
|
|
|
#include "mem.h"
|
|
|
|
/* Cast to unsigned long before casting to void * to avoid a warning from
|
|
* mmap_kmem about cutting a long long down to a void *. Not sure that
|
|
* casting is the right thing, but 32-bit UML can't have 64-bit virtual
|
|
* addresses
|
|
*/
|
|
#define __pa(virt) to_phys((void *) (unsigned long) (virt))
|
|
#define __va(phys) to_virt((unsigned long) (phys))
|
|
|
|
#define phys_to_pfn(p) ((p) >> PAGE_SHIFT)
|
|
#define pfn_to_phys(pfn) ((pfn) << PAGE_SHIFT)
|
|
|
|
#define pfn_valid(pfn) ((pfn) < max_mapnr)
|
|
#define virt_addr_valid(v) pfn_valid(phys_to_pfn(__pa(v)))
|
|
|
|
extern struct page *arch_validate(struct page *page, gfp_t mask, int order);
|
|
#define HAVE_ARCH_VALIDATE
|
|
|
|
#include <asm-generic/memory_model.h>
|
|
#include <asm-generic/page.h>
|
|
|
|
#endif
|