patch-2.3.23 linux/include/asm-arm/proc-armo/pgtable.h

Next file: linux/include/asm-arm/proc-armo/processor.h
Previous file: linux/include/asm-arm/proc-armo/param.h
Back to the patch index
Back to the overall index

diff -u --recursive --new-file v2.3.22/linux/include/asm-arm/proc-armo/pgtable.h linux/include/asm-arm/proc-armo/pgtable.h
@@ -1,105 +1,23 @@
 /*
  * linux/include/asm-arm/proc-armo/pgtable.h
  *
- * Copyright (C) 1995, 1996 Russell King
- * Modified 18/19-Oct-1997 for two-level page table
+ * Copyright (C) 1995-1999 Russell King
+ *
+ * 18-Oct-1997	RMK	Now two-level (32x32)
  */
 #ifndef __ASM_PROC_PGTABLE_H
 #define __ASM_PROC_PGTABLE_H
 
-#include <linux/config.h>
-#include <linux/slab.h>
-#include <asm/arch/memory.h>		/* For TASK_SIZE */
-
-#define LIBRARY_TEXT_START 0x0c000000
-
 /*
- * Cache flushing...
+ * PMD_SHIFT determines the size of the area a second-level page table can map
  */
-#define flush_cache_all()			do { } while (0)
-#define flush_cache_mm(mm)			do { } while (0)
-#define flush_cache_range(mm,start,end)		do { } while (0)
-#define flush_cache_page(vma,vmaddr)		do { } while (0)
-#define flush_page_to_ram(page)			do { } while (0)
-#define flush_icache_range(start,end)		do { } while (0)
-
-/*
- * TLB flushing:
- *
- *  - flush_tlb() flushes the current mm struct TLBs
- *  - flush_tlb_all() flushes all processes TLBs
- *  - flush_tlb_mm(mm) flushes the specified mm context TLB's
- *  - flush_tlb_page(vma, vmaddr) flushes one page
- *  - flush_tlb_range(mm, start, end) flushes a range of pages
- */
-#define flush_tlb()			do { } while (0)
-#define flush_tlb_all()			do { } while (0)
-#define flush_tlb_mm(mm)		do { } while (0)
-#define flush_tlb_range(mm, start, end) do { } while (0)
-#define flush_tlb_page(vma, vmaddr)	do { } while (0)
-
-/*
- * We have a mem map cache...
- */
-extern __inline__ void update_memc_all(void)
-{
-	struct task_struct *p;
-
-	p = &init_task;
-	do {
-		processor.u.armv2._update_map(p);
-		p = p->next_task;
-	} while (p != &init_task);
-
-	processor.u.armv2._remap_memc (current);
-}
-
-extern __inline__ void update_memc_task(struct task_struct *tsk)
-{
-	processor.u.armv2._update_map(tsk);
-
-	if (tsk == current)
-		processor.u.armv2._remap_memc (tsk);
-}
-
-extern __inline__ void update_memc_mm(struct mm_struct *mm)
-{
-	struct task_struct *p;
-
-	p = &init_task;
-	do {
-		if (p->mm == mm)
-			processor.u.armv2._update_map(p);
-		p = p->next_task;
-	} while (p != &init_task);
-
-	if (current->mm == mm)
-		processor.u.armv2._remap_memc (current);
-}
-
-extern __inline__ void update_memc_addr(struct mm_struct *mm, unsigned long addr, pte_t pte)
-{
-	struct task_struct *p;
-
-	p = &init_task;
-	do {
-		if (p->mm == mm)
-			processor.u.armv2._update_mmu_cache(p, addr, pte);
-		p = p->next_task;
-	} while (p != &init_task);
-
-	if (current->mm == mm)
-		processor.u.armv2._remap_memc (current);
-}
-
-#define __flush_entry_to_ram(entry)
-
-/* PMD_SHIFT determines the size of the area a second-level page table can map */
 #define PMD_SHIFT       20
 #define PMD_SIZE        (1UL << PMD_SHIFT)
 #define PMD_MASK        (~(PMD_SIZE-1))
 
-/* PGDIR_SHIFT determines what a third-level page table entry can map */
+/*
+ * PGDIR_SHIFT determines what a third-level page table entry can map
+ */
 #define PGDIR_SHIFT     20
 #define PGDIR_SIZE      (1UL << PGDIR_SHIFT)
 #define PGDIR_MASK      (~(PGDIR_SIZE-1))
@@ -107,8 +25,6 @@
 /*
  * entries per page directory level: the arm3 is one-level, so
  * we don't really have any PMD or PTE directory physically.
- *
- * 18-Oct-1997 RMK Now two-level (32x32)
  */
 #define PTRS_PER_PTE    32
 #define PTRS_PER_PMD    1
@@ -119,11 +35,11 @@
 #define VMALLOC_VMADDR(x) ((unsigned long)(x))
 #define VMALLOC_END	0x01c00000
 
-#define _PAGE_PRESENT		0x01
-#define _PAGE_READONLY		0x02
-#define _PAGE_NOT_USER		0x04
-#define _PAGE_OLD		0x08
-#define _PAGE_CLEAN		0x10
+#define _PAGE_PRESENT	0x01
+#define _PAGE_READONLY	0x02
+#define _PAGE_NOT_USER	0x04
+#define _PAGE_OLD	0x08
+#define _PAGE_CLEAN	0x10
 
 #define _PAGE_TABLE     (_PAGE_PRESENT)
 #define _PAGE_CHG_MASK  (PAGE_MASK | _PAGE_OLD | _PAGE_CLEAN)
@@ -157,47 +73,6 @@
 #define __S110  PAGE_SHARED
 #define __S111  PAGE_SHARED
 
-#undef TEST_VERIFY_AREA
-
-extern unsigned long *empty_zero_page;
-
-/*
- * BAD_PAGETABLE is used when we need a bogus page-table, while
- * BAD_PAGE is used for a bogus page.
- *
- * ZERO_PAGE is a global shared page that is always zero: used
- * for zero-mapped memory areas etc..
- */
-extern pte_t __bad_page(void);
-extern pte_t *__bad_pagetable(void);
-
-#define BAD_PAGETABLE __bad_pagetable()
-#define BAD_PAGE __bad_page()
-#define ZERO_PAGE(vaddr) ((unsigned long) empty_zero_page)
-
-/* number of bits that fit into a memory pointer */
-#define BYTES_PER_PTR			(sizeof(unsigned long))
-#define BITS_PER_PTR                    (8*BYTES_PER_PTR)
-
-/* to align the pointer to a pointer address */
-#define PTR_MASK                        (~(sizeof(void*)-1))
-
-/* sizeof(void*)==1<<SIZEOF_PTR_LOG2 */
-#define SIZEOF_PTR_LOG2                 2
-
-/* to find an entry in a page-table */
-#define PAGE_PTR(address) \
-((unsigned long)(address)>>(PAGE_SHIFT-SIZEOF_PTR_LOG2)&PTR_MASK&~PAGE_MASK)
-
-/* to set the page-dir */
-#define SET_PAGE_DIR(tsk,pgdir)						\
-do {									\
-	tsk->tss.memmap = (unsigned long)pgdir;				\
-	processor.u.armv2._update_map(tsk);				\
-	if ((tsk) == current)						\
-		processor.u.armv2._remap_memc (current);		\
-} while (0)
-
 extern unsigned long physical_start;
 extern unsigned long physical_end;
 
@@ -229,7 +104,6 @@
 extern inline int pte_exec(pte_t pte)           { return !(pte_val(pte) & _PAGE_NOT_USER);     }
 extern inline int pte_dirty(pte_t pte)          { return !(pte_val(pte) & _PAGE_CLEAN);        }
 extern inline int pte_young(pte_t pte)          { return !(pte_val(pte) & _PAGE_OLD);          }
-#define pte_cacheable(pte) 1
 
 extern inline pte_t pte_nocache(pte_t pte)	{ return pte; }
 extern inline pte_t pte_wrprotect(pte_t pte)    { pte_val(pte) |= _PAGE_READONLY;  return pte; }
@@ -322,128 +196,45 @@
  * if any.
  */
 
-#ifndef __SMP__
-#ifndef CONFIG_NO_PGT_CACHE
-extern struct pgtable_cache_struct {
-	unsigned long *pgd_cache;
-	unsigned long *pte_cache;
-	unsigned long pgtable_cache_sz;
-} quicklists;
-
-#define pmd_quicklist ((unsigned long *)0)
-#define pte_quicklist (quicklists.pte_cache)
-#define pgd_quicklist (quicklists.pgd_cache)
-#define pgtable_cache_size (quicklists.pgtable_cache_sz)
-#endif
-
-#else
-#error Pgtable caches have to be per-CPU, so that no locking is needed.
-#endif
-
-extern pgd_t *get_pgd_slow(void);
 extern void free_table(void *table);
 
-#ifndef CONFIG_NO_PGT_CACHE
-extern __inline__ pgd_t *get_pgd_fast(void)
-{
-	unsigned long *ret;
-
-	if((ret = pgd_quicklist) != NULL) {
-		pgd_quicklist = (unsigned long *)(*ret);
-		ret[0] = ret[1];
-		pgtable_cache_size--;
-	} else
-		ret = (unsigned long *)get_pgd_slow();
-	return (pgd_t *)ret;
-}
-
-extern __inline__ void free_pgd_fast(pgd_t *pgd)
-{
-	*(unsigned long *)pgd = (unsigned long) pgd_quicklist;
-	pgd_quicklist = (unsigned long *) pgd;
-	pgtable_cache_size++;
-}
-#endif
-
 /* keep this as an inline so we get type checking */
 extern __inline__ void free_pgd_slow(pgd_t *pgd)
 {
-	free_table((void *)pgd);
-}
-
-extern pte_t *get_pte_slow(pmd_t *pmd, unsigned long address_preadjusted);
-
-#ifndef CONFIG_NO_PGT_CACHE
-extern __inline__ pte_t *get_pte_fast(void)
-{
-	unsigned long *ret;
-
-	if((ret = (unsigned long *)pte_quicklist) != NULL) {
-		pte_quicklist = (unsigned long *)(*ret);
-		ret[0] = ret[1];
-		pgtable_cache_size--;
-	}
-	return (pte_t *)ret;
-}
-
-extern __inline__ void free_pte_fast(pte_t *pte)
-{
-	*(unsigned long *)pte = (unsigned long) pte_quicklist;
-	pte_quicklist = (unsigned long *) pte;
-	pgtable_cache_size++;
+	free_table(pgd);
 }
-#endif
 
 /* keep this as an inline so we get type checking */
 extern __inline__ void free_pte_slow(pte_t *pte)
 {
-	free_table((void *)pte);
-}
-
-/* We don't use pmd cache, so this is a dummy routine */
-extern __inline__ pmd_t *get_pmd_fast(void)
-{
-	return (pmd_t *)0;
-}
-
-extern __inline__ void free_pmd_fast(pmd_t *pmd)
-{
+	free_table(pte);
 }
 
 extern __inline__ void free_pmd_slow(pmd_t *pmd)
 {
 }
 
-extern void __bad_pmd(pmd_t *pmd);
-extern void __bad_pmd_kernel(pmd_t *pmd);
-
-#ifdef CONFIG_NO_PGT_CACHE
-#define pte_free_kernel(pte)    free_pte_slow(pte)
-#define pte_free(pte)           free_pte_slow(pte)
-#define pgd_free(pgd)           free_pgd_slow(pgd)
-#define pgd_alloc()             get_pgd_slow()
+#define pgd_free(pgd)           free_pgd_fast(pgd)
 
-extern __inline__ pte_t *pte_alloc(pmd_t * pmd, unsigned long address)
+extern __inline__ pgd_t *pgd_alloc(void)
 {
-	address = (address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
+	extern pgd_t *get_pgd_slow(void);
+	pgd_t *pgd;
 
-	if (pmd_none (*pmd)) {
-		return get_pte_slow(pmd, address);
-	}
-	if (pmd_bad (*pmd)) {
-		__bad_pmd(pmd);
-		return NULL;
-	}
-	return (pte_t *) pmd_page(*pmd) + address;
+	pgd = get_pgd_fast();
+	if (!pgd)
+		pgd = get_pgd_slow();
+
+	return pgd;
 }
-#else
+
 #define pte_free_kernel(pte)    free_pte_fast(pte)
 #define pte_free(pte)           free_pte_fast(pte)
-#define pgd_free(pgd)           free_pgd_fast(pgd)
-#define pgd_alloc()             get_pgd_fast()
 
 extern __inline__ pte_t *pte_alloc(pmd_t * pmd, unsigned long address)
 {
+	extern pte_t *get_pte_slow(pmd_t *pmd, unsigned long address_preadjusted);
+
 	address = (address >> PAGE_SHIFT) & (PTRS_PER_PTE - 1);
 
 	if (pmd_none (*pmd)) {
@@ -460,7 +251,6 @@
 	}
 	return (pte_t *) pmd_page(*pmd) + address;
 }
-#endif
 
 /*
  * allocating and freeing a pmd is trivial: the 1-entry pmd is
@@ -479,34 +269,4 @@
 #define pmd_alloc_kernel        pmd_alloc
 #define pte_alloc_kernel        pte_alloc
 
-extern __inline__ void set_pgdir(unsigned long address, pgd_t entry)
-{
-	struct task_struct * p;
-
-	read_lock(&tasklist_lock);
-	for_each_task(p) {
-		if (!p->mm)
-			continue;
-		*pgd_offset(p->mm,address) = entry;
-	}
-	read_unlock(&tasklist_lock);
-#ifndef CONFIG_NO_PGT_CACHE
-	{
-		pgd_t *pgd;
-		for (pgd = (pgd_t *)pgd_quicklist; pgd;
-		     pgd = (pgd_t *)*(unsigned long *)pgd)
-			pgd[address >> PGDIR_SHIFT] = entry;
-	}
-#endif
-}
-
-extern pgd_t swapper_pg_dir[PTRS_PER_PGD];
-
-#define update_mmu_cache(vma,address,pte)
-
-#define SWP_TYPE(entry) (((entry) >> 1) & 0x7f)
-#define SWP_OFFSET(entry) ((entry) >> 8)
-#define SWP_ENTRY(type,offset) (((type) << 1) | ((offset) <<  8))
-
-#endif /* __ASM_PROC_PAGE_H */
-
+#endif /* __ASM_PROC_PGTABLE_H */

FUNET's LINUX-ADM group, linux-adm@nic.funet.fi
TCL-scripts by Sam Shen (who was at: slshen@lbl.gov)