taglinefilesource code
pmd73arch/i386/kernel/vm86.cpmd_t *pmd;
pmd85arch/i386/kernel/vm86.cpmd = pmd_offset(pgd, 0xA0000);
pmd86arch/i386/kernel/vm86.cif (pmd_none(*pmd))
pmd88arch/i386/kernel/vm86.cif (pmd_bad(*pmd)) {
pmd89arch/i386/kernel/vm86.cprintk("vm86: bad pmd entry [%p]:%08lx\n", pmd, pmd_val(*pmd));
pmd90arch/i386/kernel/vm86.cpmd_clear(pmd);
pmd93arch/i386/kernel/vm86.cpte = pte_offset(pmd, 0xA0000);
pmd77arch/sparc/mm/srmmu.csrmmu_pmd_page(pmd_t pmd)
pmd81arch/sparc/mm/srmmu.cpage = (pmd_val(pmd) & (SRMMU_PTD_PTP_MASK)) << SRMMU_PTD_PTP_PADDR_SHIFT;
pmd114arch/sparc/mm/srmmu.cint srmmu_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd115arch/sparc/mm/srmmu.cint srmmu_pmd_bad(pmd_t pmd)
pmd117arch/sparc/mm/srmmu.creturn ((pmd_val(pmd)&SRMMU_ET_PTDBAD)==SRMMU_ET_PTDBAD) ||
pmd118arch/sparc/mm/srmmu.c(srmmu_pmd_page(pmd) > high_memory);
pmd121arch/sparc/mm/srmmu.cint srmmu_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & SRMMU_ET_PTD; }
pmd271arch/sparc/mm/srmmu.csrmmu_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd276arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd278arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd280arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd284arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd289arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd290arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd291arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd294arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd299arch/sparc/mm/srmmu.csrmmu_pmd_free_kernel(pmd_t *pmd)
pmd301arch/sparc/mm/srmmu.cmem_map[MAP_NR(pmd)] = 1;
pmd302arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd339arch/sparc/mm/srmmu.csrmmu_pte_alloc(pmd_t * pmd, unsigned long address)
pmd344arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd346arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd348arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd352arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd357arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd358arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd359arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd362arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd370arch/sparc/mm/srmmu.csrmmu_pmd_free(pmd_t * pmd)
pmd372arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd95arch/sparc/mm/sun4c.csun4c_pmd_page(pmd_t pmd)
pmd97arch/sparc/mm/sun4c.creturn ((pmd_val(pmd) & _SUN4C_PGD_PFN_MASK) << (_SUN4C_PGD_PAGE_SHIFT));
pmd127arch/sparc/mm/sun4c.cint sun4c_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd128arch/sparc/mm/sun4c.cint sun4c_pmd_bad(pmd_t pmd)
pmd130arch/sparc/mm/sun4c.creturn ((pmd_val(pmd) & _SUN4C_PGD_MMU_MASK) != _SUN4C_PAGE_TABLE);
pmd133arch/sparc/mm/sun4c.cint sun4c_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _SUN4C_PAGE_VALID; }
pmd248arch/sparc/mm/sun4c.csun4c_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd254arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd257arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd259arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, page);
pmd263arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd268arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd269arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd270arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd274arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd282arch/sparc/mm/sun4c.csun4c_pmd_free_kernel(pmd_t *pmd)
pmd300arch/sparc/mm/sun4c.csun4c_pte_alloc(pmd_t * pmd, unsigned long address)
pmd305arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd307arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd309arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, page);
pmd312arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd317arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd318arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd319arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd324arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd332arch/sparc/mm/sun4c.csun4c_pmd_free(pmd_t * pmd)
pmd531fs/proc/array.cstatic inline void statm_pte_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd537fs/proc/array.cif (pmd_none(*pmd))
pmd539fs/proc/array.cif (pmd_bad(*pmd)) {
pmd540fs/proc/array.cprintk("statm_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd541fs/proc/array.cpmd_clear(pmd);
pmd544fs/proc/array.cpte = pte_offset(pmd, address);
pmd572fs/proc/array.cpmd_t * pmd;
pmd582fs/proc/array.cpmd = pmd_offset(pgd, address);
pmd588fs/proc/array.cstatm_pte_range(pmd, address, end - address, pages, shared, dirty, total);
pmd590fs/proc/array.cpmd++;
pmd18include/asm-alpha/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-alpha/page.h#define pmd_val(x)  ((x).pmd)
pmd166include/asm-alpha/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd167include/asm-alpha/pgtable.h{ return PAGE_OFFSET + ((pmd_val(pmd) & _PFN_MASK) >> (32-PAGE_SHIFT)); }
pmd182include/asm-alpha/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd183include/asm-alpha/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~_PFN_MASK) != _PAGE_TABLE || pmd_page(pmd) > high_memory; }
pmd184include/asm-alpha/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_VALID; }
pmd275include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd278include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd280include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd282include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd286include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd291include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd292include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd293include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd296include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd299include/asm-alpha/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd301include/asm-alpha/pgtable.hmem_map[MAP_NR(pmd)].reserved = 0;
pmd302include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd334include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc(pmd_t *pmd, unsigned long address)
pmd337include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd339include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd341include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd344include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd349include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd350include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd351include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd354include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd357include/asm-alpha/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd359include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd18include/asm-i386/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-i386/page.h#define pmd_val(x)  ((x).pmd)
pmd154include/asm-i386/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd155include/asm-i386/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~PAGE_MASK) != _PAGE_TABLE || pmd_val(pmd) > high_memory; }
pmd156include/asm-i386/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_PRESENT; }
pmd213include/asm-i386/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd214include/asm-i386/pgtable.h{ return pmd_val(pmd) & PAGE_MASK; }
pmd245include/asm-i386/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t * pmd, unsigned long address)
pmd248include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd250include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd252include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd256include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd261include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd262include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd263include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd266include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd273include/asm-i386/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd275include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd288include/asm-i386/pgtable.hextern inline pte_t * pte_alloc(pmd_t * pmd, unsigned long address)
pmd291include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd293include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd295include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd298include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd303include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd304include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd305include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd308include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd315include/asm-i386/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd317include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd48include/asm-sparc/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd53include/asm-sparc/page.h#define pmd_val(x)      ((x).pmd)
pmd204mm/filemap.cstatic inline int filemap_sync_pte_range(pmd_t * pmd,
pmd212mm/filemap.cif (pmd_none(*pmd))
pmd214mm/filemap.cif (pmd_bad(*pmd)) {
pmd215mm/filemap.cprintk("filemap_sync_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd216mm/filemap.cpmd_clear(pmd);
pmd219mm/filemap.cpte = pte_offset(pmd, address);
pmd238mm/filemap.cpmd_t * pmd;
pmd249mm/filemap.cpmd = pmd_offset(pgd, address);
pmd257mm/filemap.cerror |= filemap_sync_pte_range(pmd, address, end - address, vma, offset, flags);
pmd259mm/filemap.cpmd++;
pmd102mm/memory.cpmd_t * pmd;
pmd111mm/memory.cpmd = pmd_offset(dir, 0);
pmd113mm/memory.cif (!pmd_inuse(pmd)) {
pmd116mm/memory.cfree_one_pmd(pmd+j);
pmd118mm/memory.cpmd_free(pmd);
pmd308mm/memory.cstatic inline void zap_pte_range(pmd_t * pmd, unsigned long address, unsigned long size)
pmd313mm/memory.cif (pmd_none(*pmd))
pmd315mm/memory.cif (pmd_bad(*pmd)) {
pmd316mm/memory.cprintk("zap_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd317mm/memory.cpmd_clear(pmd);
pmd320mm/memory.cpte = pte_offset(pmd, address);
pmd336mm/memory.cpmd_t * pmd;
pmd346mm/memory.cpmd = pmd_offset(dir, address);
pmd352mm/memory.czap_pte_range(pmd, address, end - address);
pmd354mm/memory.cpmd++;
pmd393mm/memory.cstatic inline int zeromap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size, pte_t zero_pte)
pmd402mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd407mm/memory.cpmd++;
pmd422mm/memory.cpmd_t *pmd = pmd_alloc(dir, address);
pmd424mm/memory.cif (!pmd)
pmd426mm/memory.cerror = zeromap_pmd_range(pmd, address, end - address, zero_pte);
pmd462mm/memory.cstatic inline int remap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd473mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd478mm/memory.cpmd++;
pmd492mm/memory.cpmd_t *pmd = pmd_alloc(dir, from);
pmd494mm/memory.cif (!pmd)
pmd496mm/memory.cerror = remap_pmd_range(pmd, from, end - from, offset + from, prot);
pmd527mm/memory.cpmd_t * pmd;
pmd535mm/memory.cpmd = pmd_alloc(pgd, address);
pmd536mm/memory.cif (!pmd) {
pmd541mm/memory.cpte = pte_alloc(pmd, address);
pmd1016mm/memory.cpmd_t *pmd;
pmd1020mm/memory.cpmd = pmd_alloc(pgd, address);
pmd1021mm/memory.cif (!pmd) {
pmd1025mm/memory.cpte = pte_alloc(pmd, address);
pmd1162mm/memory.cpmd_t *pmd;
pmd1166mm/memory.cpmd = pmd_alloc(pgd, address);
pmd1167mm/memory.cif (!pmd)
pmd1169mm/memory.cpte = pte_alloc(pmd, address);
pmd20mm/mprotect.cstatic inline void change_pte_range(pmd_t * pmd, unsigned long address,
pmd26mm/mprotect.cif (pmd_none(*pmd))
pmd28mm/mprotect.cif (pmd_bad(*pmd)) {
pmd29mm/mprotect.cprintk("change_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd30mm/mprotect.cpmd_clear(pmd);
pmd33mm/mprotect.cpte = pte_offset(pmd, address);
pmd50mm/mprotect.cpmd_t * pmd;
pmd60mm/mprotect.cpmd = pmd_offset(pgd, address);
pmd66mm/mprotect.cchange_pte_range(pmd, address, end - address, newprot);
pmd68mm/mprotect.cpmd++;
pmd536mm/swap.cpmd_t * pmd;
pmd547mm/swap.cpmd = pmd_offset(dir, address);
pmd554mm/swap.cint result = swap_out_pmd(tsk, vma, pmd, address, end, limit);
pmd558mm/swap.cpmd++;
pmd970mm/swap.cpmd_t * pmd;
pmd980mm/swap.cpmd = pmd_offset(dir, address);
pmd987mm/swap.cif (unuse_pmd(vma, pmd, address, end - address, offset, type, page))
pmd990mm/swap.cpmd++;
pmd41mm/vmalloc.cstatic inline void free_area_pte(pmd_t * pmd, unsigned long address, unsigned long size)
pmd46mm/vmalloc.cif (pmd_none(*pmd))
pmd48mm/vmalloc.cif (pmd_bad(*pmd)) {
pmd49mm/vmalloc.cprintk("free_area_pte: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd50mm/vmalloc.cpmd_clear(pmd);
pmd53mm/vmalloc.cpte = pte_offset(pmd, address);
pmd75mm/vmalloc.cpmd_t * pmd;
pmd85mm/vmalloc.cpmd = pmd_offset(dir, address);
pmd91mm/vmalloc.cfree_area_pte(pmd, address, end - address);
pmd93mm/vmalloc.cpmd++;
pmd133mm/vmalloc.cstatic inline int alloc_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size)
pmd142mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd148mm/vmalloc.cpmd++;
pmd160mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd161mm/vmalloc.cif (!pmd)
pmd163mm/vmalloc.cif (alloc_area_pmd(pmd, address, end - address))
pmd192mm/vmalloc.cstatic inline int remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size,
pmd203mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd208mm/vmalloc.cpmd++;
pmd221mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd222mm/vmalloc.cif (!pmd)
pmd224mm/vmalloc.cif (remap_area_pmd(pmd, address, end - address, offset + address))