taglinefilesource code
pmd73arch/i386/kernel/vm86.cpmd_t *pmd;
pmd85arch/i386/kernel/vm86.cpmd = pmd_offset(pgd, 0xA0000);
pmd86arch/i386/kernel/vm86.cif (pmd_none(*pmd))
pmd88arch/i386/kernel/vm86.cif (pmd_bad(*pmd)) {
pmd89arch/i386/kernel/vm86.cprintk("vm86: bad pmd entry [%p]:%08lx\n", pmd, pmd_val(*pmd));
pmd90arch/i386/kernel/vm86.cpmd_clear(pmd);
pmd93arch/i386/kernel/vm86.cpte = pte_offset(pmd, 0xA0000);
pmd77arch/sparc/mm/srmmu.csrmmu_pmd_page(pmd_t pmd)
pmd81arch/sparc/mm/srmmu.cpage = (pmd_val(pmd) & (SRMMU_PTD_PTP_MASK)) << SRMMU_PTD_PTP_PADDR_SHIFT;
pmd114arch/sparc/mm/srmmu.cint srmmu_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd115arch/sparc/mm/srmmu.cint srmmu_pmd_bad(pmd_t pmd)
pmd117arch/sparc/mm/srmmu.creturn ((pmd_val(pmd)&SRMMU_ET_PTDBAD)==SRMMU_ET_PTDBAD) ||
pmd118arch/sparc/mm/srmmu.c(srmmu_pmd_page(pmd) > high_memory);
pmd121arch/sparc/mm/srmmu.cint srmmu_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & SRMMU_ET_PTD; }
pmd271arch/sparc/mm/srmmu.csrmmu_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd276arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd278arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd280arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd284arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd289arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd290arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd291arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd294arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd299arch/sparc/mm/srmmu.csrmmu_pmd_free_kernel(pmd_t *pmd)
pmd301arch/sparc/mm/srmmu.cmem_map[MAP_NR(pmd)] = 1;
pmd302arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd339arch/sparc/mm/srmmu.csrmmu_pte_alloc(pmd_t * pmd, unsigned long address)
pmd344arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd346arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd348arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd352arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd357arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd358arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd359arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd362arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd370arch/sparc/mm/srmmu.csrmmu_pmd_free(pmd_t * pmd)
pmd372arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd95arch/sparc/mm/sun4c.csun4c_pmd_page(pmd_t pmd)
pmd97arch/sparc/mm/sun4c.creturn ((pmd_val(pmd) & _SUN4C_PGD_PFN_MASK) << (_SUN4C_PGD_PAGE_SHIFT));
pmd127arch/sparc/mm/sun4c.cint sun4c_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd128arch/sparc/mm/sun4c.cint sun4c_pmd_bad(pmd_t pmd)
pmd130arch/sparc/mm/sun4c.creturn ((pmd_val(pmd) & _SUN4C_PGD_MMU_MASK) != _SUN4C_PAGE_TABLE);
pmd133arch/sparc/mm/sun4c.cint sun4c_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _SUN4C_PAGE_VALID; }
pmd248arch/sparc/mm/sun4c.csun4c_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd254arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd257arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd259arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, page);
pmd263arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd268arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd269arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd270arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd274arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd282arch/sparc/mm/sun4c.csun4c_pmd_free_kernel(pmd_t *pmd)
pmd300arch/sparc/mm/sun4c.csun4c_pte_alloc(pmd_t * pmd, unsigned long address)
pmd305arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd307arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd309arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, page);
pmd312arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd317arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd318arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd319arch/sparc/mm/sun4c.csun4c_pmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd324arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd332arch/sparc/mm/sun4c.csun4c_pmd_free(pmd_t * pmd)
pmd482fs/proc/array.cstatic inline void statm_pte_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd488fs/proc/array.cif (pmd_none(*pmd))
pmd490fs/proc/array.cif (pmd_bad(*pmd)) {
pmd491fs/proc/array.cprintk("statm_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd492fs/proc/array.cpmd_clear(pmd);
pmd495fs/proc/array.cpte = pte_offset(pmd, address);
pmd523fs/proc/array.cpmd_t * pmd;
pmd533fs/proc/array.cpmd = pmd_offset(pgd, address);
pmd539fs/proc/array.cstatm_pte_range(pmd, address, end - address, pages, shared, dirty, total);
pmd541fs/proc/array.cpmd++;
pmd18include/asm-alpha/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-alpha/page.h#define pmd_val(x)  ((x).pmd)
pmd166include/asm-alpha/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd167include/asm-alpha/pgtable.h{ return PAGE_OFFSET + ((pmd_val(pmd) & _PFN_MASK) >> (32-PAGE_SHIFT)); }
pmd182include/asm-alpha/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd183include/asm-alpha/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~_PFN_MASK) != _PAGE_TABLE || pmd_page(pmd) > high_memory; }
pmd184include/asm-alpha/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_VALID; }
pmd280include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd283include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd285include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd287include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd291include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd296include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd297include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd298include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd301include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd304include/asm-alpha/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd306include/asm-alpha/pgtable.hmem_map[MAP_NR(pmd)] = 1;
pmd307include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd339include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc(pmd_t *pmd, unsigned long address)
pmd342include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd344include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd346include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd349include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd354include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd355include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd356include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd359include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd362include/asm-alpha/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd364include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd18include/asm-i386/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-i386/page.h#define pmd_val(x)  ((x).pmd)
pmd154include/asm-i386/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd155include/asm-i386/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~PAGE_MASK) != _PAGE_TABLE || pmd_val(pmd) > high_memory; }
pmd156include/asm-i386/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_PRESENT; }
pmd218include/asm-i386/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd219include/asm-i386/pgtable.h{ return pmd_val(pmd) & PAGE_MASK; }
pmd250include/asm-i386/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t * pmd, unsigned long address)
pmd253include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd255include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd257include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd261include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd266include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd267include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd268include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd271include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd278include/asm-i386/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd280include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd293include/asm-i386/pgtable.hextern inline pte_t * pte_alloc(pmd_t * pmd, unsigned long address)
pmd296include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd298include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd300include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd303include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd308include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd309include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd310include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd313include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd320include/asm-i386/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd322include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd48include/asm-sparc/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd53include/asm-sparc/page.h#define pmd_val(x)      ((x).pmd)
pmd197mm/filemap.cstatic inline int filemap_sync_pte_range(pmd_t * pmd,
pmd205mm/filemap.cif (pmd_none(*pmd))
pmd207mm/filemap.cif (pmd_bad(*pmd)) {
pmd208mm/filemap.cprintk("filemap_sync_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd209mm/filemap.cpmd_clear(pmd);
pmd212mm/filemap.cpte = pte_offset(pmd, address);
pmd231mm/filemap.cpmd_t * pmd;
pmd242mm/filemap.cpmd = pmd_offset(pgd, address);
pmd250mm/filemap.cerror |= filemap_sync_pte_range(pmd, address, end - address, vma, offset, flags);
pmd252mm/filemap.cpmd++;
pmd103mm/memory.cpmd_t * pmd;
pmd112mm/memory.cpmd = pmd_offset(dir, 0);
pmd114mm/memory.cif (!pmd_inuse(pmd)) {
pmd117mm/memory.cfree_one_pmd(pmd+j);
pmd119mm/memory.cpmd_free(pmd);
pmd328mm/memory.cstatic inline void unmap_pte_range(pmd_t * pmd, unsigned long address, unsigned long size)
pmd333mm/memory.cif (pmd_none(*pmd))
pmd335mm/memory.cif (pmd_bad(*pmd)) {
pmd336mm/memory.cprintk("unmap_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd337mm/memory.cpmd_clear(pmd);
pmd340mm/memory.cpte = pte_offset(pmd, address);
pmd356mm/memory.cpmd_t * pmd;
pmd366mm/memory.cpmd = pmd_offset(dir, address);
pmd372mm/memory.cunmap_pte_range(pmd, address, end - address);
pmd374mm/memory.cpmd++;
pmd422mm/memory.cstatic inline int zeromap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size, pte_t zero_pte)
pmd431mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd436mm/memory.cpmd++;
pmd451mm/memory.cpmd_t *pmd = pmd_alloc(dir, address);
pmd453mm/memory.cif (!pmd)
pmd455mm/memory.cerror = zeromap_pmd_range(pmd, address, end - address, zero_pte);
pmd491mm/memory.cstatic inline int remap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd502mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd507mm/memory.cpmd++;
pmd521mm/memory.cpmd_t *pmd = pmd_alloc(dir, from);
pmd523mm/memory.cif (!pmd)
pmd525mm/memory.cerror = remap_pmd_range(pmd, from, end - from, offset + from, prot);
pmd556mm/memory.cpmd_t * pmd;
pmd564mm/memory.cpmd = pmd_alloc(pgd, address);
pmd565mm/memory.cif (!pmd) {
pmd570mm/memory.cpte = pte_alloc(pmd, address);
pmd937mm/memory.cpmd_t *pmd;
pmd941mm/memory.cpmd = pmd_alloc(pgd, address);
pmd942mm/memory.cif (!pmd) {
pmd946mm/memory.cpte = pte_alloc(pmd, address);
pmd1083mm/memory.cpmd_t *pmd;
pmd1087mm/memory.cpmd = pmd_alloc(pgd, address);
pmd1088mm/memory.cif (!pmd)
pmd1090mm/memory.cpte = pte_alloc(pmd, address);
pmd20mm/mprotect.cstatic inline void change_pte_range(pmd_t * pmd, unsigned long address,
pmd26mm/mprotect.cif (pmd_none(*pmd))
pmd28mm/mprotect.cif (pmd_bad(*pmd)) {
pmd29mm/mprotect.cprintk("change_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd30mm/mprotect.cpmd_clear(pmd);
pmd33mm/mprotect.cpte = pte_offset(pmd, address);
pmd50mm/mprotect.cpmd_t * pmd;
pmd60mm/mprotect.cpmd = pmd_offset(pgd, address);
pmd66mm/mprotect.cchange_pte_range(pmd, address, end - address, newprot);
pmd68mm/mprotect.cpmd++;
pmd477mm/swap.cpmd_t * pmd;
pmd488mm/swap.cpmd = pmd_offset(dir, address);
pmd495mm/swap.cint result = swap_out_pmd(tsk, vma, pmd, address, end, limit);
pmd499mm/swap.cpmd++;
pmd913mm/swap.cpmd_t * pmd;
pmd923mm/swap.cpmd = pmd_offset(dir, address);
pmd930mm/swap.cif (unuse_pmd(vma, pmd, address, end - address, offset, type, page))
pmd933mm/swap.cpmd++;
pmd38mm/vmalloc.cstatic inline void free_area_pte(pmd_t * pmd, unsigned long address, unsigned long size)
pmd43mm/vmalloc.cif (pmd_none(*pmd))
pmd45mm/vmalloc.cif (pmd_bad(*pmd)) {
pmd46mm/vmalloc.cprintk("free_area_pte: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd47mm/vmalloc.cpmd_clear(pmd);
pmd50mm/vmalloc.cpte = pte_offset(pmd, address);
pmd72mm/vmalloc.cpmd_t * pmd;
pmd82mm/vmalloc.cpmd = pmd_offset(dir, address);
pmd88mm/vmalloc.cfree_area_pte(pmd, address, end - address);
pmd90mm/vmalloc.cpmd++;
pmd130mm/vmalloc.cstatic inline int alloc_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size)
pmd139mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd145mm/vmalloc.cpmd++;
pmd157mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd158mm/vmalloc.cif (!pmd)
pmd160mm/vmalloc.cif (alloc_area_pmd(pmd, address, end - address))
pmd189mm/vmalloc.cstatic inline int remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size,
pmd200mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd205mm/vmalloc.cpmd++;
pmd218mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd219mm/vmalloc.cif (!pmd)
pmd221mm/vmalloc.cif (remap_area_pmd(pmd, address, end - address, offset + address))