taglinefilesource code
pmd73arch/i386/kernel/vm86.cpmd_t *pmd;
pmd85arch/i386/kernel/vm86.cpmd = pmd_offset(pgd, 0xA0000);
pmd86arch/i386/kernel/vm86.cif (pmd_none(*pmd))
pmd88arch/i386/kernel/vm86.cif (pmd_bad(*pmd)) {
pmd89arch/i386/kernel/vm86.cprintk("vm86: bad pmd entry [%p]:%08lx\n", pmd, pmd_val(*pmd));
pmd90arch/i386/kernel/vm86.cpmd_clear(pmd);
pmd93arch/i386/kernel/vm86.cpte = pte_offset(pmd, 0xA0000);
pmd34arch/ppc/mm/fault.cpmd_t *pmd;
pmd58arch/ppc/mm/fault.cpmd = pmd_offset(dir, regs->dar & PAGE_MASK);
pmd59arch/ppc/mm/fault.cif (pmd && pmd_present(*pmd))
pmd61arch/ppc/mm/fault.cpte = pte_offset(pmd, regs->dar & PAGE_MASK);
pmd83arch/ppc/mm/fault.cpmd_t *pmd;
pmd102arch/ppc/mm/fault.cpmd = pmd_offset(dir, regs->dar & PAGE_MASK);
pmd103arch/ppc/mm/fault.cif (pmd && pmd_present(*pmd))
pmd105arch/ppc/mm/fault.cpte = pte_offset(pmd, regs->dar & PAGE_MASK);
pmd254arch/ppc/mm/fault.cpmd_t *pmd;
pmd259arch/ppc/mm/fault.cpmd = pmd_offset(dir, address & PAGE_MASK);
pmd260arch/ppc/mm/fault.cif (pmd && pmd_present(*pmd))
pmd262arch/ppc/mm/fault.cpte = pte_offset(pmd, address & PAGE_MASK);
pmd31arch/sparc/mm/srmmu.csrmmu_pmd_page(pmd_t pmd)
pmd35arch/sparc/mm/srmmu.cpage = (pmd_val(pmd) & (SRMMU_PTD_PTP_MASK)) << SRMMU_PTD_PTP_PADDR_SHIFT;
pmd67arch/sparc/mm/srmmu.cint srmmu_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd68arch/sparc/mm/srmmu.cint srmmu_pmd_bad(pmd_t pmd)
pmd70arch/sparc/mm/srmmu.creturn ((pmd_val(pmd)&SRMMU_ET_PTDBAD)==SRMMU_ET_PTDBAD) ||
pmd71arch/sparc/mm/srmmu.c(srmmu_pmd_page(pmd) > high_memory);
pmd74arch/sparc/mm/srmmu.cint srmmu_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & SRMMU_ET_PTD; }
pmd223arch/sparc/mm/srmmu.csrmmu_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd228arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd230arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd232arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd236arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd241arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd242arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd243arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd246arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd251arch/sparc/mm/srmmu.csrmmu_pmd_free_kernel(pmd_t *pmd)
pmd253arch/sparc/mm/srmmu.cmem_map[MAP_NR(pmd)].reserved = 0;
pmd254arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd291arch/sparc/mm/srmmu.csrmmu_pte_alloc(pmd_t * pmd, unsigned long address)
pmd296arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd298arch/sparc/mm/srmmu.cif (srmmu_pmd_none(*pmd)) {
pmd300arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, page);
pmd303arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd308arch/sparc/mm/srmmu.cif (srmmu_pmd_bad(*pmd)) {
pmd309arch/sparc/mm/srmmu.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd310arch/sparc/mm/srmmu.csrmmu_pmd_set(pmd, (pte_t *) SRMMU_ET_PTDBAD);
pmd313arch/sparc/mm/srmmu.creturn (pte_t *) srmmu_pmd_page(*pmd) + address;
pmd321arch/sparc/mm/srmmu.csrmmu_pmd_free(pmd_t * pmd)
pmd323arch/sparc/mm/srmmu.cfree_page((unsigned long) pmd);
pmd76arch/sparc/mm/sun4c.cint sun4c_pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd77arch/sparc/mm/sun4c.cint sun4c_pmd_bad(pmd_t pmd)
pmd79arch/sparc/mm/sun4c.creturn (pmd_val(pmd) & ~PAGE_MASK) != PGD_TABLE || pmd_val(pmd) > high_memory;
pmd82arch/sparc/mm/sun4c.cint sun4c_pmd_present(pmd_t pmd)  { return pmd_val(pmd) & PGD_PRESENT; }
pmd136arch/sparc/mm/sun4c.cunsigned long sun4c_pmd_page(pmd_t pmd)
pmd138arch/sparc/mm/sun4c.creturn (pmd_val(pmd) & PAGE_MASK);
pmd381arch/sparc/mm/sun4c.cpte_t *sun4c_pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd384arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd386arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd388arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) page;
pmd392arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) BAD_PAGETABLE;
pmd397arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd398arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc_kernel: %08lx\n", pmd_val(*pmd));
pmd399arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) BAD_PAGETABLE;
pmd402arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd409arch/sparc/mm/sun4c.cvoid sun4c_pmd_free_kernel(pmd_t *pmd)
pmd411arch/sparc/mm/sun4c.cpmd_val(*pmd) = 0;
pmd424arch/sparc/mm/sun4c.cpte_t *sun4c_pte_alloc(pmd_t * pmd, unsigned long address)
pmd427arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd429arch/sparc/mm/sun4c.cif (sun4c_pmd_none(*pmd)) {
pmd431arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) page;
pmd434arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) BAD_PAGETABLE;
pmd439arch/sparc/mm/sun4c.cif (sun4c_pmd_bad(*pmd)) {
pmd440arch/sparc/mm/sun4c.cprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd441arch/sparc/mm/sun4c.cpmd_val(*pmd) = PGD_TABLE | (unsigned long) BAD_PAGETABLE;
pmd444arch/sparc/mm/sun4c.creturn (pte_t *) sun4c_pmd_page(*pmd) + address;
pmd451arch/sparc/mm/sun4c.cvoid sun4c_pmd_free(pmd_t * pmd)
pmd453arch/sparc/mm/sun4c.cpmd_val(*pmd) = 0;
pmd541fs/proc/array.cstatic inline void statm_pte_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd547fs/proc/array.cif (pmd_none(*pmd))
pmd549fs/proc/array.cif (pmd_bad(*pmd)) {
pmd550fs/proc/array.cprintk("statm_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd551fs/proc/array.cpmd_clear(pmd);
pmd554fs/proc/array.cpte = pte_offset(pmd, address);
pmd582fs/proc/array.cpmd_t * pmd;
pmd592fs/proc/array.cpmd = pmd_offset(pgd, address);
pmd598fs/proc/array.cstatm_pte_range(pmd, address, end - address, pages, shared, dirty, total);
pmd600fs/proc/array.cpmd++;
pmd18include/asm-alpha/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-alpha/page.h#define pmd_val(x)  ((x).pmd)
pmd166include/asm-alpha/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd167include/asm-alpha/pgtable.h{ return PAGE_OFFSET + ((pmd_val(pmd) & _PFN_MASK) >> (32-PAGE_SHIFT)); }
pmd182include/asm-alpha/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd183include/asm-alpha/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~_PFN_MASK) != _PAGE_TABLE || pmd_page(pmd) > high_memory; }
pmd184include/asm-alpha/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_VALID; }
pmd275include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t *pmd, unsigned long address)
pmd278include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd280include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd282include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd286include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd291include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd292include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd293include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd296include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd299include/asm-alpha/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd301include/asm-alpha/pgtable.hmem_map[MAP_NR(pmd)].reserved = 0;
pmd302include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd334include/asm-alpha/pgtable.hextern inline pte_t * pte_alloc(pmd_t *pmd, unsigned long address)
pmd337include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd339include/asm-alpha/pgtable.hif (pmd_none(*pmd)) {
pmd341include/asm-alpha/pgtable.hpmd_set(pmd, page);
pmd344include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd349include/asm-alpha/pgtable.hif (pmd_bad(*pmd)) {
pmd350include/asm-alpha/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd351include/asm-alpha/pgtable.hpmd_set(pmd, (pte_t *) BAD_PAGETABLE);
pmd354include/asm-alpha/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd357include/asm-alpha/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd359include/asm-alpha/pgtable.hfree_page((unsigned long) pmd);
pmd18include/asm-i386/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd23include/asm-i386/page.h#define pmd_val(x)  ((x).pmd)
pmd154include/asm-i386/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd155include/asm-i386/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~PAGE_MASK) != _PAGE_TABLE || pmd_val(pmd) > high_memory; }
pmd156include/asm-i386/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_PRESENT; }
pmd213include/asm-i386/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd214include/asm-i386/pgtable.h{ return pmd_val(pmd) & PAGE_MASK; }
pmd245include/asm-i386/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t * pmd, unsigned long address)
pmd248include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd250include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd252include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd256include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd261include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd262include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd263include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd266include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd273include/asm-i386/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd275include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd288include/asm-i386/pgtable.hextern inline pte_t * pte_alloc(pmd_t * pmd, unsigned long address)
pmd291include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd293include/asm-i386/pgtable.hif (pmd_none(*pmd)) {
pmd295include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd298include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd303include/asm-i386/pgtable.hif (pmd_bad(*pmd)) {
pmd304include/asm-i386/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd305include/asm-i386/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd308include/asm-i386/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd315include/asm-i386/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd317include/asm-i386/pgtable.hpmd_val(*pmd) = 0;
pmd20include/asm-ppc/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd25include/asm-ppc/page.h#define pmd_val(x)  ((x).pmd)
pmd216include/asm-ppc/pgtable.hextern inline int pmd_none(pmd_t pmd)    { return !pmd_val(pmd); }
pmd217include/asm-ppc/pgtable.hextern inline int pmd_bad(pmd_t pmd)    { return (pmd_val(pmd) & ~PAGE_MASK) != _PAGE_TABLE; }
pmd218include/asm-ppc/pgtable.hextern inline int pmd_present(pmd_t pmd)  { return pmd_val(pmd) & _PAGE_PRESENT; }
pmd283include/asm-ppc/pgtable.hextern inline unsigned long pmd_page(pmd_t pmd)
pmd284include/asm-ppc/pgtable.h{ return pmd_val(pmd) & PAGE_MASK; }
pmd385include/asm-ppc/pgtable.hextern inline pte_t * pte_alloc_kernel(pmd_t * pmd, unsigned long address)
pmd388include/asm-ppc/pgtable.hif (pmd_none(*pmd)) {
pmd390include/asm-ppc/pgtable.hif (pmd_none(*pmd)) {
pmd393include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd398include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd403include/asm-ppc/pgtable.hif (pmd_bad(*pmd)) {
pmd404include/asm-ppc/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd406include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd409include/asm-ppc/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd416include/asm-ppc/pgtable.hextern inline void pmd_free_kernel(pmd_t * pmd)
pmd430include/asm-ppc/pgtable.hextern inline pte_t * pte_alloc(pmd_t * pmd, unsigned long address)
pmd433include/asm-ppc/pgtable.hif (pmd_none(*pmd)) {
pmd435include/asm-ppc/pgtable.hif (pmd_none(*pmd)) {
pmd437include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) page;
pmd440include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd445include/asm-ppc/pgtable.hif (pmd_bad(*pmd)) {
pmd446include/asm-ppc/pgtable.hprintk("Bad pmd in pte_alloc: %08lx\n", pmd_val(*pmd));
pmd447include/asm-ppc/pgtable.hpmd_val(*pmd) = _PAGE_TABLE | (unsigned long) BAD_PAGETABLE;
pmd450include/asm-ppc/pgtable.hreturn (pte_t *) pmd_page(*pmd) + address;
pmd457include/asm-ppc/pgtable.hextern inline void pmd_free(pmd_t * pmd)
pmd51include/asm-sparc/page.htypedef struct { unsigned long pmd; } pmd_t;
pmd56include/asm-sparc/page.h#define pmd_val(x)      ((x).pmd)
pmd204mm/filemap.cstatic inline int filemap_sync_pte_range(pmd_t * pmd,
pmd212mm/filemap.cif (pmd_none(*pmd))
pmd214mm/filemap.cif (pmd_bad(*pmd)) {
pmd215mm/filemap.cprintk("filemap_sync_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd216mm/filemap.cpmd_clear(pmd);
pmd219mm/filemap.cpte = pte_offset(pmd, address);
pmd238mm/filemap.cpmd_t * pmd;
pmd249mm/filemap.cpmd = pmd_offset(pgd, address);
pmd257mm/filemap.cerror |= filemap_sync_pte_range(pmd, address, end - address, vma, offset, flags);
pmd259mm/filemap.cpmd++;
pmd102mm/memory.cpmd_t * pmd;
pmd111mm/memory.cpmd = pmd_offset(dir, 0);
pmd113mm/memory.cif (!pmd_inuse(pmd)) {
pmd116mm/memory.cfree_one_pmd(pmd+j);
pmd118mm/memory.cpmd_free(pmd);
pmd308mm/memory.cstatic inline void zap_pte_range(pmd_t * pmd, unsigned long address, unsigned long size)
pmd313mm/memory.cif (pmd_none(*pmd))
pmd315mm/memory.cif (pmd_bad(*pmd)) {
pmd316mm/memory.cprintk("zap_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd317mm/memory.cpmd_clear(pmd);
pmd320mm/memory.cpte = pte_offset(pmd, address);
pmd336mm/memory.cpmd_t * pmd;
pmd346mm/memory.cpmd = pmd_offset(dir, address);
pmd352mm/memory.czap_pte_range(pmd, address, end - address);
pmd354mm/memory.cpmd++;
pmd393mm/memory.cstatic inline int zeromap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size, pte_t zero_pte)
pmd402mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd407mm/memory.cpmd++;
pmd422mm/memory.cpmd_t *pmd = pmd_alloc(dir, address);
pmd424mm/memory.cif (!pmd)
pmd426mm/memory.cerror = zeromap_pmd_range(pmd, address, end - address, zero_pte);
pmd462mm/memory.cstatic inline int remap_pmd_range(pmd_t * pmd, unsigned long address, unsigned long size,
pmd473mm/memory.cpte_t * pte = pte_alloc(pmd, address);
pmd478mm/memory.cpmd++;
pmd492mm/memory.cpmd_t *pmd = pmd_alloc(dir, from);
pmd494mm/memory.cif (!pmd)
pmd496mm/memory.cerror = remap_pmd_range(pmd, from, end - from, offset + from, prot);
pmd527mm/memory.cpmd_t * pmd;
pmd535mm/memory.cpmd = pmd_alloc(pgd, address);
pmd536mm/memory.cif (!pmd) {
pmd541mm/memory.cpte = pte_alloc(pmd, address);
pmd1016mm/memory.cpmd_t *pmd;
pmd1020mm/memory.cpmd = pmd_alloc(pgd, address);
pmd1021mm/memory.cif (!pmd) {
pmd1025mm/memory.cpte = pte_alloc(pmd, address);
pmd1162mm/memory.cpmd_t *pmd;
pmd1166mm/memory.cpmd = pmd_alloc(pgd, address);
pmd1167mm/memory.cif (!pmd)
pmd1169mm/memory.cpte = pte_alloc(pmd, address);
pmd20mm/mprotect.cstatic inline void change_pte_range(pmd_t * pmd, unsigned long address,
pmd26mm/mprotect.cif (pmd_none(*pmd))
pmd28mm/mprotect.cif (pmd_bad(*pmd)) {
pmd29mm/mprotect.cprintk("change_pte_range: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd30mm/mprotect.cpmd_clear(pmd);
pmd33mm/mprotect.cpte = pte_offset(pmd, address);
pmd50mm/mprotect.cpmd_t * pmd;
pmd60mm/mprotect.cpmd = pmd_offset(pgd, address);
pmd66mm/mprotect.cchange_pte_range(pmd, address, end - address, newprot);
pmd68mm/mprotect.cpmd++;
pmd536mm/swap.cpmd_t * pmd;
pmd547mm/swap.cpmd = pmd_offset(dir, address);
pmd554mm/swap.cint result = swap_out_pmd(tsk, vma, pmd, address, end, limit);
pmd558mm/swap.cpmd++;
pmd965mm/swap.cpmd_t * pmd;
pmd975mm/swap.cpmd = pmd_offset(dir, address);
pmd982mm/swap.cif (unuse_pmd(vma, pmd, address, end - address, offset, type, page))
pmd985mm/swap.cpmd++;
pmd41mm/vmalloc.cstatic inline void free_area_pte(pmd_t * pmd, unsigned long address, unsigned long size)
pmd46mm/vmalloc.cif (pmd_none(*pmd))
pmd48mm/vmalloc.cif (pmd_bad(*pmd)) {
pmd49mm/vmalloc.cprintk("free_area_pte: bad pmd (%08lx)\n", pmd_val(*pmd));
pmd50mm/vmalloc.cpmd_clear(pmd);
pmd53mm/vmalloc.cpte = pte_offset(pmd, address);
pmd75mm/vmalloc.cpmd_t * pmd;
pmd85mm/vmalloc.cpmd = pmd_offset(dir, address);
pmd91mm/vmalloc.cfree_area_pte(pmd, address, end - address);
pmd93mm/vmalloc.cpmd++;
pmd133mm/vmalloc.cstatic inline int alloc_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size)
pmd142mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd148mm/vmalloc.cpmd++;
pmd160mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd161mm/vmalloc.cif (!pmd)
pmd163mm/vmalloc.cif (alloc_area_pmd(pmd, address, end - address))
pmd192mm/vmalloc.cstatic inline int remap_area_pmd(pmd_t * pmd, unsigned long address, unsigned long size,
pmd203mm/vmalloc.cpte_t * pte = pte_alloc_kernel(pmd, address);
pmd208mm/vmalloc.cpmd++;
pmd221mm/vmalloc.cpmd_t *pmd = pmd_alloc_kernel(dir, address);
pmd222mm/vmalloc.cif (!pmd)
pmd224mm/vmalloc.cif (remap_area_pmd(pmd, address, end - address, offset + address))