forked from mirrors/linux
		
	1. Move page table debugging from mm.py to pgtable.py. 2. Add aarch64 kernel config and memory constants value. 3. Add below aarch64 page operation helper commands. page_to_pfn, page_to_phys, pfn_to_page, page_address, virt_to_phys, sym_to_pfn, pfn_to_kaddr, virt_to_page. 4. Only support CONFIG_SPARSEMEM_VMEMMAP=y now. Link: https://lkml.kernel.org/r/20230808083020.22254-5-Kuan-Ying.Lee@mediatek.com Signed-off-by: Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com> Cc: AngeloGioacchino Del Regno <angelogioacchino.delregno@collabora.com> Cc: Chinwen Chang <chinwen.chang@mediatek.com> Cc: Matthias Brugger <matthias.bgg@gmail.com> Cc: Qun-Wei Lin <qun-wei.lin@mediatek.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
		
			
				
	
	
		
			398 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
			
		
		
	
	
			398 lines
		
	
	
	
		
			14 KiB
		
	
	
	
		
			Python
		
	
	
	
	
	
# SPDX-License-Identifier: GPL-2.0
 | 
						|
#
 | 
						|
# Copyright (c) 2023 MediaTek Inc.
 | 
						|
#
 | 
						|
# Authors:
 | 
						|
#  Kuan-Ying Lee <Kuan-Ying.Lee@mediatek.com>
 | 
						|
#
 | 
						|
 | 
						|
import gdb
 | 
						|
import math
 | 
						|
from linux import utils, constants
 | 
						|
 | 
						|
def DIV_ROUND_UP(n,d):
 | 
						|
    return ((n) + (d) - 1) // (d)
 | 
						|
 | 
						|
def test_bit(nr, addr):
 | 
						|
    if addr.dereference() & (0x1 << nr):
 | 
						|
        return True
 | 
						|
    else:
 | 
						|
        return False
 | 
						|
 | 
						|
class page_ops():
 | 
						|
    ops = None
 | 
						|
    def __init__(self):
 | 
						|
        if not constants.LX_CONFIG_SPARSEMEM_VMEMMAP:
 | 
						|
            raise gdb.GdbError('Only support CONFIG_SPARSEMEM_VMEMMAP now')
 | 
						|
        if constants.LX_CONFIG_ARM64 and utils.is_target_arch('aarch64'):
 | 
						|
            self.ops = aarch64_page_ops()
 | 
						|
        else:
 | 
						|
            raise gdb.GdbError('Only support aarch64 now')
 | 
						|
 | 
						|
class aarch64_page_ops():
 | 
						|
    def __init__(self):
 | 
						|
        self.SUBSECTION_SHIFT = 21
 | 
						|
        self.SEBSECTION_SIZE = 1 << self.SUBSECTION_SHIFT
 | 
						|
        self.MODULES_VSIZE = 128 * 1024 * 1024
 | 
						|
 | 
						|
        if constants.LX_CONFIG_ARM64_64K_PAGES:
 | 
						|
            self.SECTION_SIZE_BITS = 29
 | 
						|
        else:
 | 
						|
            self.SECTION_SIZE_BITS = 27
 | 
						|
        self.MAX_PHYSMEM_BITS = constants.LX_CONFIG_ARM64_VA_BITS
 | 
						|
 | 
						|
        self.PAGE_SHIFT = constants.LX_CONFIG_ARM64_PAGE_SHIFT
 | 
						|
        self.PAGE_SIZE = 1 << self.PAGE_SHIFT
 | 
						|
        self.PAGE_MASK = (~(self.PAGE_SIZE - 1)) & ((1 << 64) - 1)
 | 
						|
 | 
						|
        self.VA_BITS = constants.LX_CONFIG_ARM64_VA_BITS
 | 
						|
        if self.VA_BITS > 48:
 | 
						|
            self.VA_BITS_MIN = 48
 | 
						|
            self.vabits_actual = gdb.parse_and_eval('vabits_actual')
 | 
						|
        else:
 | 
						|
            self.VA_BITS_MIN = self.VA_BITS
 | 
						|
            self.vabits_actual = self.VA_BITS
 | 
						|
        self.kimage_voffset = gdb.parse_and_eval('kimage_voffset') & ((1 << 64) - 1)
 | 
						|
 | 
						|
        self.SECTIONS_SHIFT = self.MAX_PHYSMEM_BITS - self.SECTION_SIZE_BITS
 | 
						|
 | 
						|
        if str(constants.LX_CONFIG_ARCH_FORCE_MAX_ORDER).isdigit():
 | 
						|
            self.MAX_ORDER = constants.LX_CONFIG_ARCH_FORCE_MAX_ORDER
 | 
						|
        else:
 | 
						|
            self.MAX_ORDER = 11
 | 
						|
 | 
						|
        self.MAX_ORDER_NR_PAGES = 1 << (self.MAX_ORDER - 1)
 | 
						|
        self.PFN_SECTION_SHIFT = self.SECTION_SIZE_BITS - self.PAGE_SHIFT
 | 
						|
        self.NR_MEM_SECTIONS = 1 << self.SECTIONS_SHIFT
 | 
						|
        self.PAGES_PER_SECTION = 1 << self.PFN_SECTION_SHIFT
 | 
						|
        self.PAGE_SECTION_MASK = (~(self.PAGES_PER_SECTION - 1)) & ((1 << 64) - 1)
 | 
						|
 | 
						|
        if constants.LX_CONFIG_SPARSEMEM_EXTREME:
 | 
						|
            self.SECTIONS_PER_ROOT = self.PAGE_SIZE // gdb.lookup_type("struct mem_section").sizeof
 | 
						|
        else:
 | 
						|
            self.SECTIONS_PER_ROOT = 1
 | 
						|
 | 
						|
        self.NR_SECTION_ROOTS = DIV_ROUND_UP(self.NR_MEM_SECTIONS, self.SECTIONS_PER_ROOT)
 | 
						|
        self.SECTION_ROOT_MASK = self.SECTIONS_PER_ROOT - 1
 | 
						|
        self.SUBSECTION_SHIFT = 21
 | 
						|
        self.SEBSECTION_SIZE = 1 << self.SUBSECTION_SHIFT
 | 
						|
        self.PFN_SUBSECTION_SHIFT = self.SUBSECTION_SHIFT - self.PAGE_SHIFT
 | 
						|
        self.PAGES_PER_SUBSECTION = 1 << self.PFN_SUBSECTION_SHIFT
 | 
						|
 | 
						|
        self.SECTION_HAS_MEM_MAP = 1 << int(gdb.parse_and_eval('SECTION_HAS_MEM_MAP_BIT'))
 | 
						|
        self.SECTION_IS_EARLY = 1 << int(gdb.parse_and_eval('SECTION_IS_EARLY_BIT'))
 | 
						|
 | 
						|
        self.struct_page_size = utils.get_page_type().sizeof
 | 
						|
        self.STRUCT_PAGE_MAX_SHIFT = (int)(math.log(self.struct_page_size, 2))
 | 
						|
 | 
						|
        self.PAGE_OFFSET = self._PAGE_OFFSET(self.VA_BITS)
 | 
						|
        self.MODULES_VADDR = self._PAGE_END(self.VA_BITS_MIN)
 | 
						|
        self.MODULES_END = self.MODULES_VADDR + self.MODULES_VSIZE
 | 
						|
 | 
						|
        self.VMEMMAP_SHIFT = (self.PAGE_SHIFT - self.STRUCT_PAGE_MAX_SHIFT)
 | 
						|
        self.VMEMMAP_SIZE = ((self._PAGE_END(self.VA_BITS_MIN) - self.PAGE_OFFSET) >> self.VMEMMAP_SHIFT)
 | 
						|
        self.VMEMMAP_START = (-(1 << (self.VA_BITS - self.VMEMMAP_SHIFT))) & 0xffffffffffffffff
 | 
						|
        self.VMEMMAP_END = self.VMEMMAP_START + self.VMEMMAP_SIZE
 | 
						|
 | 
						|
        self.VMALLOC_START = self.MODULES_END
 | 
						|
        self.VMALLOC_END = self.VMEMMAP_START - 256 * 1024 * 1024
 | 
						|
 | 
						|
        self.memstart_addr = gdb.parse_and_eval("memstart_addr")
 | 
						|
        self.PHYS_OFFSET = self.memstart_addr
 | 
						|
        self.vmemmap = gdb.Value(self.VMEMMAP_START).cast(utils.get_page_type().pointer()) - (self.memstart_addr >> self.PAGE_SHIFT)
 | 
						|
 | 
						|
        self.KERNEL_START = gdb.parse_and_eval("_text")
 | 
						|
        self.KERNEL_END = gdb.parse_and_eval("_end")
 | 
						|
 | 
						|
        if constants.LX_CONFIG_KASAN_GENERIC or constants.LX_CONFIG_KASAN_SW_TAGS:
 | 
						|
            if constants.LX_CONFIG_KASAN_GENERIC:
 | 
						|
                self.KASAN_SHADOW_SCALE_SHIFT = 3
 | 
						|
            else:
 | 
						|
                self.KASAN_SHADOW_SCALE_SHIFT = 4
 | 
						|
            self.KASAN_SHADOW_OFFSET = constants.LX_CONFIG_KASAN_SHADOW_OFFSET
 | 
						|
            self.KASAN_SHADOW_END = (1 << (64 - self.KASAN_SHADOW_SCALE_SHIFT)) + self.KASAN_SHADOW_OFFSET
 | 
						|
            self.PAGE_END = self.KASAN_SHADOW_END - (1 << (self.vabits_actual - self.KASAN_SHADOW_SCALE_SHIFT))
 | 
						|
        else:
 | 
						|
            self.PAGE_END = self._PAGE_END(self.VA_BITS_MIN)
 | 
						|
 | 
						|
        if constants.LX_CONFIG_NUMA and constants.LX_CONFIG_NODES_SHIFT:
 | 
						|
            self.NODE_SHIFT = constants.LX_CONFIG_NODES_SHIFT
 | 
						|
        else:
 | 
						|
            self.NODE_SHIFT = 0
 | 
						|
 | 
						|
        self.MAX_NUMNODES = 1 << self.NODE_SHIFT
 | 
						|
 | 
						|
    def SECTION_NR_TO_ROOT(self, sec):
 | 
						|
        return sec // self.SECTIONS_PER_ROOT
 | 
						|
 | 
						|
    def __nr_to_section(self, nr):
 | 
						|
        root = self.SECTION_NR_TO_ROOT(nr)
 | 
						|
        mem_section = gdb.parse_and_eval("mem_section")
 | 
						|
        return mem_section[root][nr & self.SECTION_ROOT_MASK]
 | 
						|
 | 
						|
    def pfn_to_section_nr(self, pfn):
 | 
						|
        return pfn >> self.PFN_SECTION_SHIFT
 | 
						|
 | 
						|
    def section_nr_to_pfn(self, sec):
 | 
						|
        return sec << self.PFN_SECTION_SHIFT
 | 
						|
 | 
						|
    def __pfn_to_section(self, pfn):
 | 
						|
        return self.__nr_to_section(self.pfn_to_section_nr(pfn))
 | 
						|
 | 
						|
    def pfn_to_section(self, pfn):
 | 
						|
        return self.__pfn_to_section(pfn)
 | 
						|
 | 
						|
    def subsection_map_index(self, pfn):
 | 
						|
        return (pfn & ~(self.PAGE_SECTION_MASK)) // self.PAGES_PER_SUBSECTION
 | 
						|
 | 
						|
    def pfn_section_valid(self, ms, pfn):
 | 
						|
        if constants.LX_CONFIG_SPARSEMEM_VMEMMAP:
 | 
						|
            idx = self.subsection_map_index(pfn)
 | 
						|
            return test_bit(idx, ms['usage']['subsection_map'])
 | 
						|
        else:
 | 
						|
            return True
 | 
						|
 | 
						|
    def valid_section(self, mem_section):
 | 
						|
        if mem_section != None and (mem_section['section_mem_map'] & self.SECTION_HAS_MEM_MAP):
 | 
						|
            return True
 | 
						|
        return False
 | 
						|
 | 
						|
    def early_section(self, mem_section):
 | 
						|
        if mem_section != None and (mem_section['section_mem_map'] & self.SECTION_IS_EARLY):
 | 
						|
            return True
 | 
						|
        return False
 | 
						|
 | 
						|
    def pfn_valid(self, pfn):
 | 
						|
        ms = None
 | 
						|
        if self.PHYS_PFN(self.PFN_PHYS(pfn)) != pfn:
 | 
						|
            return False
 | 
						|
        if self.pfn_to_section_nr(pfn) >= self.NR_MEM_SECTIONS:
 | 
						|
            return False
 | 
						|
        ms = self.__pfn_to_section(pfn)
 | 
						|
 | 
						|
        if not self.valid_section(ms):
 | 
						|
            return False
 | 
						|
        return self.early_section(ms) or self.pfn_section_valid(ms, pfn)
 | 
						|
 | 
						|
    def _PAGE_OFFSET(self, va):
 | 
						|
        return (-(1 << (va))) & 0xffffffffffffffff
 | 
						|
 | 
						|
    def _PAGE_END(self, va):
 | 
						|
        return (-(1 << (va - 1))) & 0xffffffffffffffff
 | 
						|
 | 
						|
    def kasan_reset_tag(self, addr):
 | 
						|
        if constants.LX_CONFIG_KASAN_SW_TAGS or constants.LX_CONFIG_KASAN_HW_TAGS:
 | 
						|
            return int(addr) | (0xff << 56)
 | 
						|
        else:
 | 
						|
            return addr
 | 
						|
 | 
						|
    def __is_lm_address(self, addr):
 | 
						|
        if (addr - self.PAGE_OFFSET) < (self.PAGE_END - self.PAGE_OFFSET):
 | 
						|
            return True
 | 
						|
        else:
 | 
						|
            return False
 | 
						|
    def __lm_to_phys(self, addr):
 | 
						|
        return addr - self.PAGE_OFFSET + self.PHYS_OFFSET
 | 
						|
 | 
						|
    def __kimg_to_phys(self, addr):
 | 
						|
        return addr - self.kimage_voffset
 | 
						|
 | 
						|
    def __virt_to_phys_nodebug(self, va):
 | 
						|
        untagged_va = self.kasan_reset_tag(va)
 | 
						|
        if self.__is_lm_address(untagged_va):
 | 
						|
            return self.__lm_to_phys(untagged_va)
 | 
						|
        else:
 | 
						|
            return self.__kimg_to_phys(untagged_va)
 | 
						|
 | 
						|
    def __virt_to_phys(self, va):
 | 
						|
        if constants.LX_CONFIG_DEBUG_VIRTUAL:
 | 
						|
            if not self.__is_lm_address(self.kasan_reset_tag(va)):
 | 
						|
                raise gdb.GdbError("Warning: virt_to_phys used for non-linear address: 0x%lx\n" % va)
 | 
						|
        return self.__virt_to_phys_nodebug(va)
 | 
						|
 | 
						|
    def virt_to_phys(self, va):
 | 
						|
        return self.__virt_to_phys(va)
 | 
						|
 | 
						|
    def PFN_PHYS(self, pfn):
 | 
						|
        return pfn << self.PAGE_SHIFT
 | 
						|
 | 
						|
    def PHYS_PFN(self, phys):
 | 
						|
        return phys >> self.PAGE_SHIFT
 | 
						|
 | 
						|
    def __phys_to_virt(self, pa):
 | 
						|
        return (pa - self.PHYS_OFFSET) | self.PAGE_OFFSET
 | 
						|
 | 
						|
    def __phys_to_pfn(self, pa):
 | 
						|
        return self.PHYS_PFN(pa)
 | 
						|
 | 
						|
    def __pfn_to_phys(self, pfn):
 | 
						|
        return self.PFN_PHYS(pfn)
 | 
						|
 | 
						|
    def __pa_symbol_nodebug(self, x):
 | 
						|
        return self.__kimg_to_phys(x)
 | 
						|
 | 
						|
    def __phys_addr_symbol(self, x):
 | 
						|
        if constants.LX_CONFIG_DEBUG_VIRTUAL:
 | 
						|
            if x < self.KERNEL_START or x > self.KERNEL_END:
 | 
						|
                raise gdb.GdbError("0x%x exceed kernel range" % x)
 | 
						|
        return self.__pa_symbol_nodebug(x)
 | 
						|
 | 
						|
    def __pa_symbol(self, x):
 | 
						|
        return self.__phys_addr_symbol(x)
 | 
						|
 | 
						|
    def __va(self, pa):
 | 
						|
        return self.__phys_to_virt(pa)
 | 
						|
 | 
						|
    def pfn_to_kaddr(self, pfn):
 | 
						|
        return self.__va(pfn << self.PAGE_SHIFT)
 | 
						|
 | 
						|
    def virt_to_pfn(self, va):
 | 
						|
        return self.__phys_to_pfn(self.__virt_to_phys(va))
 | 
						|
 | 
						|
    def sym_to_pfn(self, x):
 | 
						|
        return self.__phys_to_pfn(self.__pa_symbol(x))
 | 
						|
 | 
						|
    def page_to_pfn(self, page):
 | 
						|
        return int(page.cast(utils.get_page_type().pointer()) - self.vmemmap.cast(utils.get_page_type().pointer()))
 | 
						|
 | 
						|
    def page_to_phys(self, page):
 | 
						|
        return self.__pfn_to_phys(self.page_to_pfn(page))
 | 
						|
 | 
						|
    def pfn_to_page(self, pfn):
 | 
						|
        return (self.vmemmap + pfn).cast(utils.get_page_type().pointer())
 | 
						|
 | 
						|
    def page_to_virt(self, page):
 | 
						|
        if constants.LX_CONFIG_DEBUG_VIRTUAL:
 | 
						|
            return self.__va(self.page_to_phys(page))
 | 
						|
        else:
 | 
						|
            __idx = int((page.cast(gdb.lookup_type("unsigned long")) - self.VMEMMAP_START).cast(utils.get_ulong_type())) // self.struct_page_size
 | 
						|
            return self.PAGE_OFFSET + (__idx * self.PAGE_SIZE)
 | 
						|
 | 
						|
    def virt_to_page(self, va):
 | 
						|
        if constants.LX_CONFIG_DEBUG_VIRTUAL:
 | 
						|
            return self.pfn_to_page(self.virt_to_pfn(va))
 | 
						|
        else:
 | 
						|
            __idx = int(self.kasan_reset_tag(va) - self.PAGE_OFFSET) // self.PAGE_SIZE
 | 
						|
            addr = self.VMEMMAP_START + (__idx * self.struct_page_size)
 | 
						|
            return gdb.Value(addr).cast(utils.get_page_type().pointer())
 | 
						|
 | 
						|
    def page_address(self, page):
 | 
						|
        return self.page_to_virt(page)
 | 
						|
 | 
						|
    def folio_address(self, folio):
 | 
						|
        return self.page_address(folio['page'].address)
 | 
						|
 | 
						|
class LxPFN2Page(gdb.Command):
 | 
						|
    """PFN to struct page"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxPFN2Page, self).__init__("lx-pfn_to_page", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        pfn = int(argv[0])
 | 
						|
        page = page_ops().ops.pfn_to_page(pfn)
 | 
						|
        gdb.write("pfn_to_page(0x%x) = 0x%x\n" % (pfn, page))
 | 
						|
 | 
						|
LxPFN2Page()
 | 
						|
 | 
						|
class LxPage2PFN(gdb.Command):
 | 
						|
    """struct page to PFN"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxPage2PFN, self).__init__("lx-page_to_pfn", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        struct_page_addr = int(argv[0], 16)
 | 
						|
        page = gdb.Value(struct_page_addr).cast(utils.get_page_type().pointer())
 | 
						|
        pfn = page_ops().ops.page_to_pfn(page)
 | 
						|
        gdb.write("page_to_pfn(0x%x) = 0x%x\n" % (page, pfn))
 | 
						|
 | 
						|
LxPage2PFN()
 | 
						|
 | 
						|
class LxPageAddress(gdb.Command):
 | 
						|
    """struct page to linear mapping address"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxPageAddress, self).__init__("lx-page_address", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        struct_page_addr = int(argv[0], 16)
 | 
						|
        page = gdb.Value(struct_page_addr).cast(utils.get_page_type().pointer())
 | 
						|
        addr = page_ops().ops.page_address(page)
 | 
						|
        gdb.write("page_address(0x%x) = 0x%x\n" % (page, addr))
 | 
						|
 | 
						|
LxPageAddress()
 | 
						|
 | 
						|
class LxPage2Phys(gdb.Command):
 | 
						|
    """struct page to physical address"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxPage2Phys, self).__init__("lx-page_to_phys", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        struct_page_addr = int(argv[0], 16)
 | 
						|
        page = gdb.Value(struct_page_addr).cast(utils.get_page_type().pointer())
 | 
						|
        phys_addr = page_ops().ops.page_to_phys(page)
 | 
						|
        gdb.write("page_to_phys(0x%x) = 0x%x\n" % (page, phys_addr))
 | 
						|
 | 
						|
LxPage2Phys()
 | 
						|
 | 
						|
class LxVirt2Phys(gdb.Command):
 | 
						|
    """virtual address to physical address"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxVirt2Phys, self).__init__("lx-virt_to_phys", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        linear_addr = int(argv[0], 16)
 | 
						|
        phys_addr = page_ops().ops.virt_to_phys(linear_addr)
 | 
						|
        gdb.write("virt_to_phys(0x%x) = 0x%x\n" % (linear_addr, phys_addr))
 | 
						|
 | 
						|
LxVirt2Phys()
 | 
						|
 | 
						|
class LxVirt2Page(gdb.Command):
 | 
						|
    """virtual address to struct page"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxVirt2Page, self).__init__("lx-virt_to_page", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        linear_addr = int(argv[0], 16)
 | 
						|
        page = page_ops().ops.virt_to_page(linear_addr)
 | 
						|
        gdb.write("virt_to_page(0x%x) = 0x%x\n" % (linear_addr, page))
 | 
						|
 | 
						|
LxVirt2Page()
 | 
						|
 | 
						|
class LxSym2PFN(gdb.Command):
 | 
						|
    """symbol address to PFN"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxSym2PFN, self).__init__("lx-sym_to_pfn", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        sym_addr = int(argv[0], 16)
 | 
						|
        pfn = page_ops().ops.sym_to_pfn(sym_addr)
 | 
						|
        gdb.write("sym_to_pfn(0x%x) = %d\n" % (sym_addr, pfn))
 | 
						|
 | 
						|
LxSym2PFN()
 | 
						|
 | 
						|
class LxPFN2Kaddr(gdb.Command):
 | 
						|
    """PFN to kernel address"""
 | 
						|
 | 
						|
    def __init__(self):
 | 
						|
        super(LxPFN2Kaddr, self).__init__("lx-pfn_to_kaddr", gdb.COMMAND_USER)
 | 
						|
 | 
						|
    def invoke(self, arg, from_tty):
 | 
						|
        argv = gdb.string_to_argv(arg)
 | 
						|
        pfn = int(argv[0])
 | 
						|
        kaddr = page_ops().ops.pfn_to_kaddr(pfn)
 | 
						|
        gdb.write("pfn_to_kaddr(%d) = 0x%x\n" % (pfn, kaddr))
 | 
						|
 | 
						|
LxPFN2Kaddr()
 |