On Tue, Jun 16, 2020 at 11:30 PM Michel Lespinasse <wal...@google.com> wrote: > > On Tue, Jun 16, 2020 at 11:07 PM Stafford Horne <sho...@gmail.com> wrote: > > On Wed, Jun 17, 2020 at 02:35:39PM +0900, Stafford Horne wrote: > > > On Tue, Jun 16, 2020 at 01:47:24PM -0700, Michel Lespinasse wrote: > > > > This makes me wonder actually - maybe there is a latent bug that got > > > > exposed after my change added the rwsem_is_locked assertion to the > > > > lockdep_assert_held one. If that is the case, it may be helpful to > > > > bisect when that issue first appeared, by testing before my patchset > > > > with VM_BUG_ON(!rwsem_is_locked(&walk.mm->mmap_lock)) added to > > > > walk_page_range() / walk_page_range_novma() / walk_page_vma() ... > > > > > > Hello, > > > > > > I tried to bisect it, but I think this issue goes much further back. > > > > > > Just with the below patch booting fails all the way back to v5.7. > > > > > > What does this mean by they way, why would mmap_assert_locked() want to > > > assert > > > that the rwsem_is_locked() is not true? > > It's the opposite - VM_BUG_ON(cond) triggers if cond is true, so in > other words it asserts that cond is false. Yeah, I agree it is kinda > confusing. But in our case, it asserts that the rwsem is locked, which > is what we want. > > > The openrisc code that was walking the page ranges was not locking mm. I > > have > > added the below patch to v5.8-rc1 and it seems to work fine. I will send a > > better patch in a bit. > > > > iff --git a/arch/openrisc/kernel/dma.c b/arch/openrisc/kernel/dma.c > > index c152a68811dd..bd5f05dd9174 100644 > > --- a/arch/openrisc/kernel/dma.c > > +++ b/arch/openrisc/kernel/dma.c > > @@ -74,8 +74,10 @@ void *arch_dma_set_uncached(void *cpu_addr, size_t size) > > * We need to iterate through the pages, clearing the dcache for > > * them and setting the cache-inhibit bit. > > */ > > + mmap_read_lock(&init_mm); > > error = walk_page_range(&init_mm, va, va + size, > > &set_nocache_walk_ops, > > NULL); > > + mmap_read_unlock(&init_mm); > > if (error) > > return ERR_PTR(error); > > return cpu_addr; > > @@ -85,9 +87,11 @@ void arch_dma_clear_uncached(void *cpu_addr, size_t size) > > { > > unsigned long va = (unsigned long)cpu_addr; > > > > + mmap_read_lock(&init_mm); > > /* walk_page_range shouldn't be able to fail here */ > > WARN_ON(walk_page_range(&init_mm, va, va + size, > > &clear_nocache_walk_ops, NULL)); > > + mmap_read_unlock(&init_mm); > > } > > Thanks a lot for getting to the bottom of this. I think this is the proper > fix.
A similar patch works for RISC-V as well. Thanks for debugging it. To sum it up, mm should be locked before walk_page_range and walk_page_range_novma. Here is a diff that works for RISC-V. I will send the patch soon. diff --git a/arch/riscv/mm/pageattr.c b/arch/riscv/mm/pageattr.c index ec2c70f84994..289a9a5ea5b5 100644 --- a/arch/riscv/mm/pageattr.c +++ b/arch/riscv/mm/pageattr.c @@ -151,6 +151,7 @@ int set_memory_nx(unsigned long addr, int numpages) int set_direct_map_invalid_noflush(struct page *page) { + int ret; unsigned long start = (unsigned long)page_address(page); unsigned long end = start + PAGE_SIZE; struct pageattr_masks masks = { @@ -158,11 +159,16 @@ int set_direct_map_invalid_noflush(struct page *page) .clear_mask = __pgprot(_PAGE_PRESENT) }; - return walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); + mmap_read_lock(&init_mm); + ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); + mmap_read_unlock(&init_mm); + + return ret; } int set_direct_map_default_noflush(struct page *page) { + int ret; unsigned long start = (unsigned long)page_address(page); unsigned long end = start + PAGE_SIZE; struct pageattr_masks masks = { @@ -170,7 +176,11 @@ int set_direct_map_default_noflush(struct page *page) .clear_mask = __pgprot(0) }; - return walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); + mmap_read_lock(&init_mm); + ret = walk_page_range(&init_mm, start, end, &pageattr_ops, &masks); + mmap_read_unlock(&init_mm); + + return ret; } -- Regards, Atish