On 3/21/21 8:44 PM, Matthew Wilcox wrote:
On Mon, Mar 22, 2021 at 03:51:52AM +0100, Ingo Molnar wrote:
+++ b/mm/huge_memory.c
@@ -1794,7 +1794,7 @@ bool move_huge_pmd(struct vm_area_struct *vma, unsigned long old_addr,
/*
* Returns
* - 0 if PMD could not be locked
- * - 1 if PMD was locked but protections unchange and TLB flush unnecessary
+ * - 1 if PMD was locked but protections unchanged and TLB flush unnecessary
* - HPAGE_PMD_NR is protections changed and TLB flush necessary
s/is/if/
@@ -5306,7 +5306,7 @@ void adjust_range_if_pmd_sharing_possible(struct vm_area_struct *vma,
/*
* vma need span at least one aligned PUD size and the start,end range
- * must at least partialy within it.
+ * must at least partially within it.
* vma needs to span at least one aligned PUD size, and the range
* must be at least partially within in.
/*
* swapon tell device that all the old swap contents can be discarded,
- * to allow the swap device to optimize its wear-levelling.
+ * to allow the swap device to optimize its wear-leveling.
*/
Levelling is british english, leveling is american english. we don't
usually "correct" one into the other.
How about "labelled" (from mm/kasan/shadow.c):
@@ -384,7 +384,7 @@ static int kasan_depopulate_vmalloc_pte(pte_t *ptep, unsigned long addr,
* How does this work?
* -------------------
*
- * We have a region that is page aligned, labelled as A.
+ * We have a region that is page aligned, labeled as A.
* That might not map onto the shadow in a way that is page-aligned:
--
~Randy
Attachment:
signature.asc
Description: PGP signature