[PATCH 12/12] KVM: X86/MMU: Rename struct kvm_mmu_pages to struct kvm_mmu_page_vec

From: Lai Jiangshan
Date: Sun Jun 05 2022 - 02:44:49 EST


From: Lai Jiangshan <jiangshan.ljs@xxxxxxxxxxxx>

It is implemented as a vector and variable names for it are pvec.

Rename it to kvm_mmu_page_vec for better describing it.

Signed-off-by: Lai Jiangshan <jiangshan.ljs@xxxxxxxxxxxx>
---
arch/x86/kvm/mmu/mmu.c | 24 ++++++++++++------------
1 file changed, 12 insertions(+), 12 deletions(-)

diff --git a/arch/x86/kvm/mmu/mmu.c b/arch/x86/kvm/mmu/mmu.c
index 7cfc4bc89f60..64e0d155068c 100644
--- a/arch/x86/kvm/mmu/mmu.c
+++ b/arch/x86/kvm/mmu/mmu.c
@@ -1742,12 +1742,12 @@ static int nonpaging_sync_page(struct kvm_vcpu *vcpu,

#define KVM_PAGE_ARRAY_NR 16

-struct kvm_mmu_pages {
+struct kvm_mmu_page_vec {
struct kvm_mmu_page *sp[KVM_PAGE_ARRAY_NR];
unsigned int nr;
};

-static int mmu_pages_add(struct kvm_mmu_pages *pvec, struct kvm_mmu_page *sp)
+static int mmu_pages_add(struct kvm_mmu_page_vec *pvec, struct kvm_mmu_page *sp)
{
int i;

@@ -1768,7 +1768,7 @@ static inline void clear_unsync_child_bit(struct kvm_mmu_page *sp, int idx)
}

static int __mmu_unsync_walk_and_clear(struct kvm_mmu_page *sp,
- struct kvm_mmu_pages *pvec)
+ struct kvm_mmu_page_vec *pvec)
{
int i, ret, nr_unsync_leaf = 0;

@@ -1808,7 +1808,7 @@ static int __mmu_unsync_walk_and_clear(struct kvm_mmu_page *sp,
}

static int mmu_unsync_walk_and_clear(struct kvm_mmu_page *sp,
- struct kvm_mmu_pages *pvec)
+ struct kvm_mmu_page_vec *pvec)
{
pvec->nr = 0;
if (!sp->unsync_children)
@@ -1885,7 +1885,7 @@ static bool is_obsolete_sp(struct kvm *kvm, struct kvm_mmu_page *sp)
unlikely(sp->mmu_valid_gen != kvm->arch.mmu_valid_gen);
}

-#define for_each_sp(pvec, sp, i) \
+#define page_vec_for_each_sp(pvec, sp, i) \
for (i = 0; i < pvec.nr && ({ sp = pvec.sp[i]; 1;}); i++)

static int mmu_sync_children(struct kvm_vcpu *vcpu,
@@ -1893,14 +1893,14 @@ static int mmu_sync_children(struct kvm_vcpu *vcpu,
{
int i;
struct kvm_mmu_page *sp;
- struct kvm_mmu_pages pages;
+ struct kvm_mmu_page_vec pvec;
LIST_HEAD(invalid_list);
bool flush = false;

- while (mmu_unsync_walk_and_clear(parent, &pages)) {
+ while (mmu_unsync_walk_and_clear(parent, &pvec)) {
bool protected = false;

- for_each_sp(pages, sp, i)
+ page_vec_for_each_sp(pvec, sp, i)
protected |= kvm_vcpu_write_protect_gfn(vcpu, sp->gfn);

if (protected) {
@@ -1908,7 +1908,7 @@ static int mmu_sync_children(struct kvm_vcpu *vcpu,
flush = false;
}

- for_each_sp(pages, sp, i) {
+ page_vec_for_each_sp(pvec, sp, i) {
kvm_mmu_page_clear_unsync(vcpu->kvm, sp);
flush |= kvm_sync_page(vcpu, sp, &invalid_list) > 0;
}
@@ -2208,15 +2208,15 @@ static int mmu_zap_unsync_children(struct kvm *kvm,
struct list_head *invalid_list)
{
int i, zapped = 0;
- struct kvm_mmu_pages pages;
+ struct kvm_mmu_page_vec pvec;

if (parent->role.level == PG_LEVEL_4K)
return 0;

- while (mmu_unsync_walk_and_clear(parent, &pages)) {
+ while (mmu_unsync_walk_and_clear(parent, &pvec)) {
struct kvm_mmu_page *sp;

- for_each_sp(pages, sp, i) {
+ page_vec_for_each_sp(pvec, sp, i) {
kvm_mmu_prepare_zap_page(kvm, sp, invalid_list);
zapped++;
}
--
2.19.1.6.gb485710b