Lines Matching defs:vma_prev
116 vmg->prev = vma_prev(vmg->vmi);
662 struct vm_area_struct *vma_prev, *vma_next, *vma;
732 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
737 ASSERT_EQ(merge_new(&vmg), vma_prev);
739 ASSERT_EQ(vma_prev->vm_start, 0);
740 ASSERT_EQ(vma_prev->vm_end, 0x5000);
741 ASSERT_EQ(vma_prev->vm_pgoff, 0);
757 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
762 vmg.prev = vma_prev;
813 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
819 vmg.prev = vma_prev;
839 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
845 vmg.prev = vma_prev;
848 ASSERT_EQ(merge_existing(&vmg), vma_prev);
850 ASSERT_EQ(vma_prev->vm_start, 0);
851 ASSERT_EQ(vma_prev->vm_end, 0x5000);
852 ASSERT_EQ(vma_prev->vm_pgoff, 0);
868 struct vm_area_struct *vma_prev = alloc_and_link_vma(&mm, 0, 0x2000, 0, flags);
896 vma_prev->vm_ops = &vm_ops;
919 struct vm_area_struct *vma, *vma_prev, *vma_next;
1000 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1001 vma_prev->vm_ops = &vm_ops; /* This should have no impact. */
1005 vmg.prev = vma_prev;
1009 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1011 ASSERT_EQ(vma_prev->vm_start, 0);
1012 ASSERT_EQ(vma_prev->vm_end, 0x6000);
1013 ASSERT_EQ(vma_prev->vm_pgoff, 0);
1014 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1018 ASSERT_TRUE(vma_write_started(vma_prev));
1035 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1036 vma_prev->vm_ops = &vm_ops; /* This should have no impact. */
1039 vmg.prev = vma_prev;
1042 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1044 ASSERT_EQ(vma_prev->vm_start, 0);
1045 ASSERT_EQ(vma_prev->vm_end, 0x7000);
1046 ASSERT_EQ(vma_prev->vm_pgoff, 0);
1047 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1048 ASSERT_TRUE(vma_write_started(vma_prev));
1064 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1065 vma_prev->vm_ops = &vm_ops; /* This should have no impact. */
1069 vmg.prev = vma_prev;
1072 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1074 ASSERT_EQ(vma_prev->vm_start, 0);
1075 ASSERT_EQ(vma_prev->vm_end, 0x9000);
1076 ASSERT_EQ(vma_prev->vm_pgoff, 0);
1077 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1078 ASSERT_TRUE(vma_write_started(vma_prev));
1099 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1149 struct vm_area_struct *vma, *vma_prev, *vma_next;
1173 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1184 INIT_LIST_HEAD(&vma_prev->anon_vma_chain);
1185 list_add(&dummy_anon_vma_chain1.same_vma, &vma_prev->anon_vma_chain);
1186 ASSERT_TRUE(list_is_singular(&vma_prev->anon_vma_chain));
1187 vma_prev->anon_vma = &dummy_anon_vma;
1188 ASSERT_TRUE(is_mergeable_anon_vma(NULL, vma_prev->anon_vma, vma_prev));
1196 ASSERT_FALSE(is_mergeable_anon_vma(vma_prev->anon_vma, vma_next->anon_vma, NULL));
1199 vmg.prev = vma_prev;
1202 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1204 ASSERT_EQ(vma_prev->vm_start, 0);
1205 ASSERT_EQ(vma_prev->vm_end, 0x7000);
1206 ASSERT_EQ(vma_prev->vm_pgoff, 0);
1207 ASSERT_TRUE(vma_write_started(vma_prev));
1224 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1227 INIT_LIST_HEAD(&vma_prev->anon_vma_chain);
1228 list_add(&dummy_anon_vma_chain1.same_vma, &vma_prev->anon_vma_chain);
1229 vma_prev->anon_vma = (struct anon_vma *)1;
1236 vmg.prev = vma_prev;
1238 ASSERT_EQ(merge_new(&vmg), vma_prev);
1240 ASSERT_EQ(vma_prev->vm_start, 0);
1241 ASSERT_EQ(vma_prev->vm_end, 0x7000);
1242 ASSERT_EQ(vma_prev->vm_pgoff, 0);
1243 ASSERT_TRUE(vma_write_started(vma_prev));
1264 struct vm_area_struct *vma_prev, *vma_next, *vma;
1275 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1280 vmg.vma = vma_prev;
1286 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1287 ASSERT_TRUE(vma_prev->anon_vma->was_cloned);
1301 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1311 vmg.prev = vma_prev;
1314 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1317 ASSERT_EQ(vma_prev->vm_start, 0);
1318 ASSERT_EQ(vma_prev->vm_end, 0x8000);
1320 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1321 ASSERT_TRUE(vma_prev->anon_vma->was_cloned);
1334 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1340 vmg.prev = vma_prev;
1343 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1346 ASSERT_EQ(vma_prev->vm_start, 0);
1347 ASSERT_EQ(vma_prev->vm_end, 0x8000);
1349 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1350 ASSERT_TRUE(vma_prev->anon_vma->was_cloned);
1363 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1368 vmg.prev = vma_prev;
1371 ASSERT_EQ(merge_existing(&vmg), vma_prev);
1374 ASSERT_EQ(vma_prev->vm_start, 0);
1375 ASSERT_EQ(vma_prev->vm_end, 0x5000);
1377 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1378 ASSERT_TRUE(vma_prev->anon_vma->was_cloned);
1421 struct vm_area_struct *vma_prev, *vma;
1429 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1434 vmg.prev = vma_prev;
1443 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1456 vma_prev = alloc_and_link_vma(&mm, 0, 0x3000, 0, flags);
1461 vmg.vma = vma_prev;
1468 ASSERT_EQ(vma_prev->anon_vma, &dummy_anon_vma);
1543 struct vm_area_struct *vma_prev, *vma;
1559 vma_prev = alloc_and_link_vma(&mm, 0x3000, 0x5000, 3, flags);
1560 vmg.prev = vma_prev;
1565 ASSERT_EQ(vma, vma_prev);