LCOV - code coverage report
Current view: top level - arch/um/kernel - tlb.c (source / functions) Hit Total Coverage
Test: coverage.info Lines: 61 249 24.5 %
Date: 2023-08-24 13:40:31 Functions: 5 18 27.8 %

          Line data    Source code
       1             : // SPDX-License-Identifier: GPL-2.0
       2             : /*
       3             :  * Copyright (C) 2000 - 2007 Jeff Dike (jdike@{addtoit,linux.intel}.com)
       4             :  */
       5             : 
       6             : #include <linux/mm.h>
       7             : #include <linux/module.h>
       8             : #include <linux/sched/signal.h>
       9             : 
      10             : #include <asm/tlbflush.h>
      11             : #include <as-layout.h>
      12             : #include <mem_user.h>
      13             : #include <os.h>
      14             : #include <skas.h>
      15             : #include <kern_util.h>
      16             : 
      17             : struct host_vm_change {
      18             :         struct host_vm_op {
      19             :                 enum { NONE, MMAP, MUNMAP, MPROTECT } type;
      20             :                 union {
      21             :                         struct {
      22             :                                 unsigned long addr;
      23             :                                 unsigned long len;
      24             :                                 unsigned int prot;
      25             :                                 int fd;
      26             :                                 __u64 offset;
      27             :                         } mmap;
      28             :                         struct {
      29             :                                 unsigned long addr;
      30             :                                 unsigned long len;
      31             :                         } munmap;
      32             :                         struct {
      33             :                                 unsigned long addr;
      34             :                                 unsigned long len;
      35             :                                 unsigned int prot;
      36             :                         } mprotect;
      37             :                 } u;
      38             :         } ops[1];
      39             :         int userspace;
      40             :         int index;
      41             :         struct mm_struct *mm;
      42             :         void *data;
      43             :         int force;
      44             : };
      45             : 
      46             : #define INIT_HVC(mm, force, userspace) \
      47             :         ((struct host_vm_change) \
      48             :          { .ops         = { { .type = NONE } }, \
      49             :            .mm          = mm, \
      50             :            .data        = NULL, \
      51             :            .userspace   = userspace, \
      52             :            .index       = 0, \
      53             :            .force       = force })
      54             : 
      55             : static void report_enomem(void)
      56             : {
      57           0 :         printk(KERN_ERR "UML ran out of memory on the host side! "
      58             :                         "This can happen due to a memory limitation or "
      59             :                         "vm.max_map_count has been reached.\n");
      60             : }
      61             : 
      62         128 : static int do_ops(struct host_vm_change *hvc, int end,
      63             :                   int finished)
      64             : {
      65             :         struct host_vm_op *op;
      66         128 :         int i, ret = 0;
      67             : 
      68         256 :         for (i = 0; i < end && !ret; i++) {
      69         128 :                 op = &hvc->ops[i];
      70         128 :                 switch (op->type) {
      71             :                 case MMAP:
      72          64 :                         if (hvc->userspace)
      73           0 :                                 ret = map(&hvc->mm->context.id, op->u.mmap.addr,
      74           0 :                                           op->u.mmap.len, op->u.mmap.prot,
      75             :                                           op->u.mmap.fd,
      76             :                                           op->u.mmap.offset, finished,
      77             :                                           &hvc->data);
      78             :                         else
      79          64 :                                 map_memory(op->u.mmap.addr, op->u.mmap.offset,
      80             :                                            op->u.mmap.len, 1, 1, 1);
      81             :                         break;
      82             :                 case MUNMAP:
      83          64 :                         if (hvc->userspace)
      84           0 :                                 ret = unmap(&hvc->mm->context.id,
      85             :                                             op->u.munmap.addr,
      86             :                                             op->u.munmap.len, finished,
      87             :                                             &hvc->data);
      88             :                         else
      89         128 :                                 ret = os_unmap_memory(
      90          64 :                                         (void *) op->u.munmap.addr,
      91          64 :                                                       op->u.munmap.len);
      92             : 
      93             :                         break;
      94             :                 case MPROTECT:
      95           0 :                         if (hvc->userspace)
      96           0 :                                 ret = protect(&hvc->mm->context.id,
      97             :                                               op->u.mprotect.addr,
      98             :                                               op->u.mprotect.len,
      99             :                                               op->u.mprotect.prot,
     100             :                                               finished, &hvc->data);
     101             :                         else
     102           0 :                                 ret = os_protect_memory(
     103           0 :                                         (void *) op->u.mprotect.addr,
     104             :                                                         op->u.mprotect.len,
     105             :                                                         1, 1, 1);
     106             :                         break;
     107             :                 default:
     108           0 :                         printk(KERN_ERR "Unknown op type %d in do_ops\n",
     109             :                                op->type);
     110           0 :                         BUG();
     111             :                         break;
     112             :                 }
     113             :         }
     114             : 
     115         128 :         if (ret == -ENOMEM)
     116             :                 report_enomem();
     117             : 
     118         128 :         return ret;
     119             : }
     120             : 
     121          64 : static int add_mmap(unsigned long virt, unsigned long phys, unsigned long len,
     122             :                     unsigned int prot, struct host_vm_change *hvc)
     123             : {
     124             :         __u64 offset;
     125             :         struct host_vm_op *last;
     126          64 :         int fd = -1, ret = 0;
     127             : 
     128          64 :         if (hvc->userspace)
     129           0 :                 fd = phys_mapping(phys, &offset);
     130             :         else
     131          64 :                 offset = phys;
     132          64 :         if (hvc->index != 0) {
     133          64 :                 last = &hvc->ops[hvc->index - 1];
     134          64 :                 if ((last->type == MMAP) &&
     135           0 :                    (last->u.mmap.addr + last->u.mmap.len == virt) &&
     136           0 :                    (last->u.mmap.prot == prot) && (last->u.mmap.fd == fd) &&
     137           0 :                    (last->u.mmap.offset + last->u.mmap.len == offset)) {
     138           0 :                         last->u.mmap.len += len;
     139           0 :                         return 0;
     140             :                 }
     141             :         }
     142             : 
     143          64 :         if (hvc->index == ARRAY_SIZE(hvc->ops)) {
     144          64 :                 ret = do_ops(hvc, ARRAY_SIZE(hvc->ops), 0);
     145          64 :                 hvc->index = 0;
     146             :         }
     147             : 
     148          64 :         hvc->ops[hvc->index++] = ((struct host_vm_op)
     149             :                                   { .type       = MMAP,
     150             :                                     .u = { .mmap = { .addr      = virt,
     151             :                                                      .len       = len,
     152             :                                                      .prot      = prot,
     153             :                                                      .fd        = fd,
     154             :                                                      .offset    = offset }
     155             :                            } });
     156          64 :         return ret;
     157             : }
     158             : 
     159          64 : static int add_munmap(unsigned long addr, unsigned long len,
     160             :                       struct host_vm_change *hvc)
     161             : {
     162             :         struct host_vm_op *last;
     163          64 :         int ret = 0;
     164             : 
     165          64 :         if (hvc->index != 0) {
     166          48 :                 last = &hvc->ops[hvc->index - 1];
     167          48 :                 if ((last->type == MUNMAP) &&
     168           0 :                    (last->u.munmap.addr + last->u.mmap.len == addr)) {
     169           0 :                         last->u.munmap.len += len;
     170           0 :                         return 0;
     171             :                 }
     172             :         }
     173             : 
     174          64 :         if (hvc->index == ARRAY_SIZE(hvc->ops)) {
     175          48 :                 ret = do_ops(hvc, ARRAY_SIZE(hvc->ops), 0);
     176          48 :                 hvc->index = 0;
     177             :         }
     178             : 
     179          64 :         hvc->ops[hvc->index++] = ((struct host_vm_op)
     180             :                                   { .type       = MUNMAP,
     181             :                                     .u = { .munmap = { .addr    = addr,
     182             :                                                        .len     = len } } });
     183          64 :         return ret;
     184             : }
     185             : 
     186           0 : static int add_mprotect(unsigned long addr, unsigned long len,
     187             :                         unsigned int prot, struct host_vm_change *hvc)
     188             : {
     189             :         struct host_vm_op *last;
     190           0 :         int ret = 0;
     191             : 
     192           0 :         if (hvc->index != 0) {
     193           0 :                 last = &hvc->ops[hvc->index - 1];
     194           0 :                 if ((last->type == MPROTECT) &&
     195           0 :                    (last->u.mprotect.addr + last->u.mprotect.len == addr) &&
     196           0 :                    (last->u.mprotect.prot == prot)) {
     197           0 :                         last->u.mprotect.len += len;
     198           0 :                         return 0;
     199             :                 }
     200             :         }
     201             : 
     202           0 :         if (hvc->index == ARRAY_SIZE(hvc->ops)) {
     203           0 :                 ret = do_ops(hvc, ARRAY_SIZE(hvc->ops), 0);
     204           0 :                 hvc->index = 0;
     205             :         }
     206             : 
     207           0 :         hvc->ops[hvc->index++] = ((struct host_vm_op)
     208             :                                   { .type       = MPROTECT,
     209             :                                     .u = { .mprotect = { .addr  = addr,
     210             :                                                          .len   = len,
     211             :                                                          .prot  = prot } } });
     212           0 :         return ret;
     213             : }
     214             : 
     215             : #define ADD_ROUND(n, inc) (((n) + (inc)) & ~((inc) - 1))
     216             : 
     217           0 : static inline int update_pte_range(pmd_t *pmd, unsigned long addr,
     218             :                                    unsigned long end,
     219             :                                    struct host_vm_change *hvc)
     220             : {
     221             :         pte_t *pte;
     222           0 :         int r, w, x, prot, ret = 0;
     223             : 
     224           0 :         pte = pte_offset_kernel(pmd, addr);
     225             :         do {
     226           0 :                 r = pte_read(*pte);
     227           0 :                 w = pte_write(*pte);
     228           0 :                 x = pte_exec(*pte);
     229           0 :                 if (!pte_young(*pte)) {
     230             :                         r = 0;
     231             :                         w = 0;
     232           0 :                 } else if (!pte_dirty(*pte))
     233           0 :                         w = 0;
     234             : 
     235           0 :                 prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) |
     236           0 :                         (x ? UM_PROT_EXEC : 0));
     237           0 :                 if (hvc->force || pte_newpage(*pte)) {
     238           0 :                         if (pte_present(*pte)) {
     239           0 :                                 if (pte_newpage(*pte))
     240           0 :                                         ret = add_mmap(addr, pte_val(*pte) & PAGE_MASK,
     241             :                                                        PAGE_SIZE, prot, hvc);
     242             :                         } else
     243           0 :                                 ret = add_munmap(addr, PAGE_SIZE, hvc);
     244           0 :                 } else if (pte_newprot(*pte))
     245           0 :                         ret = add_mprotect(addr, PAGE_SIZE, prot, hvc);
     246           0 :                 *pte = pte_mkuptodate(*pte);
     247           0 :         } while (pte++, addr += PAGE_SIZE, ((addr < end) && !ret));
     248           0 :         return ret;
     249             : }
     250             : 
     251           0 : static inline int update_pmd_range(pud_t *pud, unsigned long addr,
     252             :                                    unsigned long end,
     253             :                                    struct host_vm_change *hvc)
     254             : {
     255             :         pmd_t *pmd;
     256             :         unsigned long next;
     257           0 :         int ret = 0;
     258             : 
     259           0 :         pmd = pmd_offset(pud, addr);
     260             :         do {
     261           0 :                 next = pmd_addr_end(addr, end);
     262           0 :                 if (!pmd_present(*pmd)) {
     263           0 :                         if (hvc->force || pmd_newpage(*pmd)) {
     264           0 :                                 ret = add_munmap(addr, next - addr, hvc);
     265           0 :                                 pmd_mkuptodate(*pmd);
     266             :                         }
     267             :                 }
     268           0 :                 else ret = update_pte_range(pmd, addr, next, hvc);
     269           0 :         } while (pmd++, addr = next, ((addr < end) && !ret));
     270           0 :         return ret;
     271             : }
     272             : 
     273           0 : static inline int update_pud_range(p4d_t *p4d, unsigned long addr,
     274             :                                    unsigned long end,
     275             :                                    struct host_vm_change *hvc)
     276             : {
     277             :         pud_t *pud;
     278             :         unsigned long next;
     279           0 :         int ret = 0;
     280             : 
     281           0 :         pud = pud_offset(p4d, addr);
     282             :         do {
     283           0 :                 next = pud_addr_end(addr, end);
     284           0 :                 if (!pud_present(*pud)) {
     285           0 :                         if (hvc->force || pud_newpage(*pud)) {
     286           0 :                                 ret = add_munmap(addr, next - addr, hvc);
     287           0 :                                 pud_mkuptodate(*pud);
     288             :                         }
     289             :                 }
     290           0 :                 else ret = update_pmd_range(pud, addr, next, hvc);
     291           0 :         } while (pud++, addr = next, ((addr < end) && !ret));
     292           0 :         return ret;
     293             : }
     294             : 
     295             : static inline int update_p4d_range(pgd_t *pgd, unsigned long addr,
     296             :                                    unsigned long end,
     297             :                                    struct host_vm_change *hvc)
     298             : {
     299             :         p4d_t *p4d;
     300             :         unsigned long next;
     301           0 :         int ret = 0;
     302             : 
     303           0 :         p4d = p4d_offset(pgd, addr);
     304             :         do {
     305           0 :                 next = p4d_addr_end(addr, end);
     306           0 :                 if (!p4d_present(*p4d)) {
     307             :                         if (hvc->force || p4d_newpage(*p4d)) {
     308             :                                 ret = add_munmap(addr, next - addr, hvc);
     309             :                                 p4d_mkuptodate(*p4d);
     310             :                         }
     311             :                 } else
     312           0 :                         ret = update_pud_range(p4d, addr, next, hvc);
     313           0 :         } while (p4d++, addr = next, ((addr < end) && !ret));
     314             :         return ret;
     315             : }
     316             : 
     317           0 : static void fix_range_common(struct mm_struct *mm, unsigned long start_addr,
     318             :                              unsigned long end_addr, int force)
     319             : {
     320             :         pgd_t *pgd;
     321             :         struct host_vm_change hvc;
     322           0 :         unsigned long addr = start_addr, next;
     323           0 :         int ret = 0, userspace = 1;
     324             : 
     325           0 :         hvc = INIT_HVC(mm, force, userspace);
     326           0 :         pgd = pgd_offset(mm, addr);
     327             :         do {
     328           0 :                 next = pgd_addr_end(addr, end_addr);
     329           0 :                 if (!pgd_present(*pgd)) {
     330             :                         if (force || pgd_newpage(*pgd)) {
     331             :                                 ret = add_munmap(addr, next - addr, &hvc);
     332             :                                 pgd_mkuptodate(*pgd);
     333             :                         }
     334             :                 } else
     335           0 :                         ret = update_p4d_range(pgd, addr, next, &hvc);
     336           0 :         } while (pgd++, addr = next, ((addr < end_addr) && !ret));
     337             : 
     338           0 :         if (!ret)
     339           0 :                 ret = do_ops(&hvc, hvc.index, 1);
     340             : 
     341             :         /* This is not an else because ret is modified above */
     342           0 :         if (ret) {
     343           0 :                 struct mm_id *mm_idp = &current->mm->context.id;
     344             : 
     345           0 :                 printk(KERN_ERR "fix_range_common: failed, killing current "
     346             :                        "process: %d\n", task_tgid_vnr(current));
     347           0 :                 mm_idp->kill = 1;
     348             :         }
     349           0 : }
     350             : 
     351          16 : static int flush_tlb_kernel_range_common(unsigned long start, unsigned long end)
     352             : {
     353             :         struct mm_struct *mm;
     354             :         pgd_t *pgd;
     355             :         p4d_t *p4d;
     356             :         pud_t *pud;
     357             :         pmd_t *pmd;
     358             :         pte_t *pte;
     359             :         unsigned long addr, last;
     360          16 :         int updated = 0, err = 0, force = 0, userspace = 0;
     361             :         struct host_vm_change hvc;
     362             : 
     363          16 :         mm = &init_mm;
     364          16 :         hvc = INIT_HVC(mm, force, userspace);
     365          96 :         for (addr = start; addr < end;) {
     366         128 :                 pgd = pgd_offset(mm, addr);
     367             :                 if (!pgd_present(*pgd)) {
     368             :                         last = ADD_ROUND(addr, PGDIR_SIZE);
     369             :                         if (last > end)
     370             :                                 last = end;
     371             :                         if (pgd_newpage(*pgd)) {
     372             :                                 updated = 1;
     373             :                                 err = add_munmap(addr, last - addr, &hvc);
     374             :                                 if (err < 0)
     375             :                                         panic("munmap failed, errno = %d\n",
     376             :                                               -err);
     377             :                         }
     378             :                         addr = last;
     379             :                         continue;
     380             :                 }
     381             : 
     382          64 :                 p4d = p4d_offset(pgd, addr);
     383             :                 if (!p4d_present(*p4d)) {
     384             :                         last = ADD_ROUND(addr, P4D_SIZE);
     385             :                         if (last > end)
     386             :                                 last = end;
     387             :                         if (p4d_newpage(*p4d)) {
     388             :                                 updated = 1;
     389             :                                 err = add_munmap(addr, last - addr, &hvc);
     390             :                                 if (err < 0)
     391             :                                         panic("munmap failed, errno = %d\n",
     392             :                                               -err);
     393             :                         }
     394             :                         addr = last;
     395             :                         continue;
     396             :                 }
     397             : 
     398          64 :                 pud = pud_offset(p4d, addr);
     399          64 :                 if (!pud_present(*pud)) {
     400           0 :                         last = ADD_ROUND(addr, PUD_SIZE);
     401           0 :                         if (last > end)
     402           0 :                                 last = end;
     403           0 :                         if (pud_newpage(*pud)) {
     404           0 :                                 updated = 1;
     405           0 :                                 err = add_munmap(addr, last - addr, &hvc);
     406           0 :                                 if (err < 0)
     407           0 :                                         panic("munmap failed, errno = %d\n",
     408             :                                               -err);
     409             :                         }
     410           0 :                         addr = last;
     411           0 :                         continue;
     412             :                 }
     413             : 
     414          64 :                 pmd = pmd_offset(pud, addr);
     415          64 :                 if (!pmd_present(*pmd)) {
     416           0 :                         last = ADD_ROUND(addr, PMD_SIZE);
     417           0 :                         if (last > end)
     418           0 :                                 last = end;
     419           0 :                         if (pmd_newpage(*pmd)) {
     420           0 :                                 updated = 1;
     421           0 :                                 err = add_munmap(addr, last - addr, &hvc);
     422           0 :                                 if (err < 0)
     423           0 :                                         panic("munmap failed, errno = %d\n",
     424             :                                               -err);
     425             :                         }
     426           0 :                         addr = last;
     427           0 :                         continue;
     428             :                 }
     429             : 
     430          64 :                 pte = pte_offset_kernel(pmd, addr);
     431         128 :                 if (!pte_present(*pte) || pte_newpage(*pte)) {
     432          64 :                         updated = 1;
     433          64 :                         err = add_munmap(addr, PAGE_SIZE, &hvc);
     434          64 :                         if (err < 0)
     435           0 :                                 panic("munmap failed, errno = %d\n",
     436             :                                       -err);
     437          64 :                         if (pte_present(*pte))
     438          64 :                                 err = add_mmap(addr, pte_val(*pte) & PAGE_MASK,
     439             :                                                PAGE_SIZE, 0, &hvc);
     440             :                 }
     441           0 :                 else if (pte_newprot(*pte)) {
     442           0 :                         updated = 1;
     443           0 :                         err = add_mprotect(addr, PAGE_SIZE, 0, &hvc);
     444             :                 }
     445          64 :                 addr += PAGE_SIZE;
     446             :         }
     447          16 :         if (!err)
     448          16 :                 err = do_ops(&hvc, hvc.index, 1);
     449             : 
     450          16 :         if (err < 0)
     451           0 :                 panic("flush_tlb_kernel failed, errno = %d\n", err);
     452          16 :         return updated;
     453             : }
     454             : 
     455           0 : void flush_tlb_page(struct vm_area_struct *vma, unsigned long address)
     456             : {
     457             :         pgd_t *pgd;
     458             :         p4d_t *p4d;
     459             :         pud_t *pud;
     460             :         pmd_t *pmd;
     461             :         pte_t *pte;
     462           0 :         struct mm_struct *mm = vma->vm_mm;
     463           0 :         void *flush = NULL;
     464           0 :         int r, w, x, prot, err = 0;
     465             :         struct mm_id *mm_id;
     466             : 
     467           0 :         address &= PAGE_MASK;
     468             : 
     469           0 :         pgd = pgd_offset(mm, address);
     470             :         if (!pgd_present(*pgd))
     471             :                 goto kill;
     472             : 
     473           0 :         p4d = p4d_offset(pgd, address);
     474             :         if (!p4d_present(*p4d))
     475             :                 goto kill;
     476             : 
     477           0 :         pud = pud_offset(p4d, address);
     478           0 :         if (!pud_present(*pud))
     479             :                 goto kill;
     480             : 
     481           0 :         pmd = pmd_offset(pud, address);
     482           0 :         if (!pmd_present(*pmd))
     483             :                 goto kill;
     484             : 
     485           0 :         pte = pte_offset_kernel(pmd, address);
     486             : 
     487           0 :         r = pte_read(*pte);
     488           0 :         w = pte_write(*pte);
     489           0 :         x = pte_exec(*pte);
     490           0 :         if (!pte_young(*pte)) {
     491             :                 r = 0;
     492             :                 w = 0;
     493           0 :         } else if (!pte_dirty(*pte)) {
     494           0 :                 w = 0;
     495             :         }
     496             : 
     497           0 :         mm_id = &mm->context.id;
     498           0 :         prot = ((r ? UM_PROT_READ : 0) | (w ? UM_PROT_WRITE : 0) |
     499           0 :                 (x ? UM_PROT_EXEC : 0));
     500           0 :         if (pte_newpage(*pte)) {
     501           0 :                 if (pte_present(*pte)) {
     502             :                         unsigned long long offset;
     503             :                         int fd;
     504             : 
     505           0 :                         fd = phys_mapping(pte_val(*pte) & PAGE_MASK, &offset);
     506           0 :                         err = map(mm_id, address, PAGE_SIZE, prot, fd, offset,
     507             :                                   1, &flush);
     508             :                 }
     509           0 :                 else err = unmap(mm_id, address, PAGE_SIZE, 1, &flush);
     510             :         }
     511           0 :         else if (pte_newprot(*pte))
     512           0 :                 err = protect(mm_id, address, PAGE_SIZE, prot, 1, &flush);
     513             : 
     514           0 :         if (err) {
     515           0 :                 if (err == -ENOMEM)
     516             :                         report_enomem();
     517             : 
     518             :                 goto kill;
     519             :         }
     520             : 
     521           0 :         *pte = pte_mkuptodate(*pte);
     522             : 
     523           0 :         return;
     524             : 
     525             : kill:
     526           0 :         printk(KERN_ERR "Failed to flush page for address 0x%lx\n", address);
     527           0 :         force_sig(SIGKILL);
     528             : }
     529             : 
     530           0 : void flush_tlb_all(void)
     531             : {
     532             :         /*
     533             :          * Don't bother flushing if this address space is about to be
     534             :          * destroyed.
     535             :          */
     536           0 :         if (atomic_read(&current->mm->mm_users) == 0)
     537             :                 return;
     538             : 
     539           0 :         flush_tlb_mm(current->mm);
     540             : }
     541             : 
     542          16 : void flush_tlb_kernel_range(unsigned long start, unsigned long end)
     543             : {
     544          16 :         flush_tlb_kernel_range_common(start, end);
     545          16 : }
     546             : 
     547           0 : void flush_tlb_kernel_vm(void)
     548             : {
     549           0 :         flush_tlb_kernel_range_common(start_vm, end_vm);
     550           0 : }
     551             : 
     552           0 : void __flush_tlb_one(unsigned long addr)
     553             : {
     554           0 :         flush_tlb_kernel_range_common(addr, addr + PAGE_SIZE);
     555           0 : }
     556             : 
     557             : static void fix_range(struct mm_struct *mm, unsigned long start_addr,
     558             :                       unsigned long end_addr, int force)
     559             : {
     560             :         /*
     561             :          * Don't bother flushing if this address space is about to be
     562             :          * destroyed.
     563             :          */
     564           0 :         if (atomic_read(&mm->mm_users) == 0)
     565             :                 return;
     566             : 
     567           0 :         fix_range_common(mm, start_addr, end_addr, force);
     568             : }
     569             : 
     570           0 : void flush_tlb_range(struct vm_area_struct *vma, unsigned long start,
     571             :                      unsigned long end)
     572             : {
     573           0 :         if (vma->vm_mm == NULL)
     574           0 :                 flush_tlb_kernel_range_common(start, end);
     575           0 :         else fix_range(vma->vm_mm, start, end, 0);
     576           0 : }
     577             : EXPORT_SYMBOL(flush_tlb_range);
     578             : 
     579           0 : void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start,
     580             :                         unsigned long end)
     581             : {
     582           0 :         fix_range(mm, start, end, 0);
     583           0 : }
     584             : 
     585           0 : void flush_tlb_mm(struct mm_struct *mm)
     586             : {
     587             :         struct vm_area_struct *vma;
     588           0 :         VMA_ITERATOR(vmi, mm, 0);
     589             : 
     590           0 :         for_each_vma(vmi, vma)
     591           0 :                 fix_range(mm, vma->vm_start, vma->vm_end, 0);
     592           0 : }
     593             : 
     594           0 : void force_flush_all(void)
     595             : {
     596           0 :         struct mm_struct *mm = current->mm;
     597             :         struct vm_area_struct *vma;
     598           0 :         VMA_ITERATOR(vmi, mm, 0);
     599             : 
     600             :         mmap_read_lock(mm);
     601           0 :         for_each_vma(vmi, vma)
     602           0 :                 fix_range(mm, vma->vm_start, vma->vm_end, 1);
     603           0 :         mmap_read_unlock(mm);
     604           0 : }

Generated by: LCOV version 1.14