device_pager.c 14 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229230231232233234235236237238239240241242243244245246247248249250251252253254255256257258259260261262263264265266267268269270271272273274275276277278279280281282283284285286287288289290291292293294295296297298299300301302303304305306307308309310311312313314315316317318319320321322323324325326327328329330331332333334335336337338339340341342343344345346347348349350351352353354355356357358359360361362363364365366367368369370371372373374375376377378379380381382383384385386387388389390391392393394395396397398399400401402403404405406407408409410411412413414415416417418419420421422423424425426427428429430431432433434435436437438439440441442443444445446447448449450451452453454455456457458459460461462463464465466467468469470471472473474475476477478479480481482483484485486487488489490491492493494495496497498499500501502503
  1. /*-
  2. * SPDX-License-Identifier: BSD-3-Clause
  3. *
  4. * Copyright (c) 1990 University of Utah.
  5. * Copyright (c) 1991, 1993
  6. * The Regents of the University of California. All rights reserved.
  7. *
  8. * This code is derived from software contributed to Berkeley by
  9. * the Systems Programming Group of the University of Utah Computer
  10. * Science Department.
  11. *
  12. * Redistribution and use in source and binary forms, with or without
  13. * modification, are permitted provided that the following conditions
  14. * are met:
  15. * 1. Redistributions of source code must retain the above copyright
  16. * notice, this list of conditions and the following disclaimer.
  17. * 2. Redistributions in binary form must reproduce the above copyright
  18. * notice, this list of conditions and the following disclaimer in the
  19. * documentation and/or other materials provided with the distribution.
  20. * 3. Neither the name of the University nor the names of its contributors
  21. * may be used to endorse or promote products derived from this software
  22. * without specific prior written permission.
  23. *
  24. * THIS SOFTWARE IS PROVIDED BY THE REGENTS AND CONTRIBUTORS ``AS IS'' AND
  25. * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
  26. * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
  27. * ARE DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
  28. * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
  29. * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
  30. * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
  31. * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
  32. * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
  33. * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
  34. * SUCH DAMAGE.
  35. */
  36. #include <sys/param.h>
  37. #include <sys/systm.h>
  38. #include <sys/conf.h>
  39. #include <sys/lock.h>
  40. #include <sys/proc.h>
  41. #include <sys/mutex.h>
  42. #include <sys/mman.h>
  43. #include <sys/rwlock.h>
  44. #include <sys/sx.h>
  45. #include <sys/user.h>
  46. #include <sys/vmmeter.h>
  47. #include <vm/vm.h>
  48. #include <vm/vm_param.h>
  49. #include <vm/vm_object.h>
  50. #include <vm/vm_page.h>
  51. #include <vm/vm_pager.h>
  52. #include <vm/vm_phys.h>
  53. #include <vm/uma.h>
  54. static void dev_pager_init(void);
  55. static vm_object_t dev_pager_alloc(void *, vm_ooffset_t, vm_prot_t,
  56. vm_ooffset_t, struct ucred *);
  57. static void dev_pager_dealloc(vm_object_t);
  58. static int dev_pager_getpages(vm_object_t, vm_page_t *, int, int *, int *);
  59. static void dev_pager_putpages(vm_object_t, vm_page_t *, int, int, int *);
  60. static boolean_t dev_pager_haspage(vm_object_t, vm_pindex_t, int *, int *);
  61. static void dev_pager_free_page(vm_object_t object, vm_page_t m);
  62. static int dev_pager_populate(vm_object_t object, vm_pindex_t pidx,
  63. int fault_type, vm_prot_t, vm_pindex_t *first, vm_pindex_t *last);
  64. /* list of device pager objects */
  65. static struct pagerlst dev_pager_object_list;
  66. /* protect list manipulation */
  67. static struct mtx dev_pager_mtx;
  68. const struct pagerops devicepagerops = {
  69. .pgo_kvme_type = KVME_TYPE_DEVICE,
  70. .pgo_init = dev_pager_init,
  71. .pgo_alloc = dev_pager_alloc,
  72. .pgo_dealloc = dev_pager_dealloc,
  73. .pgo_getpages = dev_pager_getpages,
  74. .pgo_putpages = dev_pager_putpages,
  75. .pgo_haspage = dev_pager_haspage,
  76. };
  77. const struct pagerops mgtdevicepagerops = {
  78. .pgo_kvme_type = KVME_TYPE_MGTDEVICE,
  79. .pgo_alloc = dev_pager_alloc,
  80. .pgo_dealloc = dev_pager_dealloc,
  81. .pgo_getpages = dev_pager_getpages,
  82. .pgo_putpages = dev_pager_putpages,
  83. .pgo_haspage = dev_pager_haspage,
  84. .pgo_populate = dev_pager_populate,
  85. };
  86. static int old_dev_pager_ctor(void *handle, vm_ooffset_t size, vm_prot_t prot,
  87. vm_ooffset_t foff, struct ucred *cred, u_short *color);
  88. static void old_dev_pager_dtor(void *handle);
  89. static int old_dev_pager_fault(vm_object_t object, vm_ooffset_t offset,
  90. int prot, vm_page_t *mres);
  91. static const struct cdev_pager_ops old_dev_pager_ops = {
  92. .cdev_pg_ctor = old_dev_pager_ctor,
  93. .cdev_pg_dtor = old_dev_pager_dtor,
  94. .cdev_pg_fault = old_dev_pager_fault
  95. };
  96. static void
  97. dev_pager_init(void)
  98. {
  99. TAILQ_INIT(&dev_pager_object_list);
  100. mtx_init(&dev_pager_mtx, "dev_pager list", NULL, MTX_DEF);
  101. }
  102. vm_object_t
  103. cdev_pager_lookup(void *handle)
  104. {
  105. vm_object_t object;
  106. again:
  107. mtx_lock(&dev_pager_mtx);
  108. object = vm_pager_object_lookup(&dev_pager_object_list, handle);
  109. if (object != NULL && object->un_pager.devp.dev == NULL) {
  110. msleep(&object->un_pager.devp.dev, &dev_pager_mtx,
  111. PVM | PDROP, "cdplkp", 0);
  112. vm_object_deallocate(object);
  113. goto again;
  114. }
  115. mtx_unlock(&dev_pager_mtx);
  116. return (object);
  117. }
  118. vm_object_t
  119. cdev_pager_allocate(void *handle, enum obj_type tp,
  120. const struct cdev_pager_ops *ops, vm_ooffset_t size, vm_prot_t prot,
  121. vm_ooffset_t foff, struct ucred *cred)
  122. {
  123. vm_object_t object;
  124. vm_pindex_t pindex;
  125. if (tp != OBJT_DEVICE && tp != OBJT_MGTDEVICE)
  126. return (NULL);
  127. KASSERT(tp == OBJT_MGTDEVICE || ops->cdev_pg_populate == NULL,
  128. ("populate on unmanaged device pager"));
  129. /*
  130. * Offset should be page aligned.
  131. */
  132. if (foff & PAGE_MASK)
  133. return (NULL);
  134. /*
  135. * Treat the mmap(2) file offset as an unsigned value for a
  136. * device mapping. This, in effect, allows a user to pass all
  137. * possible off_t values as the mapping cookie to the driver. At
  138. * this point, we know that both foff and size are a multiple
  139. * of the page size. Do a check to avoid wrap.
  140. */
  141. size = round_page(size);
  142. pindex = OFF_TO_IDX(foff) + OFF_TO_IDX(size);
  143. if (pindex > OBJ_MAX_SIZE || pindex < OFF_TO_IDX(foff) ||
  144. pindex < OFF_TO_IDX(size))
  145. return (NULL);
  146. again:
  147. mtx_lock(&dev_pager_mtx);
  148. /*
  149. * Look up pager, creating as necessary.
  150. */
  151. object = vm_pager_object_lookup(&dev_pager_object_list, handle);
  152. if (object == NULL) {
  153. vm_object_t object1;
  154. /*
  155. * Allocate object and associate it with the pager. Initialize
  156. * the object's pg_color based upon the physical address of the
  157. * device's memory.
  158. */
  159. mtx_unlock(&dev_pager_mtx);
  160. object1 = vm_object_allocate(tp, pindex);
  161. mtx_lock(&dev_pager_mtx);
  162. object = vm_pager_object_lookup(&dev_pager_object_list, handle);
  163. if (object != NULL) {
  164. object1->type = OBJT_DEAD;
  165. vm_object_deallocate(object1);
  166. object1 = NULL;
  167. if (object->un_pager.devp.dev == NULL) {
  168. msleep(&object->un_pager.devp.dev,
  169. &dev_pager_mtx, PVM | PDROP, "cdplkp", 0);
  170. vm_object_deallocate(object);
  171. goto again;
  172. }
  173. /*
  174. * We raced with other thread while allocating object.
  175. */
  176. if (pindex > object->size)
  177. object->size = pindex;
  178. KASSERT(object->type == tp,
  179. ("Inconsistent device pager type %p %d",
  180. object, tp));
  181. KASSERT(object->un_pager.devp.ops == ops,
  182. ("Inconsistent devops %p %p", object, ops));
  183. } else {
  184. u_short color;
  185. object = object1;
  186. object1 = NULL;
  187. object->handle = handle;
  188. object->un_pager.devp.ops = ops;
  189. TAILQ_INIT(&object->un_pager.devp.devp_pglist);
  190. TAILQ_INSERT_TAIL(&dev_pager_object_list, object,
  191. pager_object_list);
  192. mtx_unlock(&dev_pager_mtx);
  193. if (ops->cdev_pg_populate != NULL)
  194. vm_object_set_flag(object, OBJ_POPULATE);
  195. if (ops->cdev_pg_ctor(handle, size, prot, foff,
  196. cred, &color) != 0) {
  197. mtx_lock(&dev_pager_mtx);
  198. TAILQ_REMOVE(&dev_pager_object_list, object,
  199. pager_object_list);
  200. wakeup(&object->un_pager.devp.dev);
  201. mtx_unlock(&dev_pager_mtx);
  202. object->type = OBJT_DEAD;
  203. vm_object_deallocate(object);
  204. object = NULL;
  205. mtx_lock(&dev_pager_mtx);
  206. } else {
  207. mtx_lock(&dev_pager_mtx);
  208. object->flags |= OBJ_COLORED;
  209. object->pg_color = color;
  210. object->un_pager.devp.dev = handle;
  211. wakeup(&object->un_pager.devp.dev);
  212. }
  213. }
  214. MPASS(object1 == NULL);
  215. } else {
  216. if (object->un_pager.devp.dev == NULL) {
  217. msleep(&object->un_pager.devp.dev,
  218. &dev_pager_mtx, PVM | PDROP, "cdplkp", 0);
  219. vm_object_deallocate(object);
  220. goto again;
  221. }
  222. if (pindex > object->size)
  223. object->size = pindex;
  224. KASSERT(object->type == tp,
  225. ("Inconsistent device pager type %p %d", object, tp));
  226. }
  227. mtx_unlock(&dev_pager_mtx);
  228. return (object);
  229. }
  230. static vm_object_t
  231. dev_pager_alloc(void *handle, vm_ooffset_t size, vm_prot_t prot,
  232. vm_ooffset_t foff, struct ucred *cred)
  233. {
  234. return (cdev_pager_allocate(handle, OBJT_DEVICE, &old_dev_pager_ops,
  235. size, prot, foff, cred));
  236. }
  237. void
  238. cdev_pager_free_page(vm_object_t object, vm_page_t m)
  239. {
  240. VM_OBJECT_ASSERT_WLOCKED(object);
  241. if (object->type == OBJT_MGTDEVICE) {
  242. KASSERT((m->oflags & VPO_UNMANAGED) == 0, ("unmanaged %p", m));
  243. pmap_remove_all(m);
  244. (void)vm_page_remove(m);
  245. } else if (object->type == OBJT_DEVICE)
  246. dev_pager_free_page(object, m);
  247. }
  248. static void
  249. dev_pager_free_page(vm_object_t object, vm_page_t m)
  250. {
  251. VM_OBJECT_ASSERT_WLOCKED(object);
  252. KASSERT((object->type == OBJT_DEVICE &&
  253. (m->oflags & VPO_UNMANAGED) != 0),
  254. ("Managed device or page obj %p m %p", object, m));
  255. TAILQ_REMOVE(&object->un_pager.devp.devp_pglist, m, plinks.q);
  256. vm_page_putfake(m);
  257. }
  258. static void
  259. dev_pager_dealloc(vm_object_t object)
  260. {
  261. vm_page_t m;
  262. VM_OBJECT_WUNLOCK(object);
  263. object->un_pager.devp.ops->cdev_pg_dtor(object->un_pager.devp.dev);
  264. mtx_lock(&dev_pager_mtx);
  265. TAILQ_REMOVE(&dev_pager_object_list, object, pager_object_list);
  266. mtx_unlock(&dev_pager_mtx);
  267. VM_OBJECT_WLOCK(object);
  268. if (object->type == OBJT_DEVICE) {
  269. /*
  270. * Free up our fake pages.
  271. */
  272. while ((m = TAILQ_FIRST(&object->un_pager.devp.devp_pglist))
  273. != NULL) {
  274. if (vm_page_busy_acquire(m, VM_ALLOC_WAITFAIL) == 0)
  275. continue;
  276. dev_pager_free_page(object, m);
  277. }
  278. }
  279. object->handle = NULL;
  280. object->type = OBJT_DEAD;
  281. }
  282. static int
  283. dev_pager_getpages(vm_object_t object, vm_page_t *ma, int count, int *rbehind,
  284. int *rahead)
  285. {
  286. int error;
  287. /* Since our haspage reports zero after/before, the count is 1. */
  288. KASSERT(count == 1, ("%s: count %d", __func__, count));
  289. if (object->un_pager.devp.ops->cdev_pg_fault == NULL)
  290. return (VM_PAGER_FAIL);
  291. VM_OBJECT_WLOCK(object);
  292. error = object->un_pager.devp.ops->cdev_pg_fault(object,
  293. IDX_TO_OFF(ma[0]->pindex), PROT_READ, &ma[0]);
  294. VM_OBJECT_ASSERT_WLOCKED(object);
  295. if (error == VM_PAGER_OK) {
  296. KASSERT((object->type == OBJT_DEVICE &&
  297. (ma[0]->oflags & VPO_UNMANAGED) != 0) ||
  298. (object->type == OBJT_MGTDEVICE &&
  299. (ma[0]->oflags & VPO_UNMANAGED) == 0),
  300. ("Wrong page type %p %p", ma[0], object));
  301. if (object->type == OBJT_DEVICE) {
  302. TAILQ_INSERT_TAIL(&object->un_pager.devp.devp_pglist,
  303. ma[0], plinks.q);
  304. }
  305. if (rbehind)
  306. *rbehind = 0;
  307. if (rahead)
  308. *rahead = 0;
  309. }
  310. VM_OBJECT_WUNLOCK(object);
  311. return (error);
  312. }
  313. static int
  314. dev_pager_populate(vm_object_t object, vm_pindex_t pidx, int fault_type,
  315. vm_prot_t max_prot, vm_pindex_t *first, vm_pindex_t *last)
  316. {
  317. VM_OBJECT_ASSERT_WLOCKED(object);
  318. if (object->un_pager.devp.ops->cdev_pg_populate == NULL)
  319. return (VM_PAGER_FAIL);
  320. return (object->un_pager.devp.ops->cdev_pg_populate(object, pidx,
  321. fault_type, max_prot, first, last));
  322. }
  323. static int
  324. old_dev_pager_fault(vm_object_t object, vm_ooffset_t offset, int prot,
  325. vm_page_t *mres)
  326. {
  327. vm_paddr_t paddr;
  328. vm_page_t m_paddr, page;
  329. struct cdev *dev;
  330. struct cdevsw *csw;
  331. struct file *fpop;
  332. struct thread *td;
  333. vm_memattr_t memattr, memattr1;
  334. int ref, ret;
  335. memattr = object->memattr;
  336. VM_OBJECT_WUNLOCK(object);
  337. dev = object->handle;
  338. csw = dev_refthread(dev, &ref);
  339. if (csw == NULL) {
  340. VM_OBJECT_WLOCK(object);
  341. return (VM_PAGER_FAIL);
  342. }
  343. td = curthread;
  344. fpop = td->td_fpop;
  345. td->td_fpop = NULL;
  346. ret = csw->d_mmap(dev, offset, &paddr, prot, &memattr);
  347. td->td_fpop = fpop;
  348. dev_relthread(dev, ref);
  349. if (ret != 0) {
  350. printf(
  351. "WARNING: dev_pager_getpage: map function returns error %d", ret);
  352. VM_OBJECT_WLOCK(object);
  353. return (VM_PAGER_FAIL);
  354. }
  355. /* If "paddr" is a real page, perform a sanity check on "memattr". */
  356. if ((m_paddr = vm_phys_paddr_to_vm_page(paddr)) != NULL &&
  357. (memattr1 = pmap_page_get_memattr(m_paddr)) != memattr) {
  358. /*
  359. * For the /dev/mem d_mmap routine to return the
  360. * correct memattr, pmap_page_get_memattr() needs to
  361. * be called, which we do there.
  362. */
  363. if ((csw->d_flags & D_MEM) == 0) {
  364. printf("WARNING: Device driver %s has set "
  365. "\"memattr\" inconsistently (drv %u pmap %u).\n",
  366. csw->d_name, memattr, memattr1);
  367. }
  368. memattr = memattr1;
  369. }
  370. if (((*mres)->flags & PG_FICTITIOUS) != 0) {
  371. /*
  372. * If the passed in result page is a fake page, update it with
  373. * the new physical address.
  374. */
  375. page = *mres;
  376. VM_OBJECT_WLOCK(object);
  377. vm_page_updatefake(page, paddr, memattr);
  378. } else {
  379. /*
  380. * Replace the passed in reqpage page with our own fake page and
  381. * free up the all of the original pages.
  382. */
  383. page = vm_page_getfake(paddr, memattr);
  384. VM_OBJECT_WLOCK(object);
  385. vm_page_replace(page, object, (*mres)->pindex, *mres);
  386. *mres = page;
  387. }
  388. vm_page_valid(page);
  389. return (VM_PAGER_OK);
  390. }
  391. static void
  392. dev_pager_putpages(vm_object_t object, vm_page_t *m, int count, int flags,
  393. int *rtvals)
  394. {
  395. panic("dev_pager_putpage called");
  396. }
  397. static boolean_t
  398. dev_pager_haspage(vm_object_t object, vm_pindex_t pindex, int *before,
  399. int *after)
  400. {
  401. if (before != NULL)
  402. *before = 0;
  403. if (after != NULL)
  404. *after = 0;
  405. return (TRUE);
  406. }
  407. static int
  408. old_dev_pager_ctor(void *handle, vm_ooffset_t size, vm_prot_t prot,
  409. vm_ooffset_t foff, struct ucred *cred, u_short *color)
  410. {
  411. struct cdev *dev;
  412. struct cdevsw *csw;
  413. vm_memattr_t dummy;
  414. vm_ooffset_t off;
  415. vm_paddr_t paddr;
  416. unsigned int npages;
  417. int ref;
  418. /*
  419. * Make sure this device can be mapped.
  420. */
  421. dev = handle;
  422. csw = dev_refthread(dev, &ref);
  423. if (csw == NULL)
  424. return (ENXIO);
  425. /*
  426. * Check that the specified range of the device allows the desired
  427. * protection.
  428. *
  429. * XXX assumes VM_PROT_* == PROT_*
  430. */
  431. npages = OFF_TO_IDX(size);
  432. paddr = 0; /* Make paddr initialized for the case of size == 0. */
  433. for (off = foff; npages--; off += PAGE_SIZE) {
  434. if (csw->d_mmap(dev, off, &paddr, (int)prot, &dummy) != 0) {
  435. dev_relthread(dev, ref);
  436. return (EINVAL);
  437. }
  438. }
  439. dev_ref(dev);
  440. dev_relthread(dev, ref);
  441. *color = atop(paddr) - OFF_TO_IDX(off - PAGE_SIZE);
  442. return (0);
  443. }
  444. static void
  445. old_dev_pager_dtor(void *handle)
  446. {
  447. dev_rel(handle);
  448. }