segments.c 7.8 KB

123456789101112131415161718192021222324252627282930313233343536373839404142434445464748495051525354555657585960616263646566676869707172737475767778798081828384858687888990919293949596979899100101102103104105106107108109110111112113114115116117118119120121122123124125126127128129130131132133134135136137138139140141142143144145146147148149150151152153154155156157158159160161162163164165166167168169170171172173174175176177178179180181182183184185186187188189190191192193194195196197198199200201202203204205206207208209210211212213214215216217218219220221222223224225226227228229
  1. /*P:600
  2. * The x86 architecture has segments, which involve a table of descriptors
  3. * which can be used to do funky things with virtual address interpretation.
  4. * We originally used to use segments so the Guest couldn't alter the
  5. * Guest<->Host Switcher, and then we had to trim Guest segments, and restore
  6. * for userspace per-thread segments, but trim again for on userspace->kernel
  7. * transitions... This nightmarish creation was contained within this file,
  8. * where we knew not to tread without heavy armament and a change of underwear.
  9. *
  10. * In these modern times, the segment handling code consists of simple sanity
  11. * checks, and the worst you'll experience reading this code is butterfly-rash
  12. * from frolicking through its parklike serenity.
  13. :*/
  14. #include "lg.h"
  15. /*H:600
  16. * Segments & The Global Descriptor Table
  17. *
  18. * (That title sounds like a bad Nerdcore group. Not to suggest that there are
  19. * any good Nerdcore groups, but in high school a friend of mine had a band
  20. * called Joe Fish and the Chips, so there are definitely worse band names).
  21. *
  22. * To refresh: the GDT is a table of 8-byte values describing segments. Once
  23. * set up, these segments can be loaded into one of the 6 "segment registers".
  24. *
  25. * GDT entries are passed around as "struct desc_struct"s, which like IDT
  26. * entries are split into two 32-bit members, "a" and "b". One day, someone
  27. * will clean that up, and be declared a Hero. (No pressure, I'm just saying).
  28. *
  29. * Anyway, the GDT entry contains a base (the start address of the segment), a
  30. * limit (the size of the segment - 1), and some flags. Sounds simple, and it
  31. * would be, except those zany Intel engineers decided that it was too boring
  32. * to put the base at one end, the limit at the other, and the flags in
  33. * between. They decided to shotgun the bits at random throughout the 8 bytes,
  34. * like so:
  35. *
  36. * 0 16 40 48 52 56 63
  37. * [ limit part 1 ][ base part 1 ][ flags ][li][fl][base ]
  38. * mit ags part 2
  39. * part 2
  40. *
  41. * As a result, this file contains a certain amount of magic numeracy. Let's
  42. * begin.
  43. */
  44. /*
  45. * There are several entries we don't let the Guest set. The TSS entry is the
  46. * "Task State Segment" which controls all kinds of delicate things. The
  47. * LGUEST_CS and LGUEST_DS entries are reserved for the Switcher, and the
  48. * the Guest can't be trusted to deal with double faults.
  49. */
  50. static bool ignored_gdt(unsigned int num)
  51. {
  52. return (num == GDT_ENTRY_TSS
  53. || num == GDT_ENTRY_LGUEST_CS
  54. || num == GDT_ENTRY_LGUEST_DS
  55. || num == GDT_ENTRY_DOUBLEFAULT_TSS);
  56. }
  57. /*H:630
  58. * Once the Guest gave us new GDT entries, we fix them up a little. We
  59. * don't care if they're invalid: the worst that can happen is a General
  60. * Protection Fault in the Switcher when it restores a Guest segment register
  61. * which tries to use that entry. Then we kill the Guest for causing such a
  62. * mess: the message will be "unhandled trap 256".
  63. */
  64. static void fixup_gdt_table(struct lg_cpu *cpu, unsigned start, unsigned end)
  65. {
  66. unsigned int i;
  67. for (i = start; i < end; i++) {
  68. /*
  69. * We never copy these ones to real GDT, so we don't care what
  70. * they say
  71. */
  72. if (ignored_gdt(i))
  73. continue;
  74. /*
  75. * Segment descriptors contain a privilege level: the Guest is
  76. * sometimes careless and leaves this as 0, even though it's
  77. * running at privilege level 1. If so, we fix it here.
  78. */
  79. if (cpu->arch.gdt[i].dpl == 0)
  80. cpu->arch.gdt[i].dpl |= GUEST_PL;
  81. /*
  82. * Each descriptor has an "accessed" bit. If we don't set it
  83. * now, the CPU will try to set it when the Guest first loads
  84. * that entry into a segment register. But the GDT isn't
  85. * writable by the Guest, so bad things can happen.
  86. */
  87. cpu->arch.gdt[i].type |= 0x1;
  88. }
  89. }
  90. /*H:610
  91. * Like the IDT, we never simply use the GDT the Guest gives us. We keep
  92. * a GDT for each CPU, and copy across the Guest's entries each time we want to
  93. * run the Guest on that CPU.
  94. *
  95. * This routine is called at boot or modprobe time for each CPU to set up the
  96. * constant GDT entries: the ones which are the same no matter what Guest we're
  97. * running.
  98. */
  99. void setup_default_gdt_entries(struct lguest_ro_state *state)
  100. {
  101. struct desc_struct *gdt = state->guest_gdt;
  102. unsigned long tss = (unsigned long)&state->guest_tss;
  103. /* The Switcher segments are full 0-4G segments, privilege level 0 */
  104. gdt[GDT_ENTRY_LGUEST_CS] = FULL_EXEC_SEGMENT;
  105. gdt[GDT_ENTRY_LGUEST_DS] = FULL_SEGMENT;
  106. /*
  107. * The TSS segment refers to the TSS entry for this particular CPU.
  108. */
  109. gdt[GDT_ENTRY_TSS].a = 0;
  110. gdt[GDT_ENTRY_TSS].b = 0;
  111. gdt[GDT_ENTRY_TSS].limit0 = 0x67;
  112. gdt[GDT_ENTRY_TSS].base0 = tss & 0xFFFF;
  113. gdt[GDT_ENTRY_TSS].base1 = (tss >> 16) & 0xFF;
  114. gdt[GDT_ENTRY_TSS].base2 = tss >> 24;
  115. gdt[GDT_ENTRY_TSS].type = 0x9; /* 32-bit TSS (available) */
  116. gdt[GDT_ENTRY_TSS].p = 0x1; /* Entry is present */
  117. gdt[GDT_ENTRY_TSS].dpl = 0x0; /* Privilege level 0 */
  118. gdt[GDT_ENTRY_TSS].s = 0x0; /* system segment */
  119. }
  120. /*
  121. * This routine sets up the initial Guest GDT for booting. All entries start
  122. * as 0 (unusable).
  123. */
  124. void setup_guest_gdt(struct lg_cpu *cpu)
  125. {
  126. /*
  127. * Start with full 0-4G segments...except the Guest is allowed to use
  128. * them, so set the privilege level appropriately in the flags.
  129. */
  130. cpu->arch.gdt[GDT_ENTRY_KERNEL_CS] = FULL_EXEC_SEGMENT;
  131. cpu->arch.gdt[GDT_ENTRY_KERNEL_DS] = FULL_SEGMENT;
  132. cpu->arch.gdt[GDT_ENTRY_KERNEL_CS].dpl |= GUEST_PL;
  133. cpu->arch.gdt[GDT_ENTRY_KERNEL_DS].dpl |= GUEST_PL;
  134. }
  135. /*H:650
  136. * An optimization of copy_gdt(), for just the three "thead-local storage"
  137. * entries.
  138. */
  139. void copy_gdt_tls(const struct lg_cpu *cpu, struct desc_struct *gdt)
  140. {
  141. unsigned int i;
  142. for (i = GDT_ENTRY_TLS_MIN; i <= GDT_ENTRY_TLS_MAX; i++)
  143. gdt[i] = cpu->arch.gdt[i];
  144. }
  145. /*H:640
  146. * When the Guest is run on a different CPU, or the GDT entries have changed,
  147. * copy_gdt() is called to copy the Guest's GDT entries across to this CPU's
  148. * GDT.
  149. */
  150. void copy_gdt(const struct lg_cpu *cpu, struct desc_struct *gdt)
  151. {
  152. unsigned int i;
  153. /*
  154. * The default entries from setup_default_gdt_entries() are not
  155. * replaced. See ignored_gdt() above.
  156. */
  157. for (i = 0; i < GDT_ENTRIES; i++)
  158. if (!ignored_gdt(i))
  159. gdt[i] = cpu->arch.gdt[i];
  160. }
  161. /*H:620
  162. * This is where the Guest asks us to load a new GDT entry
  163. * (LHCALL_LOAD_GDT_ENTRY). We tweak the entry and copy it in.
  164. */
  165. void load_guest_gdt_entry(struct lg_cpu *cpu, u32 num, u32 lo, u32 hi)
  166. {
  167. /*
  168. * We assume the Guest has the same number of GDT entries as the
  169. * Host, otherwise we'd have to dynamically allocate the Guest GDT.
  170. */
  171. if (num >= ARRAY_SIZE(cpu->arch.gdt)) {
  172. kill_guest(cpu, "too many gdt entries %i", num);
  173. return;
  174. }
  175. /* Set it up, then fix it. */
  176. cpu->arch.gdt[num].a = lo;
  177. cpu->arch.gdt[num].b = hi;
  178. fixup_gdt_table(cpu, num, num+1);
  179. /*
  180. * Mark that the GDT changed so the core knows it has to copy it again,
  181. * even if the Guest is run on the same CPU.
  182. */
  183. cpu->changed |= CHANGED_GDT;
  184. }
  185. /*
  186. * This is the fast-track version for just changing the three TLS entries.
  187. * Remember that this happens on every context switch, so it's worth
  188. * optimizing. But wouldn't it be neater to have a single hypercall to cover
  189. * both cases?
  190. */
  191. void guest_load_tls(struct lg_cpu *cpu, unsigned long gtls)
  192. {
  193. struct desc_struct *tls = &cpu->arch.gdt[GDT_ENTRY_TLS_MIN];
  194. __lgread(cpu, tls, gtls, sizeof(*tls)*GDT_ENTRY_TLS_ENTRIES);
  195. fixup_gdt_table(cpu, GDT_ENTRY_TLS_MIN, GDT_ENTRY_TLS_MAX+1);
  196. /* Note that just the TLS entries have changed. */
  197. cpu->changed |= CHANGED_GDT_TLS;
  198. }
  199. /*H:660
  200. * With this, we have finished the Host.
  201. *
  202. * Five of the seven parts of our task are complete. You have made it through
  203. * the Bit of Despair (I think that's somewhere in the page table code,
  204. * myself).
  205. *
  206. * Next, we examine "make Switcher". It's short, but intense.
  207. */