diff options
Diffstat (limited to 'labs/lock.html')
| -rw-r--r-- | labs/lock.html | 148 | 
1 files changed, 148 insertions, 0 deletions
| diff --git a/labs/lock.html b/labs/lock.html new file mode 100644 index 0000000..707d6c4 --- /dev/null +++ b/labs/lock.html @@ -0,0 +1,148 @@ +<html> +<head> +<title>Lab: locks</title> +<link rel="stylesheet" href="homework.css" type="text/css" /> +</head> +<body> + +<h1>Lab: locks</h1> + +<p>In this lab you will try to avoid lock contention for certain +workloads. + +<h2>lock contention</h2> + +<p>The program user/kalloctest stresses xv6's memory allocator: three +  processes grow and shrink there address space, which will results in +  many calls to <tt>kalloc</tt> and <tt>kfree</tt>, +  respectively.  <tt>kalloc</tt> and <tt>kfree</tt> +  obtain <tt>kmem.lock</tt>.  To see if there is lock contention for +  <tt>kmem.lock</tt> replace the call to <tt>acquire</tt> +  in <tt>kalloc</tt> with the following code: + +  <pre> +    while(!tryacquire(&kmem.lock)) { +      printf("!"); +    } +  </pre> + +<p><tt>tryacquire</tt> tries to acquire <tt>kmem.lock</tt>: if the +  lock is taking it returns false (0); otherwise, it returns true (1) +  and with the lock acquired.  Your first job is to +  implement <tt>tryacquire</tt> in kernel/spinlock.c. + +<p>A few hints: +  <ul> +    <li>look at <tt>acquire</tt>. +    <li>don't forget to restore interrupts when acquision fails +    <li>Add tryacquire's signature to defs.h. +  </ul> + +<p>Run usertests to see if you didn't break anything.  Note that +  usertests never prints "!"; there is never contention +  for <tt>kmem.lock</tt>.  The caller is always able to immediately +  acquire the lock and never has to wait because some other process +  has the lock. + +<p>Now run kalloctest.  You should see quite a number of "!" on the +  console.  kalloctest causes many processes to contend on +  the <tt>kmem.lock</tt>.  This lock contention is a bit artificial, +  because qemu is simulating 3 processors, but it is likely on real +  hardware, there would be contention too. +   +<h2>Removing lock contention</h2> + +<p>The root cause of lock contention in kalloctest is that there is a +  single free list, protected by a single lock.  To remove lock +  contention, you will have to redesign the memory allocator to avoid +  a single lock and list.  The basic idea is to maintain a free list +  per CPU, each list with its own lock. Allocations and frees on each +  CPU can run in parallel, because each CPU will operate on a +  different list. +   +<p> The main challenge will be to deal with the case that one CPU runs +  out of memory, but another CPU has still free memory; in that case, +  the one CPU must "steal" part of the other CPU's free list. +  Stealing may introduce lock contention, but that may be acceptable +  because it may happen infrequently. + +<p>Your job is to implement per-CPU freelists and stealing when one +  CPU is out of memory.  Run kalloctest() to see if your +  implementation has removed lock contention. + +<p>Some hints: +  <ul> +    <li>You can use the constant <tt>NCPU</tt> in kernel/param.h +    <li>Let <tt>freerange</tt> give all free memory to the CPU +      running <tt>freerange</tt>. +    <li>The function <tt>cpuid</tt> returns the current core, but note +    that you can use it when interrupts are turned off and so you will +    need to turn on/off interrupts in your solution. +  </ul> + +<p>Run usertests to see if you don't break anything. + +<h2>More scalabale bcache lookup</h2> + + +<p>Several processes reading different files repeatedly will +  bottleneck in the buffer cache, bcache, in bio.c.  Replace the +  acquire in <tt>bget</tt> with +   +  <pre> +    while(!tryacquire(&bcache.lock)) { +      printf("!"); +    } +  </pre> + +  and run test0 from bcachetest and you will see "!"s. + +<p>Modify <tt>bget</tt> so that a lookup for a buffer that is in the +  bcache doesn't need to acquire <tt>bcache.lock</tt>.  This is more +  tricky than the kalloc assignment, because bcache buffers are truly +  shared among processes. You must maintain the invariant that a +  buffer is only once in memory. + +<p> There are several races that <tt>bcache.lock</tt> protects +against, including: +  <ul> +    <li>A <tt>brelse</tt> may set <tt>b->ref</tt> to 0, +      while concurrent <tt>bget</tt> is incrementing it. +    <li>Two <tt>bget</tt> may see <tt>b->ref = 0</tt> and one may re-use +    the buffer, while the other may replaces it with another block. +    <li>A concurrent <tt>brelse</tt> modifies the list +      that <tt>bget</tt> traverses. +  </ul> + +<p>A challenge is testing whether you code is still correct.  One way +  to do is to artificially delay certain operations +  using <tt>sleepticks</tt>.  <tt>test1</tt> trashes the buffer cache +  and exercises more code paths. + +<p>Here are some hints: +  <ul> +    <li>Read the description of buffer cache in the xv6 book (Section 7.2). +    <li>Use a simple design: i.e., don't design a lock-free implementation. +    <li>Use a simple hash table with locks per bucket. +    <li>Searching in hash table for a buffer and allocating an entry +      for that buffer when the buffer is not found must be atomic. +    <li>It is fine to acquire <tt>bcache.lock</tt> in <tt>brelse</tt> +      to update the LRU/MRU list. +  </ul> + +<p>Check that your implementation has less contention +  on <tt>test0</tt> + +<p>Make sure your implementation passes bcachetest and usertests. + +<p>Optional: +  <ul> +  <li>make the buffer cache more scalable (e.g., avoid taking +  out <tt>bcache.lock</tt> on <tt>brelse</tt>). +  <li>make lookup lock-free (Hint: use gcc's <tt>__sync_*</tt> +    functions.) How do you convince yourself that your implementation is correct? +  </ul> +   +   +</body> +</html> | 
