/*P:300 The I/O mechanism in lguest is simple yet flexible, allowing the Guest * to talk to the Launcher or directly to another Guest. It uses familiar * concepts of DMA and interrupts, plus some neat code stolen from * futexes... :*/ /* Copyright (C) 2006 Rusty Russell IBM Corporation * * This program is free software; you can redistribute it and/or modify * it under the terms of the GNU General Public License as published by * the Free Software Foundation; either version 2 of the License, or * (at your option) any later version. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, write to the Free Software * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA */ #include #include #include #include #include #include #include "lg.h" /*L:300 * I/O * * Getting data in and out of the Guest is quite an art. There are numerous * ways to do it, and they all suck differently. We try to keep things fairly * close to "real" hardware so our Guest's drivers don't look like an alien * visitation in the middle of the Linux code, and yet make sure that Guests * can talk directly to other Guests, not just the Launcher. * * To do this, the Guest gives us a key when it binds or sends DMA buffers. * The key corresponds to a "physical" address inside the Guest (ie. a virtual * address inside the Launcher process). We don't, however, use this key * directly. * * We want Guests which share memory to be able to DMA to each other: two * Launchers can mmap memory the same file, then the Guests can communicate. * Fortunately, the futex code provides us with a way to get a "union * futex_key" corresponding to the memory lying at a virtual address: if the * two processes share memory, the "union futex_key" for that memory will match * even if the memory is mapped at different addresses in each. So we always * convert the keys to "union futex_key"s to compare them. * * Before we dive into this though, we need to look at another set of helper * routines used throughout the Host kernel code to access Guest memory. :*/ static struct list_head dma_hash[61]; /* An unfortunate side effect of the Linux double-linked list implementation is * that there's no good way to statically initialize an array of linked * lists. */ void lguest_io_init(void) { unsigned int i; for (i = 0; i < ARRAY_SIZE(dma_hash); i++) INIT_LIST_HEAD(&dma_hash[i]); } /* FIXME: allow multi-page lengths. */ static int check_dma_list(struct lguest *lg, const struct lguest_dma *dma) { unsigned int i; for (i = 0; i < LGUEST_MAX_DMA_SECTIONS; i++) { if (!dma->len[i]) return 1; if (!lguest_address_ok(lg, dma->addr[i], dma->len[i])) goto kill; if (dma->len[i] > PAGE_SIZE) goto kill; /* We could do over a page, but is it worth it? */ if ((dma->addr[i] % PAGE_SIZE) + dma->len[i] > PAGE_SIZE) goto kill; } return 1; kill: kill_guest(lg, "bad DMA entry: %u@%#lx", dma->len[i], dma->addr[i]); return 0; } /*L:330 This is our hash function, using the wonderful Jenkins hash. * * The futex key is a union with three parts: an unsigned long word, a pointer, * and an int "offset". We could use jhash_2words() which takes three u32s. * (Ok, the hash functions are great: the naming sucks though). * * It's nice to be portable to 64-bit platforms, so we use the more generic * jhash2(), which takes an array of u32, the number of u32s, and an initial * u32 to roll in. This is uglier, but breaks down to almost the same code on * 32-bit platforms like this one. * * We want a position in the array, so we modulo ARRAY_SIZE(dma_hash) (ie. 61). */ static unsigned int hash(const union futex_key *key) { return jhash2((u32*)&key->both.word, (sizeof(key->both.word)+sizeof(key->both.ptr))/4, key->both.offset) % ARRAY_SIZE(dma_hash); } /* This is a convenience routine to compare two keys. It's a much bemoaned C * weakness that it doesn't allow '==' on structures or unions, so we have to * open-code it like this. */ static inline int key_eq(const union futex_key *a, const union futex_key *b) { return (a->both.word == b->both.word && a->both.ptr == b->both.ptr && a->both.offset == b->both.offset); } /*L:360 OK, when we need to actually free up a Guest's DMA array we do several * things, so we have a convenient function to do it. * * The caller must hold a read lock on dmainfo owner's current->mm->mmap_sem * for the drop_futex_key_refs(). */ static void unlink_dma(struct lguest_dma_info *dmainfo) { /* You locked this too, right? */ BUG_ON(!mutex_is_locked(&lguest_lock)); /* This is how we know that the entry is free. */ dmainfo->interrupt = 0; /* Remove it from the hash table. */ list_del(&dmainfo->list); /* Drop the references we were holding (to the inode or mm). */ drop_futex_key_refs(&dmainfo->key); } /*L:350 This is the routine which we call when the Guest asks to unregister a * DMA array attached to a given key. Returns true if the array was found. */ static int unbind_dma(struct lguest *lg, const union futex_key *key, unsigned long dmas) { int i, ret = 0; /* We don't bother with the hash table, just look through all this * Guest's DMA arrays. */ for (i = 0; i < LGUEST_MAX_DMA; i++) { /* In theory it could have more than one array on the same key, * or one array on multiple keys, so we check both */ if (key_eq(key, &lg->dma[i].key) && dmas == lg->dma[i].dmas) { unlink_dma(&lg->dma[i]); ret = 1; break; } } return ret; } /*L:340 BIND_DMA: this is the hypercall which sets up an array of "struct * lguest_dma" for receiving I/O. * * The Guest wants to bind an array of "struct lguest_dma"s to a particular key * to receive input. This only happens when the Guest is setting up a new * device, so it doesn't have to be very fast. * * It returns 1 on a successful registration (it can fail if we hit the limit * of registrations for this Guest). */ int bind_dma(struct lguest *lg, unsigned long ukey, unsigned long dmas, u16 numdmas, u8 interrupt) { unsigned int i; int ret = 0; union futex_key key; /* Futex code needs the mmap_sem. */ struct rw_semaphore *fshared = ¤t->mm->mmap_sem; /* Invalid interrupt? (We could kill the guest here). */ if (interrupt >= LGUEST_IRQS) return 0; /* We need to grab the Big Lguest Lock, because other Guests may be * trying to look through this Guest's DMAs to send something while * we're doing this. */ mutex_lock(&lguest_lock); down_read(fshared); if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) { kill_guest(lg, "bad dma key %#lx", ukey); goto unlock; } /* We want to keep this key valid once we drop mmap_sem, so we have to * hold a reference. */ get_futex_key_refs(&key); /* If the Guest specified an interrupt of 0, that means they want to * unregister this array of "struct lguest_dma"s. */ if (interrupt == 0) ret = unbind_dma(lg, &key, dmas); else { /* Look through this Guest's dma array for an unused entry. */ for (i = 0; i < LGUEST_MAX_DMA; i++) { /* If the interrupt is non-zero, the entry is already * used. */ if (lg->dma[i].interrupt) continue; /* OK, a free one! Fill on our details. */ lg->dma[i].dmas = dmas; lg->dma[i].num_dmas = numdmas; lg->dma[i].next_dma = 0; lg->dma[i].key = key; lg->dma[i].guestid = lg->guestid; lg->dma[i].interrupt = interrupt; /* Now we add it to the hash table: the position * depends on the futex key that we got. */ list_add(&lg->dma[i].list, &dma_hash[hash(&key)]); /* Success! */ ret = 1; goto unlock; } } /* If we didn't find a slot to put the key in, drop the reference * again. */ drop_futex_key_refs(&key); unlock: /* Unlock and out. */ up_read(fshared); mutex_unlock(&lguest_lock); return ret; } /*L:385 Note that our routines to access a different Guest's memory are called * lgread_other() and lgwrite_other(): these names emphasize that they are only * used when the Guest is *not* the current Guest. * * The interface for copying from another process's memory is called * access_process_vm(), with a final argument of 0 for a read, and 1 for a * write. * * We need lgread_other() to read the destination Guest's "struct lguest_dma" * array. */ static int lgread_other(struct lguest *lg, void *buf, u32 addr, unsigned bytes) { if (!lguest_address_ok(lg, addr, bytes) || access_process_vm(lg->tsk, addr, buf, bytes, 0) != bytes) { memset(buf, 0, bytes); kill_guest(lg, "bad address in registered DMA struct"); return 0; } return 1; } /* "lgwrite()" to another Guest: used to update the destination "used_len" once * we've transferred data into the buffer. */ static int lgwrite_other(struct lguest *lg, u32 addr, const void *buf, unsigned bytes) { if (!lguest_address_ok(lg, addr, bytes) || (access_process_vm(lg->tsk, addr, (void *)buf, bytes, 1) != bytes)) { kill_guest(lg, "bad address writing to registered DMA"); return 0; } return 1; } /*L:400 This is the generic engine which copies from a source "struct * lguest_dma" from this Guest into another Guest's "struct lguest_dma". The * destination Guest's pages have already been mapped, as contained in the * pages array. * * If you're wondering if there's a nice "copy from one process to another" * routine, so was I. But Linux isn't really set up to copy between two * unrelated processes, so we have to write it ourselves. */ static u32 copy_data(struct lguest *srclg, const struct lguest_dma *src, const struct lguest_dma *dst, struct page *pages[]) { unsigned int totlen, si, di, srcoff, dstoff; void *maddr = NULL; /* We return the total length transferred. */ totlen = 0; /* We keep indexes into the source and destination "struct lguest_dma", * and an offset within each region. */ si = di = 0; srcoff = dstoff = 0; /* We loop until the source or destination is exhausted. */ while (si < LGUEST_MAX_DMA_SECTIONS && src->len[si] && di < LGUEST_MAX_DMA_SECTIONS && dst->len[di]) { /* We can only transfer the rest of the src buffer, or as much * as will fit into the destination buffer. */ u32 len = min(src->len[si] - srcoff, dst->len[di] - dstoff); /* For systems using "highmem" we need to use kmap() to access * the page we want. We often use the same page over and over, * so rather than kmap() it on every loop, we set the maddr * pointer to NULL when we need to move to the next * destination page. */ if (!maddr) maddr = kmap(pages[di]); /* Copy directly from (this Guest's) source address to the * destination Guest's kmap()ed buffer. Note that maddr points * to the start of the page: we need to add the offset of the * destination address and offset within the buffer. */ /* FIXME: This is not completely portable. I looked at * copy_to_user_page(), and some arch's seem to need special * flushes. x86 is fine. */ if (copy_from_user(maddr + (dst->addr[di] + dstoff)%PAGE_SIZE, (void __user *)src->addr[si], len) != 0) { /* If a copy failed, it's the source's fault. */ kill_guest(srclg, "bad address in sending DMA"); totlen = 0; break; } /* Increment the total and src & dst offsets */ totlen += len; srcoff += len; dstoff += len; /* Presumably we reached the end of the src or dest buffers: */ if (srcoff == src->len[si]) { /* Move to the next buffer at offset 0 */ si++; srcoff = 0; } if (dstoff == dst->len[di]) { /* We need to unmap that destination page and reset * maddr ready for the next one. */ kunmap(pages[di]); maddr = NULL; di++; dstoff = 0; } } /* If we still had a page mapped at the end, unmap now. */ if (maddr) kunmap(pages[di]); return totlen; } /*L:390 This is how we transfer a "struct lguest_dma" from the source Guest * (the current Guest which called SEND_DMA) to another Guest. */ static u32 do_dma(struct lguest *srclg, const struct lguest_dma *src, struct lguest *dstlg, const struct lguest_dma *dst) { int i; u32 ret; struct page *pages[LGUEST_MAX_DMA_SECTIONS]; /* We check that both source and destination "struct lguest_dma"s are * within the bounds of the source and destination Guests */ if (!check_dma_list(dstlg, dst) || !check_dma_list(srclg, src)) return 0; /* We need to map the pages which correspond to each parts of * destination buffer. */ for (i = 0; i < LGUEST_MAX_DMA_SECTIONS; i++) { if (dst->len[i] == 0) break; /* get_user_pages() is a complicated function, especially since * we only want a single page. But it works, and returns the * number of pages. Note that we're holding the destination's * mmap_sem, as get_user_pages() requires. */ if (get_user_pages(dstlg->tsk, dstlg->mm, dst->addr[i], 1, 1, 1, pages+i, NULL) != 1) { /* This means the destination gave us a bogus buffer */ kill_guest(dstlg, "Error mapping DMA pages"); ret = 0; goto drop_pages; } } /* Now copy the data until we run out of src or dst. */ ret = copy_data(srclg, src, dst, pages); drop_pages: while (--i >= 0) put_page(pages[i]); return ret; } /*L:380 Transferring data from one Guest to another is not as simple as I'd * like. We've found the "struct lguest_dma_info" bound to the same address as * the send, we need to copy into it. * * This function returns true if the destination array was empty. */ static int dma_transfer(struct lguest *srclg, unsigned long udma, struct lguest_dma_info *dst) { struct lguest_dma dst_dma, src_dma; struct lguest *dstlg; u32 i, dma = 0; /* From the "struct lguest_dma_info" we found in the hash, grab the * Guest. */ dstlg = &lguests[dst->guestid]; /* Read in the source "struct lguest_dma" handed to SEND_DMA. */ lgread(srclg, &src_dma, udma, sizeof(src_dma)); /* We need the destination's mmap_sem, and we already hold the source's * mmap_sem for the futex key lookup. Normally this would suggest that * we could deadlock if the destination Guest was trying to send to * this source Guest at the same time, which is another reason that all * I/O is done under the big lguest_lock. */ down_read(&dstlg->mm->mmap_sem); /* Look through the destination DMA array for an available buffer. */ for (i = 0; i < dst->num_dmas; i++) { /* We keep a "next_dma" pointer which often helps us avoid * looking at lots of previously-filled entries. */ dma = (dst->next_dma + i) % dst->num_dmas; if (!lgread_other(dstlg, &dst_dma, dst->dmas + dma * sizeof(struct lguest_dma), sizeof(dst_dma))) { goto fail; } if (!dst_dma.used_len) break; } /* If we found a buffer, we do the actual data copy. */ if (i != dst->num_dmas) { unsigned long used_lenp; unsigned int ret; ret = do_dma(srclg, &src_dma, dstlg, &dst_dma); /* Put used length in the source "struct lguest_dma"'s used_len * field. It's a little tricky to figure out where that is, * though. */ lgwrite_u32(srclg, udma+offsetof(struct lguest_dma, used_len), ret); /* Tranferring 0 bytes is OK if the source buffer was empty. */ if (ret == 0 && src_dma.len[0] != 0) goto fail; /* The destination Guest might be running on a different CPU: * we have to make sure that it will see the "used_len" field * change to non-zero *after* it sees the data we copied into * the buffer. Hence a write memory barrier. */ wmb(); /* Figuring out where the destination's used_len field for this * "struct lguest_dma" in the array is also a little ugly. */ used_lenp = dst->dmas + dma * sizeof(struct lguest_dma) + offsetof(struct lguest_dma, used_len); lgwrite_other(dstlg, used_lenp, &ret, sizeof(ret)); /* Move the cursor for next time. */ dst->next_dma++; } up_read(&dstlg->mm->mmap_sem); /* We trigger the destination interrupt, even if the destination was * empty and we didn't transfer anything: this gives them a chance to * wake up and refill. */ set_bit(dst->interrupt, dstlg->irqs_pending); /* Wake up the destination process. */ wake_up_process(dstlg->tsk); /* If we passed the last "struct lguest_dma", the receive had no * buffers left. */ return i == dst->num_dmas; fail: up_read(&dstlg->mm->mmap_sem); return 0; } /*L:370 This is the counter-side to the BIND_DMA hypercall; the SEND_DMA * hypercall. We find out who's listening, and send to them. */ void send_dma(struct lguest *lg, unsigned long ukey, unsigned long udma) { union futex_key key; int empty = 0; struct rw_semaphore *fshared = ¤t->mm->mmap_sem; again: mutex_lock(&lguest_lock); down_read(fshared); /* Get the futex key for the key the Guest gave us */ if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) { kill_guest(lg, "bad sending DMA key"); goto unlock; } /* Since the key must be a multiple of 4, the futex key uses the lower * bit of the "offset" field (which would always be 0) to indicate a * mapping which is shared with other processes (ie. Guests). */ if (key.shared.offset & 1) { struct lguest_dma_info *i; /* Look through the hash for other Guests. */ list_for_each_entry(i, &dma_hash[hash(&key)], list) { /* Don't send to ourselves. */ if (i->guestid == lg->guestid) continue; if (!key_eq(&key, &i->key)) continue; /* If dma_transfer() tells us the destination has no * available buffers, we increment "empty". */ empty += dma_transfer(lg, udma, i); break; } /* If the destination is empty, we release our locks and * give the destination Guest a brief chance to restock. */ if (empty == 1) { /* Give any recipients one chance to restock. */ up_read(¤t->mm->mmap_sem); mutex_unlock(&lguest_lock); /* Next time, we won't try again. */ empty++; goto again; } } else { /* Private mapping: Guest is sending to its Launcher. We set * the "dma_is_pending" flag so that the main loop will exit * and the Launcher's read() from /dev/lguest will return. */ lg->dma_is_pending = 1; lg->pending_dma = udma; lg->pending_key = ukey; } unlock: up_read(fshared); mutex_unlock(&lguest_lock); } /*:*/ void release_all_dma(struct lguest *lg) { unsigned int i; BUG_ON(!mutex_is_locked(&lguest_lock)); down_read(&lg->mm->mmap_sem); for (i = 0; i < LGUEST_MAX_DMA; i++) { if (lg->dma[i].interrupt) unlink_dma(&lg->dma[i]); } up_read(&lg->mm->mmap_sem); } /*M:007 We only return a single DMA buffer to the Launcher, but it would be * more efficient to return a pointer to the entire array of DMA buffers, which * it can cache and choose one whenever it wants. * * Currently the Launcher uses a write to /dev/lguest, and the return value is * the address of the DMA structure with the interrupt number placed in * dma->used_len. If we wanted to return the entire array, we need to return * the address, array size and interrupt number: this seems to require an * ioctl(). :*/ /*L:320 This routine looks for a DMA buffer registered by the Guest on the * given key (using the BIND_DMA hypercall). */ unsigned long get_dma_buffer(struct lguest *lg, unsigned long ukey, unsigned long *interrupt) { unsigned long ret = 0; union futex_key key; struct lguest_dma_info *i; struct rw_semaphore *fshared = ¤t->mm->mmap_sem; /* Take the Big Lguest Lock to stop other Guests sending this Guest DMA * at the same time. */ mutex_lock(&lguest_lock); /* To match between Guests sharing the same underlying memory we steal * code from the futex infrastructure. This requires that we hold the * "mmap_sem" for our process (the Launcher), and pass it to the futex * code. */ down_read(fshared); /* This can fail if it's not a valid address, or if the address is not * divisible by 4 (the futex code needs that, we don't really). */ if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) { kill_guest(lg, "bad registered DMA buffer"); goto unlock; } /* Search the hash table for matching entries (the Launcher can only * send to its own Guest for the moment, so the entry must be for this * Guest) */ list_for_each_entry(i, &dma_hash[hash(&key)], list) { if (key_eq(&key, &i->key) && i->guestid == lg->guestid) { unsigned int j; /* Look through the registered DMA array for an * available buffer. */ for (j = 0; j < i->num_dmas; j++) { struct lguest_dma dma; ret = i->dmas + j * sizeof(struct lguest_dma); lgread(lg, &dma, ret, sizeof(dma)); if (dma.used_len == 0) break; } /* Store the interrupt the Guest wants when the buffer * is used. */ *interrupt = i->interrupt; break; } } unlock: up_read(fshared); mutex_unlock(&lguest_lock); return ret; } /*:*/ /*L:410 This really has completed the Launcher. Not only have we now finished * the longest chapter in our journey, but this also means we are over halfway * through! * * Enough prevaricating around the bush: it is time for us to dive into the * core of the Host, in "make Host". */