6d14bfe77b
It's void __user *, not void * __user... Signed-off-by: Al Viro <viro@zeniv.linux.org.uk> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
399 lines
9.6 KiB
C
399 lines
9.6 KiB
C
/* Simple I/O model for guests, based on shared memory.
|
|
* Copyright (C) 2006 Rusty Russell IBM Corporation
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA 02110-1301 USA
|
|
*/
|
|
#include <linux/types.h>
|
|
#include <linux/futex.h>
|
|
#include <linux/jhash.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/highmem.h>
|
|
#include <linux/uaccess.h>
|
|
#include "lg.h"
|
|
|
|
static struct list_head dma_hash[61];
|
|
|
|
void lguest_io_init(void)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < ARRAY_SIZE(dma_hash); i++)
|
|
INIT_LIST_HEAD(&dma_hash[i]);
|
|
}
|
|
|
|
/* FIXME: allow multi-page lengths. */
|
|
static int check_dma_list(struct lguest *lg, const struct lguest_dma *dma)
|
|
{
|
|
unsigned int i;
|
|
|
|
for (i = 0; i < LGUEST_MAX_DMA_SECTIONS; i++) {
|
|
if (!dma->len[i])
|
|
return 1;
|
|
if (!lguest_address_ok(lg, dma->addr[i], dma->len[i]))
|
|
goto kill;
|
|
if (dma->len[i] > PAGE_SIZE)
|
|
goto kill;
|
|
/* We could do over a page, but is it worth it? */
|
|
if ((dma->addr[i] % PAGE_SIZE) + dma->len[i] > PAGE_SIZE)
|
|
goto kill;
|
|
}
|
|
return 1;
|
|
|
|
kill:
|
|
kill_guest(lg, "bad DMA entry: %u@%#lx", dma->len[i], dma->addr[i]);
|
|
return 0;
|
|
}
|
|
|
|
static unsigned int hash(const union futex_key *key)
|
|
{
|
|
return jhash2((u32*)&key->both.word,
|
|
(sizeof(key->both.word)+sizeof(key->both.ptr))/4,
|
|
key->both.offset)
|
|
% ARRAY_SIZE(dma_hash);
|
|
}
|
|
|
|
static inline int key_eq(const union futex_key *a, const union futex_key *b)
|
|
{
|
|
return (a->both.word == b->both.word
|
|
&& a->both.ptr == b->both.ptr
|
|
&& a->both.offset == b->both.offset);
|
|
}
|
|
|
|
/* Must hold read lock on dmainfo owner's current->mm->mmap_sem */
|
|
static void unlink_dma(struct lguest_dma_info *dmainfo)
|
|
{
|
|
BUG_ON(!mutex_is_locked(&lguest_lock));
|
|
dmainfo->interrupt = 0;
|
|
list_del(&dmainfo->list);
|
|
drop_futex_key_refs(&dmainfo->key);
|
|
}
|
|
|
|
static int unbind_dma(struct lguest *lg,
|
|
const union futex_key *key,
|
|
unsigned long dmas)
|
|
{
|
|
int i, ret = 0;
|
|
|
|
for (i = 0; i < LGUEST_MAX_DMA; i++) {
|
|
if (key_eq(key, &lg->dma[i].key) && dmas == lg->dma[i].dmas) {
|
|
unlink_dma(&lg->dma[i]);
|
|
ret = 1;
|
|
break;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
int bind_dma(struct lguest *lg,
|
|
unsigned long ukey, unsigned long dmas, u16 numdmas, u8 interrupt)
|
|
{
|
|
unsigned int i;
|
|
int ret = 0;
|
|
union futex_key key;
|
|
struct rw_semaphore *fshared = ¤t->mm->mmap_sem;
|
|
|
|
if (interrupt >= LGUEST_IRQS)
|
|
return 0;
|
|
|
|
mutex_lock(&lguest_lock);
|
|
down_read(fshared);
|
|
if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) {
|
|
kill_guest(lg, "bad dma key %#lx", ukey);
|
|
goto unlock;
|
|
}
|
|
get_futex_key_refs(&key);
|
|
|
|
if (interrupt == 0)
|
|
ret = unbind_dma(lg, &key, dmas);
|
|
else {
|
|
for (i = 0; i < LGUEST_MAX_DMA; i++) {
|
|
if (lg->dma[i].interrupt)
|
|
continue;
|
|
|
|
lg->dma[i].dmas = dmas;
|
|
lg->dma[i].num_dmas = numdmas;
|
|
lg->dma[i].next_dma = 0;
|
|
lg->dma[i].key = key;
|
|
lg->dma[i].guestid = lg->guestid;
|
|
lg->dma[i].interrupt = interrupt;
|
|
list_add(&lg->dma[i].list, &dma_hash[hash(&key)]);
|
|
ret = 1;
|
|
goto unlock;
|
|
}
|
|
}
|
|
drop_futex_key_refs(&key);
|
|
unlock:
|
|
up_read(fshared);
|
|
mutex_unlock(&lguest_lock);
|
|
return ret;
|
|
}
|
|
|
|
/* lgread from another guest */
|
|
static int lgread_other(struct lguest *lg,
|
|
void *buf, u32 addr, unsigned bytes)
|
|
{
|
|
if (!lguest_address_ok(lg, addr, bytes)
|
|
|| access_process_vm(lg->tsk, addr, buf, bytes, 0) != bytes) {
|
|
memset(buf, 0, bytes);
|
|
kill_guest(lg, "bad address in registered DMA struct");
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
/* lgwrite to another guest */
|
|
static int lgwrite_other(struct lguest *lg, u32 addr,
|
|
const void *buf, unsigned bytes)
|
|
{
|
|
if (!lguest_address_ok(lg, addr, bytes)
|
|
|| (access_process_vm(lg->tsk, addr, (void *)buf, bytes, 1)
|
|
!= bytes)) {
|
|
kill_guest(lg, "bad address writing to registered DMA");
|
|
return 0;
|
|
}
|
|
return 1;
|
|
}
|
|
|
|
static u32 copy_data(struct lguest *srclg,
|
|
const struct lguest_dma *src,
|
|
const struct lguest_dma *dst,
|
|
struct page *pages[])
|
|
{
|
|
unsigned int totlen, si, di, srcoff, dstoff;
|
|
void *maddr = NULL;
|
|
|
|
totlen = 0;
|
|
si = di = 0;
|
|
srcoff = dstoff = 0;
|
|
while (si < LGUEST_MAX_DMA_SECTIONS && src->len[si]
|
|
&& di < LGUEST_MAX_DMA_SECTIONS && dst->len[di]) {
|
|
u32 len = min(src->len[si] - srcoff, dst->len[di] - dstoff);
|
|
|
|
if (!maddr)
|
|
maddr = kmap(pages[di]);
|
|
|
|
/* FIXME: This is not completely portable, since
|
|
archs do different things for copy_to_user_page. */
|
|
if (copy_from_user(maddr + (dst->addr[di] + dstoff)%PAGE_SIZE,
|
|
(void __user *)src->addr[si], len) != 0) {
|
|
kill_guest(srclg, "bad address in sending DMA");
|
|
totlen = 0;
|
|
break;
|
|
}
|
|
|
|
totlen += len;
|
|
srcoff += len;
|
|
dstoff += len;
|
|
if (srcoff == src->len[si]) {
|
|
si++;
|
|
srcoff = 0;
|
|
}
|
|
if (dstoff == dst->len[di]) {
|
|
kunmap(pages[di]);
|
|
maddr = NULL;
|
|
di++;
|
|
dstoff = 0;
|
|
}
|
|
}
|
|
|
|
if (maddr)
|
|
kunmap(pages[di]);
|
|
|
|
return totlen;
|
|
}
|
|
|
|
/* Src is us, ie. current. */
|
|
static u32 do_dma(struct lguest *srclg, const struct lguest_dma *src,
|
|
struct lguest *dstlg, const struct lguest_dma *dst)
|
|
{
|
|
int i;
|
|
u32 ret;
|
|
struct page *pages[LGUEST_MAX_DMA_SECTIONS];
|
|
|
|
if (!check_dma_list(dstlg, dst) || !check_dma_list(srclg, src))
|
|
return 0;
|
|
|
|
/* First get the destination pages */
|
|
for (i = 0; i < LGUEST_MAX_DMA_SECTIONS; i++) {
|
|
if (dst->len[i] == 0)
|
|
break;
|
|
if (get_user_pages(dstlg->tsk, dstlg->mm,
|
|
dst->addr[i], 1, 1, 1, pages+i, NULL)
|
|
!= 1) {
|
|
kill_guest(dstlg, "Error mapping DMA pages");
|
|
ret = 0;
|
|
goto drop_pages;
|
|
}
|
|
}
|
|
|
|
/* Now copy until we run out of src or dst. */
|
|
ret = copy_data(srclg, src, dst, pages);
|
|
|
|
drop_pages:
|
|
while (--i >= 0)
|
|
put_page(pages[i]);
|
|
return ret;
|
|
}
|
|
|
|
static int dma_transfer(struct lguest *srclg,
|
|
unsigned long udma,
|
|
struct lguest_dma_info *dst)
|
|
{
|
|
struct lguest_dma dst_dma, src_dma;
|
|
struct lguest *dstlg;
|
|
u32 i, dma = 0;
|
|
|
|
dstlg = &lguests[dst->guestid];
|
|
/* Get our dma list. */
|
|
lgread(srclg, &src_dma, udma, sizeof(src_dma));
|
|
|
|
/* We can't deadlock against them dmaing to us, because this
|
|
* is all under the lguest_lock. */
|
|
down_read(&dstlg->mm->mmap_sem);
|
|
|
|
for (i = 0; i < dst->num_dmas; i++) {
|
|
dma = (dst->next_dma + i) % dst->num_dmas;
|
|
if (!lgread_other(dstlg, &dst_dma,
|
|
dst->dmas + dma * sizeof(struct lguest_dma),
|
|
sizeof(dst_dma))) {
|
|
goto fail;
|
|
}
|
|
if (!dst_dma.used_len)
|
|
break;
|
|
}
|
|
if (i != dst->num_dmas) {
|
|
unsigned long used_lenp;
|
|
unsigned int ret;
|
|
|
|
ret = do_dma(srclg, &src_dma, dstlg, &dst_dma);
|
|
/* Put used length in src. */
|
|
lgwrite_u32(srclg,
|
|
udma+offsetof(struct lguest_dma, used_len), ret);
|
|
if (ret == 0 && src_dma.len[0] != 0)
|
|
goto fail;
|
|
|
|
/* Make sure destination sees contents before length. */
|
|
wmb();
|
|
used_lenp = dst->dmas
|
|
+ dma * sizeof(struct lguest_dma)
|
|
+ offsetof(struct lguest_dma, used_len);
|
|
lgwrite_other(dstlg, used_lenp, &ret, sizeof(ret));
|
|
dst->next_dma++;
|
|
}
|
|
up_read(&dstlg->mm->mmap_sem);
|
|
|
|
/* Do this last so dst doesn't simply sleep on lock. */
|
|
set_bit(dst->interrupt, dstlg->irqs_pending);
|
|
wake_up_process(dstlg->tsk);
|
|
return i == dst->num_dmas;
|
|
|
|
fail:
|
|
up_read(&dstlg->mm->mmap_sem);
|
|
return 0;
|
|
}
|
|
|
|
void send_dma(struct lguest *lg, unsigned long ukey, unsigned long udma)
|
|
{
|
|
union futex_key key;
|
|
int empty = 0;
|
|
struct rw_semaphore *fshared = ¤t->mm->mmap_sem;
|
|
|
|
again:
|
|
mutex_lock(&lguest_lock);
|
|
down_read(fshared);
|
|
if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) {
|
|
kill_guest(lg, "bad sending DMA key");
|
|
goto unlock;
|
|
}
|
|
/* Shared mapping? Look for other guests... */
|
|
if (key.shared.offset & 1) {
|
|
struct lguest_dma_info *i;
|
|
list_for_each_entry(i, &dma_hash[hash(&key)], list) {
|
|
if (i->guestid == lg->guestid)
|
|
continue;
|
|
if (!key_eq(&key, &i->key))
|
|
continue;
|
|
|
|
empty += dma_transfer(lg, udma, i);
|
|
break;
|
|
}
|
|
if (empty == 1) {
|
|
/* Give any recipients one chance to restock. */
|
|
up_read(¤t->mm->mmap_sem);
|
|
mutex_unlock(&lguest_lock);
|
|
empty++;
|
|
goto again;
|
|
}
|
|
} else {
|
|
/* Private mapping: tell our userspace. */
|
|
lg->dma_is_pending = 1;
|
|
lg->pending_dma = udma;
|
|
lg->pending_key = ukey;
|
|
}
|
|
unlock:
|
|
up_read(fshared);
|
|
mutex_unlock(&lguest_lock);
|
|
}
|
|
|
|
void release_all_dma(struct lguest *lg)
|
|
{
|
|
unsigned int i;
|
|
|
|
BUG_ON(!mutex_is_locked(&lguest_lock));
|
|
|
|
down_read(&lg->mm->mmap_sem);
|
|
for (i = 0; i < LGUEST_MAX_DMA; i++) {
|
|
if (lg->dma[i].interrupt)
|
|
unlink_dma(&lg->dma[i]);
|
|
}
|
|
up_read(&lg->mm->mmap_sem);
|
|
}
|
|
|
|
/* Userspace wants a dma buffer from this guest. */
|
|
unsigned long get_dma_buffer(struct lguest *lg,
|
|
unsigned long ukey, unsigned long *interrupt)
|
|
{
|
|
unsigned long ret = 0;
|
|
union futex_key key;
|
|
struct lguest_dma_info *i;
|
|
struct rw_semaphore *fshared = ¤t->mm->mmap_sem;
|
|
|
|
mutex_lock(&lguest_lock);
|
|
down_read(fshared);
|
|
if (get_futex_key((u32 __user *)ukey, fshared, &key) != 0) {
|
|
kill_guest(lg, "bad registered DMA buffer");
|
|
goto unlock;
|
|
}
|
|
list_for_each_entry(i, &dma_hash[hash(&key)], list) {
|
|
if (key_eq(&key, &i->key) && i->guestid == lg->guestid) {
|
|
unsigned int j;
|
|
for (j = 0; j < i->num_dmas; j++) {
|
|
struct lguest_dma dma;
|
|
|
|
ret = i->dmas + j * sizeof(struct lguest_dma);
|
|
lgread(lg, &dma, ret, sizeof(dma));
|
|
if (dma.used_len == 0)
|
|
break;
|
|
}
|
|
*interrupt = i->interrupt;
|
|
break;
|
|
}
|
|
}
|
|
unlock:
|
|
up_read(fshared);
|
|
mutex_unlock(&lguest_lock);
|
|
return ret;
|
|
}
|
|
|