shithub: riscv

ref: 78808ca314949f40d9f6aa7abd4aa0e3f6e4b88d
dir: /sys/src/9/xen/xensystem.c/

View raw version
/*
 * xensystem.c
 *
 * TODO: we could handle mmu updates more efficiently by
 * using a multicall.
 * XXX perhaps we should check return values and panic on failure?
 */
#include	"u.h"
#include	"../port/lib.h"
#include	"mem.h"
#include	"dat.h"
#include	"fns.h"
#include	"io.h"
#include	"ureg.h"

#define LOG(a)

/*
 * These functions replace all the inlines that are used on Linux systems
 */

/* in xen.s */
int xencall1(int op);
int xencall2(int op, ulong arg1);
int xencall3(int op, ulong arg1, ulong arg2);
int xencall4(int op, ulong arg1, ulong arg2, ulong arg3);
int xencall5(int op, ulong arg1, ulong arg2, ulong arg3, ulong arg4);
int xencall6(int op, ulong arg1, ulong arg2, ulong arg3, ulong arg4, ulong arg5);

int
HYPERVISOR_update_va_mapping(ulong va, uvlong newval, ulong flags)
{
	int ret;

	ret = xencall5(__HYPERVISOR_update_va_mapping, va, newval, newval>>32, flags);
	if(ret < 0)
		panic("update_va_mapping failed");
	return ret;
}

long
HYPERVISOR_set_timer_op(uvlong timeout)
{
	ulong hi, lo;

	hi = timeout>>32;
	lo = timeout;
	return xencall3(__HYPERVISOR_set_timer_op, lo, hi);
}

int 
HYPERVISOR_set_trap_table(trap_info_t *table)
{
	return xencall2(__HYPERVISOR_set_trap_table, (ulong)table);
}

int
HYPERVISOR_mmu_update(mmu_update_t *req, int count,
	int *success_count, domid_t domid)
{
	return xencall5(__HYPERVISOR_mmu_update, (ulong)req, count, (ulong)success_count, domid);
}

int
HYPERVISOR_mmuext_op(struct mmuext_op *op, int count, int *scount, domid_t domid)
{
	return xencall5(__HYPERVISOR_mmuext_op, (ulong)op, count, (ulong)scount, domid);
}

int 
HYPERVISOR_set_gdt(unsigned long *frame_list, int entries)
{
	return xencall3(__HYPERVISOR_set_gdt, (ulong)frame_list, entries);
}

int
HYPERVISOR_stack_switch(ulong ss, ulong esp)
{
	return xencall3(__HYPERVISOR_stack_switch, ss, esp);
}

/* XXX match evfunc and fsfunc prototypes? */
int
HYPERVISOR_set_callbacks(ulong evss, ulong evfunc, ulong fsss, ulong fsfunc)
{
	return xencall5(__HYPERVISOR_set_callbacks, evss, evfunc, fsss, fsfunc);
}

int
HYPERVISOR_fpu_taskswitch(void)
{
	return xencall1(__HYPERVISOR_fpu_taskswitch);
}

int
HYPERVISOR_yield(void)
{
	return xencall3(__HYPERVISOR_sched_op, SCHEDOP_yield, 0);
}

int
HYPERVISOR_block(void)
{
	return xencall3(__HYPERVISOR_sched_op, SCHEDOP_block, 0);
}

int 
HYPERVISOR_shutdown(int reboot)
{
	sched_shutdown_t arg;

	arg.reason = reboot? SHUTDOWN_reboot : SHUTDOWN_poweroff;
	return xencall3(__HYPERVISOR_sched_op, SCHEDOP_shutdown, (ulong)&arg);
}

int
HYPERVISOR_multicall(void *call_list, int nr_calls)
{
	return xencall3(__HYPERVISOR_multicall, (ulong)call_list, nr_calls);
}


int 
HYPERVISOR_event_channel_op(void *op)
{
	return xencall2(__HYPERVISOR_event_channel_op, (ulong)op);
}

int
HYPERVISOR_xen_version(int cmd, void *arg)
{
	return xencall3(__HYPERVISOR_xen_version, cmd, (ulong)arg);
}

int
HYPERVISOR_console_io(int cmd, int count, char *str)
{
	return xencall4(__HYPERVISOR_console_io, cmd, count, (ulong)str);
}

int
HYPERVISOR_grant_table_op(int cmd, gnttab_setup_table_t *setup, int count)
{
	return xencall4(__HYPERVISOR_grant_table_op, cmd, (ulong)setup, count);
}

int
HYPERVISOR_memory_op(int cmd, struct xen_memory_reservation *arg)
{
	return xencall3(__HYPERVISOR_memory_op, cmd, (ulong)arg);
}

/* 
 * XXX this comment is leftover from old code.  revisit and update.
 *
 * The use of 'barrier' in the following reflects their use as local-lock
 * operations. Reentrancy must be prevented (e.g., __cli()) /before/ following
 * critical operations are executed. All critical operatiosn must complete
 * /before/ reentrancy is permitted (e.g., __sti()). Alpha architecture also
 * includes these barriers, for example.
 */

/*
 * conversions to machine page numbers, pages and addresses
 */
#define MFN(pa)		(patomfn[(pa)>>PGSHIFT])
#define MFNPG(pa)		((uvlong)MFN(pa)<<PGSHIFT)
#define PA2MA(pa)		(MFNPG(pa) | PGOFF(pa))
#define VA2MA(va)		PA2MA(PADDR(va))
#define VA2MFN(va)		MFN(PADDR(va))

ulong hypervisor_virt_start;
ulong xentop;
start_info_t *xenstart;
shared_info_t *HYPERVISOR_shared_info;
ulong *patomfn;
ulong *matopfn;

int
xenpdptpin(ulong va)
{
	struct mmuext_op op;
	ulong mfn;

	mfn = MFN(PADDR(va));
	LOG(dprint("pdptpin %lux %lux\n", va, mfn);)
	print("pdptpin %lux %lux\n", va, mfn);
	/* mark page readonly first */
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID, UVMF_INVLPG|UVMF_LOCAL);

	/*  L3 here refers to page directory pointer table (PAE mode) */
	op.cmd = MMUEXT_PIN_L3_TABLE;
	op.arg1.mfn = mfn;
	if (HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF) == 0)
		return 1;
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID|PTEWRITE, UVMF_INVLPG|UVMF_LOCAL);
	return 0;
}

int
xenpgdpin(ulong va)
{
	struct mmuext_op op;
	ulong mfn;

	mfn = MFN(PADDR(va));
	LOG(dprint("pdpin %lux %lux\n", va, mfn);)
	/* mark page readonly first */
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID, UVMF_INVLPG|UVMF_LOCAL);

	/* to confuse you, L2 here refers to page directories */
	op.cmd = MMUEXT_PIN_L2_TABLE;
	op.arg1.mfn = mfn;
	if (HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF) == 0)
		return 1;
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID|PTEWRITE, UVMF_INVLPG|UVMF_LOCAL);
	return 0;
}

int
xenptpin(ulong va)
{
	struct mmuext_op op;
	ulong mfn;

	mfn = MFN(PADDR(va));
	LOG(dprint("pin %lux %lux\n", va, mfn);)
	/* mark page readonly first */
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID, UVMF_INVLPG|UVMF_LOCAL);

	/* to confuse you, L1 here refers to page tables */
	op.cmd = MMUEXT_PIN_L1_TABLE;
	op.arg1.mfn = mfn;
	if (HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF) == 0)
		return 1;
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID|PTEWRITE, UVMF_INVLPG|UVMF_LOCAL);
	return 0;
}

void
xenptunpin(ulong va)
{
	struct mmuext_op op;
	ulong mfn;

	mfn = MFN(PADDR(va));
	LOG(dprint("unpin %lux %lux\n", va, mfn);)
	op.cmd = MMUEXT_UNPIN_TABLE;
	op.arg1.mfn = mfn;
	if(HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF)<0)
		panic("xenptunpin va=%lux called from %lux", va, getcallerpc(&va));

	/* mark page read-write */
	HYPERVISOR_update_va_mapping(va, ((uvlong)mfn<<PGSHIFT)|PTEVALID|PTEWRITE, UVMF_INVLPG|UVMF_LOCAL);
}

void
xenptswitch(ulong pa)
{
	struct mmuext_op op;

	op.cmd = MMUEXT_NEW_BASEPTR;
	op.arg1.mfn = MFN(pa);
	if(HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF)<0)
		panic("xenptswitch");
}

void
xentlbflush(void)
{
	struct mmuext_op op;

	op.cmd = MMUEXT_TLB_FLUSH_LOCAL;
	HYPERVISOR_mmuext_op(&op, 1, 0, DOMID_SELF);
}

/* update a pte using a machine page frame number */
void 
xenupdatema(ulong *ptr, uvlong val)
{
	mmu_update_t u;

	u.ptr = VA2MA(ptr);
	u.val = val;
	if(HYPERVISOR_mmu_update(&u, 1, 0, DOMID_SELF) < 0)
		panic("xenupdatema - pte %lux value %llux (was %llux) called from %lux", (ulong)ptr, val, *(uvlong*)ptr, getcallerpc(&ptr));
}

/* update a pte using a guest "physical" page number */
void 
xenupdate(ulong *ptr, ulong val)
{
	mmu_update_t u;

	u.ptr = VA2MA(ptr);
	u.val = PA2MA(val);
	if(HYPERVISOR_mmu_update(&u, 1, 0, DOMID_SELF) < 0)
		panic("xenupdate - pte %lux value %lux (%llux) called from %lux", (ulong)ptr, val, PA2MA(val), getcallerpc(&ptr));
}

void
acceptframe(int ref, void *va)
{
	ulong mfn;

	mfn = xengrantend(ref);
	if (mfn == 0)
		panic("can't accept page frame");
	LOG(dprint("acceptframe ref %d va %lux mfn %lux\n", ref, (ulong)va, mfn);)
	VA2MFN(va) = mfn;
	mmumapframe((ulong)va, mfn);
}

int
donateframe(int domid, void *va)
{
	ulong mfn;
	int ref;
	ulong *pte;
	struct xen_memory_reservation mem;

	mfn = VA2MFN(va);
	ref = xengrant(domid, mfn, GTF_accept_transfer);
	LOG(dprint("grant transfer %lux (%lux) -> %d\n", (ulong)va, mfn, ref);)
	pte = mmuwalk(m->pdb, (ulong)va, 2, 0);
	xenupdatema(pte, 0);
	set_xen_guest_handle(mem.extent_start, &mfn);
	mem.nr_extents = 1;
	mem.extent_order = 0;
	mem.address_bits = 0;
	mem.domid = DOMID_SELF;
	if (HYPERVISOR_memory_op(XENMEM_decrease_reservation, &mem) != 1)
		panic("XENMEM_decrease_reservation");
	VA2MFN(va) = ~0;
	return ref;
}

int
shareframe(int domid, void *va, int write)
{
	ulong mfn;
	int ref;
	int flags;

	mfn = VA2MFN(va);
	flags = GTF_permit_access;
	if (!write)
		flags |= GTF_readonly;
	ref = xengrant(domid, mfn, flags);
	LOG(dprint("grant shared %lux (%lux) -> %d\n", (ulong)va, mfn, ref);)
	return ref;
}

/*
 * Upcall from hypervisor, entered with evtchn_upcall_pending masked.
 */
void
xenupcall(Ureg *ureg)
{
	vcpu_info_t *vcpu;
	shared_info_t *s;
	ulong sel1, sel2, n1, n2, port;

	ureg->ecode = 0;
	s = HYPERVISOR_shared_info;
	vcpu = &HYPERVISOR_shared_info->vcpu_info[0];
	for (;;) {
		vcpu->evtchn_upcall_pending = 0;
		sel1 = xchgl((uint*)&vcpu->evtchn_pending_sel, 0);
		while(sel1) {
			n1 = ffs(sel1);
			sel1 &= ~(1<<n1);
			sel2 = xchgl((uint*)&s->evtchn_pending[n1], 0);
			while(sel2) {
				n2 = ffs(sel2);
				sel2 &= ~(1<<n2);
				port = (n1<<5) + n2;
				ureg->trap = 100+port;
				trap(ureg);
			}
		}
		if (vcpu->evtchn_upcall_pending)
			continue;
		vcpu->evtchn_upcall_mask = 0;
		if (vcpu->evtchn_upcall_pending == 0)
			break;
		vcpu->evtchn_upcall_mask = 1;
	}
	
}

/*
 * tbdf field is abused to distinguish virqs from channels:
 *
 * tbdf=BUSUNKNOWN -> irq is a virq to be bound to a channel
 * tbdf=0 -> irq is a channel number
 */
int
xenintrenable(Vctl *v)
{
	evtchn_op_t op;
	uint port;

	/* XXX locking? */
	if (v->tbdf != BUSUNKNOWN) {
		op.cmd = EVTCHNOP_bind_virq;
		op.u.bind_virq.virq = v->irq;
		op.u.bind_virq.vcpu = m->machno;
		if(HYPERVISOR_event_channel_op(&op) != 0)
			panic("xenintrenable: bind %d failed", v->irq);
		port = op.u.bind_virq.port;
	} else
		port = v->irq;
	if (port > 155)
		return -1;
	HYPERVISOR_shared_info->evtchn_mask[port/32] &= ~(1<<(port%32));
	if(0)print("xenintrenable %s: irq %d port %d mask[%d] = %#lux\n", v->name, v->irq, port, port/32, HYPERVISOR_shared_info->evtchn_mask[port/32]);
	return 100+port;
}

int
xenintrdisable(int irq)
{
	USED(irq);
	panic("xenintrdisable notyet\n");
	return 0;
}

int
xenintrvecno(int irq)
{
	return irq;
}

int
islo(void)
{
	vcpu_info_t *cpu;

	cpu = &HYPERVISOR_shared_info->vcpu_info[m->machno];	// XXX m->shared
	return (cpu->evtchn_upcall_mask == 0);
}

/*
 * Note: Portable code expects spllo <= spl* <= spldone for
 * accounting purposes.  Lets hope the compiler doesn't reorder
 * us.
 */
int 
spllo(void)
{
	vcpu_info_t *cpu = &HYPERVISOR_shared_info->vcpu_info[m->machno];	// XXX m->shared

	if(cpu->evtchn_upcall_mask == 0)
		return 0;
	m->splpc = 0;
	cpu->evtchn_upcall_mask = 0;

	/*
	 * If an event arrived while masked off,
	 * use a dummy call to trigger delivery
	 */
	if (cpu->evtchn_upcall_pending)
		HYPERVISOR_xen_version(0, 0);

	return 1;
}

int 
splhi(void)
{
	ulong dummy;
	vcpu_info_t *cpu = &HYPERVISOR_shared_info->vcpu_info[m->machno];	// XXX m->shared
	int oldmask;

	oldmask = xchgb(&cpu->evtchn_upcall_mask, 1);
   	if (cpu->evtchn_upcall_mask != 1)
		panic("xchgb");
	/* XXX ad-hoc ¨getcallerpc" because we have no arguments */
	m->splpc = (&dummy)[1];
	return oldmask;
}

void
splx(int x)
{
	if(x)
		splhi();
	else
		spllo();
}

/* marker for profiling in portable code */
void
spldone(void)
{
}

/* allocate an event channel */
int
xenchanalloc(int dom)
{
	evtchn_op_t op;

	op.cmd = EVTCHNOP_alloc_unbound;
	op.u.alloc_unbound.dom = DOMID_SELF;
	op.u.alloc_unbound.remote_dom = dom;
	if (HYPERVISOR_event_channel_op(&op) != 0)
		panic("xenchanalloc");
	return op.u.alloc_unbound.port;
}

/* notify over an event channel */
void
xenchannotify(int port)
{
	evtchn_op_t op;

	op.cmd = EVTCHNOP_send;
	op.u.send.port = port;
	HYPERVISOR_event_channel_op(&op);
}

void
halt(void)
{
	extern int nrdy;

	splhi();
	if (nrdy) {
		spllo();
		return;
	}
	HYPERVISOR_block();
}

void
mb(void)
{
	coherence();
}