Merge branch 'master' of /pub/scm/linux/kernel/git/torvalds/linux-2.6
This commit is contained in:
commit
5f65f150fb
11 changed files with 44 additions and 25 deletions
|
@ -440,7 +440,7 @@ int kvmppc_44x_emul_tlbwe(struct kvm_vcpu *vcpu, u8 ra, u8 rs, u8 ws)
|
||||||
unsigned int gtlb_index;
|
unsigned int gtlb_index;
|
||||||
|
|
||||||
gtlb_index = kvmppc_get_gpr(vcpu, ra);
|
gtlb_index = kvmppc_get_gpr(vcpu, ra);
|
||||||
if (gtlb_index > KVM44x_GUEST_TLB_SIZE) {
|
if (gtlb_index >= KVM44x_GUEST_TLB_SIZE) {
|
||||||
printk("%s: index %d\n", __func__, gtlb_index);
|
printk("%s: index %d\n", __func__, gtlb_index);
|
||||||
kvmppc_dump_vcpu(vcpu);
|
kvmppc_dump_vcpu(vcpu);
|
||||||
return EMULATE_FAIL;
|
return EMULATE_FAIL;
|
||||||
|
|
|
@ -2067,7 +2067,7 @@ static int cpuid_interception(struct vcpu_svm *svm)
|
||||||
static int iret_interception(struct vcpu_svm *svm)
|
static int iret_interception(struct vcpu_svm *svm)
|
||||||
{
|
{
|
||||||
++svm->vcpu.stat.nmi_window_exits;
|
++svm->vcpu.stat.nmi_window_exits;
|
||||||
svm->vmcb->control.intercept &= ~(1UL << INTERCEPT_IRET);
|
svm->vmcb->control.intercept &= ~(1ULL << INTERCEPT_IRET);
|
||||||
svm->vcpu.arch.hflags |= HF_IRET_MASK;
|
svm->vcpu.arch.hflags |= HF_IRET_MASK;
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
@ -2479,7 +2479,7 @@ static void svm_inject_nmi(struct kvm_vcpu *vcpu)
|
||||||
|
|
||||||
svm->vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI;
|
svm->vmcb->control.event_inj = SVM_EVTINJ_VALID | SVM_EVTINJ_TYPE_NMI;
|
||||||
vcpu->arch.hflags |= HF_NMI_MASK;
|
vcpu->arch.hflags |= HF_NMI_MASK;
|
||||||
svm->vmcb->control.intercept |= (1UL << INTERCEPT_IRET);
|
svm->vmcb->control.intercept |= (1ULL << INTERCEPT_IRET);
|
||||||
++vcpu->stat.nmi_injections;
|
++vcpu->stat.nmi_injections;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -2539,10 +2539,10 @@ static void svm_set_nmi_mask(struct kvm_vcpu *vcpu, bool masked)
|
||||||
|
|
||||||
if (masked) {
|
if (masked) {
|
||||||
svm->vcpu.arch.hflags |= HF_NMI_MASK;
|
svm->vcpu.arch.hflags |= HF_NMI_MASK;
|
||||||
svm->vmcb->control.intercept |= (1UL << INTERCEPT_IRET);
|
svm->vmcb->control.intercept |= (1ULL << INTERCEPT_IRET);
|
||||||
} else {
|
} else {
|
||||||
svm->vcpu.arch.hflags &= ~HF_NMI_MASK;
|
svm->vcpu.arch.hflags &= ~HF_NMI_MASK;
|
||||||
svm->vmcb->control.intercept &= ~(1UL << INTERCEPT_IRET);
|
svm->vmcb->control.intercept &= ~(1ULL << INTERCEPT_IRET);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -2703,8 +2703,7 @@ static int vmx_nmi_allowed(struct kvm_vcpu *vcpu)
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
return !(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
|
return !(vmcs_read32(GUEST_INTERRUPTIBILITY_INFO) &
|
||||||
(GUEST_INTR_STATE_STI | GUEST_INTR_STATE_MOV_SS |
|
(GUEST_INTR_STATE_MOV_SS | GUEST_INTR_STATE_NMI));
|
||||||
GUEST_INTR_STATE_NMI));
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu)
|
static bool vmx_get_nmi_mask(struct kvm_vcpu *vcpu)
|
||||||
|
|
|
@ -1712,6 +1712,7 @@ static int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
|
||||||
if (copy_from_user(cpuid_entries, entries,
|
if (copy_from_user(cpuid_entries, entries,
|
||||||
cpuid->nent * sizeof(struct kvm_cpuid_entry)))
|
cpuid->nent * sizeof(struct kvm_cpuid_entry)))
|
||||||
goto out_free;
|
goto out_free;
|
||||||
|
vcpu_load(vcpu);
|
||||||
for (i = 0; i < cpuid->nent; i++) {
|
for (i = 0; i < cpuid->nent; i++) {
|
||||||
vcpu->arch.cpuid_entries[i].function = cpuid_entries[i].function;
|
vcpu->arch.cpuid_entries[i].function = cpuid_entries[i].function;
|
||||||
vcpu->arch.cpuid_entries[i].eax = cpuid_entries[i].eax;
|
vcpu->arch.cpuid_entries[i].eax = cpuid_entries[i].eax;
|
||||||
|
@ -1729,6 +1730,7 @@ static int kvm_vcpu_ioctl_set_cpuid(struct kvm_vcpu *vcpu,
|
||||||
r = 0;
|
r = 0;
|
||||||
kvm_apic_set_version(vcpu);
|
kvm_apic_set_version(vcpu);
|
||||||
kvm_x86_ops->cpuid_update(vcpu);
|
kvm_x86_ops->cpuid_update(vcpu);
|
||||||
|
vcpu_put(vcpu);
|
||||||
|
|
||||||
out_free:
|
out_free:
|
||||||
vfree(cpuid_entries);
|
vfree(cpuid_entries);
|
||||||
|
@ -1749,9 +1751,11 @@ static int kvm_vcpu_ioctl_set_cpuid2(struct kvm_vcpu *vcpu,
|
||||||
if (copy_from_user(&vcpu->arch.cpuid_entries, entries,
|
if (copy_from_user(&vcpu->arch.cpuid_entries, entries,
|
||||||
cpuid->nent * sizeof(struct kvm_cpuid_entry2)))
|
cpuid->nent * sizeof(struct kvm_cpuid_entry2)))
|
||||||
goto out;
|
goto out;
|
||||||
|
vcpu_load(vcpu);
|
||||||
vcpu->arch.cpuid_nent = cpuid->nent;
|
vcpu->arch.cpuid_nent = cpuid->nent;
|
||||||
kvm_apic_set_version(vcpu);
|
kvm_apic_set_version(vcpu);
|
||||||
kvm_x86_ops->cpuid_update(vcpu);
|
kvm_x86_ops->cpuid_update(vcpu);
|
||||||
|
vcpu_put(vcpu);
|
||||||
return 0;
|
return 0;
|
||||||
|
|
||||||
out:
|
out:
|
||||||
|
|
|
@ -1875,6 +1875,7 @@ got_driver:
|
||||||
*/
|
*/
|
||||||
if (filp->f_op == &hung_up_tty_fops)
|
if (filp->f_op == &hung_up_tty_fops)
|
||||||
filp->f_op = &tty_fops;
|
filp->f_op = &tty_fops;
|
||||||
|
unlock_kernel();
|
||||||
goto retry_open;
|
goto retry_open;
|
||||||
}
|
}
|
||||||
unlock_kernel();
|
unlock_kernel();
|
||||||
|
|
|
@ -349,6 +349,9 @@ int wm831x_auxadc_read(struct wm831x *wm831x, enum wm831x_auxadc input)
|
||||||
goto disable;
|
goto disable;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* If an interrupt arrived late clean up after it */
|
||||||
|
try_wait_for_completion(&wm831x->auxadc_done);
|
||||||
|
|
||||||
/* Ignore the result to allow us to soldier on without IRQ hookup */
|
/* Ignore the result to allow us to soldier on without IRQ hookup */
|
||||||
wait_for_completion_timeout(&wm831x->auxadc_done, msecs_to_jiffies(5));
|
wait_for_completion_timeout(&wm831x->auxadc_done, msecs_to_jiffies(5));
|
||||||
|
|
||||||
|
|
|
@ -363,6 +363,10 @@ int wm8350_read_auxadc(struct wm8350 *wm8350, int channel, int scale, int vref)
|
||||||
reg |= 1 << channel | WM8350_AUXADC_POLL;
|
reg |= 1 << channel | WM8350_AUXADC_POLL;
|
||||||
wm8350_reg_write(wm8350, WM8350_DIGITISER_CONTROL_1, reg);
|
wm8350_reg_write(wm8350, WM8350_DIGITISER_CONTROL_1, reg);
|
||||||
|
|
||||||
|
/* If a late IRQ left the completion signalled then consume
|
||||||
|
* the completion. */
|
||||||
|
try_wait_for_completion(&wm8350->auxadc_done);
|
||||||
|
|
||||||
/* We ignore the result of the completion and just check for a
|
/* We ignore the result of the completion and just check for a
|
||||||
* conversion result, allowing us to soldier on if the IRQ
|
* conversion result, allowing us to soldier on if the IRQ
|
||||||
* infrastructure is not set up for the chip. */
|
* infrastructure is not set up for the chip. */
|
||||||
|
|
|
@ -120,7 +120,8 @@
|
||||||
#define MX2_UCR3_RXDMUXSEL (1<<2) /* RXD Muxed Input Select, on mx2/mx3 */
|
#define MX2_UCR3_RXDMUXSEL (1<<2) /* RXD Muxed Input Select, on mx2/mx3 */
|
||||||
#define UCR3_INVT (1<<1) /* Inverted Infrared transmission */
|
#define UCR3_INVT (1<<1) /* Inverted Infrared transmission */
|
||||||
#define UCR3_BPEN (1<<0) /* Preset registers enable */
|
#define UCR3_BPEN (1<<0) /* Preset registers enable */
|
||||||
#define UCR4_CTSTL_32 (32<<10) /* CTS trigger level (32 chars) */
|
#define UCR4_CTSTL_SHF 10 /* CTS trigger level shift */
|
||||||
|
#define UCR4_CTSTL_MASK 0x3F /* CTS trigger is 6 bits wide */
|
||||||
#define UCR4_INVR (1<<9) /* Inverted infrared reception */
|
#define UCR4_INVR (1<<9) /* Inverted infrared reception */
|
||||||
#define UCR4_ENIRI (1<<8) /* Serial infrared interrupt enable */
|
#define UCR4_ENIRI (1<<8) /* Serial infrared interrupt enable */
|
||||||
#define UCR4_WKEN (1<<7) /* Wake interrupt enable */
|
#define UCR4_WKEN (1<<7) /* Wake interrupt enable */
|
||||||
|
@ -591,6 +592,9 @@ static int imx_setup_ufcr(struct imx_port *sport, unsigned int mode)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* half the RX buffer size */
|
||||||
|
#define CTSTL 16
|
||||||
|
|
||||||
static int imx_startup(struct uart_port *port)
|
static int imx_startup(struct uart_port *port)
|
||||||
{
|
{
|
||||||
struct imx_port *sport = (struct imx_port *)port;
|
struct imx_port *sport = (struct imx_port *)port;
|
||||||
|
@ -607,6 +611,10 @@ static int imx_startup(struct uart_port *port)
|
||||||
if (USE_IRDA(sport))
|
if (USE_IRDA(sport))
|
||||||
temp |= UCR4_IRSC;
|
temp |= UCR4_IRSC;
|
||||||
|
|
||||||
|
/* set the trigger level for CTS */
|
||||||
|
temp &= ~(UCR4_CTSTL_MASK<< UCR4_CTSTL_SHF);
|
||||||
|
temp |= CTSTL<< UCR4_CTSTL_SHF;
|
||||||
|
|
||||||
writel(temp & ~UCR4_DREN, sport->port.membase + UCR4);
|
writel(temp & ~UCR4_DREN, sport->port.membase + UCR4);
|
||||||
|
|
||||||
if (USE_IRDA(sport)) {
|
if (USE_IRDA(sport)) {
|
||||||
|
|
|
@ -1467,7 +1467,7 @@ mpc52xx_uart_init(void)
|
||||||
/*
|
/*
|
||||||
* Map the PSC FIFO Controller and init if on MPC512x.
|
* Map the PSC FIFO Controller and init if on MPC512x.
|
||||||
*/
|
*/
|
||||||
if (psc_ops->fifoc_init) {
|
if (psc_ops && psc_ops->fifoc_init) {
|
||||||
ret = psc_ops->fifoc_init();
|
ret = psc_ops->fifoc_init();
|
||||||
if (ret)
|
if (ret)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
|
@ -197,7 +197,7 @@ int kvm_ioapic_set_irq(struct kvm_ioapic *ioapic, int irq, int level)
|
||||||
union kvm_ioapic_redirect_entry entry;
|
union kvm_ioapic_redirect_entry entry;
|
||||||
int ret = 1;
|
int ret = 1;
|
||||||
|
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
if (irq >= 0 && irq < IOAPIC_NUM_PINS) {
|
if (irq >= 0 && irq < IOAPIC_NUM_PINS) {
|
||||||
entry = ioapic->redirtbl[irq];
|
entry = ioapic->redirtbl[irq];
|
||||||
level ^= entry.fields.polarity;
|
level ^= entry.fields.polarity;
|
||||||
|
@ -214,7 +214,7 @@ int kvm_ioapic_set_irq(struct kvm_ioapic *ioapic, int irq, int level)
|
||||||
}
|
}
|
||||||
trace_kvm_ioapic_set_irq(entry.bits, irq, ret == 0);
|
trace_kvm_ioapic_set_irq(entry.bits, irq, ret == 0);
|
||||||
}
|
}
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
|
|
||||||
return ret;
|
return ret;
|
||||||
}
|
}
|
||||||
|
@ -238,9 +238,9 @@ static void __kvm_ioapic_update_eoi(struct kvm_ioapic *ioapic, int vector,
|
||||||
* is dropped it will be put into irr and will be delivered
|
* is dropped it will be put into irr and will be delivered
|
||||||
* after ack notifier returns.
|
* after ack notifier returns.
|
||||||
*/
|
*/
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
kvm_notify_acked_irq(ioapic->kvm, KVM_IRQCHIP_IOAPIC, i);
|
kvm_notify_acked_irq(ioapic->kvm, KVM_IRQCHIP_IOAPIC, i);
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
|
|
||||||
if (trigger_mode != IOAPIC_LEVEL_TRIG)
|
if (trigger_mode != IOAPIC_LEVEL_TRIG)
|
||||||
continue;
|
continue;
|
||||||
|
@ -259,9 +259,9 @@ void kvm_ioapic_update_eoi(struct kvm *kvm, int vector, int trigger_mode)
|
||||||
smp_rmb();
|
smp_rmb();
|
||||||
if (!test_bit(vector, ioapic->handled_vectors))
|
if (!test_bit(vector, ioapic->handled_vectors))
|
||||||
return;
|
return;
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
__kvm_ioapic_update_eoi(ioapic, vector, trigger_mode);
|
__kvm_ioapic_update_eoi(ioapic, vector, trigger_mode);
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
}
|
}
|
||||||
|
|
||||||
static inline struct kvm_ioapic *to_ioapic(struct kvm_io_device *dev)
|
static inline struct kvm_ioapic *to_ioapic(struct kvm_io_device *dev)
|
||||||
|
@ -287,7 +287,7 @@ static int ioapic_mmio_read(struct kvm_io_device *this, gpa_t addr, int len,
|
||||||
ASSERT(!(addr & 0xf)); /* check alignment */
|
ASSERT(!(addr & 0xf)); /* check alignment */
|
||||||
|
|
||||||
addr &= 0xff;
|
addr &= 0xff;
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
switch (addr) {
|
switch (addr) {
|
||||||
case IOAPIC_REG_SELECT:
|
case IOAPIC_REG_SELECT:
|
||||||
result = ioapic->ioregsel;
|
result = ioapic->ioregsel;
|
||||||
|
@ -301,7 +301,7 @@ static int ioapic_mmio_read(struct kvm_io_device *this, gpa_t addr, int len,
|
||||||
result = 0;
|
result = 0;
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
|
|
||||||
switch (len) {
|
switch (len) {
|
||||||
case 8:
|
case 8:
|
||||||
|
@ -338,7 +338,7 @@ static int ioapic_mmio_write(struct kvm_io_device *this, gpa_t addr, int len,
|
||||||
}
|
}
|
||||||
|
|
||||||
addr &= 0xff;
|
addr &= 0xff;
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
switch (addr) {
|
switch (addr) {
|
||||||
case IOAPIC_REG_SELECT:
|
case IOAPIC_REG_SELECT:
|
||||||
ioapic->ioregsel = data;
|
ioapic->ioregsel = data;
|
||||||
|
@ -356,7 +356,7 @@ static int ioapic_mmio_write(struct kvm_io_device *this, gpa_t addr, int len,
|
||||||
default:
|
default:
|
||||||
break;
|
break;
|
||||||
}
|
}
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -386,7 +386,7 @@ int kvm_ioapic_init(struct kvm *kvm)
|
||||||
ioapic = kzalloc(sizeof(struct kvm_ioapic), GFP_KERNEL);
|
ioapic = kzalloc(sizeof(struct kvm_ioapic), GFP_KERNEL);
|
||||||
if (!ioapic)
|
if (!ioapic)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
mutex_init(&ioapic->lock);
|
spin_lock_init(&ioapic->lock);
|
||||||
kvm->arch.vioapic = ioapic;
|
kvm->arch.vioapic = ioapic;
|
||||||
kvm_ioapic_reset(ioapic);
|
kvm_ioapic_reset(ioapic);
|
||||||
kvm_iodevice_init(&ioapic->dev, &ioapic_mmio_ops);
|
kvm_iodevice_init(&ioapic->dev, &ioapic_mmio_ops);
|
||||||
|
@ -419,9 +419,9 @@ int kvm_get_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state)
|
||||||
if (!ioapic)
|
if (!ioapic)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
memcpy(state, ioapic, sizeof(struct kvm_ioapic_state));
|
memcpy(state, ioapic, sizeof(struct kvm_ioapic_state));
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -431,9 +431,9 @@ int kvm_set_ioapic(struct kvm *kvm, struct kvm_ioapic_state *state)
|
||||||
if (!ioapic)
|
if (!ioapic)
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
|
|
||||||
mutex_lock(&ioapic->lock);
|
spin_lock(&ioapic->lock);
|
||||||
memcpy(ioapic, state, sizeof(struct kvm_ioapic_state));
|
memcpy(ioapic, state, sizeof(struct kvm_ioapic_state));
|
||||||
update_handled_vectors(ioapic);
|
update_handled_vectors(ioapic);
|
||||||
mutex_unlock(&ioapic->lock);
|
spin_unlock(&ioapic->lock);
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
|
@ -45,7 +45,7 @@ struct kvm_ioapic {
|
||||||
struct kvm_io_device dev;
|
struct kvm_io_device dev;
|
||||||
struct kvm *kvm;
|
struct kvm *kvm;
|
||||||
void (*ack_notifier)(void *opaque, int irq);
|
void (*ack_notifier)(void *opaque, int irq);
|
||||||
struct mutex lock;
|
spinlock_t lock;
|
||||||
DECLARE_BITMAP(handled_vectors, 256);
|
DECLARE_BITMAP(handled_vectors, 256);
|
||||||
};
|
};
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue