Merge branch 'drm-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/airlied/drm-2.6
* 'drm-fixes' of git://git.kernel.org/pub/scm/linux/kernel/git/airlied/drm-2.6: drm/i915: Add missing mutex_lock(&dev->struct_mutex) drm/i915: fix WC mapping in non-GEM i915 code. drm/i915: Fix regression in 95ca9d drm/i915: Retire requests from i915_gem_busy_ioctl. drm/i915: suspend/resume GEM when KMS is active drm/i915: Don't let a device flush to prepare buffers clear new write_domains. drm/i915: Cut two args to set_to_gpu_domain that confused this tricky path.
This commit is contained in:
commit
8310509252
5 changed files with 67 additions and 27 deletions
|
@ -202,7 +202,7 @@ static int i915_initialize(struct drm_device * dev, drm_i915_init_t * init)
|
||||||
dev_priv->ring.map.flags = 0;
|
dev_priv->ring.map.flags = 0;
|
||||||
dev_priv->ring.map.mtrr = 0;
|
dev_priv->ring.map.mtrr = 0;
|
||||||
|
|
||||||
drm_core_ioremap(&dev_priv->ring.map, dev);
|
drm_core_ioremap_wc(&dev_priv->ring.map, dev);
|
||||||
|
|
||||||
if (dev_priv->ring.map.handle == NULL) {
|
if (dev_priv->ring.map.handle == NULL) {
|
||||||
i915_dma_cleanup(dev);
|
i915_dma_cleanup(dev);
|
||||||
|
|
|
@ -27,6 +27,7 @@
|
||||||
*
|
*
|
||||||
*/
|
*/
|
||||||
|
|
||||||
|
#include <linux/device.h>
|
||||||
#include "drmP.h"
|
#include "drmP.h"
|
||||||
#include "drm.h"
|
#include "drm.h"
|
||||||
#include "i915_drm.h"
|
#include "i915_drm.h"
|
||||||
|
@ -66,6 +67,12 @@ static int i915_suspend(struct drm_device *dev, pm_message_t state)
|
||||||
|
|
||||||
i915_save_state(dev);
|
i915_save_state(dev);
|
||||||
|
|
||||||
|
/* If KMS is active, we do the leavevt stuff here */
|
||||||
|
if (drm_core_check_feature(dev, DRIVER_MODESET) && i915_gem_idle(dev)) {
|
||||||
|
dev_err(&dev->pdev->dev, "GEM idle failed, aborting suspend\n");
|
||||||
|
return -EBUSY;
|
||||||
|
}
|
||||||
|
|
||||||
intel_opregion_free(dev);
|
intel_opregion_free(dev);
|
||||||
|
|
||||||
if (state.event == PM_EVENT_SUSPEND) {
|
if (state.event == PM_EVENT_SUSPEND) {
|
||||||
|
@ -79,6 +86,9 @@ static int i915_suspend(struct drm_device *dev, pm_message_t state)
|
||||||
|
|
||||||
static int i915_resume(struct drm_device *dev)
|
static int i915_resume(struct drm_device *dev)
|
||||||
{
|
{
|
||||||
|
struct drm_i915_private *dev_priv = dev->dev_private;
|
||||||
|
int ret = 0;
|
||||||
|
|
||||||
pci_set_power_state(dev->pdev, PCI_D0);
|
pci_set_power_state(dev->pdev, PCI_D0);
|
||||||
pci_restore_state(dev->pdev);
|
pci_restore_state(dev->pdev);
|
||||||
if (pci_enable_device(dev->pdev))
|
if (pci_enable_device(dev->pdev))
|
||||||
|
@ -89,7 +99,18 @@ static int i915_resume(struct drm_device *dev)
|
||||||
|
|
||||||
intel_opregion_init(dev);
|
intel_opregion_init(dev);
|
||||||
|
|
||||||
return 0;
|
/* KMS EnterVT equivalent */
|
||||||
|
if (drm_core_check_feature(dev, DRIVER_MODESET)) {
|
||||||
|
mutex_lock(&dev->struct_mutex);
|
||||||
|
dev_priv->mm.suspended = 0;
|
||||||
|
|
||||||
|
ret = i915_gem_init_ringbuffer(dev);
|
||||||
|
if (ret != 0)
|
||||||
|
ret = -1;
|
||||||
|
mutex_unlock(&dev->struct_mutex);
|
||||||
|
}
|
||||||
|
|
||||||
|
return ret;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct vm_operations_struct i915_gem_vm_ops = {
|
static struct vm_operations_struct i915_gem_vm_ops = {
|
||||||
|
|
|
@ -618,6 +618,7 @@ int i915_gem_init_ringbuffer(struct drm_device *dev);
|
||||||
void i915_gem_cleanup_ringbuffer(struct drm_device *dev);
|
void i915_gem_cleanup_ringbuffer(struct drm_device *dev);
|
||||||
int i915_gem_do_init(struct drm_device *dev, unsigned long start,
|
int i915_gem_do_init(struct drm_device *dev, unsigned long start,
|
||||||
unsigned long end);
|
unsigned long end);
|
||||||
|
int i915_gem_idle(struct drm_device *dev);
|
||||||
int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
|
int i915_gem_fault(struct vm_area_struct *vma, struct vm_fault *vmf);
|
||||||
int i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj,
|
int i915_gem_object_set_to_gtt_domain(struct drm_gem_object *obj,
|
||||||
int write);
|
int write);
|
||||||
|
|
|
@ -34,10 +34,6 @@
|
||||||
|
|
||||||
#define I915_GEM_GPU_DOMAINS (~(I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT))
|
#define I915_GEM_GPU_DOMAINS (~(I915_GEM_DOMAIN_CPU | I915_GEM_DOMAIN_GTT))
|
||||||
|
|
||||||
static void
|
|
||||||
i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj,
|
|
||||||
uint32_t read_domains,
|
|
||||||
uint32_t write_domain);
|
|
||||||
static void i915_gem_object_flush_gpu_write_domain(struct drm_gem_object *obj);
|
static void i915_gem_object_flush_gpu_write_domain(struct drm_gem_object *obj);
|
||||||
static void i915_gem_object_flush_gtt_write_domain(struct drm_gem_object *obj);
|
static void i915_gem_object_flush_gtt_write_domain(struct drm_gem_object *obj);
|
||||||
static void i915_gem_object_flush_cpu_write_domain(struct drm_gem_object *obj);
|
static void i915_gem_object_flush_cpu_write_domain(struct drm_gem_object *obj);
|
||||||
|
@ -2021,30 +2017,28 @@ i915_gem_object_set_to_cpu_domain(struct drm_gem_object *obj, int write)
|
||||||
* drm_agp_chipset_flush
|
* drm_agp_chipset_flush
|
||||||
*/
|
*/
|
||||||
static void
|
static void
|
||||||
i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj,
|
i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj)
|
||||||
uint32_t read_domains,
|
|
||||||
uint32_t write_domain)
|
|
||||||
{
|
{
|
||||||
struct drm_device *dev = obj->dev;
|
struct drm_device *dev = obj->dev;
|
||||||
struct drm_i915_gem_object *obj_priv = obj->driver_private;
|
struct drm_i915_gem_object *obj_priv = obj->driver_private;
|
||||||
uint32_t invalidate_domains = 0;
|
uint32_t invalidate_domains = 0;
|
||||||
uint32_t flush_domains = 0;
|
uint32_t flush_domains = 0;
|
||||||
|
|
||||||
BUG_ON(read_domains & I915_GEM_DOMAIN_CPU);
|
BUG_ON(obj->pending_read_domains & I915_GEM_DOMAIN_CPU);
|
||||||
BUG_ON(write_domain == I915_GEM_DOMAIN_CPU);
|
BUG_ON(obj->pending_write_domain == I915_GEM_DOMAIN_CPU);
|
||||||
|
|
||||||
#if WATCH_BUF
|
#if WATCH_BUF
|
||||||
DRM_INFO("%s: object %p read %08x -> %08x write %08x -> %08x\n",
|
DRM_INFO("%s: object %p read %08x -> %08x write %08x -> %08x\n",
|
||||||
__func__, obj,
|
__func__, obj,
|
||||||
obj->read_domains, read_domains,
|
obj->read_domains, obj->pending_read_domains,
|
||||||
obj->write_domain, write_domain);
|
obj->write_domain, obj->pending_write_domain);
|
||||||
#endif
|
#endif
|
||||||
/*
|
/*
|
||||||
* If the object isn't moving to a new write domain,
|
* If the object isn't moving to a new write domain,
|
||||||
* let the object stay in multiple read domains
|
* let the object stay in multiple read domains
|
||||||
*/
|
*/
|
||||||
if (write_domain == 0)
|
if (obj->pending_write_domain == 0)
|
||||||
read_domains |= obj->read_domains;
|
obj->pending_read_domains |= obj->read_domains;
|
||||||
else
|
else
|
||||||
obj_priv->dirty = 1;
|
obj_priv->dirty = 1;
|
||||||
|
|
||||||
|
@ -2054,15 +2048,17 @@ i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj,
|
||||||
* any read domains which differ from the old
|
* any read domains which differ from the old
|
||||||
* write domain
|
* write domain
|
||||||
*/
|
*/
|
||||||
if (obj->write_domain && obj->write_domain != read_domains) {
|
if (obj->write_domain &&
|
||||||
|
obj->write_domain != obj->pending_read_domains) {
|
||||||
flush_domains |= obj->write_domain;
|
flush_domains |= obj->write_domain;
|
||||||
invalidate_domains |= read_domains & ~obj->write_domain;
|
invalidate_domains |=
|
||||||
|
obj->pending_read_domains & ~obj->write_domain;
|
||||||
}
|
}
|
||||||
/*
|
/*
|
||||||
* Invalidate any read caches which may have
|
* Invalidate any read caches which may have
|
||||||
* stale data. That is, any new read domains.
|
* stale data. That is, any new read domains.
|
||||||
*/
|
*/
|
||||||
invalidate_domains |= read_domains & ~obj->read_domains;
|
invalidate_domains |= obj->pending_read_domains & ~obj->read_domains;
|
||||||
if ((flush_domains | invalidate_domains) & I915_GEM_DOMAIN_CPU) {
|
if ((flush_domains | invalidate_domains) & I915_GEM_DOMAIN_CPU) {
|
||||||
#if WATCH_BUF
|
#if WATCH_BUF
|
||||||
DRM_INFO("%s: CPU domain flush %08x invalidate %08x\n",
|
DRM_INFO("%s: CPU domain flush %08x invalidate %08x\n",
|
||||||
|
@ -2071,9 +2067,15 @@ i915_gem_object_set_to_gpu_domain(struct drm_gem_object *obj,
|
||||||
i915_gem_clflush_object(obj);
|
i915_gem_clflush_object(obj);
|
||||||
}
|
}
|
||||||
|
|
||||||
if ((write_domain | flush_domains) != 0)
|
/* The actual obj->write_domain will be updated with
|
||||||
obj->write_domain = write_domain;
|
* pending_write_domain after we emit the accumulated flush for all
|
||||||
obj->read_domains = read_domains;
|
* of our domain changes in execbuffers (which clears objects'
|
||||||
|
* write_domains). So if we have a current write domain that we
|
||||||
|
* aren't changing, set pending_write_domain to that.
|
||||||
|
*/
|
||||||
|
if (flush_domains == 0 && obj->pending_write_domain == 0)
|
||||||
|
obj->pending_write_domain = obj->write_domain;
|
||||||
|
obj->read_domains = obj->pending_read_domains;
|
||||||
|
|
||||||
dev->invalidate_domains |= invalidate_domains;
|
dev->invalidate_domains |= invalidate_domains;
|
||||||
dev->flush_domains |= flush_domains;
|
dev->flush_domains |= flush_domains;
|
||||||
|
@ -2583,9 +2585,7 @@ i915_gem_execbuffer(struct drm_device *dev, void *data,
|
||||||
struct drm_gem_object *obj = object_list[i];
|
struct drm_gem_object *obj = object_list[i];
|
||||||
|
|
||||||
/* Compute new gpu domains and update invalidate/flush */
|
/* Compute new gpu domains and update invalidate/flush */
|
||||||
i915_gem_object_set_to_gpu_domain(obj,
|
i915_gem_object_set_to_gpu_domain(obj);
|
||||||
obj->pending_read_domains,
|
|
||||||
obj->pending_write_domain);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
i915_verify_inactive(dev, __FILE__, __LINE__);
|
i915_verify_inactive(dev, __FILE__, __LINE__);
|
||||||
|
@ -2604,6 +2604,12 @@ i915_gem_execbuffer(struct drm_device *dev, void *data,
|
||||||
(void)i915_add_request(dev, dev->flush_domains);
|
(void)i915_add_request(dev, dev->flush_domains);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
for (i = 0; i < args->buffer_count; i++) {
|
||||||
|
struct drm_gem_object *obj = object_list[i];
|
||||||
|
|
||||||
|
obj->write_domain = obj->pending_write_domain;
|
||||||
|
}
|
||||||
|
|
||||||
i915_verify_inactive(dev, __FILE__, __LINE__);
|
i915_verify_inactive(dev, __FILE__, __LINE__);
|
||||||
|
|
||||||
#if WATCH_COHERENCY
|
#if WATCH_COHERENCY
|
||||||
|
@ -2866,6 +2872,13 @@ i915_gem_busy_ioctl(struct drm_device *dev, void *data,
|
||||||
return -EBADF;
|
return -EBADF;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* Update the active list for the hardware's current position.
|
||||||
|
* Otherwise this only updates on a delayed timer or when irqs are
|
||||||
|
* actually unmasked, and our working set ends up being larger than
|
||||||
|
* required.
|
||||||
|
*/
|
||||||
|
i915_gem_retire_requests(dev);
|
||||||
|
|
||||||
obj_priv = obj->driver_private;
|
obj_priv = obj->driver_private;
|
||||||
/* Don't count being on the flushing list against the object being
|
/* Don't count being on the flushing list against the object being
|
||||||
* done. Otherwise, a buffer left on the flushing list but not getting
|
* done. Otherwise, a buffer left on the flushing list but not getting
|
||||||
|
@ -2967,7 +2980,7 @@ i915_gem_evict_from_list(struct drm_device *dev, struct list_head *head)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int
|
int
|
||||||
i915_gem_idle(struct drm_device *dev)
|
i915_gem_idle(struct drm_device *dev)
|
||||||
{
|
{
|
||||||
drm_i915_private_t *dev_priv = dev->dev_private;
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||||||
|
@ -3130,16 +3143,20 @@ static void
|
||||||
i915_gem_cleanup_hws(struct drm_device *dev)
|
i915_gem_cleanup_hws(struct drm_device *dev)
|
||||||
{
|
{
|
||||||
drm_i915_private_t *dev_priv = dev->dev_private;
|
drm_i915_private_t *dev_priv = dev->dev_private;
|
||||||
struct drm_gem_object *obj = dev_priv->hws_obj;
|
struct drm_gem_object *obj;
|
||||||
struct drm_i915_gem_object *obj_priv = obj->driver_private;
|
struct drm_i915_gem_object *obj_priv;
|
||||||
|
|
||||||
if (dev_priv->hws_obj == NULL)
|
if (dev_priv->hws_obj == NULL)
|
||||||
return;
|
return;
|
||||||
|
|
||||||
|
obj = dev_priv->hws_obj;
|
||||||
|
obj_priv = obj->driver_private;
|
||||||
|
|
||||||
kunmap(obj_priv->page_list[0]);
|
kunmap(obj_priv->page_list[0]);
|
||||||
i915_gem_object_unpin(obj);
|
i915_gem_object_unpin(obj);
|
||||||
drm_gem_object_unreference(obj);
|
drm_gem_object_unreference(obj);
|
||||||
dev_priv->hws_obj = NULL;
|
dev_priv->hws_obj = NULL;
|
||||||
|
|
||||||
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
||||||
dev_priv->hw_status_page = NULL;
|
dev_priv->hw_status_page = NULL;
|
||||||
|
|
||||||
|
|
|
@ -1008,6 +1008,7 @@ static int intel_crtc_cursor_set(struct drm_crtc *crtc,
|
||||||
temp = CURSOR_MODE_DISABLE;
|
temp = CURSOR_MODE_DISABLE;
|
||||||
addr = 0;
|
addr = 0;
|
||||||
bo = NULL;
|
bo = NULL;
|
||||||
|
mutex_lock(&dev->struct_mutex);
|
||||||
goto finish;
|
goto finish;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
Loading…
Add table
Reference in a new issue