mirror of
https://gitlab.freedesktop.org/mesa/drm.git
synced 2026-01-25 13:50:26 +01:00
Buffer object reply fill in.
Lindent of drm_bo.c drm_ttm.c
This commit is contained in:
parent
de144ba23c
commit
033bda07e9
5 changed files with 188 additions and 163 deletions
|
|
@ -955,11 +955,13 @@ typedef struct drm_buffer_object{
|
|||
unsigned long num_pages;
|
||||
unsigned long buffer_start;
|
||||
drm_bo_type_t type;
|
||||
unsigned long offset;
|
||||
|
||||
atomic_t mapped;
|
||||
uint32_t map_flags;
|
||||
uint32_t flags;
|
||||
uint32_t mask;
|
||||
uint32_t mask_hint;
|
||||
uint32_t hint;
|
||||
|
||||
drm_mm_node_t *vram;
|
||||
drm_mm_node_t *tt;
|
||||
|
|
|
|||
|
|
@ -55,7 +55,7 @@
|
|||
* 2.) Refer to ttm locking orders.
|
||||
*/
|
||||
|
||||
int drm_fence_buffer_objects(drm_file_t *priv)
|
||||
int drm_fence_buffer_objects(drm_file_t * priv)
|
||||
{
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
drm_buffer_manager_t *bm = &dev->bm;
|
||||
|
|
@ -69,7 +69,7 @@ int drm_fence_buffer_objects(drm_file_t *priv)
|
|||
mutex_lock(&bm->bm_mutex);
|
||||
|
||||
list_for_each_entry(entry, &bm->unfenced, head) {
|
||||
BUG_ON(!entry->unfenced);
|
||||
BUG_ON(!entry->unfenced);
|
||||
fence_flags |= entry->fence_flags;
|
||||
count++;
|
||||
}
|
||||
|
|
@ -85,7 +85,7 @@ int drm_fence_buffer_objects(drm_file_t *priv)
|
|||
mutex_unlock(&bm->bm_mutex);
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
||||
|
||||
ret = drm_fence_object_init(dev, fence_flags, 1, fence);
|
||||
if (ret) {
|
||||
drm_free(fence, sizeof(*fence), DRM_MEM_FENCE);
|
||||
|
|
@ -120,24 +120,23 @@ int drm_fence_buffer_objects(drm_file_t *priv)
|
|||
* dev locked.
|
||||
*/
|
||||
|
||||
|
||||
static int drm_move_tt_to_local(drm_buffer_object_t *buf, int lazy)
|
||||
static int drm_move_tt_to_local(drm_buffer_object_t * buf, int lazy)
|
||||
{
|
||||
drm_device_t *dev = buf->dev;
|
||||
drm_buffer_manager_t *bm = &dev->bm;
|
||||
int ret = 0;
|
||||
|
||||
|
||||
BUG_ON(!buf->tt);
|
||||
|
||||
if (buf->fence) {
|
||||
ret = drm_fence_object_wait(dev, buf->fence, lazy, !lazy,
|
||||
ret = drm_fence_object_wait(dev, buf->fence, lazy, !lazy,
|
||||
buf->fence_flags);
|
||||
if (ret)
|
||||
return ret;
|
||||
drm_fence_usage_deref_unlocked(dev, buf->fence);
|
||||
buf->fence = NULL;
|
||||
}
|
||||
|
||||
|
||||
drm_unbind_ttm_region(buf->ttm_region);
|
||||
drm_mm_put_block(&bm->tt_manager, buf->tt);
|
||||
buf->tt = NULL;
|
||||
|
|
@ -147,10 +146,9 @@ static int drm_move_tt_to_local(drm_buffer_object_t *buf, int lazy)
|
|||
return 0;
|
||||
}
|
||||
|
||||
|
||||
static void drm_bo_destroy_locked(drm_device_t *dev, drm_buffer_object_t *bo)
|
||||
static void drm_bo_destroy_locked(drm_device_t * dev, drm_buffer_object_t * bo)
|
||||
{
|
||||
|
||||
|
||||
drm_buffer_manager_t *bm = &dev->bm;
|
||||
|
||||
BUG_ON(bo->unfenced);
|
||||
|
|
@ -189,15 +187,14 @@ static void drm_bo_destroy_locked(drm_device_t *dev, drm_buffer_object_t *bo)
|
|||
drm_free(bo, sizeof(*bo), DRM_MEM_BUFOBJ);
|
||||
}
|
||||
|
||||
|
||||
void drm_bo_usage_deref_locked(drm_device_t *dev, drm_buffer_object_t *bo)
|
||||
void drm_bo_usage_deref_locked(drm_device_t * dev, drm_buffer_object_t * bo)
|
||||
{
|
||||
if (atomic_dec_and_test(&bo->usage)) {
|
||||
drm_bo_destroy_locked(dev, bo);
|
||||
}
|
||||
}
|
||||
|
||||
void drm_bo_usage_deref_unlocked(drm_device_t *dev, drm_buffer_object_t *bo)
|
||||
void drm_bo_usage_deref_unlocked(drm_device_t * dev, drm_buffer_object_t * bo)
|
||||
{
|
||||
if (atomic_dec_and_test(&bo->usage)) {
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
|
|
@ -207,20 +204,20 @@ void drm_bo_usage_deref_unlocked(drm_device_t *dev, drm_buffer_object_t *bo)
|
|||
}
|
||||
}
|
||||
|
||||
static void drm_bo_base_deref_locked(drm_file_t *priv, drm_user_object_t *uo)
|
||||
static void drm_bo_base_deref_locked(drm_file_t * priv, drm_user_object_t * uo)
|
||||
{
|
||||
drm_bo_usage_deref_locked(priv->head->dev,
|
||||
drm_user_object_entry(uo, drm_buffer_object_t, base));
|
||||
drm_bo_usage_deref_locked(priv->head->dev,
|
||||
drm_user_object_entry(uo, drm_buffer_object_t,
|
||||
base));
|
||||
}
|
||||
|
||||
|
||||
static int drm_bo_new_flags(drm_bo_driver_t *driver,
|
||||
static int drm_bo_new_flags(drm_bo_driver_t * driver,
|
||||
uint32_t flags, uint32_t new_mask, uint32_t hint,
|
||||
int init, uint32_t *n_flags)
|
||||
int init, uint32_t * n_flags)
|
||||
{
|
||||
uint32_t new_flags;
|
||||
uint32_t new_props;
|
||||
|
||||
|
||||
if (!(flags & new_mask & DRM_BO_MASK_MEM) || init) {
|
||||
|
||||
/*
|
||||
|
|
@ -234,10 +231,11 @@ static int drm_bo_new_flags(drm_bo_driver_t *driver,
|
|||
DRM_ERROR("Invalid buffer object memory flags\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
|
||||
if (new_flags & DRM_BO_FLAG_MEM_LOCAL) {
|
||||
if ((hint & DRM_BO_HINT_AVOID_LOCAL) &&
|
||||
new_flags & (DRM_BO_FLAG_MEM_VRAM | DRM_BO_FLAG_MEM_TT)) {
|
||||
if ((hint & DRM_BO_HINT_AVOID_LOCAL) &&
|
||||
new_flags & (DRM_BO_FLAG_MEM_VRAM |
|
||||
DRM_BO_FLAG_MEM_TT)) {
|
||||
new_flags &= ~DRM_BO_FLAG_MEM_LOCAL;
|
||||
} else {
|
||||
new_flags = DRM_BO_FLAG_MEM_LOCAL;
|
||||
|
|
@ -254,7 +252,7 @@ static int drm_bo_new_flags(drm_bo_driver_t *driver,
|
|||
} else {
|
||||
new_flags = flags & DRM_BO_MASK_MEM;
|
||||
}
|
||||
|
||||
|
||||
new_props = new_mask & (DRM_BO_FLAG_EXE | DRM_BO_FLAG_WRITE |
|
||||
DRM_BO_FLAG_READ);
|
||||
|
||||
|
|
@ -268,14 +266,16 @@ static int drm_bo_new_flags(drm_bo_driver_t *driver,
|
|||
if (hint & DRM_BO_HINT_BIND_CACHED) {
|
||||
new_flags |= DRM_BO_FLAG_CACHED;
|
||||
if (((new_flags & DRM_BO_FLAG_MEM_TT) && !driver->cached_tt) ||
|
||||
((new_flags & DRM_BO_FLAG_MEM_VRAM) && !driver->cached_vram))
|
||||
((new_flags & DRM_BO_FLAG_MEM_VRAM)
|
||||
&& !driver->cached_vram))
|
||||
new_flags &= ~DRM_BO_FLAG_CACHED;
|
||||
}
|
||||
|
||||
|
||||
if ((new_flags & DRM_BO_FLAG_NO_EVICT) &&
|
||||
((flags ^ new_flags) & DRM_BO_FLAG_CACHED)) {
|
||||
if (flags & DRM_BO_FLAG_CACHED) {
|
||||
DRM_ERROR("Cannot change caching policy of pinned buffer\n");
|
||||
DRM_ERROR
|
||||
("Cannot change caching policy of pinned buffer\n");
|
||||
return -EINVAL;
|
||||
} else {
|
||||
new_flags &= ~DRM_BO_FLAG_CACHED;
|
||||
|
|
@ -285,12 +285,10 @@ static int drm_bo_new_flags(drm_bo_driver_t *driver,
|
|||
*n_flags = new_flags;
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
|
||||
#if 0
|
||||
|
||||
static int drm_bo_evict(drm_device_t *dev, drm_buffer_object_t *buf, int tt);
|
||||
static int drm_bo_evict(drm_device_t * dev, drm_buffer_object_t * buf, int tt);
|
||||
{
|
||||
int ret;
|
||||
if (tt) {
|
||||
|
|
@ -300,8 +298,8 @@ static int drm_bo_evict(drm_device_t *dev, drm_buffer_object_t *buf, int tt);
|
|||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
int drm_bo_alloc_space(drm_device_t *dev, int tt, drm_buffer_object_t *buf)
|
||||
|
||||
int drm_bo_alloc_space(drm_device_t * dev, int tt, drm_buffer_object_t * buf)
|
||||
{
|
||||
drm_mm_node_t *node;
|
||||
drm_buffer_manager_t *bm = &dev->bm;
|
||||
|
|
@ -315,13 +313,14 @@ int drm_bo_alloc_space(drm_device_t *dev, int tt, drm_buffer_object_t *buf)
|
|||
if (node)
|
||||
break;
|
||||
|
||||
if (lru->next == lru)
|
||||
if (lru->next == lru)
|
||||
break;
|
||||
|
||||
if (tt) {
|
||||
bo = list_entry(lru->next, drm_buffer_object_t, tt_lru);
|
||||
} else {
|
||||
bo = list_entry(lru->next, drm_buffer_object_t, vram_lru);
|
||||
bo = list_entry(lru->next, drm_buffer_object_t,
|
||||
vram_lru);
|
||||
}
|
||||
|
||||
drm_bo_evict(dev, bo, tt);
|
||||
|
|
@ -344,22 +343,20 @@ int drm_bo_alloc_space(drm_device_t *dev, int tt, drm_buffer_object_t *buf)
|
|||
return 0;
|
||||
}
|
||||
#endif
|
||||
|
||||
|
||||
/*
|
||||
* Call dev->struct_mutex locked.
|
||||
*/
|
||||
|
||||
|
||||
drm_buffer_object_t *drm_lookup_buffer_object(drm_file_t *priv, uint32_t handle,
|
||||
int check_owner)
|
||||
drm_buffer_object_t *drm_lookup_buffer_object(drm_file_t * priv,
|
||||
uint32_t handle, int check_owner)
|
||||
{
|
||||
drm_user_object_t *uo;
|
||||
drm_buffer_object_t *bo;
|
||||
|
||||
uo = drm_lookup_user_object(priv, handle);
|
||||
|
||||
if (!uo || (uo->type != drm_buffer_type))
|
||||
if (!uo || (uo->type != drm_buffer_type))
|
||||
return NULL;
|
||||
|
||||
if (check_owner && priv != uo->owner) {
|
||||
|
|
@ -371,15 +368,15 @@ drm_buffer_object_t *drm_lookup_buffer_object(drm_file_t *priv, uint32_t handle,
|
|||
atomic_inc(&bo->usage);
|
||||
return bo;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Call bo->mutex locked.
|
||||
* Wait until the buffer is idle.
|
||||
*/
|
||||
|
||||
static int drm_bo_wait(drm_device_t *dev, drm_buffer_object_t *bo, int lazy)
|
||||
static int drm_bo_wait(drm_device_t * dev, drm_buffer_object_t * bo, int lazy)
|
||||
{
|
||||
|
||||
|
||||
drm_fence_object_t *fence = bo->fence;
|
||||
int ret;
|
||||
|
||||
|
|
@ -399,7 +396,9 @@ static int drm_bo_wait(drm_device_t *dev, drm_buffer_object_t *bo, int lazy)
|
|||
|
||||
atomic_inc(&fence->usage);
|
||||
mutex_unlock(&bo->mutex);
|
||||
ret = drm_fence_object_wait(dev, fence, lazy, !lazy, bo->fence_flags);
|
||||
ret =
|
||||
drm_fence_object_wait(dev, fence, lazy, !lazy,
|
||||
bo->fence_flags);
|
||||
mutex_lock(&bo->mutex);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
|
@ -410,7 +409,7 @@ static int drm_bo_wait(drm_device_t *dev, drm_buffer_object_t *bo, int lazy)
|
|||
bo->fence = NULL;
|
||||
}
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
}
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
|
@ -419,7 +418,7 @@ static int drm_bo_wait(drm_device_t *dev, drm_buffer_object_t *bo, int lazy)
|
|||
* Returns 1 if the buffer is currently rendered to or from. 0 otherwise.
|
||||
*/
|
||||
|
||||
static int drm_bo_busy(drm_device_t *dev, drm_buffer_object_t *bo)
|
||||
static int drm_bo_busy(drm_device_t * dev, drm_buffer_object_t * bo)
|
||||
{
|
||||
drm_fence_object_t *fence = bo->fence;
|
||||
|
||||
|
|
@ -439,7 +438,6 @@ static int drm_bo_busy(drm_device_t *dev, drm_buffer_object_t *bo)
|
|||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Wait for buffer idle and register that we've mapped the buffer.
|
||||
|
|
@ -448,13 +446,12 @@ static int drm_bo_busy(drm_device_t *dev, drm_buffer_object_t *bo)
|
|||
* unregistered.
|
||||
*/
|
||||
|
||||
|
||||
static int drm_buffer_object_map(drm_file_t *priv, uint32_t handle, int wait)
|
||||
static int drm_buffer_object_map(drm_file_t * priv, uint32_t handle, int wait)
|
||||
{
|
||||
drm_buffer_object_t *bo;
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
int ret;
|
||||
|
||||
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
bo = drm_lookup_buffer_object(priv, handle, 1);
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
|
|
@ -465,8 +462,7 @@ static int drm_buffer_object_map(drm_file_t *priv, uint32_t handle, int wait)
|
|||
mutex_lock(&bo->mutex);
|
||||
|
||||
if (!wait) {
|
||||
if ((atomic_read(&bo->mapped) == 0) &&
|
||||
drm_bo_busy(dev, bo)) {
|
||||
if ((atomic_read(&bo->mapped) == 0) && drm_bo_busy(dev, bo)) {
|
||||
mutex_unlock(&bo->mutex);
|
||||
ret = -EBUSY;
|
||||
goto out;
|
||||
|
|
@ -486,14 +482,13 @@ static int drm_buffer_object_map(drm_file_t *priv, uint32_t handle, int wait)
|
|||
atomic_inc(&bo->mapped);
|
||||
}
|
||||
mutex_unlock(&bo->mutex);
|
||||
|
||||
out:
|
||||
drm_bo_usage_deref_unlocked(dev,bo);
|
||||
|
||||
out:
|
||||
drm_bo_usage_deref_unlocked(dev, bo);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
static int drm_buffer_object_unmap(drm_file_t *priv, uint32_t handle)
|
||||
static int drm_buffer_object_unmap(drm_file_t * priv, uint32_t handle)
|
||||
{
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
drm_buffer_object_t *bo;
|
||||
|
|
@ -516,21 +511,22 @@ static int drm_buffer_object_unmap(drm_file_t *priv, uint32_t handle)
|
|||
|
||||
drm_remove_ref_object(priv, ro);
|
||||
drm_bo_usage_deref_locked(dev, bo);
|
||||
out:
|
||||
out:
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Call struct-sem locked.
|
||||
*/
|
||||
|
||||
static void drm_buffer_user_object_unmap(drm_file_t *priv, drm_user_object_t *uo,
|
||||
static void drm_buffer_user_object_unmap(drm_file_t * priv,
|
||||
drm_user_object_t * uo,
|
||||
drm_ref_t action)
|
||||
{
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
drm_buffer_object_t *bo = drm_user_object_entry(uo, drm_buffer_object_t, base);
|
||||
drm_buffer_object_t *bo =
|
||||
drm_user_object_entry(uo, drm_buffer_object_t, base);
|
||||
|
||||
BUG_ON(action != _DRM_REF_TYPE1);
|
||||
|
||||
|
|
@ -545,33 +541,32 @@ static void drm_buffer_user_object_unmap(drm_file_t *priv, drm_user_object_t *uo
|
|||
}
|
||||
}
|
||||
|
||||
static int drm_buffer_object_validate(drm_device_t *dev, drm_buffer_object_t *bo)
|
||||
static int drm_buffer_object_validate(drm_device_t * dev,
|
||||
drm_buffer_object_t * bo)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* Call bo->mutex locked.
|
||||
*/
|
||||
|
||||
static int drm_bo_add_ttm(drm_file_t *priv, drm_buffer_object_t *bo, uint32_t hint,
|
||||
uint32_t ttm_handle)
|
||||
|
||||
static int drm_bo_add_ttm(drm_file_t * priv, drm_buffer_object_t * bo,
|
||||
uint32_t hint, uint32_t ttm_handle)
|
||||
{
|
||||
drm_device_t *dev = bo->dev;
|
||||
drm_ttm_object_t *to = NULL;
|
||||
drm_ttm_t *ttm;
|
||||
int ret=0;
|
||||
int ret = 0;
|
||||
uint32_t ttm_flags = 0;
|
||||
|
||||
bo->ttm_object = NULL;
|
||||
bo->ttm_region = NULL;
|
||||
|
||||
switch(bo->type) {
|
||||
switch (bo->type) {
|
||||
case drm_bo_type_dc:
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
ret = drm_ttm_object_create(dev, bo->num_pages*PAGE_SIZE,
|
||||
ret = drm_ttm_object_create(dev, bo->num_pages * PAGE_SIZE,
|
||||
ttm_flags, &to);
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
break;
|
||||
|
|
@ -579,11 +574,11 @@ static int drm_bo_add_ttm(drm_file_t *priv, drm_buffer_object_t *bo, uint32_t hi
|
|||
mutex_lock(&dev->struct_mutex);
|
||||
to = drm_lookup_ttm_object(priv, ttm_handle, 1);
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
if (!to)
|
||||
if (!to)
|
||||
ret = -EINVAL;
|
||||
break;
|
||||
case drm_bo_type_user:
|
||||
|
||||
case drm_bo_type_fake:
|
||||
break;
|
||||
default:
|
||||
ret = -EINVAL;
|
||||
|
|
@ -597,8 +592,8 @@ static int drm_bo_add_ttm(drm_file_t *priv, drm_buffer_object_t *bo, uint32_t hi
|
|||
bo->ttm_object = to;
|
||||
ttm = drm_ttm_from_object(to);
|
||||
ret = drm_create_ttm_region(ttm, bo->buffer_start >> PAGE_SHIFT,
|
||||
bo->num_pages,
|
||||
hint & DRM_BO_HINT_BIND_CACHED,
|
||||
bo->num_pages,
|
||||
hint & DRM_BO_HINT_BIND_CACHED,
|
||||
&bo->ttm_region);
|
||||
if (ret) {
|
||||
drm_ttm_object_deref_unlocked(dev, to);
|
||||
|
|
@ -606,28 +601,27 @@ static int drm_bo_add_ttm(drm_file_t *priv, drm_buffer_object_t *bo, uint32_t hi
|
|||
}
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
int drm_buffer_object_create(drm_file_t *priv,
|
||||
int drm_buffer_object_create(drm_file_t * priv,
|
||||
unsigned long size,
|
||||
drm_bo_type_t type,
|
||||
uint32_t ttm_handle,
|
||||
uint32_t mask,
|
||||
uint32_t hint,
|
||||
unsigned long buffer_start,
|
||||
drm_buffer_object_t **buf_obj)
|
||||
drm_buffer_object_t ** buf_obj)
|
||||
{
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
drm_buffer_object_t *bo;
|
||||
int ret = 0;
|
||||
uint32_t new_flags;
|
||||
unsigned long num_pages;
|
||||
|
||||
|
||||
if (buffer_start & ~PAGE_MASK) {
|
||||
DRM_ERROR("Invalid buffer object start.\n");
|
||||
return -EINVAL;
|
||||
}
|
||||
num_pages = (size + PAGE_SIZE -1) >> PAGE_SHIFT;
|
||||
num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
|
||||
if (num_pages == 0) {
|
||||
DRM_ERROR("Illegal buffer object size.\n");
|
||||
return -EINVAL;
|
||||
|
|
@ -653,14 +647,14 @@ int drm_buffer_object_create(drm_file_t *priv,
|
|||
|
||||
ret = drm_bo_new_flags(dev->driver->bo_driver, bo->flags, mask, hint,
|
||||
1, &new_flags);
|
||||
if (ret)
|
||||
if (ret)
|
||||
goto out_err;
|
||||
ret = drm_bo_add_ttm(priv, bo, new_flags, ttm_handle);
|
||||
if (ret)
|
||||
if (ret)
|
||||
goto out_err;
|
||||
|
||||
bo->mask = mask;
|
||||
bo->mask_hint = hint;
|
||||
bo->hint = hint;
|
||||
|
||||
ret = drm_buffer_object_validate(dev, bo);
|
||||
if (ret)
|
||||
|
|
@ -669,14 +663,14 @@ int drm_buffer_object_create(drm_file_t *priv,
|
|||
mutex_unlock(&bo->mutex);
|
||||
*buf_obj = bo;
|
||||
return 0;
|
||||
|
||||
out_err:
|
||||
|
||||
out_err:
|
||||
mutex_unlock(&bo->mutex);
|
||||
drm_free(bo, sizeof(*bo), DRM_MEM_BUFOBJ);
|
||||
return ret;
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int drm_bo_add_user_object(drm_file_t *priv, drm_buffer_object_t *bo,
|
||||
static int drm_bo_add_user_object(drm_file_t * priv, drm_buffer_object_t * bo,
|
||||
int shareable)
|
||||
{
|
||||
drm_device_t *dev = priv->head->dev;
|
||||
|
|
@ -691,13 +685,31 @@ static int drm_bo_add_user_object(drm_file_t *priv, drm_buffer_object_t *bo,
|
|||
bo->base.type = drm_buffer_type;
|
||||
bo->base.ref_struct_locked = NULL;
|
||||
bo->base.unref = drm_buffer_user_object_unmap;
|
||||
|
||||
out:
|
||||
|
||||
out:
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
||||
|
||||
static void drm_bo_fill_rep_arg(const drm_buffer_object_t * bo,
|
||||
drm_bo_arg_reply_t * rep)
|
||||
{
|
||||
rep->handle = bo->base.hash.key;
|
||||
rep->flags = bo->flags;
|
||||
rep->size = bo->num_pages * PAGE_SIZE;
|
||||
rep->offset = bo->offset;
|
||||
|
||||
if (bo->ttm_object) {
|
||||
rep->arg_handle = bo->ttm_object->map_list.user_token;
|
||||
} else {
|
||||
rep->arg_handle = 0;
|
||||
}
|
||||
|
||||
rep->map_flags = bo->map_flags;
|
||||
rep->mask = bo->mask;
|
||||
rep->hint = bo->hint;
|
||||
rep->buffer_start = bo->buffer_start;
|
||||
}
|
||||
|
||||
int drm_bo_ioctl(DRM_IOCTL_ARGS)
|
||||
{
|
||||
|
|
@ -714,34 +726,45 @@ int drm_bo_ioctl(DRM_IOCTL_ARGS)
|
|||
rep.ret = 0;
|
||||
rep.handled = 0;
|
||||
switch (req->op) {
|
||||
case drm_bo_create: {
|
||||
unsigned long buffer_start = req->buffer_start;
|
||||
rep.ret = drm_buffer_object_create(priv, req->size,
|
||||
req->type, req->arg_handle,
|
||||
req->mask, req->hint,
|
||||
buffer_start,
|
||||
&entry);
|
||||
if (rep.ret)
|
||||
case drm_bo_create:{
|
||||
unsigned long buffer_start = req->buffer_start;
|
||||
rep.ret =
|
||||
drm_buffer_object_create(priv, req->size,
|
||||
req->type,
|
||||
req->arg_handle,
|
||||
req->mask,
|
||||
req->hint,
|
||||
buffer_start,
|
||||
&entry);
|
||||
if (rep.ret)
|
||||
break;
|
||||
|
||||
rep.ret =
|
||||
drm_bo_add_user_object(priv, entry,
|
||||
req->
|
||||
mask &
|
||||
DRM_BO_FLAG_SHAREABLE);
|
||||
if (rep.ret)
|
||||
drm_bo_usage_deref_unlocked(dev, entry);
|
||||
|
||||
mutex_lock(&entry->mutex);
|
||||
drm_bo_fill_rep_arg(entry, &rep);
|
||||
mutex_unlock(&entry->mutex);
|
||||
break;
|
||||
|
||||
rep.ret = drm_bo_add_user_object(priv, entry, req->mask &
|
||||
DRM_BO_FLAG_SHAREABLE);
|
||||
if (rep.ret)
|
||||
drm_bo_usage_deref_unlocked(dev, entry);
|
||||
break;
|
||||
}
|
||||
}
|
||||
case drm_bo_unmap:
|
||||
rep.ret = drm_buffer_object_unmap(priv, req->handle);
|
||||
break;
|
||||
case drm_bo_map:
|
||||
rep.ret = drm_buffer_object_map(priv, req->handle,
|
||||
!(req->hint &
|
||||
rep.ret = drm_buffer_object_map(priv, req->handle,
|
||||
!(req->hint &
|
||||
DRM_BO_HINT_DONT_BLOCK));
|
||||
break;
|
||||
case drm_bo_destroy:
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
uo = drm_lookup_user_object(priv, req->handle);
|
||||
if (!uo || (uo->type != drm_buffer_type) || uo->owner != priv) {
|
||||
if (!uo || (uo->type != drm_buffer_type)
|
||||
|| uo->owner != priv) {
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
rep.ret = -EINVAL;
|
||||
break;
|
||||
|
|
@ -749,19 +772,24 @@ int drm_bo_ioctl(DRM_IOCTL_ARGS)
|
|||
rep.ret = drm_remove_user_object(priv, uo);
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
break;
|
||||
case drm_bo_reference:
|
||||
rep.ret = drm_user_object_ref(priv, req->handle,
|
||||
case drm_bo_reference:
|
||||
rep.ret = drm_user_object_ref(priv, req->handle,
|
||||
drm_buffer_type, &uo);
|
||||
if (rep.ret)
|
||||
break;
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
uo = drm_lookup_user_object(priv, req->handle);
|
||||
entry = drm_user_object_entry(uo, drm_buffer_object_t, base);
|
||||
entry =
|
||||
drm_user_object_entry(uo, drm_buffer_object_t,
|
||||
base);
|
||||
atomic_dec(&entry->usage);
|
||||
mutex_unlock(&dev->struct_mutex);
|
||||
mutex_lock(&entry->mutex);
|
||||
drm_bo_fill_rep_arg(entry, &rep);
|
||||
mutex_unlock(&entry->mutex);
|
||||
break;
|
||||
case drm_bo_unreference:
|
||||
rep.ret = drm_user_object_unref(priv, req->handle,
|
||||
rep.ret = drm_user_object_unref(priv, req->handle,
|
||||
drm_buffer_type);
|
||||
break;
|
||||
default:
|
||||
|
|
@ -777,7 +805,3 @@ int drm_bo_ioctl(DRM_IOCTL_ARGS)
|
|||
return 0;
|
||||
|
||||
}
|
||||
|
||||
|
||||
|
||||
|
||||
|
|
|
|||
|
|
@ -42,7 +42,6 @@ typedef struct drm_val_action {
|
|||
int validated;
|
||||
} drm_val_action_t;
|
||||
|
||||
|
||||
/*
|
||||
* We may be manipulating other processes page tables, so for each TTM, keep track of
|
||||
* which mm_structs are currently mapping the ttm so that we can take the appropriate
|
||||
|
|
@ -204,7 +203,7 @@ int drm_destroy_ttm(drm_ttm_t * ttm)
|
|||
ttm->destroy = 1;
|
||||
DRM_ERROR("VMAs are still alive. Skipping destruction.\n");
|
||||
return -EBUSY;
|
||||
}
|
||||
}
|
||||
|
||||
DRM_ERROR("Destroying a ttm\n");
|
||||
if (ttm->be_list) {
|
||||
|
|
@ -263,7 +262,7 @@ int drm_destroy_ttm(drm_ttm_t * ttm)
|
|||
* FIXME: Avoid using vmalloc for the page- and page_flags tables?
|
||||
*/
|
||||
|
||||
static drm_ttm_t *drm_init_ttm(struct drm_device * dev, unsigned long size)
|
||||
static drm_ttm_t *drm_init_ttm(struct drm_device *dev, unsigned long size)
|
||||
{
|
||||
|
||||
drm_ttm_t *ttm;
|
||||
|
|
@ -354,7 +353,8 @@ static int drm_ttm_lock_mmap_sem(drm_ttm_t * ttm)
|
|||
DRM_MEM_TTM);
|
||||
cur_count = shared_count + 10;
|
||||
mm_list =
|
||||
drm_alloc(sizeof(*mm_list) * cur_count, DRM_MEM_TTM);
|
||||
drm_alloc(sizeof(*mm_list) * cur_count,
|
||||
DRM_MEM_TTM);
|
||||
if (!mm_list)
|
||||
return -ENOMEM;
|
||||
}
|
||||
|
|
@ -489,7 +489,7 @@ void drm_destroy_ttm_region(drm_ttm_backend_list_t * entry)
|
|||
if (be->needs_cache_adjust(be)) {
|
||||
int ret = drm_ttm_lock_mmap_sem(ttm);
|
||||
drm_ttm_lock_mm(ttm, 0, 1);
|
||||
unmap_vma_pages(ttm, entry->page_offset,
|
||||
unmap_vma_pages(ttm, entry->page_offset,
|
||||
entry->num_pages);
|
||||
drm_ttm_unlock_mm(ttm, 0, 1);
|
||||
drm_set_caching(ttm, entry->page_offset,
|
||||
|
|
@ -542,7 +542,8 @@ int drm_create_ttm_region(drm_ttm_t * ttm, unsigned long page_offset,
|
|||
if (!entry)
|
||||
return -ENOMEM;
|
||||
|
||||
be = ttm->dev->driver->bo_driver->create_ttm_backend_entry(ttm->dev, cached);
|
||||
be = ttm->dev->driver->bo_driver->create_ttm_backend_entry(ttm->dev,
|
||||
cached);
|
||||
if (!be) {
|
||||
drm_free(entry, sizeof(*entry), DRM_MEM_TTM);
|
||||
DRM_ERROR("Couldn't create backend.\n");
|
||||
|
|
@ -750,11 +751,10 @@ int drm_user_create_region(drm_device_t * dev, unsigned long start, int len,
|
|||
return 0;
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* dev->struct_mutex locked.
|
||||
*/
|
||||
static void drm_ttm_object_remove(drm_device_t *dev, drm_ttm_object_t *object)
|
||||
static void drm_ttm_object_remove(drm_device_t * dev, drm_ttm_object_t * object)
|
||||
{
|
||||
drm_map_list_t *list = &object->map_list;
|
||||
drm_local_map_t *map;
|
||||
|
|
@ -765,7 +765,7 @@ static void drm_ttm_object_remove(drm_device_t *dev, drm_ttm_object_t *object)
|
|||
map = list->map;
|
||||
|
||||
if (map) {
|
||||
drm_ttm_t *ttm = (drm_ttm_t *)map->offset;
|
||||
drm_ttm_t *ttm = (drm_ttm_t *) map->offset;
|
||||
if (ttm) {
|
||||
if (drm_destroy_ttm(ttm) != -EBUSY) {
|
||||
drm_free(map, sizeof(*map), DRM_MEM_TTM);
|
||||
|
|
@ -778,15 +778,14 @@ static void drm_ttm_object_remove(drm_device_t *dev, drm_ttm_object_t *object)
|
|||
drm_free(object, sizeof(*object), DRM_MEM_TTM);
|
||||
}
|
||||
|
||||
|
||||
void drm_ttm_object_deref_locked(drm_device_t *dev, drm_ttm_object_t *to)
|
||||
void drm_ttm_object_deref_locked(drm_device_t * dev, drm_ttm_object_t * to)
|
||||
{
|
||||
if (atomic_dec_and_test(&to->usage)) {
|
||||
drm_ttm_object_remove(dev, to);
|
||||
}
|
||||
}
|
||||
|
||||
void drm_ttm_object_deref_unlocked(drm_device_t *dev, drm_ttm_object_t *to)
|
||||
void drm_ttm_object_deref_unlocked(drm_device_t * dev, drm_ttm_object_t * to)
|
||||
{
|
||||
if (atomic_dec_and_test(&to->usage)) {
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
|
|
@ -796,26 +795,25 @@ void drm_ttm_object_deref_unlocked(drm_device_t *dev, drm_ttm_object_t *to)
|
|||
}
|
||||
}
|
||||
|
||||
|
||||
/*
|
||||
* dev->struct_mutex locked.
|
||||
*/
|
||||
static void drm_ttm_user_deref_locked(drm_file_t *priv, drm_user_object_t *base)
|
||||
static void drm_ttm_user_deref_locked(drm_file_t * priv,
|
||||
drm_user_object_t * base)
|
||||
{
|
||||
drm_ttm_object_deref_locked(priv->head->dev,
|
||||
drm_user_object_entry(base, drm_ttm_object_t,
|
||||
drm_user_object_entry(base,
|
||||
drm_ttm_object_t,
|
||||
base));
|
||||
}
|
||||
|
||||
|
||||
|
||||
/*
|
||||
* Create a ttm and add it to the drm book-keeping.
|
||||
* dev->struct_mutex locked.
|
||||
*/
|
||||
|
||||
int drm_ttm_object_create(drm_device_t *dev, unsigned long size,
|
||||
uint32_t flags, drm_ttm_object_t **ttm_object)
|
||||
int drm_ttm_object_create(drm_device_t * dev, unsigned long size,
|
||||
uint32_t flags, drm_ttm_object_t ** ttm_object)
|
||||
{
|
||||
drm_ttm_object_t *object;
|
||||
drm_map_list_t *list;
|
||||
|
|
@ -823,11 +821,11 @@ int drm_ttm_object_create(drm_device_t *dev, unsigned long size,
|
|||
drm_ttm_t *ttm;
|
||||
|
||||
object = drm_calloc(1, sizeof(*object), DRM_MEM_TTM);
|
||||
if (!object)
|
||||
if (!object)
|
||||
return -ENOMEM;
|
||||
object->flags = flags;
|
||||
list = &object->map_list;
|
||||
|
||||
|
||||
list->map = drm_calloc(1, sizeof(*map), DRM_MEM_TTM);
|
||||
if (!list->map) {
|
||||
drm_ttm_object_remove(dev, object);
|
||||
|
|
@ -847,9 +845,9 @@ int drm_ttm_object_create(drm_device_t *dev, unsigned long size,
|
|||
map->flags = _DRM_REMOVABLE;
|
||||
map->size = ttm->num_pages * PAGE_SIZE;
|
||||
map->handle = (void *)object;
|
||||
|
||||
if (drm_ht_just_insert_please(&dev->map_hash, &list->hash,
|
||||
(unsigned long) map->handle,
|
||||
|
||||
if (drm_ht_just_insert_please(&dev->map_hash, &list->hash,
|
||||
(unsigned long)map->handle,
|
||||
32 - PAGE_SHIFT - 3, PAGE_SHIFT,
|
||||
DRM_MAP_HASH_OFFSET)) {
|
||||
drm_ttm_object_remove(dev, object);
|
||||
|
|
@ -863,7 +861,7 @@ int drm_ttm_object_create(drm_device_t *dev, unsigned long size,
|
|||
return 0;
|
||||
}
|
||||
|
||||
drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t *priv, uint32_t handle,
|
||||
drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t * priv, uint32_t handle,
|
||||
int check_owner)
|
||||
{
|
||||
drm_user_object_t *uo;
|
||||
|
|
@ -871,7 +869,7 @@ drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t *priv, uint32_t handle,
|
|||
|
||||
uo = drm_lookup_user_object(priv, handle);
|
||||
|
||||
if (!uo || (uo->type != drm_ttm_type))
|
||||
if (!uo || (uo->type != drm_ttm_type))
|
||||
return NULL;
|
||||
|
||||
if (check_owner && priv != uo->owner) {
|
||||
|
|
@ -884,10 +882,9 @@ drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t *priv, uint32_t handle,
|
|||
return to;
|
||||
}
|
||||
|
||||
|
||||
int drm_ttm_ioctl(DRM_IOCTL_ARGS)
|
||||
{
|
||||
DRM_DEVICE;
|
||||
DRM_DEVICE;
|
||||
drm_ttm_arg_t arg;
|
||||
drm_ttm_object_t *entry;
|
||||
drm_user_object_t *uo;
|
||||
|
|
@ -895,8 +892,8 @@ int drm_ttm_ioctl(DRM_IOCTL_ARGS)
|
|||
int ret;
|
||||
|
||||
DRM_COPY_FROM_USER_IOCTL(arg, (void __user *)data, sizeof(arg));
|
||||
|
||||
switch(arg.op) {
|
||||
|
||||
switch (arg.op) {
|
||||
case drm_ttm_create:
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
size = arg.size;
|
||||
|
|
@ -905,7 +902,7 @@ int drm_ttm_ioctl(DRM_IOCTL_ARGS)
|
|||
mutex_unlock(&dev->struct_mutex);
|
||||
return ret;
|
||||
}
|
||||
ret = drm_add_user_object(priv, &entry->base,
|
||||
ret = drm_add_user_object(priv, &entry->base,
|
||||
arg.flags & DRM_TTM_FLAG_SHAREABLE);
|
||||
if (ret) {
|
||||
drm_ttm_object_remove(dev, entry);
|
||||
|
|
@ -923,7 +920,7 @@ int drm_ttm_ioctl(DRM_IOCTL_ARGS)
|
|||
if (ret)
|
||||
return ret;
|
||||
mutex_lock(&dev->struct_mutex);
|
||||
entry = drm_lookup_ttm_object(priv, arg.handle , 0);
|
||||
entry = drm_lookup_ttm_object(priv, arg.handle, 0);
|
||||
break;
|
||||
case drm_ttm_unreference:
|
||||
return drm_user_object_unref(priv, arg.handle, drm_ttm_type);
|
||||
|
|
|
|||
|
|
@ -98,7 +98,7 @@ typedef struct drm_ttm {
|
|||
drm_ttm_backend_list_t *be_list;
|
||||
atomic_t vma_count;
|
||||
int mmap_sem_locked;
|
||||
int destroy;
|
||||
int destroy;
|
||||
} drm_ttm_t;
|
||||
|
||||
typedef struct drm_ttm_object {
|
||||
|
|
@ -107,18 +107,18 @@ typedef struct drm_ttm_object {
|
|||
uint32_t flags;
|
||||
drm_map_list_t map_list;
|
||||
} drm_ttm_object_t;
|
||||
|
||||
extern int drm_ttm_object_create(struct drm_device *dev, unsigned long size,
|
||||
uint32_t flags, drm_ttm_object_t **ttm_object);
|
||||
extern void drm_ttm_object_deref_locked(struct drm_device *dev, drm_ttm_object_t *to);
|
||||
extern void drm_ttm_object_deref_unlocked(struct drm_device *dev, drm_ttm_object_t *to);
|
||||
extern drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t *priv, uint32_t handle,
|
||||
|
||||
extern int drm_ttm_object_create(struct drm_device *dev, unsigned long size,
|
||||
uint32_t flags,
|
||||
drm_ttm_object_t ** ttm_object);
|
||||
extern void drm_ttm_object_deref_locked(struct drm_device *dev,
|
||||
drm_ttm_object_t * to);
|
||||
extern void drm_ttm_object_deref_unlocked(struct drm_device *dev,
|
||||
drm_ttm_object_t * to);
|
||||
extern drm_ttm_object_t *drm_lookup_ttm_object(drm_file_t * priv,
|
||||
uint32_t handle,
|
||||
int check_owner);
|
||||
|
||||
|
||||
|
||||
|
||||
|
||||
/*
|
||||
* Bind a part of the ttm starting at page_offset size n_pages into the GTT, at
|
||||
* aperture offset aper_offset. The region handle will be used to reference this
|
||||
|
|
@ -166,7 +166,7 @@ extern int drm_ttm_add_mm_to_list(drm_ttm_t * ttm, struct mm_struct *mm);
|
|||
extern void drm_ttm_delete_mm(drm_ttm_t * ttm, struct mm_struct *mm);
|
||||
extern int drm_ttm_ioctl(DRM_IOCTL_ARGS);
|
||||
|
||||
static __inline__ drm_ttm_t *drm_ttm_from_object(drm_ttm_object_t *to)
|
||||
static __inline__ drm_ttm_t *drm_ttm_from_object(drm_ttm_object_t * to)
|
||||
{
|
||||
return (drm_ttm_t *) to->map_list.map->offset;
|
||||
}
|
||||
|
|
|
|||
|
|
@ -702,7 +702,8 @@ typedef struct drm_ttm_arg {
|
|||
typedef enum {
|
||||
drm_bo_type_ttm,
|
||||
drm_bo_type_dc,
|
||||
drm_bo_type_user
|
||||
drm_bo_type_user,
|
||||
drm_bo_type_fake
|
||||
}drm_bo_type_t;
|
||||
|
||||
|
||||
|
|
@ -823,12 +824,13 @@ typedef union drm_bo_arg{
|
|||
|
||||
/**
|
||||
* Device specific ioctls should only be in their respective headers
|
||||
* The device specific ioctl range is from 0x40 to 0x79.
|
||||
* The device specific ioctl range is from 0x40 to 0x99.
|
||||
* Generic IOCTLS restart at 0xA0.
|
||||
*
|
||||
* \sa drmCommandNone(), drmCommandRead(), drmCommandWrite(), and
|
||||
* drmCommandReadWrite().
|
||||
*/
|
||||
#define DRM_COMMAND_BASE 0x40
|
||||
#define DRM_COMMAND_END 0x80
|
||||
#define DRM_COMMAND_END 0xA0
|
||||
|
||||
#endif
|
||||
|
|
|
|||
Loading…
Add table
Reference in a new issue