mirror of
https://gitlab.freedesktop.org/mesa/drm.git
synced 2025-12-25 00:10:11 +01:00
DRM_IOCTL_VIA_DMA_INIT DRM_IOCTL_VIA_CMDBUFFER DRM_IOCTL_VIA_FLUSH
The first ioctl sets up an area in AGP memory that will be used as the ring
buffer. The second ioctl copies a command buffer from user space memory
to the ring buffer. The third ioctl waits for engine idle until it
returns.
The motivation for this patch is to avoid the wait for engine idle call
before each buffer flush in the current DRI driver. With this patch,
the DRI driver can continue to flush its buffer as long as there is
free space in the ring buffer.
This patch adds an additional copy operation on the command buffer. This
buffer copying is necessary to support multiple DRI clients rendering
simultaneously. Otherwise, more CPU time will be spent in the busy loop
waiting for engine idle between DRI context switch. Even in the single
client case, the tradeoff is reasonable in comparision to the kernel
call to check for free buffer space for the client to render directly
to the ring buffer.
1198 lines
32 KiB
C
1198 lines
32 KiB
C
/**
|
|
* \file drm_drv.h
|
|
* Generic driver template
|
|
*
|
|
* \author Rickard E. (Rik) Faith <faith@valinux.com>
|
|
* \author Gareth Hughes <gareth@valinux.com>
|
|
*
|
|
* To use this template, you must at least define the following (samples
|
|
* given for the MGA driver):
|
|
*
|
|
* \code
|
|
* #define DRIVER_AUTHOR "VA Linux Systems, Inc."
|
|
*
|
|
* #define DRIVER_NAME "mga"
|
|
* #define DRIVER_DESC "Matrox G200/G400"
|
|
* #define DRIVER_DATE "20001127"
|
|
*
|
|
* #define DRIVER_MAJOR 2
|
|
* #define DRIVER_MINOR 0
|
|
* #define DRIVER_PATCHLEVEL 2
|
|
*
|
|
* #define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( mga_ioctls )
|
|
*
|
|
* #define DRM(x) mga_##x
|
|
* \endcode
|
|
*/
|
|
|
|
/*
|
|
* Created: Thu Nov 23 03:10:50 2000 by gareth@valinux.com
|
|
*
|
|
* Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas.
|
|
* Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California.
|
|
* All Rights Reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#ifndef __MUST_HAVE_AGP
|
|
#define __MUST_HAVE_AGP 0
|
|
#endif
|
|
#ifndef __HAVE_CTX_BITMAP
|
|
#define __HAVE_CTX_BITMAP 0
|
|
#endif
|
|
#ifndef __HAVE_IRQ
|
|
#define __HAVE_IRQ 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_QUEUE
|
|
#define __HAVE_DMA_QUEUE 0
|
|
#endif
|
|
#ifndef __HAVE_MULTIPLE_DMA_QUEUES
|
|
#define __HAVE_MULTIPLE_DMA_QUEUES 0
|
|
#endif
|
|
#ifndef __HAVE_COUNTERS
|
|
#define __HAVE_COUNTERS 0
|
|
#endif
|
|
#ifndef __HAVE_SG
|
|
#define __HAVE_SG 0
|
|
#endif
|
|
|
|
#ifndef DRIVER_IOCTLS
|
|
#define DRIVER_IOCTLS
|
|
#endif
|
|
|
|
static void __exit drm_cleanup( drm_device_t *dev );
|
|
|
|
/** Stub information */
|
|
struct drm_stub_info {
|
|
int (*info_register)(const char *name, struct file_operations *fops,
|
|
drm_device_t *dev);
|
|
int (*info_unregister)(int minor);
|
|
struct class_simple *drm_class;
|
|
int *info_count;
|
|
struct proc_dir_entry *proc_root;
|
|
};
|
|
extern struct drm_stub_info DRM(stub_info);
|
|
|
|
#ifndef MODULE
|
|
/** Use an additional macro to avoid preprocessor troubles */
|
|
#define DRM_OPTIONS_FUNC DRM(options)
|
|
/**
|
|
* Called by the kernel to parse command-line options passed via the
|
|
* boot-loader (e.g., LILO). It calls the insmod option routine,
|
|
* parse_options().
|
|
*/
|
|
static int __init DRM(options)( char *str )
|
|
{
|
|
DRM(parse_options)( str );
|
|
return 1;
|
|
}
|
|
|
|
__setup( DRIVER_NAME "=", DRM_OPTIONS_FUNC );
|
|
#undef DRM_OPTIONS_FUNC
|
|
#endif
|
|
|
|
#define MAX_DEVICES 4
|
|
drm_device_t DRM(device)[MAX_DEVICES];
|
|
int DRM(numdevs) = 0;
|
|
int DRM(fb_loaded) = 0;
|
|
|
|
struct file_operations DRM(fops) = {
|
|
.owner = THIS_MODULE,
|
|
.open = DRM(open),
|
|
.flush = DRM(flush),
|
|
.release = DRM(release),
|
|
.ioctl = DRM(ioctl),
|
|
.mmap = DRM(mmap),
|
|
.fasync = DRM(fasync),
|
|
.poll = DRM(poll),
|
|
.read = DRM(read),
|
|
};
|
|
|
|
/** Ioctl table */
|
|
drm_ioctl_desc_t DRM(ioctls)[] = {
|
|
[DRM_IOCTL_NR(DRM_IOCTL_VERSION)] = { DRM(version), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_UNIQUE)] = { DRM(getunique), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_MAGIC)] = { DRM(getmagic), 0, 0 },
|
|
#if __HAVE_IRQ
|
|
[DRM_IOCTL_NR(DRM_IOCTL_IRQ_BUSID)] = { DRM(irq_by_busid), 0, 1 },
|
|
#endif
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_MAP)] = { DRM(getmap), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_CLIENT)] = { DRM(getclient), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_STATS)] = { DRM(getstats), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_VERSION)] = { DRM(setversion), 0, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_UNIQUE)] = { DRM(setunique), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_BLOCK)] = { DRM(noop), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_UNBLOCK)] = { DRM(noop), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AUTH_MAGIC)] = { DRM(authmagic), 1, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_MAP)] = { DRM(addmap), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_MAP)] = { DRM(rmmap), 1, 0 },
|
|
|
|
#if __HAVE_CTX_BITMAP
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_SAREA_CTX)] = { DRM(setsareactx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_SAREA_CTX)] = { DRM(getsareactx), 1, 0 },
|
|
#endif
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_CTX)] = { DRM(addctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_CTX)] = { DRM(rmctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MOD_CTX)] = { DRM(modctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_CTX)] = { DRM(getctx), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SWITCH_CTX)] = { DRM(switchctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_NEW_CTX)] = { DRM(newctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RES_CTX)] = { DRM(resctx), 1, 0 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_DRAW)] = { DRM(adddraw), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_DRAW)] = { DRM(rmdraw), 1, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_LOCK)] = { DRM(lock), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_UNLOCK)] = { DRM(unlock), 1, 0 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_FINISH)] = { DRM(noop), 1, 0 },
|
|
|
|
#if __HAVE_DMA
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_BUFS)] = { DRM(addbufs), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MARK_BUFS)] = { DRM(markbufs), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_INFO_BUFS)] = { DRM(infobufs), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MAP_BUFS)] = { DRM(mapbufs), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_FREE_BUFS)] = { DRM(freebufs), 1, 0 },
|
|
/* The DRM_IOCTL_DMA ioctl should be defined by the driver. */
|
|
#endif
|
|
#if __HAVE_IRQ || __HAVE_DMA
|
|
[DRM_IOCTL_NR(DRM_IOCTL_CONTROL)] = { DRM(control), 1, 1 },
|
|
#endif
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ACQUIRE)] = { DRM(agp_acquire), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_RELEASE)] = { DRM(agp_release), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ENABLE)] = { DRM(agp_enable), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_INFO)] = { DRM(agp_info), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ALLOC)] = { DRM(agp_alloc), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_FREE)] = { DRM(agp_free), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_BIND)] = { DRM(agp_bind), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_UNBIND)] = { DRM(agp_unbind), 1, 1 },
|
|
#endif
|
|
|
|
#if __HAVE_SG
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SG_ALLOC)] = { DRM(sg_alloc), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SG_FREE)] = { DRM(sg_free), 1, 1 },
|
|
#endif
|
|
|
|
#if __HAVE_VBL_IRQ
|
|
[DRM_IOCTL_NR(DRM_IOCTL_WAIT_VBLANK)] = { DRM(wait_vblank), 0, 0 },
|
|
#endif
|
|
|
|
DRIVER_IOCTLS
|
|
};
|
|
|
|
#define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( DRM(ioctls) )
|
|
|
|
#ifdef MODULE
|
|
static char *drm_opts = NULL;
|
|
#endif
|
|
|
|
MODULE_AUTHOR( DRIVER_AUTHOR );
|
|
MODULE_DESCRIPTION( DRIVER_DESC );
|
|
MODULE_PARM( drm_opts, "s" );
|
|
MODULE_LICENSE("GPL and additional rights");
|
|
|
|
static int DRM(setup)( drm_device_t *dev )
|
|
{
|
|
int i;
|
|
|
|
if (dev->fn_tbl.presetup)
|
|
dev->fn_tbl.presetup(dev);
|
|
|
|
atomic_set( &dev->ioctl_count, 0 );
|
|
atomic_set( &dev->vma_count, 0 );
|
|
dev->buf_use = 0;
|
|
atomic_set( &dev->buf_alloc, 0 );
|
|
|
|
#if __HAVE_DMA
|
|
i = DRM(dma_setup)( dev );
|
|
if ( i < 0 )
|
|
return i;
|
|
#endif
|
|
|
|
dev->counters = 6 + __HAVE_COUNTERS;
|
|
dev->types[0] = _DRM_STAT_LOCK;
|
|
dev->types[1] = _DRM_STAT_OPENS;
|
|
dev->types[2] = _DRM_STAT_CLOSES;
|
|
dev->types[3] = _DRM_STAT_IOCTLS;
|
|
dev->types[4] = _DRM_STAT_LOCKS;
|
|
dev->types[5] = _DRM_STAT_UNLOCKS;
|
|
#ifdef __HAVE_COUNTER6
|
|
dev->types[6] = __HAVE_COUNTER6;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER7
|
|
dev->types[7] = __HAVE_COUNTER7;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER8
|
|
dev->types[8] = __HAVE_COUNTER8;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER9
|
|
dev->types[9] = __HAVE_COUNTER9;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER10
|
|
dev->types[10] = __HAVE_COUNTER10;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER11
|
|
dev->types[11] = __HAVE_COUNTER11;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER12
|
|
dev->types[12] = __HAVE_COUNTER12;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER13
|
|
dev->types[13] = __HAVE_COUNTER13;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER14
|
|
dev->types[14] = __HAVE_COUNTER14;
|
|
#endif
|
|
|
|
for ( i = 0 ; i < DRM_ARRAY_SIZE(dev->counts) ; i++ )
|
|
atomic_set( &dev->counts[i], 0 );
|
|
|
|
for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) {
|
|
dev->magiclist[i].head = NULL;
|
|
dev->magiclist[i].tail = NULL;
|
|
}
|
|
|
|
dev->maplist = DRM(alloc)(sizeof(*dev->maplist),
|
|
DRM_MEM_MAPS);
|
|
if(dev->maplist == NULL) return -ENOMEM;
|
|
memset(dev->maplist, 0, sizeof(*dev->maplist));
|
|
INIT_LIST_HEAD(&dev->maplist->head);
|
|
|
|
dev->ctxlist = DRM(alloc)(sizeof(*dev->ctxlist),
|
|
DRM_MEM_CTXLIST);
|
|
if(dev->ctxlist == NULL) return -ENOMEM;
|
|
memset(dev->ctxlist, 0, sizeof(*dev->ctxlist));
|
|
INIT_LIST_HEAD(&dev->ctxlist->head);
|
|
|
|
dev->vmalist = NULL;
|
|
dev->sigdata.lock = dev->lock.hw_lock = NULL;
|
|
init_waitqueue_head( &dev->lock.lock_queue );
|
|
dev->queue_count = 0;
|
|
dev->queue_reserved = 0;
|
|
dev->queue_slots = 0;
|
|
dev->queuelist = NULL;
|
|
dev->irq_enabled = 0;
|
|
dev->context_flag = 0;
|
|
dev->interrupt_flag = 0;
|
|
dev->dma_flag = 0;
|
|
dev->last_context = 0;
|
|
dev->last_switch = 0;
|
|
dev->last_checked = 0;
|
|
init_waitqueue_head( &dev->context_wait );
|
|
dev->if_version = 0;
|
|
|
|
dev->ctx_start = 0;
|
|
dev->lck_start = 0;
|
|
|
|
dev->buf_rp = dev->buf;
|
|
dev->buf_wp = dev->buf;
|
|
dev->buf_end = dev->buf + DRM_BSZ;
|
|
dev->buf_async = NULL;
|
|
init_waitqueue_head( &dev->buf_readers );
|
|
init_waitqueue_head( &dev->buf_writers );
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
/*
|
|
* The kernel's context could be created here, but is now created
|
|
* in drm_dma_enqueue. This is more resource-efficient for
|
|
* hardware that does not do DMA, but may mean that
|
|
* drm_select_queue fails between the time the interrupt is
|
|
* initialized and the time the queues are initialized.
|
|
*/
|
|
if (dev->fn_tbl.postsetup)
|
|
dev->fn_tbl.postsetup(dev);
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* Take down the DRM device.
|
|
*
|
|
* \param dev DRM device structure.
|
|
*
|
|
* Frees every resource in \p dev.
|
|
*
|
|
* \sa drm_device and setup().
|
|
*/
|
|
static int DRM(takedown)( drm_device_t *dev )
|
|
{
|
|
drm_magic_entry_t *pt, *next;
|
|
drm_map_t *map;
|
|
drm_map_list_t *r_list;
|
|
struct list_head *list, *list_next;
|
|
drm_vma_entry_t *vma, *vma_next;
|
|
int i;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
if (dev->fn_tbl.pretakedown)
|
|
dev->fn_tbl.pretakedown(dev);
|
|
#if __HAVE_IRQ
|
|
if ( dev->irq_enabled ) DRM(irq_uninstall)( dev );
|
|
#endif
|
|
|
|
down( &dev->struct_sem );
|
|
del_timer( &dev->timer );
|
|
|
|
if ( dev->devname ) {
|
|
DRM(free)( dev->devname, strlen( dev->devname ) + 1,
|
|
DRM_MEM_DRIVER );
|
|
dev->devname = NULL;
|
|
}
|
|
|
|
if ( dev->unique ) {
|
|
DRM(free)( dev->unique, strlen( dev->unique ) + 1,
|
|
DRM_MEM_DRIVER );
|
|
dev->unique = NULL;
|
|
dev->unique_len = 0;
|
|
}
|
|
/* Clear pid list */
|
|
for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) {
|
|
for ( pt = dev->magiclist[i].head ; pt ; pt = next ) {
|
|
next = pt->next;
|
|
DRM(free)( pt, sizeof(*pt), DRM_MEM_MAGIC );
|
|
}
|
|
dev->magiclist[i].head = dev->magiclist[i].tail = NULL;
|
|
}
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
/* Clear AGP information */
|
|
if ( dev->agp ) {
|
|
drm_agp_mem_t *entry;
|
|
drm_agp_mem_t *nexte;
|
|
|
|
/* Remove AGP resources, but leave dev->agp
|
|
intact until drv_cleanup is called. */
|
|
for ( entry = dev->agp->memory ; entry ; entry = nexte ) {
|
|
nexte = entry->next;
|
|
if ( entry->bound ) DRM(unbind_agp)( entry->memory );
|
|
DRM(free_agp)( entry->memory, entry->pages );
|
|
DRM(free)( entry, sizeof(*entry), DRM_MEM_AGPLISTS );
|
|
}
|
|
dev->agp->memory = NULL;
|
|
|
|
if ( dev->agp->acquired ) DRM(agp_do_release)();
|
|
|
|
dev->agp->acquired = 0;
|
|
dev->agp->enabled = 0;
|
|
}
|
|
#endif
|
|
|
|
/* Clear vma list (only built for debugging) */
|
|
if ( dev->vmalist ) {
|
|
for ( vma = dev->vmalist ; vma ; vma = vma_next ) {
|
|
vma_next = vma->next;
|
|
DRM(free)( vma, sizeof(*vma), DRM_MEM_VMAS );
|
|
}
|
|
dev->vmalist = NULL;
|
|
}
|
|
|
|
if( dev->maplist ) {
|
|
list_for_each_safe( list, list_next, &dev->maplist->head ) {
|
|
r_list = (drm_map_list_t *)list;
|
|
|
|
if ( ( map = r_list->map ) ) {
|
|
switch ( map->type ) {
|
|
case _DRM_REGISTERS:
|
|
case _DRM_FRAME_BUFFER:
|
|
#if __REALLY_HAVE_MTRR
|
|
if ( map->mtrr >= 0 ) {
|
|
int retcode;
|
|
retcode = mtrr_del( map->mtrr,
|
|
map->offset,
|
|
map->size );
|
|
DRM_DEBUG( "mtrr_del=%d\n", retcode );
|
|
}
|
|
#endif
|
|
DRM(ioremapfree)( map->handle, map->size, dev );
|
|
break;
|
|
case _DRM_SHM:
|
|
vfree(map->handle);
|
|
break;
|
|
|
|
case _DRM_AGP:
|
|
/* Do nothing here, because this is all
|
|
* handled in the AGP/GART driver.
|
|
*/
|
|
break;
|
|
case _DRM_SCATTER_GATHER:
|
|
/* Handle it, but do nothing, if HAVE_SG
|
|
* isn't defined.
|
|
*/
|
|
#if __HAVE_SG
|
|
if(dev->sg) {
|
|
DRM(sg_cleanup)(dev->sg);
|
|
dev->sg = NULL;
|
|
}
|
|
#endif
|
|
break;
|
|
}
|
|
DRM(free)(map, sizeof(*map), DRM_MEM_MAPS);
|
|
}
|
|
list_del( list );
|
|
DRM(free)(r_list, sizeof(*r_list), DRM_MEM_MAPS);
|
|
}
|
|
DRM(free)(dev->maplist, sizeof(*dev->maplist), DRM_MEM_MAPS);
|
|
dev->maplist = NULL;
|
|
}
|
|
|
|
#if __HAVE_DMA_QUEUE || __HAVE_MULTIPLE_DMA_QUEUES
|
|
if ( dev->queuelist ) {
|
|
for ( i = 0 ; i < dev->queue_count ; i++ ) {
|
|
if (dev->fn_tbl.waitlist_destroy)
|
|
dev->fn_tbl.waitlist_destroy( &dev->queuelist[i]->waitlist);
|
|
|
|
if ( dev->queuelist[i] ) {
|
|
DRM(free)( dev->queuelist[i],
|
|
sizeof(*dev->queuelist[0]),
|
|
DRM_MEM_QUEUES );
|
|
dev->queuelist[i] = NULL;
|
|
}
|
|
}
|
|
DRM(free)( dev->queuelist,
|
|
dev->queue_slots * sizeof(*dev->queuelist),
|
|
DRM_MEM_QUEUES );
|
|
dev->queuelist = NULL;
|
|
}
|
|
dev->queue_count = 0;
|
|
#endif
|
|
|
|
#if __HAVE_DMA
|
|
DRM(dma_takedown)( dev );
|
|
#endif
|
|
if ( dev->lock.hw_lock ) {
|
|
dev->sigdata.lock = dev->lock.hw_lock = NULL; /* SHM removed */
|
|
dev->lock.filp = NULL;
|
|
wake_up_interruptible( &dev->lock.lock_queue );
|
|
}
|
|
|
|
up( &dev->struct_sem );
|
|
|
|
return 0;
|
|
}
|
|
|
|
#include "drm_pciids.h"
|
|
|
|
static struct pci_device_id DRM(pciidlist)[] = {
|
|
DRM(PCI_IDS)
|
|
};
|
|
|
|
static int drm_probe(struct pci_dev *pdev, const struct pci_device_id *ent)
|
|
{
|
|
drm_device_t *dev;
|
|
int retcode;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
if (DRM(numdevs) >= MAX_DEVICES)
|
|
return -ENODEV;
|
|
|
|
dev = &(DRM(device)[DRM(numdevs)]);
|
|
if (DRM(fb_loaded)==0)
|
|
pci_set_drvdata(pdev, dev);
|
|
|
|
memset( (void *)dev, 0, sizeof(*dev) );
|
|
dev->count_lock = SPIN_LOCK_UNLOCKED;
|
|
init_timer( &dev->timer );
|
|
sema_init( &dev->struct_sem, 1 );
|
|
sema_init( &dev->ctxlist_sem, 1 );
|
|
|
|
dev->name = DRIVER_NAME;
|
|
|
|
dev->pdev = pdev;
|
|
pci_enable_device(pdev);
|
|
#ifdef __alpha__
|
|
dev->hose = pdev->sysdata;
|
|
dev->pci_domain = dev->hose->bus->number;
|
|
#else
|
|
dev->pci_domain = 0;
|
|
#endif
|
|
dev->pci_bus = pdev->bus->number;
|
|
dev->pci_slot = PCI_SLOT(pdev->devfn);
|
|
dev->pci_func = PCI_FUNC(pdev->devfn);
|
|
dev->irq = pdev->irq;
|
|
|
|
/* dev_priv_size can be changed by a driver in driver_register_fns */
|
|
dev->dev_priv_size = sizeof(u32);
|
|
DRM(driver_register_fns)(dev);
|
|
|
|
if (dev->fn_tbl.preinit)
|
|
if ((retcode = dev->fn_tbl.preinit(dev, ent->driver_data)))
|
|
goto error_out_unreg;
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
dev->agp = DRM(agp_init)();
|
|
#if __MUST_HAVE_AGP
|
|
if ( dev->agp == NULL ) {
|
|
DRM_ERROR( "Cannot initialize the agpgart module.\n" );
|
|
retcode = -EINVAL;
|
|
goto error_out_unreg;
|
|
}
|
|
#endif
|
|
#if __REALLY_HAVE_MTRR
|
|
if (dev->agp)
|
|
dev->agp->agp_mtrr = mtrr_add( dev->agp->agp_info.aper_base,
|
|
dev->agp->agp_info.aper_size*1024*1024,
|
|
MTRR_TYPE_WRCOMB,
|
|
1 );
|
|
#endif
|
|
#endif
|
|
|
|
#if __HAVE_CTX_BITMAP
|
|
retcode = DRM(ctxbitmap_init)( dev );
|
|
if( retcode ) {
|
|
DRM_ERROR( "Cannot allocate memory for context bitmap.\n" );
|
|
goto error_out_unreg;
|
|
}
|
|
#endif
|
|
if ((dev->minor = DRM(stub_register)(DRIVER_NAME, &DRM(fops),dev)) < 0)
|
|
{
|
|
retcode = -EPERM;
|
|
goto error_out;
|
|
}
|
|
|
|
dev->device = MKDEV(DRM_MAJOR, dev->minor );
|
|
|
|
DRM(numdevs)++; /* no errors, mark it reserved */
|
|
|
|
DRM_INFO( "Initialized %s %d.%d.%d %s on minor %d: %s\n",
|
|
DRIVER_NAME,
|
|
DRIVER_MAJOR,
|
|
DRIVER_MINOR,
|
|
DRIVER_PATCHLEVEL,
|
|
DRIVER_DATE,
|
|
dev->minor,
|
|
pci_pretty_name(pdev)
|
|
);
|
|
|
|
/* drivers add secondary heads here if needed */
|
|
if (dev->fn_tbl.postinit)
|
|
if ((retcode = dev->fn_tbl.postinit(dev, ent->driver_data)))
|
|
goto error_out_unreg;
|
|
|
|
return 0;
|
|
|
|
error_out_unreg:
|
|
DRM(stub_unregister)(dev->minor);
|
|
DRM(takedown)(dev);
|
|
error_out:
|
|
return retcode;
|
|
}
|
|
|
|
static void __exit drm_cleanup_pci(struct pci_dev *pdev)
|
|
{
|
|
drm_device_t *dev = pci_get_drvdata(pdev);
|
|
|
|
pci_set_drvdata(pdev, NULL);
|
|
drm_cleanup(dev);
|
|
}
|
|
|
|
static struct pci_driver drm_driver = {
|
|
.name = DRIVER_NAME,
|
|
.id_table = DRM(pciidlist),
|
|
.probe = drm_probe,
|
|
.remove = __devexit_p(drm_cleanup_pci),
|
|
};
|
|
|
|
/**
|
|
* Module initialization. Called via init_module at module load time, or via
|
|
* linux/init/main.c (this is not currently supported).
|
|
*
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* Initializes an array of drm_device structures, and attempts to
|
|
* initialize all available devices, using consecutive minors, registering the
|
|
* stubs and initializing the AGP device.
|
|
*
|
|
* Expands the \c DRIVER_PREINIT and \c DRIVER_POST_INIT macros before and
|
|
* after the initialization for driver customization.
|
|
*/
|
|
static int __init drm_init( void )
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
struct pci_driver *pdriver = NULL;
|
|
int i;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
#ifdef MODULE
|
|
DRM(parse_options)( drm_opts );
|
|
#endif
|
|
|
|
DRM(mem_init)();
|
|
|
|
for (i=0; DRM(pciidlist)[i].vendor != 0; i++) {
|
|
pdev = pci_get_subsys(DRM(pciidlist[i]).vendor, DRM(pciidlist[i]).device, DRM(pciidlist[i]).subvendor, DRM(pciidlist[i]).subdevice, NULL);
|
|
if (pdev)
|
|
{
|
|
pdriver = pci_dev_driver(pdev);
|
|
if (pdriver)
|
|
{
|
|
DRM(fb_loaded)=1;
|
|
drm_probe(pdev, &DRM(pciidlist[i]));
|
|
}
|
|
else
|
|
pci_dev_put(pdev);
|
|
}
|
|
}
|
|
|
|
if (DRM(fb_loaded)==0)
|
|
pci_register_driver(&drm_driver);
|
|
else
|
|
DRM_INFO("Used old pci detect: framebuffer loaded\n");
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Called via cleanup_module() at module unload time.
|
|
*
|
|
* Cleans up all DRM device, calling takedown().
|
|
*
|
|
* \sa drm_init().
|
|
*/
|
|
static void __exit drm_cleanup( drm_device_t *dev )
|
|
{
|
|
DRM_DEBUG( "\n" );
|
|
if (!dev) {
|
|
DRM_ERROR("cleanup called no dev\n");
|
|
return;
|
|
}
|
|
|
|
DRM(takedown)(dev);
|
|
|
|
if (DRM(fb_loaded)==0)
|
|
pci_disable_device(dev->pdev);
|
|
|
|
DRM(numdevs)--;
|
|
if ( DRM(stub_unregister)(dev->minor) ) {
|
|
DRM_ERROR( "Cannot unload module\n" );
|
|
} else {
|
|
DRM_DEBUG( "minor %d unregistered\n", dev->minor);
|
|
}
|
|
#if __HAVE_CTX_BITMAP
|
|
DRM(ctxbitmap_cleanup)( dev );
|
|
#endif
|
|
#if __REALLY_HAVE_AGP && __REALLY_HAVE_MTRR
|
|
if ( dev->agp && dev->agp->agp_mtrr >= 0) {
|
|
int retval;
|
|
retval = mtrr_del( dev->agp->agp_mtrr,
|
|
dev->agp->agp_info.aper_base,
|
|
dev->agp->agp_info.aper_size*1024*1024 );
|
|
DRM_DEBUG( "mtrr_del=%d\n", retval );
|
|
}
|
|
#endif
|
|
#if __REALLY_HAVE_AGP
|
|
if ( dev->agp ) {
|
|
DRM(agp_uninit)();
|
|
DRM(free)( dev->agp, sizeof(*dev->agp), DRM_MEM_AGPLISTS );
|
|
dev->agp = NULL;
|
|
}
|
|
#endif
|
|
if (dev->fn_tbl.postcleanup)
|
|
dev->fn_tbl.postcleanup(dev);
|
|
}
|
|
|
|
static void __exit drm_exit (void)
|
|
{
|
|
DRM_DEBUG( "\n" );
|
|
if (DRM(fb_loaded)==1)
|
|
{
|
|
int i;
|
|
drm_device_t *dev;
|
|
|
|
for (i = DRM(numdevs) - 1; i >= 0; i--) {
|
|
dev = &(DRM(device)[i]);
|
|
/* release the pci driver */
|
|
if (dev->pdev)
|
|
pci_dev_put(dev->pdev);
|
|
drm_cleanup(dev);
|
|
}
|
|
}
|
|
else
|
|
pci_unregister_driver(&drm_driver);
|
|
DRM_INFO( "Module unloaded\n" );
|
|
}
|
|
|
|
|
|
module_init( drm_init );
|
|
module_exit( drm_exit );
|
|
|
|
|
|
/**
|
|
* Get version information
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_version structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Fills in the version information in \p arg.
|
|
*/
|
|
int DRM(version)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_version_t __user *argp = (void __user *)arg;
|
|
drm_version_t version;
|
|
int len;
|
|
|
|
if ( copy_from_user( &version, argp, sizeof(version) ) )
|
|
return -EFAULT;
|
|
|
|
#define DRM_COPY( name, value ) \
|
|
len = strlen( value ); \
|
|
if ( len > name##_len ) len = name##_len; \
|
|
name##_len = strlen( value ); \
|
|
if ( len && name ) { \
|
|
if ( copy_to_user( name, value, len ) ) \
|
|
return -EFAULT; \
|
|
}
|
|
|
|
version.version_major = DRIVER_MAJOR;
|
|
version.version_minor = DRIVER_MINOR;
|
|
version.version_patchlevel = DRIVER_PATCHLEVEL;
|
|
|
|
DRM_COPY( version.name, DRIVER_NAME );
|
|
DRM_COPY( version.date, DRIVER_DATE );
|
|
DRM_COPY( version.desc, DRIVER_DESC );
|
|
|
|
if ( copy_to_user( argp, &version, sizeof(version) ) )
|
|
return -EFAULT;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Open file.
|
|
*
|
|
* \param inode device inode
|
|
* \param filp file pointer.
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* Searches the DRM device with the same minor number, calls open_helper(), and
|
|
* increments the device open count. If the open count was previous at zero,
|
|
* i.e., it's the first that the device is open, then calls setup().
|
|
*/
|
|
int DRM(open)( struct inode *inode, struct file *filp )
|
|
{
|
|
drm_device_t *dev = NULL;
|
|
int retcode = 0;
|
|
int i;
|
|
|
|
for (i = 0; i < DRM(numdevs); i++) {
|
|
if (iminor(inode) == DRM(device)[i].minor) {
|
|
dev = &(DRM(device)[i]);
|
|
break;
|
|
}
|
|
}
|
|
if (!dev) {
|
|
return -ENODEV;
|
|
}
|
|
|
|
retcode = DRM(open_helper)( inode, filp, dev );
|
|
if ( !retcode ) {
|
|
atomic_inc( &dev->counts[_DRM_STAT_OPENS] );
|
|
spin_lock( &dev->count_lock );
|
|
if ( !dev->open_count++ ) {
|
|
spin_unlock( &dev->count_lock );
|
|
return DRM(setup)( dev );
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
}
|
|
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Release file.
|
|
*
|
|
* \param inode device inode
|
|
* \param filp file pointer.
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* If the hardware lock is held then free it, and take it again for the kernel
|
|
* context since it's necessary to reclaim buffers. Unlink the file private
|
|
* data from its list and free it. Decreases the open count and if it reaches
|
|
* zero calls takedown().
|
|
*/
|
|
int DRM(release)( struct inode *inode, struct file *filp )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev;
|
|
int retcode = 0;
|
|
|
|
lock_kernel();
|
|
dev = priv->dev;
|
|
|
|
DRM_DEBUG( "open_count = %d\n", dev->open_count );
|
|
|
|
if (dev->fn_tbl.prerelease)
|
|
dev->fn_tbl.prerelease(dev, filp);
|
|
|
|
/* ========================================================
|
|
* Begin inline drm_release
|
|
*/
|
|
|
|
DRM_DEBUG( "pid = %d, device = 0x%lx, open_count = %d\n",
|
|
current->pid, (long)old_encode_dev(dev->device), dev->open_count );
|
|
|
|
if ( priv->lock_count && dev->lock.hw_lock &&
|
|
_DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock) &&
|
|
dev->lock.filp == filp ) {
|
|
DRM_DEBUG( "File %p released, freeing lock for context %d\n",
|
|
filp,
|
|
_DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) );
|
|
|
|
if (dev->fn_tbl.release)
|
|
dev->fn_tbl.release(dev, filp);
|
|
|
|
DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
_DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) );
|
|
|
|
/* FIXME: may require heavy-handed reset of
|
|
hardware at this point, possibly
|
|
processed via a callback to the X
|
|
server. */
|
|
}
|
|
else if ( dev->fn_tbl.release && priv->lock_count && dev->lock.hw_lock ) {
|
|
/* The lock is required to reclaim buffers */
|
|
DECLARE_WAITQUEUE( entry, current );
|
|
|
|
add_wait_queue( &dev->lock.lock_queue, &entry );
|
|
for (;;) {
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
if ( !dev->lock.hw_lock ) {
|
|
/* Device has been unregistered */
|
|
retcode = -EINTR;
|
|
break;
|
|
}
|
|
if ( DRM(lock_take)( &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT ) ) {
|
|
dev->lock.filp = filp;
|
|
dev->lock.lock_time = jiffies;
|
|
atomic_inc( &dev->counts[_DRM_STAT_LOCKS] );
|
|
break; /* Got lock */
|
|
}
|
|
/* Contention */
|
|
schedule();
|
|
if ( signal_pending( current ) ) {
|
|
retcode = -ERESTARTSYS;
|
|
break;
|
|
}
|
|
}
|
|
current->state = TASK_RUNNING;
|
|
remove_wait_queue( &dev->lock.lock_queue, &entry );
|
|
if( !retcode ) {
|
|
if (dev->fn_tbl.release)
|
|
dev->fn_tbl.release(dev, filp);
|
|
DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT );
|
|
}
|
|
}
|
|
|
|
#if __HAVE_DMA
|
|
DRM(reclaim_buffers)( filp );
|
|
#endif
|
|
|
|
DRM(fasync)( -1, filp, 0 );
|
|
|
|
down( &dev->ctxlist_sem );
|
|
if ( !list_empty( &dev->ctxlist->head ) ) {
|
|
drm_ctx_list_t *pos, *n;
|
|
|
|
list_for_each_entry_safe( pos, n, &dev->ctxlist->head, head ) {
|
|
if ( pos->tag == priv &&
|
|
pos->handle != DRM_KERNEL_CONTEXT ) {
|
|
if (dev->fn_tbl.context_dtor)
|
|
dev->fn_tbl.context_dtor(dev, pos->handle);
|
|
#if __HAVE_CTX_BITMAP
|
|
DRM(ctxbitmap_free)( dev, pos->handle );
|
|
#endif
|
|
list_del( &pos->head );
|
|
DRM(free)( pos, sizeof(*pos), DRM_MEM_CTXLIST );
|
|
--dev->ctx_count;
|
|
}
|
|
}
|
|
}
|
|
up( &dev->ctxlist_sem );
|
|
|
|
down( &dev->struct_sem );
|
|
if ( priv->remove_auth_on_close == 1 ) {
|
|
drm_file_t *temp = dev->file_first;
|
|
while ( temp ) {
|
|
temp->authenticated = 0;
|
|
temp = temp->next;
|
|
}
|
|
}
|
|
if ( priv->prev ) {
|
|
priv->prev->next = priv->next;
|
|
} else {
|
|
dev->file_first = priv->next;
|
|
}
|
|
if ( priv->next ) {
|
|
priv->next->prev = priv->prev;
|
|
} else {
|
|
dev->file_last = priv->prev;
|
|
}
|
|
up( &dev->struct_sem );
|
|
|
|
DRM(free)( priv, sizeof(*priv), DRM_MEM_FILES );
|
|
|
|
/* ========================================================
|
|
* End inline drm_release
|
|
*/
|
|
|
|
atomic_inc( &dev->counts[_DRM_STAT_CLOSES] );
|
|
spin_lock( &dev->count_lock );
|
|
if ( !--dev->open_count ) {
|
|
if ( atomic_read( &dev->ioctl_count ) || dev->blocked ) {
|
|
DRM_ERROR( "Device busy: %d %d\n",
|
|
atomic_read( &dev->ioctl_count ),
|
|
dev->blocked );
|
|
spin_unlock( &dev->count_lock );
|
|
unlock_kernel();
|
|
return -EBUSY;
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
unlock_kernel();
|
|
return DRM(takedown)( dev );
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
|
|
unlock_kernel();
|
|
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Called whenever a process performs an ioctl on /dev/drm.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Looks up the ioctl function in the ::ioctls table, checking for root
|
|
* previleges if so required, and dispatches to the respective function.
|
|
*/
|
|
int DRM(ioctl)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
drm_ioctl_desc_t *ioctl;
|
|
drm_ioctl_t *func;
|
|
int nr = DRM_IOCTL_NR(cmd);
|
|
int retcode = 0;
|
|
|
|
atomic_inc( &dev->ioctl_count );
|
|
atomic_inc( &dev->counts[_DRM_STAT_IOCTLS] );
|
|
++priv->ioctl_count;
|
|
|
|
DRM_DEBUG( "pid=%d, cmd=0x%02x, nr=0x%02x, dev 0x%lx, auth=%d\n",
|
|
current->pid, cmd, nr, (long)old_encode_dev(dev->device),
|
|
priv->authenticated );
|
|
|
|
if ( nr >= DRIVER_IOCTL_COUNT ) {
|
|
retcode = -EINVAL;
|
|
} else {
|
|
ioctl = &DRM(ioctls)[nr];
|
|
func = ioctl->func;
|
|
|
|
if ( !func ) {
|
|
DRM_DEBUG( "no function\n" );
|
|
retcode = -EINVAL;
|
|
} else if ( ( ioctl->root_only && !capable( CAP_SYS_ADMIN ) )||
|
|
( ioctl->auth_needed && !priv->authenticated ) ) {
|
|
retcode = -EACCES;
|
|
} else {
|
|
retcode = func( inode, filp, cmd, arg );
|
|
}
|
|
}
|
|
|
|
atomic_dec( &dev->ioctl_count );
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Lock ioctl.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_lock structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Add the current task to the lock wait queue, and attempt to take to lock.
|
|
*/
|
|
int DRM(lock)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
DECLARE_WAITQUEUE( entry, current );
|
|
drm_lock_t lock;
|
|
int ret = 0;
|
|
#if __HAVE_MULTIPLE_DMA_QUEUES
|
|
drm_queue_t *q;
|
|
#endif
|
|
|
|
++priv->lock_count;
|
|
|
|
if ( copy_from_user( &lock, (drm_lock_t __user *)arg, sizeof(lock) ) )
|
|
return -EFAULT;
|
|
|
|
if ( lock.context == DRM_KERNEL_CONTEXT ) {
|
|
DRM_ERROR( "Process %d using kernel context %d\n",
|
|
current->pid, lock.context );
|
|
return -EINVAL;
|
|
}
|
|
|
|
DRM_DEBUG( "%d (pid %d) requests lock (0x%08x), flags = 0x%08x\n",
|
|
lock.context, current->pid,
|
|
dev->lock.hw_lock->lock, lock.flags );
|
|
|
|
#if __HAVE_DMA_QUEUE
|
|
if ( lock.context < 0 )
|
|
return -EINVAL;
|
|
#elif __HAVE_MULTIPLE_DMA_QUEUES
|
|
if ( lock.context < 0 || lock.context >= dev->queue_count )
|
|
return -EINVAL;
|
|
q = dev->queuelist[lock.context];
|
|
#endif
|
|
|
|
if (dev->fn_tbl.dma_flush_block_and_flush)
|
|
ret = dev->fn_tbl.dma_flush_block_and_flush(dev, lock.context, lock.flags);
|
|
if ( !ret ) {
|
|
add_wait_queue( &dev->lock.lock_queue, &entry );
|
|
for (;;) {
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
if ( !dev->lock.hw_lock ) {
|
|
/* Device has been unregistered */
|
|
ret = -EINTR;
|
|
break;
|
|
}
|
|
if ( DRM(lock_take)( &dev->lock.hw_lock->lock,
|
|
lock.context ) ) {
|
|
dev->lock.filp = filp;
|
|
dev->lock.lock_time = jiffies;
|
|
atomic_inc( &dev->counts[_DRM_STAT_LOCKS] );
|
|
break; /* Got lock */
|
|
}
|
|
|
|
/* Contention */
|
|
schedule();
|
|
if ( signal_pending( current ) ) {
|
|
ret = -ERESTARTSYS;
|
|
break;
|
|
}
|
|
}
|
|
current->state = TASK_RUNNING;
|
|
remove_wait_queue( &dev->lock.lock_queue, &entry );
|
|
}
|
|
|
|
if (dev->fn_tbl.dma_flush_unblock)
|
|
dev->fn_tbl.dma_flush_unblock(dev, lock.context, lock.flags);
|
|
|
|
if ( !ret ) {
|
|
sigemptyset( &dev->sigmask );
|
|
sigaddset( &dev->sigmask, SIGSTOP );
|
|
sigaddset( &dev->sigmask, SIGTSTP );
|
|
sigaddset( &dev->sigmask, SIGTTIN );
|
|
sigaddset( &dev->sigmask, SIGTTOU );
|
|
dev->sigdata.context = lock.context;
|
|
dev->sigdata.lock = dev->lock.hw_lock;
|
|
block_all_signals( DRM(notifier),
|
|
&dev->sigdata, &dev->sigmask );
|
|
|
|
if (dev->fn_tbl.dma_ready && (lock.flags & _DRM_LOCK_READY))
|
|
dev->fn_tbl.dma_ready(dev);
|
|
|
|
if ( dev->fn_tbl.dma_quiescent && (lock.flags & _DRM_LOCK_QUIESCENT ))
|
|
return dev->fn_tbl.dma_quiescent(dev);
|
|
|
|
|
|
if ( dev->fn_tbl.kernel_context_switch && dev->last_context != lock.context ) {
|
|
dev->fn_tbl.kernel_context_switch(dev, dev->last_context,
|
|
lock.context);
|
|
}
|
|
}
|
|
|
|
DRM_DEBUG( "%d %s\n", lock.context, ret ? "interrupted" : "has lock" );
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Unlock ioctl.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_lock structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Transfer and free the lock.
|
|
*/
|
|
int DRM(unlock)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
drm_lock_t lock;
|
|
|
|
if ( copy_from_user( &lock, (drm_lock_t __user *)arg, sizeof(lock) ) )
|
|
return -EFAULT;
|
|
|
|
if ( lock.context == DRM_KERNEL_CONTEXT ) {
|
|
DRM_ERROR( "Process %d using kernel context %d\n",
|
|
current->pid, lock.context );
|
|
return -EINVAL;
|
|
}
|
|
|
|
atomic_inc( &dev->counts[_DRM_STAT_UNLOCKS] );
|
|
|
|
if (dev->fn_tbl.kernel_context_switch_unlock)
|
|
dev->fn_tbl.kernel_context_switch_unlock(dev);
|
|
else
|
|
{
|
|
DRM(lock_transfer)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT );
|
|
|
|
if (dev->fn_tbl.dma_schedule)
|
|
dev->fn_tbl.dma_schedule(dev, 1);
|
|
|
|
if ( DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT ) ) {
|
|
DRM_ERROR( "\n" );
|
|
}
|
|
}
|
|
|
|
unblock_all_signals();
|
|
return 0;
|
|
}
|