mirror of
https://gitlab.freedesktop.org/mesa/drm.git
synced 2025-12-20 21:00:12 +01:00
a busid that doesn't correspond to the device the DRM is attached to.
This is a breaking of backwards-compatibility only for the
multiple-DRI-head case with X Servers that don't use interface 1.1.
- Move irq_busid to drm_irq.h and make it only return the IRQ for the
current device. Retains compatibility with previous X Servers, cleans
up unnecessary code. This means no irq_busid on !__HAVE_IRQ, but can be
changed if necessary.
- Bump interface version to 1.2. This version when set signifies that the
control ioctl should ignore the irq number passed in and enable the
interrupt handler for the attached device. Otherwise it errors out when
the passed-in irq is not equal to the device's.
- Store the highest version the interface has been set to in the device.
- Fix a recursion on DRM_LOCK in irq_uninstall on FreeBSD. This leaves
irq_uninstall being done without the lock in some cases, but it was
racey anyways.
1184 lines
31 KiB
C
1184 lines
31 KiB
C
/**
|
|
* \file drm_drv.h
|
|
* Generic driver template
|
|
*
|
|
* \author Rickard E. (Rik) Faith <faith@valinux.com>
|
|
* \author Gareth Hughes <gareth@valinux.com>
|
|
*
|
|
* To use this template, you must at least define the following (samples
|
|
* given for the MGA driver):
|
|
*
|
|
* \code
|
|
* #define DRIVER_AUTHOR "VA Linux Systems, Inc."
|
|
*
|
|
* #define DRIVER_NAME "mga"
|
|
* #define DRIVER_DESC "Matrox G200/G400"
|
|
* #define DRIVER_DATE "20001127"
|
|
*
|
|
* #define DRIVER_MAJOR 2
|
|
* #define DRIVER_MINOR 0
|
|
* #define DRIVER_PATCHLEVEL 2
|
|
*
|
|
* #define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( mga_ioctls )
|
|
*
|
|
* #define DRM(x) mga_##x
|
|
* \endcode
|
|
*/
|
|
|
|
/*
|
|
* Created: Thu Nov 23 03:10:50 2000 by gareth@valinux.com
|
|
*
|
|
* Copyright 1999, 2000 Precision Insight, Inc., Cedar Park, Texas.
|
|
* Copyright 2000 VA Linux Systems, Inc., Sunnyvale, California.
|
|
* All Rights Reserved.
|
|
*
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
* to deal in the Software without restriction, including without limitation
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
*
|
|
* The above copyright notice and this permission notice (including the next
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
* Software.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
* VA LINUX SYSTEMS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM, DAMAGES OR
|
|
* OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE,
|
|
* ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR
|
|
* OTHER DEALINGS IN THE SOFTWARE.
|
|
*/
|
|
|
|
#ifndef __MUST_HAVE_AGP
|
|
#define __MUST_HAVE_AGP 0
|
|
#endif
|
|
#ifndef __HAVE_CTX_BITMAP
|
|
#define __HAVE_CTX_BITMAP 0
|
|
#endif
|
|
#ifndef __HAVE_IRQ
|
|
#define __HAVE_IRQ 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_QUEUE
|
|
#define __HAVE_DMA_QUEUE 0
|
|
#endif
|
|
#ifndef __HAVE_MULTIPLE_DMA_QUEUES
|
|
#define __HAVE_MULTIPLE_DMA_QUEUES 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_SCHEDULE
|
|
#define __HAVE_DMA_SCHEDULE 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_FLUSH
|
|
#define __HAVE_DMA_FLUSH 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_READY
|
|
#define __HAVE_DMA_READY 0
|
|
#endif
|
|
#ifndef __HAVE_DMA_QUIESCENT
|
|
#define __HAVE_DMA_QUIESCENT 0
|
|
#endif
|
|
#ifndef __HAVE_RELEASE
|
|
#define __HAVE_RELEASE 0
|
|
#endif
|
|
#ifndef __HAVE_COUNTERS
|
|
#define __HAVE_COUNTERS 0
|
|
#endif
|
|
#ifndef __HAVE_SG
|
|
#define __HAVE_SG 0
|
|
#endif
|
|
/* __HAVE_KERNEL_CTX_SWITCH isn't used by any of the drm modules in
|
|
* the DRI cvs tree, but it is required by the kernel tree's sparc
|
|
* driver.
|
|
*/
|
|
#ifndef __HAVE_KERNEL_CTX_SWITCH
|
|
#define __HAVE_KERNEL_CTX_SWITCH 0
|
|
#endif
|
|
#ifndef __HAVE_DRIVER_FOPS_READ
|
|
#define __HAVE_DRIVER_FOPS_READ 0
|
|
#endif
|
|
#ifndef __HAVE_DRIVER_FOPS_POLL
|
|
#define __HAVE_DRIVER_FOPS_POLL 0
|
|
#endif
|
|
|
|
#ifndef DRIVER_PREINIT
|
|
#define DRIVER_PREINIT()
|
|
#endif
|
|
#ifndef DRIVER_POSTINIT
|
|
#define DRIVER_POSTINIT()
|
|
#endif
|
|
#ifndef DRIVER_PRERELEASE
|
|
#define DRIVER_PRERELEASE()
|
|
#endif
|
|
#ifndef DRIVER_PRETAKEDOWN
|
|
#define DRIVER_PRETAKEDOWN()
|
|
#endif
|
|
#ifndef DRIVER_POSTCLEANUP
|
|
#define DRIVER_POSTCLEANUP()
|
|
#endif
|
|
#ifndef DRIVER_PRESETUP
|
|
#define DRIVER_PRESETUP()
|
|
#endif
|
|
#ifndef DRIVER_POSTSETUP
|
|
#define DRIVER_POSTSETUP()
|
|
#endif
|
|
#ifndef DRIVER_IOCTLS
|
|
#define DRIVER_IOCTLS
|
|
#endif
|
|
#ifndef DRIVER_OPEN_HELPER
|
|
#define DRIVER_OPEN_HELPER( priv, dev )
|
|
#endif
|
|
#ifndef DRIVER_FOPS
|
|
#define DRIVER_FOPS \
|
|
static struct file_operations DRM(fops) = { \
|
|
.owner = THIS_MODULE, \
|
|
.open = DRM(open), \
|
|
.flush = DRM(flush), \
|
|
.release = DRM(release), \
|
|
.ioctl = DRM(ioctl), \
|
|
.mmap = DRM(mmap), \
|
|
.fasync = DRM(fasync), \
|
|
.poll = DRM(poll), \
|
|
.read = DRM(read), \
|
|
}
|
|
#endif
|
|
|
|
#ifndef MODULE
|
|
/** Use an additional macro to avoid preprocessor troubles */
|
|
#define DRM_OPTIONS_FUNC DRM(options)
|
|
/**
|
|
* Called by the kernel to parse command-line options passed via the
|
|
* boot-loader (e.g., LILO). It calls the insmod option routine,
|
|
* parse_options().
|
|
*/
|
|
static int __init DRM(options)( char *str )
|
|
{
|
|
DRM(parse_options)( str );
|
|
return 1;
|
|
}
|
|
|
|
__setup( DRIVER_NAME "=", DRM_OPTIONS_FUNC );
|
|
#undef DRM_OPTIONS_FUNC
|
|
#endif
|
|
|
|
#define MAX_DEVICES 4
|
|
static drm_device_t DRM(device)[MAX_DEVICES];
|
|
static int DRM(numdevs) = 0;
|
|
|
|
DRIVER_FOPS;
|
|
|
|
/** Ioctl table */
|
|
static drm_ioctl_desc_t DRM(ioctls)[] = {
|
|
[DRM_IOCTL_NR(DRM_IOCTL_VERSION)] = { DRM(version), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_UNIQUE)] = { DRM(getunique), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_MAGIC)] = { DRM(getmagic), 0, 0 },
|
|
#if __HAVE_IRQ
|
|
[DRM_IOCTL_NR(DRM_IOCTL_IRQ_BUSID)] = { DRM(irq_by_busid), 0, 1 },
|
|
#endif
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_MAP)] = { DRM(getmap), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_CLIENT)] = { DRM(getclient), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_STATS)] = { DRM(getstats), 0, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_VERSION)] = { DRM(setversion), 0, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_UNIQUE)] = { DRM(setunique), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_BLOCK)] = { DRM(noop), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_UNBLOCK)] = { DRM(noop), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AUTH_MAGIC)] = { DRM(authmagic), 1, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_MAP)] = { DRM(addmap), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_MAP)] = { DRM(rmmap), 1, 0 },
|
|
|
|
#if __HAVE_CTX_BITMAP
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SET_SAREA_CTX)] = { DRM(setsareactx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_SAREA_CTX)] = { DRM(getsareactx), 1, 0 },
|
|
#endif
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_CTX)] = { DRM(addctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_CTX)] = { DRM(rmctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MOD_CTX)] = { DRM(modctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_GET_CTX)] = { DRM(getctx), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SWITCH_CTX)] = { DRM(switchctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_NEW_CTX)] = { DRM(newctx), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RES_CTX)] = { DRM(resctx), 1, 0 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_DRAW)] = { DRM(adddraw), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_RM_DRAW)] = { DRM(rmdraw), 1, 1 },
|
|
|
|
[DRM_IOCTL_NR(DRM_IOCTL_LOCK)] = { DRM(lock), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_UNLOCK)] = { DRM(unlock), 1, 0 },
|
|
|
|
#if __HAVE_DMA_FLUSH
|
|
/* Gamma only, really */
|
|
[DRM_IOCTL_NR(DRM_IOCTL_FINISH)] = { DRM(finish), 1, 0 },
|
|
#else
|
|
[DRM_IOCTL_NR(DRM_IOCTL_FINISH)] = { DRM(noop), 1, 0 },
|
|
#endif
|
|
|
|
#if __HAVE_DMA
|
|
[DRM_IOCTL_NR(DRM_IOCTL_ADD_BUFS)] = { DRM(addbufs), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MARK_BUFS)] = { DRM(markbufs), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_INFO_BUFS)] = { DRM(infobufs), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_MAP_BUFS)] = { DRM(mapbufs), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_FREE_BUFS)] = { DRM(freebufs), 1, 0 },
|
|
/* The DRM_IOCTL_DMA ioctl should be defined by the driver. */
|
|
#endif
|
|
#if __HAVE_IRQ || __HAVE_DMA
|
|
[DRM_IOCTL_NR(DRM_IOCTL_CONTROL)] = { DRM(control), 1, 1 },
|
|
#endif
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ACQUIRE)] = { DRM(agp_acquire), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_RELEASE)] = { DRM(agp_release), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ENABLE)] = { DRM(agp_enable), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_INFO)] = { DRM(agp_info), 1, 0 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_ALLOC)] = { DRM(agp_alloc), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_FREE)] = { DRM(agp_free), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_BIND)] = { DRM(agp_bind), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_AGP_UNBIND)] = { DRM(agp_unbind), 1, 1 },
|
|
#endif
|
|
|
|
#if __HAVE_SG
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SG_ALLOC)] = { DRM(sg_alloc), 1, 1 },
|
|
[DRM_IOCTL_NR(DRM_IOCTL_SG_FREE)] = { DRM(sg_free), 1, 1 },
|
|
#endif
|
|
|
|
#if __HAVE_VBL_IRQ
|
|
[DRM_IOCTL_NR(DRM_IOCTL_WAIT_VBLANK)] = { DRM(wait_vblank), 0, 0 },
|
|
#endif
|
|
|
|
DRIVER_IOCTLS
|
|
};
|
|
|
|
#define DRIVER_IOCTL_COUNT DRM_ARRAY_SIZE( DRM(ioctls) )
|
|
|
|
#ifdef MODULE
|
|
static char *drm_opts = NULL;
|
|
#endif
|
|
|
|
MODULE_AUTHOR( DRIVER_AUTHOR );
|
|
MODULE_DESCRIPTION( DRIVER_DESC );
|
|
MODULE_PARM( drm_opts, "s" );
|
|
MODULE_LICENSE("GPL and additional rights");
|
|
|
|
static int DRM(setup)( drm_device_t *dev )
|
|
{
|
|
int i;
|
|
|
|
DRIVER_PRESETUP();
|
|
atomic_set( &dev->ioctl_count, 0 );
|
|
atomic_set( &dev->vma_count, 0 );
|
|
dev->buf_use = 0;
|
|
atomic_set( &dev->buf_alloc, 0 );
|
|
|
|
#if __HAVE_DMA
|
|
i = DRM(dma_setup)( dev );
|
|
if ( i < 0 )
|
|
return i;
|
|
#endif
|
|
|
|
dev->counters = 6 + __HAVE_COUNTERS;
|
|
dev->types[0] = _DRM_STAT_LOCK;
|
|
dev->types[1] = _DRM_STAT_OPENS;
|
|
dev->types[2] = _DRM_STAT_CLOSES;
|
|
dev->types[3] = _DRM_STAT_IOCTLS;
|
|
dev->types[4] = _DRM_STAT_LOCKS;
|
|
dev->types[5] = _DRM_STAT_UNLOCKS;
|
|
#ifdef __HAVE_COUNTER6
|
|
dev->types[6] = __HAVE_COUNTER6;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER7
|
|
dev->types[7] = __HAVE_COUNTER7;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER8
|
|
dev->types[8] = __HAVE_COUNTER8;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER9
|
|
dev->types[9] = __HAVE_COUNTER9;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER10
|
|
dev->types[10] = __HAVE_COUNTER10;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER11
|
|
dev->types[11] = __HAVE_COUNTER11;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER12
|
|
dev->types[12] = __HAVE_COUNTER12;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER13
|
|
dev->types[13] = __HAVE_COUNTER13;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER14
|
|
dev->types[14] = __HAVE_COUNTER14;
|
|
#endif
|
|
#ifdef __HAVE_COUNTER15
|
|
dev->types[14] = __HAVE_COUNTER14;
|
|
#endif
|
|
|
|
for ( i = 0 ; i < DRM_ARRAY_SIZE(dev->counts) ; i++ )
|
|
atomic_set( &dev->counts[i], 0 );
|
|
|
|
for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) {
|
|
dev->magiclist[i].head = NULL;
|
|
dev->magiclist[i].tail = NULL;
|
|
}
|
|
|
|
dev->maplist = DRM(alloc)(sizeof(*dev->maplist),
|
|
DRM_MEM_MAPS);
|
|
if(dev->maplist == NULL) return -ENOMEM;
|
|
memset(dev->maplist, 0, sizeof(*dev->maplist));
|
|
INIT_LIST_HEAD(&dev->maplist->head);
|
|
|
|
dev->vmalist = NULL;
|
|
dev->sigdata.lock = dev->lock.hw_lock = NULL;
|
|
init_waitqueue_head( &dev->lock.lock_queue );
|
|
dev->queue_count = 0;
|
|
dev->queue_reserved = 0;
|
|
dev->queue_slots = 0;
|
|
dev->queuelist = NULL;
|
|
dev->irq_enabled = 0;
|
|
dev->context_flag = 0;
|
|
dev->interrupt_flag = 0;
|
|
dev->dma_flag = 0;
|
|
dev->last_context = 0;
|
|
dev->last_switch = 0;
|
|
dev->last_checked = 0;
|
|
init_waitqueue_head( &dev->context_wait );
|
|
dev->if_version = 0;
|
|
|
|
dev->ctx_start = 0;
|
|
dev->lck_start = 0;
|
|
|
|
dev->buf_rp = dev->buf;
|
|
dev->buf_wp = dev->buf;
|
|
dev->buf_end = dev->buf + DRM_BSZ;
|
|
dev->buf_async = NULL;
|
|
init_waitqueue_head( &dev->buf_readers );
|
|
init_waitqueue_head( &dev->buf_writers );
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
/*
|
|
* The kernel's context could be created here, but is now created
|
|
* in drm_dma_enqueue. This is more resource-efficient for
|
|
* hardware that does not do DMA, but may mean that
|
|
* drm_select_queue fails between the time the interrupt is
|
|
* initialized and the time the queues are initialized.
|
|
*/
|
|
DRIVER_POSTSETUP();
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* Take down the DRM device.
|
|
*
|
|
* \param dev DRM device structure.
|
|
*
|
|
* Frees every resource in \p dev.
|
|
*
|
|
* \sa drm_device and setup().
|
|
*/
|
|
static int DRM(takedown)( drm_device_t *dev )
|
|
{
|
|
drm_magic_entry_t *pt, *next;
|
|
drm_map_t *map;
|
|
drm_map_list_t *r_list;
|
|
struct list_head *list, *list_next;
|
|
drm_vma_entry_t *vma, *vma_next;
|
|
int i;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
DRIVER_PRETAKEDOWN();
|
|
#if __HAVE_IRQ
|
|
if ( dev->irq_enabled ) DRM(irq_uninstall)( dev );
|
|
#endif
|
|
|
|
down( &dev->struct_sem );
|
|
del_timer( &dev->timer );
|
|
|
|
if ( dev->devname ) {
|
|
DRM(free)( dev->devname, strlen( dev->devname ) + 1,
|
|
DRM_MEM_DRIVER );
|
|
dev->devname = NULL;
|
|
}
|
|
|
|
if ( dev->unique ) {
|
|
DRM(free)( dev->unique, strlen( dev->unique ) + 1,
|
|
DRM_MEM_DRIVER );
|
|
dev->unique = NULL;
|
|
dev->unique_len = 0;
|
|
}
|
|
/* Clear pid list */
|
|
for ( i = 0 ; i < DRM_HASH_SIZE ; i++ ) {
|
|
for ( pt = dev->magiclist[i].head ; pt ; pt = next ) {
|
|
next = pt->next;
|
|
DRM(free)( pt, sizeof(*pt), DRM_MEM_MAGIC );
|
|
}
|
|
dev->magiclist[i].head = dev->magiclist[i].tail = NULL;
|
|
}
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
/* Clear AGP information */
|
|
if ( dev->agp ) {
|
|
drm_agp_mem_t *entry;
|
|
drm_agp_mem_t *nexte;
|
|
|
|
/* Remove AGP resources, but leave dev->agp
|
|
intact until drv_cleanup is called. */
|
|
for ( entry = dev->agp->memory ; entry ; entry = nexte ) {
|
|
nexte = entry->next;
|
|
if ( entry->bound ) DRM(unbind_agp)( entry->memory );
|
|
DRM(free_agp)( entry->memory, entry->pages );
|
|
DRM(free)( entry, sizeof(*entry), DRM_MEM_AGPLISTS );
|
|
}
|
|
dev->agp->memory = NULL;
|
|
|
|
if ( dev->agp->acquired ) DRM(agp_do_release)();
|
|
|
|
dev->agp->acquired = 0;
|
|
dev->agp->enabled = 0;
|
|
}
|
|
#endif
|
|
|
|
/* Clear vma list (only built for debugging) */
|
|
if ( dev->vmalist ) {
|
|
for ( vma = dev->vmalist ; vma ; vma = vma_next ) {
|
|
vma_next = vma->next;
|
|
DRM(free)( vma, sizeof(*vma), DRM_MEM_VMAS );
|
|
}
|
|
dev->vmalist = NULL;
|
|
}
|
|
|
|
if( dev->maplist ) {
|
|
list_for_each_safe( list, list_next, &dev->maplist->head ) {
|
|
r_list = (drm_map_list_t *)list;
|
|
|
|
if ( ( map = r_list->map ) ) {
|
|
switch ( map->type ) {
|
|
case _DRM_REGISTERS:
|
|
case _DRM_FRAME_BUFFER:
|
|
#if __REALLY_HAVE_MTRR
|
|
if ( map->mtrr >= 0 ) {
|
|
int retcode;
|
|
retcode = mtrr_del( map->mtrr,
|
|
map->offset,
|
|
map->size );
|
|
DRM_DEBUG( "mtrr_del=%d\n", retcode );
|
|
}
|
|
#endif
|
|
DRM(ioremapfree)( map->handle, map->size, dev );
|
|
break;
|
|
case _DRM_SHM:
|
|
vfree(map->handle);
|
|
break;
|
|
|
|
case _DRM_AGP:
|
|
/* Do nothing here, because this is all
|
|
* handled in the AGP/GART driver.
|
|
*/
|
|
break;
|
|
case _DRM_SCATTER_GATHER:
|
|
/* Handle it, but do nothing, if HAVE_SG
|
|
* isn't defined.
|
|
*/
|
|
#if __HAVE_SG
|
|
if(dev->sg) {
|
|
DRM(sg_cleanup)(dev->sg);
|
|
dev->sg = NULL;
|
|
}
|
|
#endif
|
|
break;
|
|
}
|
|
DRM(free)(map, sizeof(*map), DRM_MEM_MAPS);
|
|
}
|
|
list_del( list );
|
|
DRM(free)(r_list, sizeof(*r_list), DRM_MEM_MAPS);
|
|
}
|
|
DRM(free)(dev->maplist, sizeof(*dev->maplist), DRM_MEM_MAPS);
|
|
dev->maplist = NULL;
|
|
}
|
|
|
|
#if __HAVE_DMA_QUEUE || __HAVE_MULTIPLE_DMA_QUEUES
|
|
if ( dev->queuelist ) {
|
|
for ( i = 0 ; i < dev->queue_count ; i++ ) {
|
|
#if __HAVE_DMA_WAITLIST
|
|
DRM(waitlist_destroy)( &dev->queuelist[i]->waitlist );
|
|
#endif
|
|
if ( dev->queuelist[i] ) {
|
|
DRM(free)( dev->queuelist[i],
|
|
sizeof(*dev->queuelist[0]),
|
|
DRM_MEM_QUEUES );
|
|
dev->queuelist[i] = NULL;
|
|
}
|
|
}
|
|
DRM(free)( dev->queuelist,
|
|
dev->queue_slots * sizeof(*dev->queuelist),
|
|
DRM_MEM_QUEUES );
|
|
dev->queuelist = NULL;
|
|
}
|
|
dev->queue_count = 0;
|
|
#endif
|
|
|
|
#if __HAVE_DMA
|
|
DRM(dma_takedown)( dev );
|
|
#endif
|
|
if ( dev->lock.hw_lock ) {
|
|
dev->sigdata.lock = dev->lock.hw_lock = NULL; /* SHM removed */
|
|
dev->lock.filp = 0;
|
|
wake_up_interruptible( &dev->lock.lock_queue );
|
|
}
|
|
up( &dev->struct_sem );
|
|
|
|
return 0;
|
|
}
|
|
|
|
static drm_pci_id_list_t DRM(pciidlist)[] = {
|
|
DRIVER_PCI_IDS
|
|
};
|
|
|
|
static int DRM(probe)(struct pci_dev *pdev)
|
|
{
|
|
drm_device_t *dev;
|
|
#if __HAVE_CTX_BITMAP
|
|
int retcode;
|
|
#endif
|
|
int i;
|
|
char *desc = NULL;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
for (i = 0; DRM(pciidlist)[i].vendor != 0; i++) {
|
|
if ((DRM(pciidlist)[i].vendor == pdev->vendor) &&
|
|
(DRM(pciidlist)[i].device == pdev->device)) {
|
|
desc = DRM(pciidlist)[i].name;
|
|
}
|
|
}
|
|
if (desc == NULL)
|
|
return -ENODEV;
|
|
|
|
if (DRM(numdevs) >= MAX_DEVICES)
|
|
return -ENODEV;
|
|
|
|
dev = &(DRM(device)[DRM(numdevs)]);
|
|
|
|
memset( (void *)dev, 0, sizeof(*dev) );
|
|
dev->count_lock = SPIN_LOCK_UNLOCKED;
|
|
init_timer( &dev->timer );
|
|
sema_init( &dev->struct_sem, 1 );
|
|
|
|
if ((dev->minor = DRM(stub_register)(DRIVER_NAME, &DRM(fops),dev)) < 0)
|
|
return -EPERM;
|
|
dev->device = MKDEV(DRM_MAJOR, dev->minor );
|
|
dev->name = DRIVER_NAME;
|
|
|
|
dev->pdev = pdev;
|
|
#ifdef __alpha__
|
|
dev->hose = pdev->sysdata;
|
|
dev->pci_domain = dev->hose->bus->number;
|
|
#else
|
|
dev->pci_domain = 0;
|
|
#endif
|
|
dev->pci_bus = pdev->bus->number;
|
|
dev->pci_slot = PCI_SLOT(pdev->devfn);
|
|
dev->pci_func = PCI_FUNC(pdev->devfn);
|
|
dev->irq = pdev->irq;
|
|
|
|
DRIVER_PREINIT();
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
dev->agp = DRM(agp_init)();
|
|
#if __MUST_HAVE_AGP
|
|
if ( dev->agp == NULL ) {
|
|
DRM_ERROR( "Cannot initialize the agpgart module.\n" );
|
|
DRM(stub_unregister)(dev->minor);
|
|
DRM(takedown)( dev );
|
|
return -ENOMEM;
|
|
}
|
|
#endif
|
|
#if __REALLY_HAVE_MTRR
|
|
if (dev->agp)
|
|
dev->agp->agp_mtrr = mtrr_add( dev->agp->agp_info.aper_base,
|
|
dev->agp->agp_info.aper_size*1024*1024,
|
|
MTRR_TYPE_WRCOMB,
|
|
1 );
|
|
#endif
|
|
#endif
|
|
|
|
#if __HAVE_CTX_BITMAP
|
|
retcode = DRM(ctxbitmap_init)( dev );
|
|
if( retcode ) {
|
|
DRM_ERROR( "Cannot allocate memory for context bitmap.\n" );
|
|
DRM(stub_unregister)(dev->minor);
|
|
DRM(takedown)( dev );
|
|
return retcode;
|
|
}
|
|
#endif
|
|
DRM(numdevs)++; /* no errors, mark it reserved */
|
|
|
|
DRM_INFO( "Initialized %s %d.%d.%d %s on minor %d: %s\n",
|
|
DRIVER_NAME,
|
|
DRIVER_MAJOR,
|
|
DRIVER_MINOR,
|
|
DRIVER_PATCHLEVEL,
|
|
DRIVER_DATE,
|
|
dev->minor,
|
|
desc );
|
|
|
|
DRIVER_POSTINIT();
|
|
|
|
return 0;
|
|
}
|
|
|
|
|
|
/**
|
|
* Module initialization. Called via init_module at module load time, or via
|
|
* linux/init/main.c (this is not currently supported).
|
|
*
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* Initializes an array of drm_device structures, and attempts to
|
|
* initialize all available devices, using consecutive minors, registering the
|
|
* stubs and initializing the AGP device.
|
|
*
|
|
* Expands the \c DRIVER_PREINIT and \c DRIVER_POST_INIT macros before and
|
|
* after the initialization for driver customization.
|
|
*/
|
|
static int __init drm_init( void )
|
|
{
|
|
struct pci_dev *pdev = NULL;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
#ifdef MODULE
|
|
DRM(parse_options)( drm_opts );
|
|
#endif
|
|
|
|
DRM(mem_init)();
|
|
|
|
while ((pdev = pci_find_device(PCI_ANY_ID, PCI_ANY_ID, pdev)) != NULL) {
|
|
DRM(probe)(pdev);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Called via cleanup_module() at module unload time.
|
|
*
|
|
* Cleans up all DRM device, calling takedown().
|
|
*
|
|
* \sa drm_init().
|
|
*/
|
|
static void __exit drm_cleanup( void )
|
|
{
|
|
drm_device_t *dev;
|
|
int i;
|
|
|
|
DRM_DEBUG( "\n" );
|
|
|
|
for (i = DRM(numdevs) - 1; i >= 0; i--) {
|
|
dev = &(DRM(device)[i]);
|
|
if ( DRM(stub_unregister)(dev->minor) ) {
|
|
DRM_ERROR( "Cannot unload module\n" );
|
|
} else {
|
|
DRM_DEBUG("minor %d unregistered\n", dev->minor);
|
|
if (i == 0) {
|
|
DRM_INFO( "Module unloaded\n" );
|
|
}
|
|
}
|
|
#if __HAVE_CTX_BITMAP
|
|
DRM(ctxbitmap_cleanup)( dev );
|
|
#endif
|
|
|
|
#if __REALLY_HAVE_AGP && __REALLY_HAVE_MTRR
|
|
if ( dev->agp && dev->agp->agp_mtrr >= 0) {
|
|
int retval;
|
|
retval = mtrr_del( dev->agp->agp_mtrr,
|
|
dev->agp->agp_info.aper_base,
|
|
dev->agp->agp_info.aper_size*1024*1024 );
|
|
DRM_DEBUG( "mtrr_del=%d\n", retval );
|
|
}
|
|
#endif
|
|
|
|
DRM(takedown)( dev );
|
|
|
|
#if __REALLY_HAVE_AGP
|
|
if ( dev->agp ) {
|
|
DRM(agp_uninit)();
|
|
DRM(free)( dev->agp, sizeof(*dev->agp), DRM_MEM_AGPLISTS );
|
|
dev->agp = NULL;
|
|
}
|
|
#endif
|
|
}
|
|
DRIVER_POSTCLEANUP();
|
|
DRM(numdevs) = 0;
|
|
}
|
|
|
|
module_init( drm_init );
|
|
module_exit( drm_cleanup );
|
|
|
|
|
|
/**
|
|
* Get version information
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_version structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Fills in the version information in \p arg.
|
|
*/
|
|
int DRM(version)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_version_t version;
|
|
int len;
|
|
|
|
if ( copy_from_user( &version,
|
|
(drm_version_t *)arg,
|
|
sizeof(version) ) )
|
|
return -EFAULT;
|
|
|
|
#define DRM_COPY( name, value ) \
|
|
len = strlen( value ); \
|
|
if ( len > name##_len ) len = name##_len; \
|
|
name##_len = strlen( value ); \
|
|
if ( len && name ) { \
|
|
if ( copy_to_user( name, value, len ) ) \
|
|
return -EFAULT; \
|
|
}
|
|
|
|
version.version_major = DRIVER_MAJOR;
|
|
version.version_minor = DRIVER_MINOR;
|
|
version.version_patchlevel = DRIVER_PATCHLEVEL;
|
|
|
|
DRM_COPY( version.name, DRIVER_NAME );
|
|
DRM_COPY( version.date, DRIVER_DATE );
|
|
DRM_COPY( version.desc, DRIVER_DESC );
|
|
|
|
if ( copy_to_user( (drm_version_t *)arg,
|
|
&version,
|
|
sizeof(version) ) )
|
|
return -EFAULT;
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* Open file.
|
|
*
|
|
* \param inode device inode
|
|
* \param filp file pointer.
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* Searches the DRM device with the same minor number, calls open_helper(), and
|
|
* increments the device open count. If the open count was previous at zero,
|
|
* i.e., it's the first that the device is open, then calls setup().
|
|
*/
|
|
int DRM(open)( struct inode *inode, struct file *filp )
|
|
{
|
|
drm_device_t *dev = NULL;
|
|
int retcode = 0;
|
|
int i;
|
|
|
|
for (i = 0; i < DRM(numdevs); i++) {
|
|
if (minor(inode->i_rdev) == DRM(device)[i].minor) {
|
|
dev = &(DRM(device)[i]);
|
|
break;
|
|
}
|
|
}
|
|
if (!dev) {
|
|
return -ENODEV;
|
|
}
|
|
|
|
retcode = DRM(open_helper)( inode, filp, dev );
|
|
if ( !retcode ) {
|
|
atomic_inc( &dev->counts[_DRM_STAT_OPENS] );
|
|
spin_lock( &dev->count_lock );
|
|
if ( !dev->open_count++ ) {
|
|
spin_unlock( &dev->count_lock );
|
|
return DRM(setup)( dev );
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
}
|
|
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Release file.
|
|
*
|
|
* \param inode device inode
|
|
* \param filp file pointer.
|
|
* \return zero on success or a negative number on failure.
|
|
*
|
|
* If the hardware lock is held then free it, and take it again for the kernel
|
|
* context since it's necessary to reclaim buffers. Unlink the file private
|
|
* data from its list and free it. Decreases the open count and if it reaches
|
|
* zero calls takedown().
|
|
*/
|
|
int DRM(release)( struct inode *inode, struct file *filp )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev;
|
|
int retcode = 0;
|
|
|
|
lock_kernel();
|
|
dev = priv->dev;
|
|
|
|
DRM_DEBUG( "open_count = %d\n", dev->open_count );
|
|
|
|
DRIVER_PRERELEASE();
|
|
|
|
/* ========================================================
|
|
* Begin inline drm_release
|
|
*/
|
|
|
|
DRM_DEBUG( "pid = %d, device = 0x%lx, open_count = %d\n",
|
|
current->pid, (long)dev->device, dev->open_count );
|
|
|
|
if ( priv->lock_count && dev->lock.hw_lock &&
|
|
_DRM_LOCK_IS_HELD(dev->lock.hw_lock->lock) &&
|
|
dev->lock.filp == filp ) {
|
|
DRM_DEBUG( "File %p released, freeing lock for context %d\n",
|
|
filp,
|
|
_DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) );
|
|
#if __HAVE_RELEASE
|
|
DRIVER_RELEASE();
|
|
#endif
|
|
DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
_DRM_LOCKING_CONTEXT(dev->lock.hw_lock->lock) );
|
|
|
|
/* FIXME: may require heavy-handed reset of
|
|
hardware at this point, possibly
|
|
processed via a callback to the X
|
|
server. */
|
|
}
|
|
#if __HAVE_RELEASE
|
|
else if ( priv->lock_count && dev->lock.hw_lock ) {
|
|
/* The lock is required to reclaim buffers */
|
|
DECLARE_WAITQUEUE( entry, current );
|
|
|
|
add_wait_queue( &dev->lock.lock_queue, &entry );
|
|
for (;;) {
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
if ( !dev->lock.hw_lock ) {
|
|
/* Device has been unregistered */
|
|
retcode = -EINTR;
|
|
break;
|
|
}
|
|
if ( DRM(lock_take)( &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT ) ) {
|
|
dev->lock.filp = filp;
|
|
dev->lock.lock_time = jiffies;
|
|
atomic_inc( &dev->counts[_DRM_STAT_LOCKS] );
|
|
break; /* Got lock */
|
|
}
|
|
/* Contention */
|
|
schedule();
|
|
if ( signal_pending( current ) ) {
|
|
retcode = -ERESTARTSYS;
|
|
break;
|
|
}
|
|
}
|
|
current->state = TASK_RUNNING;
|
|
remove_wait_queue( &dev->lock.lock_queue, &entry );
|
|
if( !retcode ) {
|
|
DRIVER_RELEASE();
|
|
DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT );
|
|
}
|
|
}
|
|
#elif __HAVE_DMA
|
|
DRM(reclaim_buffers)( filp );
|
|
#endif
|
|
|
|
DRM(fasync)( -1, filp, 0 );
|
|
|
|
down( &dev->struct_sem );
|
|
if ( priv->remove_auth_on_close == 1 ) {
|
|
drm_file_t *temp = dev->file_first;
|
|
while ( temp ) {
|
|
temp->authenticated = 0;
|
|
temp = temp->next;
|
|
}
|
|
}
|
|
if ( priv->prev ) {
|
|
priv->prev->next = priv->next;
|
|
} else {
|
|
dev->file_first = priv->next;
|
|
}
|
|
if ( priv->next ) {
|
|
priv->next->prev = priv->prev;
|
|
} else {
|
|
dev->file_last = priv->prev;
|
|
}
|
|
up( &dev->struct_sem );
|
|
|
|
DRM(free)( priv, sizeof(*priv), DRM_MEM_FILES );
|
|
|
|
/* ========================================================
|
|
* End inline drm_release
|
|
*/
|
|
|
|
atomic_inc( &dev->counts[_DRM_STAT_CLOSES] );
|
|
spin_lock( &dev->count_lock );
|
|
if ( !--dev->open_count ) {
|
|
if ( atomic_read( &dev->ioctl_count ) || dev->blocked ) {
|
|
DRM_ERROR( "Device busy: %d %d\n",
|
|
atomic_read( &dev->ioctl_count ),
|
|
dev->blocked );
|
|
spin_unlock( &dev->count_lock );
|
|
unlock_kernel();
|
|
return -EBUSY;
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
unlock_kernel();
|
|
return DRM(takedown)( dev );
|
|
}
|
|
spin_unlock( &dev->count_lock );
|
|
|
|
unlock_kernel();
|
|
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Called whenever a process performs an ioctl on /dev/drm.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Looks up the ioctl function in the ::ioctls table, checking for root
|
|
* previleges if so required, and dispatches to the respective function.
|
|
*/
|
|
int DRM(ioctl)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
drm_ioctl_desc_t *ioctl;
|
|
drm_ioctl_t *func;
|
|
int nr = DRM_IOCTL_NR(cmd);
|
|
int retcode = 0;
|
|
|
|
atomic_inc( &dev->ioctl_count );
|
|
atomic_inc( &dev->counts[_DRM_STAT_IOCTLS] );
|
|
++priv->ioctl_count;
|
|
|
|
DRM_DEBUG( "pid=%d, cmd=0x%02x, nr=0x%02x, dev 0x%lx, auth=%d\n",
|
|
current->pid, cmd, nr, (long)dev->device,
|
|
priv->authenticated );
|
|
|
|
if ( nr >= DRIVER_IOCTL_COUNT ) {
|
|
retcode = -EINVAL;
|
|
} else {
|
|
ioctl = &DRM(ioctls)[nr];
|
|
func = ioctl->func;
|
|
|
|
if ( !func ) {
|
|
DRM_DEBUG( "no function\n" );
|
|
retcode = -EINVAL;
|
|
} else if ( ( ioctl->root_only && !capable( CAP_SYS_ADMIN ) )||
|
|
( ioctl->auth_needed && !priv->authenticated ) ) {
|
|
retcode = -EACCES;
|
|
} else {
|
|
retcode = func( inode, filp, cmd, arg );
|
|
}
|
|
}
|
|
|
|
atomic_dec( &dev->ioctl_count );
|
|
return retcode;
|
|
}
|
|
|
|
/**
|
|
* Lock ioctl.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_lock structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Add the current task to the lock wait queue, and attempt to take to lock.
|
|
*/
|
|
int DRM(lock)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
DECLARE_WAITQUEUE( entry, current );
|
|
drm_lock_t lock;
|
|
int ret = 0;
|
|
#if __HAVE_MULTIPLE_DMA_QUEUES
|
|
drm_queue_t *q;
|
|
#endif
|
|
|
|
++priv->lock_count;
|
|
|
|
if ( copy_from_user( &lock, (drm_lock_t *)arg, sizeof(lock) ) )
|
|
return -EFAULT;
|
|
|
|
if ( lock.context == DRM_KERNEL_CONTEXT ) {
|
|
DRM_ERROR( "Process %d using kernel context %d\n",
|
|
current->pid, lock.context );
|
|
return -EINVAL;
|
|
}
|
|
|
|
DRM_DEBUG( "%d (pid %d) requests lock (0x%08x), flags = 0x%08x\n",
|
|
lock.context, current->pid,
|
|
dev->lock.hw_lock->lock, lock.flags );
|
|
|
|
#if __HAVE_DMA_QUEUE
|
|
if ( lock.context < 0 )
|
|
return -EINVAL;
|
|
#elif __HAVE_MULTIPLE_DMA_QUEUES
|
|
if ( lock.context < 0 || lock.context >= dev->queue_count )
|
|
return -EINVAL;
|
|
q = dev->queuelist[lock.context];
|
|
#endif
|
|
|
|
#if __HAVE_DMA_FLUSH
|
|
ret = DRM(flush_block_and_flush)( dev, lock.context, lock.flags );
|
|
#endif
|
|
if ( !ret ) {
|
|
add_wait_queue( &dev->lock.lock_queue, &entry );
|
|
for (;;) {
|
|
current->state = TASK_INTERRUPTIBLE;
|
|
if ( !dev->lock.hw_lock ) {
|
|
/* Device has been unregistered */
|
|
ret = -EINTR;
|
|
break;
|
|
}
|
|
if ( DRM(lock_take)( &dev->lock.hw_lock->lock,
|
|
lock.context ) ) {
|
|
dev->lock.filp = filp;
|
|
dev->lock.lock_time = jiffies;
|
|
atomic_inc( &dev->counts[_DRM_STAT_LOCKS] );
|
|
break; /* Got lock */
|
|
}
|
|
|
|
/* Contention */
|
|
schedule();
|
|
if ( signal_pending( current ) ) {
|
|
ret = -ERESTARTSYS;
|
|
break;
|
|
}
|
|
}
|
|
current->state = TASK_RUNNING;
|
|
remove_wait_queue( &dev->lock.lock_queue, &entry );
|
|
}
|
|
|
|
#if __HAVE_DMA_FLUSH
|
|
DRM(flush_unblock)( dev, lock.context, lock.flags ); /* cleanup phase */
|
|
#endif
|
|
|
|
if ( !ret ) {
|
|
sigemptyset( &dev->sigmask );
|
|
sigaddset( &dev->sigmask, SIGSTOP );
|
|
sigaddset( &dev->sigmask, SIGTSTP );
|
|
sigaddset( &dev->sigmask, SIGTTIN );
|
|
sigaddset( &dev->sigmask, SIGTTOU );
|
|
dev->sigdata.context = lock.context;
|
|
dev->sigdata.lock = dev->lock.hw_lock;
|
|
block_all_signals( DRM(notifier),
|
|
&dev->sigdata, &dev->sigmask );
|
|
|
|
#if __HAVE_DMA_READY
|
|
if ( lock.flags & _DRM_LOCK_READY ) {
|
|
DRIVER_DMA_READY();
|
|
}
|
|
#endif
|
|
#if __HAVE_DMA_QUIESCENT
|
|
if ( lock.flags & _DRM_LOCK_QUIESCENT ) {
|
|
DRIVER_DMA_QUIESCENT();
|
|
}
|
|
#endif
|
|
/* __HAVE_KERNEL_CTX_SWITCH isn't used by any of the
|
|
* drm modules in the DRI cvs tree, but it is required
|
|
* by the Sparc driver.
|
|
*/
|
|
#if __HAVE_KERNEL_CTX_SWITCH
|
|
if ( dev->last_context != lock.context ) {
|
|
DRM(context_switch)(dev, dev->last_context,
|
|
lock.context);
|
|
}
|
|
#endif
|
|
}
|
|
|
|
DRM_DEBUG( "%d %s\n", lock.context, ret ? "interrupted" : "has lock" );
|
|
|
|
return ret;
|
|
}
|
|
|
|
/**
|
|
* Unlock ioctl.
|
|
*
|
|
* \param inode device inode.
|
|
* \param filp file pointer.
|
|
* \param cmd command.
|
|
* \param arg user argument, pointing to a drm_lock structure.
|
|
* \return zero on success or negative number on failure.
|
|
*
|
|
* Transfer and free the lock.
|
|
*/
|
|
int DRM(unlock)( struct inode *inode, struct file *filp,
|
|
unsigned int cmd, unsigned long arg )
|
|
{
|
|
drm_file_t *priv = filp->private_data;
|
|
drm_device_t *dev = priv->dev;
|
|
drm_lock_t lock;
|
|
|
|
if ( copy_from_user( &lock, (drm_lock_t *)arg, sizeof(lock) ) )
|
|
return -EFAULT;
|
|
|
|
if ( lock.context == DRM_KERNEL_CONTEXT ) {
|
|
DRM_ERROR( "Process %d using kernel context %d\n",
|
|
current->pid, lock.context );
|
|
return -EINVAL;
|
|
}
|
|
|
|
atomic_inc( &dev->counts[_DRM_STAT_UNLOCKS] );
|
|
|
|
/* __HAVE_KERNEL_CTX_SWITCH isn't used by any of the drm
|
|
* modules in the DRI cvs tree, but it is required by the
|
|
* Sparc driver.
|
|
*/
|
|
#if __HAVE_KERNEL_CTX_SWITCH
|
|
/* We no longer really hold it, but if we are the next
|
|
* agent to request it then we should just be able to
|
|
* take it immediately and not eat the ioctl.
|
|
*/
|
|
dev->lock.filp = 0;
|
|
{
|
|
__volatile__ unsigned int *plock = &dev->lock.hw_lock->lock;
|
|
unsigned int old, new, prev, ctx;
|
|
|
|
ctx = lock.context;
|
|
do {
|
|
old = *plock;
|
|
new = ctx;
|
|
prev = cmpxchg(plock, old, new);
|
|
} while (prev != old);
|
|
}
|
|
wake_up_interruptible(&dev->lock.lock_queue);
|
|
#else
|
|
DRM(lock_transfer)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT );
|
|
#if __HAVE_DMA_SCHEDULE
|
|
DRM(dma_schedule)( dev, 1 );
|
|
#endif
|
|
|
|
if ( DRM(lock_free)( dev, &dev->lock.hw_lock->lock,
|
|
DRM_KERNEL_CONTEXT ) ) {
|
|
DRM_ERROR( "\n" );
|
|
}
|
|
#endif /* !__HAVE_KERNEL_CTX_SWITCH */
|
|
|
|
unblock_all_signals();
|
|
return 0;
|
|
}
|