/**********************************************************
* Copyright 2009-2015 VMware, Inc. All rights reserved.
*
* Permission is hereby granted, free of charge, to any person
* obtaining a copy of this software and associated documentation
* files (the "Software"), to deal in the Software without
* restriction, including without limitation the rights to use, copy,
* modify, merge, publish, distribute, sublicense, and/or sell copies
* of the Software, and to permit persons to whom the Software is
* furnished to do so, subject to the following conditions:
*
* The above copyright notice and this permission notice shall be
* included in all copies or substantial portions of the Software.
*
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
* SOFTWARE.
*
**********************************************************/
/**
* @file
*
* Wrappers for DRM ioctl functionlaity used by the rest of the vmw
* drm winsys.
*
* Based on svgaicd_escape.c
*/
#include "svga_cmd.h"
#include "util/u_memory.h"
#include "util/u_math.h"
#include "svgadump/svga_dump.h"
#include "state_tracker/drm_driver.h"
#include "vmw_screen.h"
#include "vmw_context.h"
#include "vmw_fence.h"
#include "xf86drm.h"
#include "vmwgfx_drm.h"
#include "svga3d_caps.h"
#include "svga3d_reg.h"
#include "os/os_mman.h"
#include <errno.h>
#include <unistd.h>
#define VMW_MAX_DEFAULT_TEXTURE_SIZE (128 * 1024 * 1024)
#define VMW_FENCE_TIMEOUT_SECONDS 60
struct vmw_region
{
uint32_t handle;
uint64_t map_handle;
void *data;
uint32_t map_count;
int drm_fd;
uint32_t size;
};
uint32_t
vmw_region_size(struct vmw_region *region)
{
return region->size;
}
#if defined(__DragonFly__) || defined(__FreeBSD__) || \
defined(__NetBSD__) || defined(__OpenBSD__)
#define ERESTART EINTR
#endif
uint32
vmw_ioctl_context_create(struct vmw_winsys_screen *vws)
{
struct drm_vmw_context_arg c_arg;
int ret;
VMW_FUNC;
ret = drmCommandRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_CONTEXT,
&c_arg, sizeof(c_arg));
if (ret)
return -1;
vmw_printf("Context id is %d\n", c_arg.cid);
return c_arg.cid;
}
uint32
vmw_ioctl_extended_context_create(struct vmw_winsys_screen *vws,
boolean vgpu10)
{
union drm_vmw_extended_context_arg c_arg;
int ret;
VMW_FUNC;
memset(&c_arg, 0, sizeof(c_arg));
c_arg.req = (vgpu10 ? drm_vmw_context_vgpu10 : drm_vmw_context_legacy);
ret = drmCommandWriteRead(vws->ioctl.drm_fd,
DRM_VMW_CREATE_EXTENDED_CONTEXT,
&c_arg, sizeof(c_arg));
if (ret)
return -1;
vmw_printf("Context id is %d\n", c_arg.cid);
return c_arg.rep.cid;
}
void
vmw_ioctl_context_destroy(struct vmw_winsys_screen *vws, uint32 cid)
{
struct drm_vmw_context_arg c_arg;
VMW_FUNC;
memset(&c_arg, 0, sizeof(c_arg));
c_arg.cid = cid;
(void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_CONTEXT,
&c_arg, sizeof(c_arg));
}
uint32
vmw_ioctl_surface_create(struct vmw_winsys_screen *vws,
SVGA3dSurfaceFlags flags,
SVGA3dSurfaceFormat format,
unsigned usage,
SVGA3dSize size,
uint32_t numFaces, uint32_t numMipLevels,
unsigned sampleCount)
{
union drm_vmw_surface_create_arg s_arg;
struct drm_vmw_surface_create_req *req = &s_arg.req;
struct drm_vmw_surface_arg *rep = &s_arg.rep;
struct drm_vmw_size sizes[DRM_VMW_MAX_SURFACE_FACES*
DRM_VMW_MAX_MIP_LEVELS];
struct drm_vmw_size *cur_size;
uint32_t iFace;
uint32_t iMipLevel;
int ret;
vmw_printf("%s flags %d format %d\n", __FUNCTION__, flags, format);
memset(&s_arg, 0, sizeof(s_arg));
req->flags = (uint32_t) flags;
req->scanout = !!(usage & SVGA_SURFACE_USAGE_SCANOUT);
req->format = (uint32_t) format;
req->shareable = !!(usage & SVGA_SURFACE_USAGE_SHARED);
assert(numFaces * numMipLevels < DRM_VMW_MAX_SURFACE_FACES*
DRM_VMW_MAX_MIP_LEVELS);
cur_size = sizes;
for (iFace = 0; iFace < numFaces; ++iFace) {
SVGA3dSize mipSize = size;
req->mip_levels[iFace] = numMipLevels;
for (iMipLevel = 0; iMipLevel < numMipLevels; ++iMipLevel) {
cur_size->width = mipSize.width;
cur_size->height = mipSize.height;
cur_size->depth = mipSize.depth;
mipSize.width = MAX2(mipSize.width >> 1, 1);
mipSize.height = MAX2(mipSize.height >> 1, 1);
mipSize.depth = MAX2(mipSize.depth >> 1, 1);
cur_size++;
}
}
for (iFace = numFaces; iFace < SVGA3D_MAX_SURFACE_FACES; ++iFace) {
req->mip_levels[iFace] = 0;
}
req->size_addr = (unsigned long)&sizes;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_SURFACE,
&s_arg, sizeof(s_arg));
if (ret)
return -1;
vmw_printf("Surface id is %d\n", rep->sid);
return rep->sid;
}
uint32
vmw_ioctl_gb_surface_create(struct vmw_winsys_screen *vws,
SVGA3dSurfaceFlags flags,
SVGA3dSurfaceFormat format,
unsigned usage,
SVGA3dSize size,
uint32_t numFaces,
uint32_t numMipLevels,
unsigned sampleCount,
uint32_t buffer_handle,
struct vmw_region **p_region)
{
union drm_vmw_gb_surface_create_arg s_arg;
struct drm_vmw_gb_surface_create_req *req = &s_arg.req;
struct drm_vmw_gb_surface_create_rep *rep = &s_arg.rep;
struct vmw_region *region = NULL;
int ret;
vmw_printf("%s flags %d format %d\n", __FUNCTION__, flags, format);
if (p_region) {
region = CALLOC_STRUCT(vmw_region);
if (!region)
return SVGA3D_INVALID_ID;
}
memset(&s_arg, 0, sizeof(s_arg));
req->svga3d_flags = (uint32_t) flags;
if (usage & SVGA_SURFACE_USAGE_SCANOUT)
req->drm_surface_flags |= drm_vmw_surface_flag_scanout;
req->format = (uint32_t) format;
if (usage & SVGA_SURFACE_USAGE_SHARED)
req->drm_surface_flags |= drm_vmw_surface_flag_shareable;
req->drm_surface_flags |= drm_vmw_surface_flag_create_buffer;
req->base_size.width = size.width;
req->base_size.height = size.height;
req->base_size.depth = size.depth;
req->mip_levels = numMipLevels;
req->multisample_count = 0;
req->autogen_filter = SVGA3D_TEX_FILTER_NONE;
if (vws->base.have_vgpu10) {
req->array_size = numFaces;
req->multisample_count = sampleCount;
} else {
assert(numFaces * numMipLevels < DRM_VMW_MAX_SURFACE_FACES*
DRM_VMW_MAX_MIP_LEVELS);
req->array_size = 0;
}
if (buffer_handle)
req->buffer_handle = buffer_handle;
else
req->buffer_handle = SVGA3D_INVALID_ID;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GB_SURFACE_CREATE,
&s_arg, sizeof(s_arg));
if (ret)
goto out_fail_create;
if (p_region) {
region->handle = rep->buffer_handle;
region->map_handle = rep->buffer_map_handle;
region->drm_fd = vws->ioctl.drm_fd;
region->size = rep->backup_size;
*p_region = region;
}
vmw_printf("Surface id is %d\n", rep->sid);
return rep->handle;
out_fail_create:
FREE(region);
return SVGA3D_INVALID_ID;
}
/**
* vmw_ioctl_surface_req - Fill in a struct surface_req
*
* @vws: Winsys screen
* @whandle: Surface handle
* @req: The struct surface req to fill in
* @needs_unref: This call takes a kernel surface reference that needs to
* be unreferenced.
*
* Returns 0 on success, negative error type otherwise.
* Fills in the surface_req structure according to handle type and kernel
* capabilities.
*/
static int
vmw_ioctl_surface_req(const struct vmw_winsys_screen *vws,
const struct winsys_handle *whandle,
struct drm_vmw_surface_arg *req,
boolean *needs_unref)
{
int ret;
switch(whandle->type) {
case DRM_API_HANDLE_TYPE_SHARED:
case DRM_API_HANDLE_TYPE_KMS:
*needs_unref = FALSE;
req->handle_type = DRM_VMW_HANDLE_LEGACY;
req->sid = whandle->handle;
break;
case DRM_API_HANDLE_TYPE_FD:
if (!vws->ioctl.have_drm_2_6) {
uint32_t handle;
ret = drmPrimeFDToHandle(vws->ioctl.drm_fd, whandle->handle, &handle);
if (ret) {
vmw_error("Failed to get handle from prime fd %d.\n",
(int) whandle->handle);
return -EINVAL;
}
*needs_unref = TRUE;
req->handle_type = DRM_VMW_HANDLE_LEGACY;
req->sid = handle;
} else {
*needs_unref = FALSE;
req->handle_type = DRM_VMW_HANDLE_PRIME;
req->sid = whandle->handle;
}
break;
default:
vmw_error("Attempt to import unsupported handle type %d.\n",
whandle->type);
return -EINVAL;
}
return 0;
}
/**
* vmw_ioctl_gb_surface_ref - Put a reference on a guest-backed surface and
* get surface information
*
* @vws: Screen to register the reference on
* @handle: Kernel handle of the guest-backed surface
* @flags: flags used when the surface was created
* @format: Format used when the surface was created
* @numMipLevels: Number of mipmap levels of the surface
* @p_region: On successful return points to a newly allocated
* struct vmw_region holding a reference to the surface backup buffer.
*
* Returns 0 on success, a system error on failure.
*/
int
vmw_ioctl_gb_surface_ref(struct vmw_winsys_screen *vws,
const struct winsys_handle *whandle,
SVGA3dSurfaceFlags *flags,
SVGA3dSurfaceFormat *format,
uint32_t *numMipLevels,
uint32_t *handle,
struct vmw_region **p_region)
{
union drm_vmw_gb_surface_reference_arg s_arg;
struct drm_vmw_surface_arg *req = &s_arg.req;
struct drm_vmw_gb_surface_ref_rep *rep = &s_arg.rep;
struct vmw_region *region = NULL;
boolean needs_unref = FALSE;
int ret;
vmw_printf("%s flags %d format %d\n", __FUNCTION__, flags, format);
assert(p_region != NULL);
region = CALLOC_STRUCT(vmw_region);
if (!region)
return -ENOMEM;
memset(&s_arg, 0, sizeof(s_arg));
ret = vmw_ioctl_surface_req(vws, whandle, req, &needs_unref);
if (ret)
goto out_fail_req;
*handle = req->sid;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GB_SURFACE_REF,
&s_arg, sizeof(s_arg));
if (ret)
goto out_fail_ref;
region->handle = rep->crep.buffer_handle;
region->map_handle = rep->crep.buffer_map_handle;
region->drm_fd = vws->ioctl.drm_fd;
region->size = rep->crep.backup_size;
*p_region = region;
*handle = rep->crep.handle;
*flags = rep->creq.svga3d_flags;
*format = rep->creq.format;
*numMipLevels = rep->creq.mip_levels;
if (needs_unref)
vmw_ioctl_surface_destroy(vws, *handle);
return 0;
out_fail_ref:
if (needs_unref)
vmw_ioctl_surface_destroy(vws, *handle);
out_fail_req:
FREE(region);
return ret;
}
void
vmw_ioctl_surface_destroy(struct vmw_winsys_screen *vws, uint32 sid)
{
struct drm_vmw_surface_arg s_arg;
VMW_FUNC;
memset(&s_arg, 0, sizeof(s_arg));
s_arg.sid = sid;
(void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_SURFACE,
&s_arg, sizeof(s_arg));
}
void
vmw_ioctl_command(struct vmw_winsys_screen *vws, int32_t cid,
uint32_t throttle_us, void *commands, uint32_t size,
struct pipe_fence_handle **pfence, int32_t imported_fence_fd,
uint32_t flags)
{
struct drm_vmw_execbuf_arg arg;
struct drm_vmw_fence_rep rep;
int ret;
int argsize;
#ifdef DEBUG
{
static boolean firsttime = TRUE;
static boolean debug = FALSE;
static boolean skip = FALSE;
if (firsttime) {
debug = debug_get_bool_option("SVGA_DUMP_CMD", FALSE);
skip = debug_get_bool_option("SVGA_SKIP_CMD", FALSE);
}
if (debug) {
VMW_FUNC;
svga_dump_commands(commands, size);
}
firsttime = FALSE;
if (skip) {
size = 0;
}
}
#endif
memset(&arg, 0, sizeof(arg));
memset(&rep, 0, sizeof(rep));
if (flags & SVGA_HINT_FLAG_EXPORT_FENCE_FD) {
arg.flags |= DRM_VMW_EXECBUF_FLAG_EXPORT_FENCE_FD;
}
if (imported_fence_fd != -1) {
arg.flags |= DRM_VMW_EXECBUF_FLAG_IMPORT_FENCE_FD;
}
rep.error = -EFAULT;
if (pfence)
arg.fence_rep = (unsigned long)&rep;
arg.commands = (unsigned long)commands;
arg.command_size = size;
arg.throttle_us = throttle_us;
arg.version = vws->ioctl.drm_execbuf_version;
arg.context_handle = (vws->base.have_vgpu10 ? cid : SVGA3D_INVALID_ID);
/* Older DRM module requires this to be zero */
if (vws->base.have_fence_fd)
arg.imported_fence_fd = imported_fence_fd;
/* In DRM_VMW_EXECBUF_VERSION 1, the drm_vmw_execbuf_arg structure ends with
* the flags field. The structure size sent to drmCommandWrite must match
* the drm_execbuf_version. Otherwise, an invalid value will be returned.
*/
argsize = vws->ioctl.drm_execbuf_version > 1 ? sizeof(arg) :
offsetof(struct drm_vmw_execbuf_arg, context_handle);
do {
ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_EXECBUF, &arg, argsize);
} while(ret == -ERESTART);
if (ret) {
vmw_error("%s error %s.\n", __FUNCTION__, strerror(-ret));
abort();
}
if (rep.error) {
/*
* Kernel has already synced, or caller requested no fence.
*/
if (pfence)
*pfence = NULL;
} else {
if (pfence) {
vmw_fences_signal(vws->fence_ops, rep.passed_seqno, rep.seqno,
TRUE);
/* Older DRM module will set this to zero, but -1 is the proper FD
* to use for no Fence FD support */
if (!vws->base.have_fence_fd)
rep.fd = -1;
*pfence = vmw_fence_create(vws->fence_ops, rep.handle,
rep.seqno, rep.mask, rep.fd);
if (*pfence == NULL) {
/*
* Fence creation failed. Need to sync.
*/
(void) vmw_ioctl_fence_finish(vws, rep.handle, rep.mask);
vmw_ioctl_fence_unref(vws, rep.handle);
}
}
}
}
struct vmw_region *
vmw_ioctl_region_create(struct vmw_winsys_screen *vws, uint32_t size)
{
struct vmw_region *region;
union drm_vmw_alloc_dmabuf_arg arg;
struct drm_vmw_alloc_dmabuf_req *req = &arg.req;
struct drm_vmw_dmabuf_rep *rep = &arg.rep;
int ret;
vmw_printf("%s: size = %u\n", __FUNCTION__, size);
region = CALLOC_STRUCT(vmw_region);
if (!region)
goto out_err1;
memset(&arg, 0, sizeof(arg));
req->size = size;
do {
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_ALLOC_DMABUF, &arg,
sizeof(arg));
} while (ret == -ERESTART);
if (ret) {
vmw_error("IOCTL failed %d: %s\n", ret, strerror(-ret));
goto out_err1;
}
region->data = NULL;
region->handle = rep->handle;
region->map_handle = rep->map_handle;
region->map_count = 0;
region->size = size;
region->drm_fd = vws->ioctl.drm_fd;
vmw_printf(" gmrId = %u, offset = %u\n",
region->ptr.gmrId, region->ptr.offset);
return region;
out_err1:
FREE(region);
return NULL;
}
void
vmw_ioctl_region_destroy(struct vmw_region *region)
{
struct drm_vmw_unref_dmabuf_arg arg;
vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
region->ptr.gmrId, region->ptr.offset);
if (region->data) {
os_munmap(region->data, region->size);
region->data = NULL;
}
memset(&arg, 0, sizeof(arg));
arg.handle = region->handle;
drmCommandWrite(region->drm_fd, DRM_VMW_UNREF_DMABUF, &arg, sizeof(arg));
FREE(region);
}
SVGAGuestPtr
vmw_ioctl_region_ptr(struct vmw_region *region)
{
SVGAGuestPtr ptr = {region->handle, 0};
return ptr;
}
void *
vmw_ioctl_region_map(struct vmw_region *region)
{
void *map;
vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
region->ptr.gmrId, region->ptr.offset);
if (region->data == NULL) {
map = os_mmap(NULL, region->size, PROT_READ | PROT_WRITE, MAP_SHARED,
region->drm_fd, region->map_handle);
if (map == MAP_FAILED) {
vmw_error("%s: Map failed.\n", __FUNCTION__);
return NULL;
}
region->data = map;
}
++region->map_count;
return region->data;
}
void
vmw_ioctl_region_unmap(struct vmw_region *region)
{
vmw_printf("%s: gmrId = %u, offset = %u\n", __FUNCTION__,
region->ptr.gmrId, region->ptr.offset);
--region->map_count;
}
/**
* vmw_ioctl_syncforcpu - Synchronize a buffer object for CPU usage
*
* @region: Pointer to a struct vmw_region representing the buffer object.
* @dont_block: Dont wait for GPU idle, but rather return -EBUSY if the
* GPU is busy with the buffer object.
* @readonly: Hint that the CPU access is read-only.
* @allow_cs: Allow concurrent command submission while the buffer is
* synchronized for CPU. If FALSE command submissions referencing the
* buffer will block until a corresponding call to vmw_ioctl_releasefromcpu.
*
* This function idles any GPU activities touching the buffer and blocks
* command submission of commands referencing the buffer, even from
* other processes.
*/
int
vmw_ioctl_syncforcpu(struct vmw_region *region,
boolean dont_block,
boolean readonly,
boolean allow_cs)
{
struct drm_vmw_synccpu_arg arg;
memset(&arg, 0, sizeof(arg));
arg.op = drm_vmw_synccpu_grab;
arg.handle = region->handle;
arg.flags = drm_vmw_synccpu_read;
if (!readonly)
arg.flags |= drm_vmw_synccpu_write;
if (dont_block)
arg.flags |= drm_vmw_synccpu_dontblock;
if (allow_cs)
arg.flags |= drm_vmw_synccpu_allow_cs;
return drmCommandWrite(region->drm_fd, DRM_VMW_SYNCCPU, &arg, sizeof(arg));
}
/**
* vmw_ioctl_releasefromcpu - Undo a previous syncforcpu.
*
* @region: Pointer to a struct vmw_region representing the buffer object.
* @readonly: Should hold the same value as the matching syncforcpu call.
* @allow_cs: Should hold the same value as the matching syncforcpu call.
*/
void
vmw_ioctl_releasefromcpu(struct vmw_region *region,
boolean readonly,
boolean allow_cs)
{
struct drm_vmw_synccpu_arg arg;
memset(&arg, 0, sizeof(arg));
arg.op = drm_vmw_synccpu_release;
arg.handle = region->handle;
arg.flags = drm_vmw_synccpu_read;
if (!readonly)
arg.flags |= drm_vmw_synccpu_write;
if (allow_cs)
arg.flags |= drm_vmw_synccpu_allow_cs;
(void) drmCommandWrite(region->drm_fd, DRM_VMW_SYNCCPU, &arg, sizeof(arg));
}
void
vmw_ioctl_fence_unref(struct vmw_winsys_screen *vws,
uint32_t handle)
{
struct drm_vmw_fence_arg arg;
int ret;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_FENCE_UNREF,
&arg, sizeof(arg));
if (ret != 0)
vmw_error("%s Failed\n", __FUNCTION__);
}
static inline uint32_t
vmw_drm_fence_flags(uint32_t flags)
{
uint32_t dflags = 0;
if (flags & SVGA_FENCE_FLAG_EXEC)
dflags |= DRM_VMW_FENCE_FLAG_EXEC;
if (flags & SVGA_FENCE_FLAG_QUERY)
dflags |= DRM_VMW_FENCE_FLAG_QUERY;
return dflags;
}
int
vmw_ioctl_fence_signalled(struct vmw_winsys_screen *vws,
uint32_t handle,
uint32_t flags)
{
struct drm_vmw_fence_signaled_arg arg;
uint32_t vflags = vmw_drm_fence_flags(flags);
int ret;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.flags = vflags;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_FENCE_SIGNALED,
&arg, sizeof(arg));
if (ret != 0)
return ret;
vmw_fences_signal(vws->fence_ops, arg.passed_seqno, 0, FALSE);
return (arg.signaled) ? 0 : -1;
}
int
vmw_ioctl_fence_finish(struct vmw_winsys_screen *vws,
uint32_t handle,
uint32_t flags)
{
struct drm_vmw_fence_wait_arg arg;
uint32_t vflags = vmw_drm_fence_flags(flags);
int ret;
memset(&arg, 0, sizeof(arg));
arg.handle = handle;
arg.timeout_us = VMW_FENCE_TIMEOUT_SECONDS*1000000;
arg.lazy = 0;
arg.flags = vflags;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_FENCE_WAIT,
&arg, sizeof(arg));
if (ret != 0)
vmw_error("%s Failed\n", __FUNCTION__);
return 0;
}
uint32
vmw_ioctl_shader_create(struct vmw_winsys_screen *vws,
SVGA3dShaderType type,
uint32 code_len)
{
struct drm_vmw_shader_create_arg sh_arg;
int ret;
VMW_FUNC;
memset(&sh_arg, 0, sizeof(sh_arg));
sh_arg.size = code_len;
sh_arg.buffer_handle = SVGA3D_INVALID_ID;
sh_arg.shader_handle = SVGA3D_INVALID_ID;
switch (type) {
case SVGA3D_SHADERTYPE_VS:
sh_arg.shader_type = drm_vmw_shader_type_vs;
break;
case SVGA3D_SHADERTYPE_PS:
sh_arg.shader_type = drm_vmw_shader_type_ps;
break;
default:
assert(!"Invalid shader type.");
break;
}
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_CREATE_SHADER,
&sh_arg, sizeof(sh_arg));
if (ret)
return SVGA3D_INVALID_ID;
return sh_arg.shader_handle;
}
void
vmw_ioctl_shader_destroy(struct vmw_winsys_screen *vws, uint32 shid)
{
struct drm_vmw_shader_arg sh_arg;
VMW_FUNC;
memset(&sh_arg, 0, sizeof(sh_arg));
sh_arg.handle = shid;
(void)drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_UNREF_SHADER,
&sh_arg, sizeof(sh_arg));
}
static int
vmw_ioctl_parse_caps(struct vmw_winsys_screen *vws,
const uint32_t *cap_buffer)
{
int i;
if (vws->base.have_gb_objects) {
for (i = 0; i < vws->ioctl.num_cap_3d; ++i) {
vws->ioctl.cap_3d[i].has_cap = TRUE;
vws->ioctl.cap_3d[i].result.u = cap_buffer[i];
}
return 0;
} else {
const uint32 *capsBlock;
const SVGA3dCapsRecord *capsRecord = NULL;
uint32 offset;
const SVGA3dCapPair *capArray;
int numCaps, index;
/*
* Search linearly through the caps block records for the specified type.
*/
capsBlock = cap_buffer;
for (offset = 0; capsBlock[offset] != 0; offset += capsBlock[offset]) {
const SVGA3dCapsRecord *record;
assert(offset < SVGA_FIFO_3D_CAPS_SIZE);
record = (const SVGA3dCapsRecord *) (capsBlock + offset);
if ((record->header.type >= SVGA3DCAPS_RECORD_DEVCAPS_MIN) &&
(record->header.type <= SVGA3DCAPS_RECORD_DEVCAPS_MAX) &&
(!capsRecord || (record->header.type > capsRecord->header.type))) {
capsRecord = record;
}
}
if(!capsRecord)
return -1;
/*
* Calculate the number of caps from the size of the record.
*/
capArray = (const SVGA3dCapPair *) capsRecord->data;
numCaps = (int) ((capsRecord->header.length * sizeof(uint32) -
sizeof capsRecord->header) / (2 * sizeof(uint32)));
for (i = 0; i < numCaps; i++) {
index = capArray[i][0];
if (index < vws->ioctl.num_cap_3d) {
vws->ioctl.cap_3d[index].has_cap = TRUE;
vws->ioctl.cap_3d[index].result.u = capArray[i][1];
} else {
debug_printf("Unknown devcaps seen: %d\n", index);
}
}
}
return 0;
}
boolean
vmw_ioctl_init(struct vmw_winsys_screen *vws)
{
struct drm_vmw_getparam_arg gp_arg;
struct drm_vmw_get_3d_cap_arg cap_arg;
unsigned int size;
int ret;
uint32_t *cap_buffer;
drmVersionPtr version;
boolean drm_gb_capable;
boolean have_drm_2_5;
VMW_FUNC;
version = drmGetVersion(vws->ioctl.drm_fd);
if (!version)
goto out_no_version;
have_drm_2_5 = version->version_major > 2 ||
(version->version_major == 2 && version->version_minor > 4);
vws->ioctl.have_drm_2_6 = version->version_major > 2 ||
(version->version_major == 2 && version->version_minor > 5);
vws->ioctl.have_drm_2_9 = version->version_major > 2 ||
(version->version_major == 2 && version->version_minor > 8);
vws->ioctl.drm_execbuf_version = vws->ioctl.have_drm_2_9 ? 2 : 1;
drm_gb_capable = have_drm_2_5;
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_3D;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret || gp_arg.value == 0) {
vmw_error("No 3D enabled (%i, %s).\n", ret, strerror(-ret));
goto out_no_3d;
}
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_FIFO_HW_VERSION;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret) {
vmw_error("Failed to get fifo hw version (%i, %s).\n",
ret, strerror(-ret));
goto out_no_3d;
}
vws->ioctl.hwversion = gp_arg.value;
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_HW_CAPS;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret)
vws->base.have_gb_objects = FALSE;
else
vws->base.have_gb_objects =
!!(gp_arg.value & (uint64_t) SVGA_CAP_GBOBJECTS);
if (vws->base.have_gb_objects && !drm_gb_capable)
goto out_no_3d;
vws->base.have_vgpu10 = FALSE;
if (vws->base.have_gb_objects) {
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_3D_CAPS_SIZE;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret)
size = SVGA_FIFO_3D_CAPS_SIZE * sizeof(uint32_t);
else
size = gp_arg.value;
if (vws->base.have_gb_objects)
vws->ioctl.num_cap_3d = size / sizeof(uint32_t);
else
vws->ioctl.num_cap_3d = SVGA3D_DEVCAP_MAX;
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_MAX_MOB_MEMORY;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret) {
/* Just guess a large enough value. */
vws->ioctl.max_mob_memory = 256*1024*1024;
} else {
vws->ioctl.max_mob_memory = gp_arg.value;
}
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_MAX_MOB_SIZE;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret || gp_arg.value == 0) {
vws->ioctl.max_texture_size = VMW_MAX_DEFAULT_TEXTURE_SIZE;
} else {
vws->ioctl.max_texture_size = gp_arg.value;
}
/* Never early flush surfaces, mobs do accounting. */
vws->ioctl.max_surface_memory = -1;
if (vws->ioctl.have_drm_2_9) {
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_VGPU10;
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (ret == 0 && gp_arg.value != 0) {
const char *vgpu10_val;
debug_printf("Have VGPU10 interface and hardware.\n");
vws->base.have_vgpu10 = TRUE;
vgpu10_val = getenv("SVGA_VGPU10");
if (vgpu10_val && strcmp(vgpu10_val, "0") == 0) {
debug_printf("Disabling VGPU10 interface.\n");
vws->base.have_vgpu10 = FALSE;
} else {
debug_printf("Enabling VGPU10 interface.\n");
}
}
}
} else {
vws->ioctl.num_cap_3d = SVGA3D_DEVCAP_MAX;
memset(&gp_arg, 0, sizeof(gp_arg));
gp_arg.param = DRM_VMW_PARAM_MAX_SURF_MEMORY;
if (have_drm_2_5)
ret = drmCommandWriteRead(vws->ioctl.drm_fd, DRM_VMW_GET_PARAM,
&gp_arg, sizeof(gp_arg));
if (!have_drm_2_5 || ret) {
/* Just guess a large enough value, around 800mb. */
vws->ioctl.max_surface_memory = 0x30000000;
} else {
vws->ioctl.max_surface_memory = gp_arg.value;
}
vws->ioctl.max_texture_size = VMW_MAX_DEFAULT_TEXTURE_SIZE;
size = SVGA_FIFO_3D_CAPS_SIZE * sizeof(uint32_t);
}
debug_printf("VGPU10 interface is %s.\n",
vws->base.have_vgpu10 ? "on" : "off");
cap_buffer = calloc(1, size);
if (!cap_buffer) {
debug_printf("Failed alloc fifo 3D caps buffer.\n");
goto out_no_3d;
}
vws->ioctl.cap_3d = calloc(vws->ioctl.num_cap_3d,
sizeof(*vws->ioctl.cap_3d));
if (!vws->ioctl.cap_3d) {
debug_printf("Failed alloc fifo 3D caps buffer.\n");
goto out_no_caparray;
}
memset(&cap_arg, 0, sizeof(cap_arg));
cap_arg.buffer = (uint64_t) (unsigned long) (cap_buffer);
cap_arg.max_size = size;
ret = drmCommandWrite(vws->ioctl.drm_fd, DRM_VMW_GET_3D_CAP,
&cap_arg, sizeof(cap_arg));
if (ret) {
debug_printf("Failed to get 3D capabilities"
" (%i, %s).\n", ret, strerror(-ret));
goto out_no_caps;
}
ret = vmw_ioctl_parse_caps(vws, cap_buffer);
if (ret) {
debug_printf("Failed to parse 3D capabilities"
" (%i, %s).\n", ret, strerror(-ret));
goto out_no_caps;
}
if (((version->version_major == 2 && version->version_minor >= 10)
|| version->version_major > 2) && vws->base.have_vgpu10) {
/* support for these commands didn't make it into vmwgfx kernel
* modules before 2.10.
*/
vws->base.have_generate_mipmap_cmd = TRUE;
vws->base.have_set_predication_cmd = TRUE;
}
if (version->version_major == 2 && version->version_minor >= 14) {
vws->base.have_fence_fd = TRUE;
}
free(cap_buffer);
drmFreeVersion(version);
vmw_printf("%s OK\n", __FUNCTION__);
return TRUE;
out_no_caps:
free(vws->ioctl.cap_3d);
out_no_caparray:
free(cap_buffer);
out_no_3d:
drmFreeVersion(version);
out_no_version:
vws->ioctl.num_cap_3d = 0;
debug_printf("%s Failed\n", __FUNCTION__);
return FALSE;
}
void
vmw_ioctl_cleanup(struct vmw_winsys_screen *vws)
{
VMW_FUNC;
}