/************************************************************************** * * Copyright 2006 Tungsten Graphics, Inc., Cedar Park, Texas. * All Rights Reserved. * * Permission is hereby granted, free of charge, to any person obtaining a * copy of this software and associated documentation files (the * "Software"), to deal in the Software without restriction, including * without limitation the rights to use, copy, modify, merge, publish, * distribute, sub license, and/or sell copies of the Software, and to * permit persons to whom the Software is furnished to do so, subject to * the following conditions: * * The above copyright notice and this permission notice (including the * next paragraph) shall be included in all copies or substantial portions * of the Software. * * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS * OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. * IN NO EVENT SHALL TUNGSTEN GRAPHICS AND/OR ITS SUPPLIERS BE LIABLE FOR * ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, * TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE * SOFTWARE OR THE USE OR OTHER DEALINGS IN THE SOFTWARE. * **************************************************************************/ #ifndef INTEL_BATCHBUFFER_H #define INTEL_BATCHBUFFER_H #include "mtypes.h" #include "bufmgr.h" struct intel_context; #define BATCH_SZ (16 * 1024) #define BATCH_REFILL 4096 #define BATCH_RESERVED 16 #define INTEL_BATCH_NO_CLIPRECTS 0x1 #define INTEL_BATCH_CLIPRECTS 0x2 struct intel_batchbuffer { struct intel_context *intel; struct buffer *buffer; GLuint flags; unsigned long offset; GLubyte *map; GLubyte *ptr; }; struct intel_batchbuffer *intel_batchbuffer_alloc( struct intel_context *intel ); void intel_batchbuffer_free( struct intel_batchbuffer *batch ); GLboolean intel_batchbuffer_flush( struct intel_batchbuffer *batch ); void intel_batchbuffer_unmap( struct intel_batchbuffer *batch ); GLubyte *intel_batchbuffer_map( struct intel_batchbuffer *batch ); /* Unlike bmBufferData, this currently requires the buffer be mapped. * Consider it a convenience function wrapping multple * intel_buffer_dword() calls. */ void intel_batchbuffer_data(struct intel_batchbuffer *batch, const void *data, GLuint bytes, GLuint flags); void intel_batchbuffer_release_space(struct intel_batchbuffer *batch, GLuint bytes); /* Inline functions - might actually be better off with these * non-inlined. Certainly better off switching all command packets to * be passed as structs rather than dwords, but that's a little bit of * work... */ static inline GLuint intel_batchbuffer_space( struct intel_batchbuffer *batch ) { return (BATCH_SZ - BATCH_RESERVED) - (batch->ptr - (batch->map + batch->offset)); } static inline void intel_batchbuffer_emit_dword(struct intel_batchbuffer *batch, GLuint dword) { assert(batch->map); assert(intel_batchbuffer_space(batch) >= 4); *(GLuint *)(batch->ptr) = dword; batch->ptr += 4; } static inline void intel_batchbuffer_require_space(struct intel_batchbuffer *batch, GLuint sz, GLuint flags) { assert(sz < BATCH_SZ - 8); if (intel_batchbuffer_space(batch) < sz || (batch->flags != 0 && flags != 0 && batch->flags != flags)) intel_batchbuffer_flush(batch); batch->flags |= flags; } void intel_batchbuffer_align( struct intel_batchbuffer *batch, GLuint align, GLuint sz ); /* Here are the crusty old macros, to be removed: */ #define BATCH_LOCALS #define BEGIN_BATCH(n, flags) intel_batchbuffer_require_space(intel->batch, n*4, flags) #define OUT_BATCH(d) intel_batchbuffer_emit_dword(intel->batch, d) #define ADVANCE_BATCH() do { } while(0) #endif