#ifndef INTEL_BATCHBUFFER_H #define INTEL_BATCHBUFFER_H #include #include "intel_bufmgr.h" #define BATCH_SZ 4096 #define BATCH_RESERVED 16 struct intel_batchbuffer { drm_intel_bufmgr *bufmgr; uint32_t devid; drm_intel_bo *bo; uint8_t buffer[BATCH_SZ]; uint8_t *ptr; uint8_t *state; }; struct intel_batchbuffer *intel_batchbuffer_alloc(drm_intel_bufmgr *bufmgr, uint32_t devid); void intel_batchbuffer_free(struct intel_batchbuffer *batch); void intel_batchbuffer_flush(struct intel_batchbuffer *batch); void intel_batchbuffer_flush_on_ring(struct intel_batchbuffer *batch, int ring); void intel_batchbuffer_flush_with_context(struct intel_batchbuffer *batch, drm_intel_context *context); void intel_batchbuffer_reset(struct intel_batchbuffer *batch); void intel_batchbuffer_data(struct intel_batchbuffer *batch, const void *data, unsigned int bytes); void intel_batchbuffer_emit_reloc(struct intel_batchbuffer *batch, drm_intel_bo *buffer, uint32_t delta, uint32_t read_domains, uint32_t write_domain, int fenced); /* Inline functions - might actually be better off with these * non-inlined. Certainly better off switching all command packets to * be passed as structs rather than dwords, but that's a little bit of * work... */ #pragma GCC diagnostic ignored "-Winline" static inline unsigned int intel_batchbuffer_space(struct intel_batchbuffer *batch) { return (BATCH_SZ - BATCH_RESERVED) - (batch->ptr - batch->buffer); } static inline void intel_batchbuffer_emit_dword(struct intel_batchbuffer *batch, uint32_t dword) { assert(intel_batchbuffer_space(batch) >= 4); *(uint32_t *) (batch->ptr) = dword; batch->ptr += 4; } static inline void intel_batchbuffer_require_space(struct intel_batchbuffer *batch, unsigned int sz) { assert(sz < BATCH_SZ - BATCH_RESERVED); if (intel_batchbuffer_space(batch) < sz) intel_batchbuffer_flush(batch); } /** * BEGIN_BATCH: * @n: number of DWORDS to emit * * Prepares a batch to emit @n DWORDS, flushing it if there's not enough space * available. * * This macro needs a pointer to an #intel_batchbuffer structure called batch in * scope. */ #define BEGIN_BATCH(n) do { \ intel_batchbuffer_require_space(batch, (n)*4); \ } while (0) /** * OUT_BATCH: * @d: DWORD to emit * * Emits @d into a batch. * * This macro needs a pointer to an #intel_batchbuffer structure called batch in * scope. */ #define OUT_BATCH(d) intel_batchbuffer_emit_dword(batch, d) /** * OUT_RELOC_FENCED: * @buf: relocation target libdrm buffer object * @read_domains: gem domain bits for the relocation * @write_domain: gem domain bit for the relocation * @delta: delta value to add to @buffer's gpu address * * Emits a fenced relocation into a batch. * * This macro needs a pointer to an #intel_batchbuffer structure called batch in * scope. */ #define OUT_RELOC_FENCED(buf, read_domains, write_domain, delta) do { \ assert((delta) >= 0); \ intel_batchbuffer_emit_reloc(batch, buf, delta, \ read_domains, write_domain, 1); \ } while (0) /** * OUT_RELOC_FENCED: * @buf: relocation target libdrm buffer object * @read_domains: gem domain bits for the relocation * @write_domain: gem domain bit for the relocation * @delta: delta value to add to @buffer's gpu address * * Emits a normal, unfenced relocation into a batch. * * This macro needs a pointer to an #intel_batchbuffer structure called batch in * scope. */ #define OUT_RELOC(buf, read_domains, write_domain, delta) do { \ assert((delta) >= 0); \ intel_batchbuffer_emit_reloc(batch, buf, delta, \ read_domains, write_domain, 0); \ } while (0) /** * ADVANCE_BATCH: * * Completes the batch command emission sequence started with #BEGIN_BATCH. * * This macro needs a pointer to an #intel_batchbuffer structure called batch in * scope. */ #define ADVANCE_BATCH() do { \ } while(0) #define BLIT_COPY_BATCH_START(devid, flags) do { \ if (intel_gen(devid) >= 8) { \ BEGIN_BATCH(10); \ OUT_BATCH(XY_SRC_COPY_BLT_CMD | \ XY_SRC_COPY_BLT_WRITE_ALPHA | \ XY_SRC_COPY_BLT_WRITE_RGB | \ (flags) | 8); \ } else { \ BEGIN_BATCH(8); \ OUT_BATCH(XY_SRC_COPY_BLT_CMD | \ XY_SRC_COPY_BLT_WRITE_ALPHA | \ XY_SRC_COPY_BLT_WRITE_RGB | \ (flags) | 6); \ } \ } while(0) #define COLOR_BLIT_COPY_BATCH_START(devid, flags) do { \ if (intel_gen(devid) >= 8) { \ BEGIN_BATCH(8); \ OUT_BATCH(MI_NOOP); \ OUT_BATCH(XY_COLOR_BLT_CMD_NOLEN | 0x5 | \ COLOR_BLT_WRITE_ALPHA | \ XY_COLOR_BLT_WRITE_RGB); \ } else { \ BEGIN_BATCH(6); \ OUT_BATCH(XY_COLOR_BLT_CMD_NOLEN | 0x4 | \ COLOR_BLT_WRITE_ALPHA | \ XY_COLOR_BLT_WRITE_RGB); \ } \ } while(0) /** * BLIT_RELOC_UDW: * @devid: pci device id of the drm device * * Emits the upper relocation DWORD on gen8+ and nothing on earlier generations. */ #define BLIT_RELOC_UDW(devid) do { \ if (intel_gen(devid) >= 8) { \ OUT_BATCH(0); \ } \ } while(0) void intel_blt_copy(struct intel_batchbuffer *batch, drm_intel_bo *src_bo, int src_x1, int src_y1, int src_pitch, drm_intel_bo *dst_bo, int dst_x1, int dst_y1, int dst_pitch, int width, int height, int bpp); void intel_copy_bo(struct intel_batchbuffer *batch, drm_intel_bo *dst_bo, drm_intel_bo *src_bo, long int size); #endif