diff options
Diffstat (limited to 'src/mesa/drivers/dri/i965/intel_buffer_objects.c')
-rw-r--r-- | src/mesa/drivers/dri/i965/intel_buffer_objects.c | 82 |
1 files changed, 41 insertions, 41 deletions
diff --git a/src/mesa/drivers/dri/i965/intel_buffer_objects.c b/src/mesa/drivers/dri/i965/intel_buffer_objects.c index 56da2da08a8..9ce13c2ee62 100644 --- a/src/mesa/drivers/dri/i965/intel_buffer_objects.c +++ b/src/mesa/drivers/dri/i965/intel_buffer_objects.c @@ -42,20 +42,20 @@ /** * Map a buffer object; issue performance warnings if mapping causes stalls. * - * This matches the drm_intel_bo_map API, but takes an additional human-readable + * This matches the drm_bacon_bo_map API, but takes an additional human-readable * name for the buffer object to use in the performance debug message. */ int brw_bo_map(struct brw_context *brw, - drm_intel_bo *bo, int write_enable, + drm_bacon_bo *bo, int write_enable, const char *bo_name) { - if (likely(!brw->perf_debug) || !drm_intel_bo_busy(bo)) - return drm_intel_bo_map(bo, write_enable); + if (likely(!brw->perf_debug) || !drm_bacon_bo_busy(bo)) + return drm_bacon_bo_map(bo, write_enable); double start_time = get_time(); - int ret = drm_intel_bo_map(bo, write_enable); + int ret = drm_bacon_bo_map(bo, write_enable); perf_debug("CPU mapping a busy %s BO stalled and took %.03f ms.\n", bo_name, (get_time() - start_time) * 1000); @@ -64,14 +64,14 @@ brw_bo_map(struct brw_context *brw, } int -brw_bo_map_gtt(struct brw_context *brw, drm_intel_bo *bo, const char *bo_name) +brw_bo_map_gtt(struct brw_context *brw, drm_bacon_bo *bo, const char *bo_name) { - if (likely(!brw->perf_debug) || !drm_intel_bo_busy(bo)) - return drm_intel_gem_bo_map_gtt(bo); + if (likely(!brw->perf_debug) || !drm_bacon_bo_busy(bo)) + return drm_bacon_gem_bo_map_gtt(bo); double start_time = get_time(); - int ret = drm_intel_gem_bo_map_gtt(bo); + int ret = drm_bacon_gem_bo_map_gtt(bo); perf_debug("GTT mapping a busy %s BO stalled and took %.03f ms.\n", bo_name, (get_time() - start_time) * 1000); @@ -94,12 +94,12 @@ mark_buffer_inactive(struct intel_buffer_object *intel_obj) intel_obj->gpu_active_end = 0; } -/** Allocates a new drm_intel_bo to store the data for the buffer object. */ +/** Allocates a new drm_bacon_bo to store the data for the buffer object. */ static void alloc_buffer_object(struct brw_context *brw, struct intel_buffer_object *intel_obj) { - intel_obj->buffer = drm_intel_bo_alloc(brw->bufmgr, "bufferobj", + intel_obj->buffer = drm_bacon_bo_alloc(brw->bufmgr, "bufferobj", intel_obj->Base.Size, 64); /* the buffer might be bound as a uniform buffer, need to update it @@ -119,7 +119,7 @@ alloc_buffer_object(struct brw_context *brw, static void release_buffer(struct intel_buffer_object *intel_obj) { - drm_intel_bo_unreference(intel_obj->buffer); + drm_bacon_bo_unreference(intel_obj->buffer); intel_obj->buffer = NULL; } @@ -166,7 +166,7 @@ brw_delete_buffer(struct gl_context * ctx, struct gl_buffer_object *obj) */ _mesa_buffer_unmap_all_mappings(ctx, obj); - drm_intel_bo_unreference(intel_obj->buffer); + drm_bacon_bo_unreference(intel_obj->buffer); _mesa_delete_buffer_object(ctx, obj); } @@ -213,7 +213,7 @@ brw_buffer_data(struct gl_context *ctx, return false; if (data != NULL) - drm_intel_bo_subdata(intel_obj->buffer, 0, size, data); + drm_bacon_bo_subdata(intel_obj->buffer, 0, size, data); } return true; @@ -257,9 +257,9 @@ brw_buffer_subdata(struct gl_context *ctx, if (offset + size <= intel_obj->gpu_active_start || intel_obj->gpu_active_end <= offset) { if (brw->has_llc) { - drm_intel_gem_bo_map_unsynchronized(intel_obj->buffer); + drm_bacon_gem_bo_map_unsynchronized(intel_obj->buffer); memcpy(intel_obj->buffer->virtual + offset, data, size); - drm_intel_bo_unmap(intel_obj->buffer); + drm_bacon_bo_unmap(intel_obj->buffer); if (intel_obj->gpu_active_end > intel_obj->gpu_active_start) intel_obj->prefer_stall_to_blit = true; @@ -270,13 +270,13 @@ brw_buffer_subdata(struct gl_context *ctx, } busy = - drm_intel_bo_busy(intel_obj->buffer) || - drm_intel_bo_references(brw->batch.bo, intel_obj->buffer); + drm_bacon_bo_busy(intel_obj->buffer) || + drm_bacon_bo_references(brw->batch.bo, intel_obj->buffer); if (busy) { if (size == intel_obj->Base.Size) { /* Replace the current busy bo so the subdata doesn't stall. */ - drm_intel_bo_unreference(intel_obj->buffer); + drm_bacon_bo_unreference(intel_obj->buffer); alloc_buffer_object(brw, intel_obj); } else if (!intel_obj->prefer_stall_to_blit) { perf_debug("Using a blit copy to avoid stalling on " @@ -285,17 +285,17 @@ brw_buffer_subdata(struct gl_context *ctx, (long)offset, (long)offset + size, (long)(size/1024), intel_obj->gpu_active_start, intel_obj->gpu_active_end); - drm_intel_bo *temp_bo = - drm_intel_bo_alloc(brw->bufmgr, "subdata temp", size, 64); + drm_bacon_bo *temp_bo = + drm_bacon_bo_alloc(brw->bufmgr, "subdata temp", size, 64); - drm_intel_bo_subdata(temp_bo, 0, size, data); + drm_bacon_bo_subdata(temp_bo, 0, size, data); intel_emit_linear_blit(brw, intel_obj->buffer, offset, temp_bo, 0, size); - drm_intel_bo_unreference(temp_bo); + drm_bacon_bo_unreference(temp_bo); return; } else { perf_debug("Stalling on glBufferSubData(%ld, %ld) (%ldkb) to a busy " @@ -308,7 +308,7 @@ brw_buffer_subdata(struct gl_context *ctx, } } - drm_intel_bo_subdata(intel_obj->buffer, offset, size, data); + drm_bacon_bo_subdata(intel_obj->buffer, offset, size, data); mark_buffer_inactive(intel_obj); } @@ -330,10 +330,10 @@ brw_get_buffer_subdata(struct gl_context *ctx, struct brw_context *brw = brw_context(ctx); assert(intel_obj); - if (drm_intel_bo_references(brw->batch.bo, intel_obj->buffer)) { + if (drm_bacon_bo_references(brw->batch.bo, intel_obj->buffer)) { intel_batchbuffer_flush(brw); } - drm_intel_bo_get_subdata(intel_obj->buffer, offset, size, data); + drm_bacon_bo_get_subdata(intel_obj->buffer, offset, size, data); mark_buffer_inactive(intel_obj); } @@ -389,18 +389,18 @@ brw_map_buffer_range(struct gl_context *ctx, * achieve the required synchronization. */ if (!(access & GL_MAP_UNSYNCHRONIZED_BIT)) { - if (drm_intel_bo_references(brw->batch.bo, intel_obj->buffer)) { + if (drm_bacon_bo_references(brw->batch.bo, intel_obj->buffer)) { if (access & GL_MAP_INVALIDATE_BUFFER_BIT) { - drm_intel_bo_unreference(intel_obj->buffer); + drm_bacon_bo_unreference(intel_obj->buffer); alloc_buffer_object(brw, intel_obj); } else { perf_debug("Stalling on the GPU for mapping a busy buffer " "object\n"); intel_batchbuffer_flush(brw); } - } else if (drm_intel_bo_busy(intel_obj->buffer) && + } else if (drm_bacon_bo_busy(intel_obj->buffer) && (access & GL_MAP_INVALIDATE_BUFFER_BIT)) { - drm_intel_bo_unreference(intel_obj->buffer); + drm_bacon_bo_unreference(intel_obj->buffer); alloc_buffer_object(brw, intel_obj); } } @@ -416,14 +416,14 @@ brw_map_buffer_range(struct gl_context *ctx, */ if (!(access & (GL_MAP_UNSYNCHRONIZED_BIT | GL_MAP_PERSISTENT_BIT)) && (access & GL_MAP_INVALIDATE_RANGE_BIT) && - drm_intel_bo_busy(intel_obj->buffer)) { + drm_bacon_bo_busy(intel_obj->buffer)) { /* Ensure that the base alignment of the allocation meets the alignment * guarantees the driver has advertised to the application. */ const unsigned alignment = ctx->Const.MinMapBufferAlignment; intel_obj->map_extra[index] = (uintptr_t) offset % alignment; - intel_obj->range_map_bo[index] = drm_intel_bo_alloc(brw->bufmgr, + intel_obj->range_map_bo[index] = drm_bacon_bo_alloc(brw->bufmgr, "BO blit temp", length + intel_obj->map_extra[index], @@ -432,7 +432,7 @@ brw_map_buffer_range(struct gl_context *ctx, brw_bo_map(brw, intel_obj->range_map_bo[index], (access & GL_MAP_WRITE_BIT) != 0, "range-map"); } else { - drm_intel_gem_bo_map_gtt(intel_obj->range_map_bo[index]); + drm_bacon_gem_bo_map_gtt(intel_obj->range_map_bo[index]); } obj->Mappings[index].Pointer = intel_obj->range_map_bo[index]->virtual + intel_obj->map_extra[index]; @@ -441,13 +441,13 @@ brw_map_buffer_range(struct gl_context *ctx, if (access & GL_MAP_UNSYNCHRONIZED_BIT) { if (!brw->has_llc && brw->perf_debug && - drm_intel_bo_busy(intel_obj->buffer)) { + drm_bacon_bo_busy(intel_obj->buffer)) { perf_debug("MapBufferRange with GL_MAP_UNSYNCHRONIZED_BIT stalling (it's actually synchronized on non-LLC platforms)\n"); } - drm_intel_gem_bo_map_unsynchronized(intel_obj->buffer); + drm_bacon_gem_bo_map_unsynchronized(intel_obj->buffer); } else if (!brw->has_llc && (!(access & GL_MAP_READ_BIT) || (access & GL_MAP_PERSISTENT_BIT))) { - drm_intel_gem_bo_map_gtt(intel_obj->buffer); + drm_bacon_gem_bo_map_gtt(intel_obj->buffer); mark_buffer_inactive(intel_obj); } else { brw_bo_map(brw, intel_obj->buffer, (access & GL_MAP_WRITE_BIT) != 0, @@ -543,7 +543,7 @@ brw_unmap_buffer(struct gl_context *ctx, assert(intel_obj); assert(obj->Mappings[index].Pointer); if (intel_obj->range_map_bo[index] != NULL) { - drm_intel_bo_unmap(intel_obj->range_map_bo[index]); + drm_bacon_bo_unmap(intel_obj->range_map_bo[index]); if (!(obj->Mappings[index].AccessFlags & GL_MAP_FLUSH_EXPLICIT_BIT)) { intel_emit_linear_blit(brw, @@ -562,10 +562,10 @@ brw_unmap_buffer(struct gl_context *ctx, */ brw_emit_mi_flush(brw); - drm_intel_bo_unreference(intel_obj->range_map_bo[index]); + drm_bacon_bo_unreference(intel_obj->range_map_bo[index]); intel_obj->range_map_bo[index] = NULL; } else if (intel_obj->buffer != NULL) { - drm_intel_bo_unmap(intel_obj->buffer); + drm_bacon_bo_unmap(intel_obj->buffer); } obj->Mappings[index].Pointer = NULL; obj->Mappings[index].Offset = 0; @@ -581,7 +581,7 @@ brw_unmap_buffer(struct gl_context *ctx, * Anywhere that uses buffer objects in the pipeline should be using this to * mark the range of the buffer that is being accessed by the pipeline. */ -drm_intel_bo * +drm_bacon_bo * intel_bufferobj_buffer(struct brw_context *brw, struct intel_buffer_object *intel_obj, uint32_t offset, uint32_t size) @@ -615,7 +615,7 @@ brw_copy_buffer_subdata(struct gl_context *ctx, struct brw_context *brw = brw_context(ctx); struct intel_buffer_object *intel_src = intel_buffer_object(src); struct intel_buffer_object *intel_dst = intel_buffer_object(dst); - drm_intel_bo *src_bo, *dst_bo; + drm_bacon_bo *src_bo, *dst_bo; if (size == 0) return; |