ADVANCE_BCS_BATCH(batch);
}
+#if 0
static void
gen6_mfd_aes_state(VADriverContextP ctx,
struct decode_state *decode_state,
OUT_BCS_BATCH(batch, MFX_WAIT | (1 << 8));
ADVANCE_BCS_BATCH(batch);
}
+#endif
static void
gen6_mfd_avc_img_state(VADriverContextP ctx,
struct gen6_encoder_context *gen6_encoder_context)
{
struct gen6_vme_context *vme_context = &gen6_encoder_context->vme_context;
- unsigned char *constant_buffer;
+ // unsigned char *constant_buffer;
dri_bo_map(vme_context->curbe.bo, 1);
assert(vme_context->curbe.bo->virtual);
- constant_buffer = vme_context->curbe.bo->virtual;
+ // constant_buffer = vme_context->curbe.bo->virtual;
/*TODO copy buffer into CURB*/
ADVANCE_BCS_BATCH(batch);
}
+#if 0
static void
gen7_mfd_aes_state(VADriverContextP ctx,
struct decode_state *decode_state,
{
/* FIXME */
}
+#endif
static void
gen7_mfd_qm_state(VADriverContextP ctx,
intel_batchbuffer_data(batch, qm_buffer, 16 * 4);
ADVANCE_BCS_BATCH(batch);
}
+
+#if 0
static void
gen7_mfd_wait(VADriverContextP ctx,
struct decode_state *decode_state,
OUT_BCS_BATCH(batch, MFX_WAIT | (1 << 8));
ADVANCE_BCS_BATCH(batch);
}
+#endif
static void
gen7_mfd_avc_img_state(VADriverContextP ctx,
struct gen7_mfd_context *gen7_mfd_context)
{
struct intel_batchbuffer *batch = gen7_mfd_context->base.batch;
- int qm_present_flag;
int img_struct;
int mbaff_frame_flag;
unsigned int width_in_mbs, height_in_mbs;
pic_param = (VAPictureParameterBufferH264 *)decode_state->pic_param->buffer;
assert(!(pic_param->CurrPic.flags & VA_PICTURE_H264_INVALID));
- if (decode_state->iq_matrix && decode_state->iq_matrix->buffer)
- qm_present_flag = 1;
- else
- qm_present_flag = 0; /* built-in QM matrices */
-
if (pic_param->CurrPic.flags & VA_PICTURE_H264_TOP_FIELD)
img_struct = 1;
else if (pic_param->CurrPic.flags & VA_PICTURE_H264_BOTTOM_FIELD)
{
struct i965_driver_data *i965 = i965_driver_data(ctx);
struct object_context *obj_context = CONTEXT(context);
- VAStatus vaStatus;
+ VAStatus vaStatus = VA_STATUS_ERROR_UNKNOWN;
int i;
for (i = 0; i < num_buffers; i++) {
{
struct i965_driver_data *i965 = i965_driver_data(ctx);
struct object_context *obj_context = CONTEXT(context);
- VAStatus vaStatus;
+ VAStatus vaStatus = VA_STATUS_ERROR_UNKNOWN;
int i;
for (i = 0; i < num_buffers; i++) {
{
struct i965_driver_data *i965 = i965_driver_data(ctx);
struct object_context *obj_context = CONTEXT(context);
- VAStatus vaStatus;
+ VAStatus vaStatus = VA_STATUS_ERROR_UNKNOWN;
int i;
for (i = 0; i < num_buffers; i++) {
VAImage *out_image) /* out */
{
struct i965_driver_data *i965 = i965_driver_data(ctx);
- struct i965_render_state *render_state = &i965->render_state;
struct object_image *obj_image;
struct object_surface *obj_surface;
VAImageID image_id;
{
struct i965_driver_data *i965 = i965_driver_data(ctx);
struct object_surface *obj_surface = SURFACE(surface);
- int w, h;
- dri_bo *region;
dri_bo *subpic_region;
struct object_subpic *obj_subpic = SUBPIC(obj_surface->subpic);
struct object_image *obj_image = IMAGE(obj_subpic->image);
assert(obj_surface);
assert(obj_surface->bo);
- w = obj_surface->width;
- h = obj_surface->height;
- region = obj_surface->bo;
subpic_region = obj_image->bo;
/*subpicture surface*/
i965_render_src_surface_state(ctx, 1, subpic_region, 0, obj_subpic->width, obj_subpic->height, obj_subpic->pitch, obj_subpic->format);
#include "intel_driver.h"
#include "intel_batchbuffer_dump.h"
+#ifdef I965_DEBUG
+
#define BUFFER_FAIL(_count, _len, _name) do { \
fprintf(gout, "Buffer size too small in %s (%d < %d)\n", \
(_name), (_count), (_len)); \
return failures;
}
+
+#endif
#define OPCODE_MI_FLUSH 0x04
#define OPCODE_MI_BATCH_BUFFER_END 0x0A
+#ifdef I965_DEBUG
+
int intel_batchbuffer_dump(unsigned int *data, unsigned int offset, int count, unsigned int device);
+#endif
+
#endif /* _INTEL_BATCHBUFFER_DUMP_H_ */