Commit c70dc788 authored by Kristian Høgsberg's avatar Kristian Høgsberg Committed by Stefan Richter

firewire: Fix dualbuffer iso receive mode and drop buffer fill mode.

The dualbuffer DMA setup did not account for the iso trailer word
and thus didn't  work correctly.  With this fixed we can drop the
dual buffer fallback mode.
Signed-off-by: default avatarKristian Høgsberg <krh@redhat.com>
Signed-off-by: default avatarStefan Richter <stefanr@s5r6.in-berlin.de>
parent d60d7f1d
...@@ -541,20 +541,32 @@ static int ioctl_create_iso_context(struct client *client, void __user *arg) ...@@ -541,20 +541,32 @@ static int ioctl_create_iso_context(struct client *client, void __user *arg)
if (copy_from_user(&request, arg, sizeof request)) if (copy_from_user(&request, arg, sizeof request))
return -EFAULT; return -EFAULT;
if (request.type > FW_ISO_CONTEXT_RECEIVE)
return -EINVAL;
if (request.channel > 63) if (request.channel > 63)
return -EINVAL; return -EINVAL;
if (request.sync > 15) switch (request.type) {
return -EINVAL; case FW_ISO_CONTEXT_RECEIVE:
if (request.sync > 15)
return -EINVAL;
if (request.tags == 0 || request.tags > 15) if (request.tags == 0 || request.tags > 15)
return -EINVAL; return -EINVAL;
if (request.header_size < 4 || (request.header_size & 3))
return -EINVAL;
if (request.speed > SCODE_3200) break;
case FW_ISO_CONTEXT_TRANSMIT:
if (request.speed > SCODE_3200)
return -EINVAL;
break;
default:
return -EINVAL; return -EINVAL;
}
client->iso_context = fw_iso_context_create(client->device->card, client->iso_context = fw_iso_context_create(client->device->card,
request.type, request.type,
......
...@@ -1233,24 +1233,6 @@ ohci_get_bus_time(struct fw_card *card) ...@@ -1233,24 +1233,6 @@ ohci_get_bus_time(struct fw_card *card)
return bus_time; return bus_time;
} }
static int handle_ir_bufferfill_packet(struct context *context,
struct descriptor *d,
struct descriptor *last)
{
struct iso_context *ctx =
container_of(context, struct iso_context, context);
if (d->res_count > 0)
return 0;
if (le16_to_cpu(last->control) & descriptor_irq_always)
ctx->base.callback(&ctx->base,
le16_to_cpu(last->res_count),
0, NULL, ctx->base.callback_data);
return 1;
}
static int handle_ir_dualbuffer_packet(struct context *context, static int handle_ir_dualbuffer_packet(struct context *context,
struct descriptor *d, struct descriptor *d,
struct descriptor *last) struct descriptor *last)
...@@ -1258,19 +1240,33 @@ static int handle_ir_dualbuffer_packet(struct context *context, ...@@ -1258,19 +1240,33 @@ static int handle_ir_dualbuffer_packet(struct context *context,
struct iso_context *ctx = struct iso_context *ctx =
container_of(context, struct iso_context, context); container_of(context, struct iso_context, context);
struct db_descriptor *db = (struct db_descriptor *) d; struct db_descriptor *db = (struct db_descriptor *) d;
__le32 *ir_header;
size_t header_length; size_t header_length;
void *p, *end;
int i;
if (db->first_res_count > 0 && db->second_res_count > 0) if (db->first_res_count > 0 && db->second_res_count > 0)
/* This descriptor isn't done yet, stop iteration. */ /* This descriptor isn't done yet, stop iteration. */
return 0; return 0;
header_length = db->first_req_count - db->first_res_count; header_length = le16_to_cpu(db->first_req_count) -
if (ctx->header_length + header_length <= PAGE_SIZE) le16_to_cpu(db->first_res_count);
memcpy(ctx->header + ctx->header_length, db + 1, header_length);
ctx->header_length += header_length; i = ctx->header_length;
p = db + 1;
end = p + header_length;
while (p < end && i + ctx->base.header_size <= PAGE_SIZE) {
memcpy(ctx->header + i, p + 4, ctx->base.header_size);
i += ctx->base.header_size;
p += ctx->base.header_size + 4;
}
ctx->header_length = i;
if (le16_to_cpu(db->control) & descriptor_irq_always) { if (le16_to_cpu(db->control) & descriptor_irq_always) {
ctx->base.callback(&ctx->base, 0, ir_header = (__le32 *) (db + 1);
ctx->base.callback(&ctx->base,
le32_to_cpu(ir_header[0]) & 0xffff,
ctx->header_length, ctx->header, ctx->header_length, ctx->header,
ctx->base.callback_data); ctx->base.callback_data);
ctx->header_length = 0; ctx->header_length = 0;
...@@ -1315,12 +1311,10 @@ ohci_allocate_iso_context(struct fw_card *card, int type, ...@@ -1315,12 +1311,10 @@ ohci_allocate_iso_context(struct fw_card *card, int type,
} else { } else {
mask = &ohci->ir_context_mask; mask = &ohci->ir_context_mask;
list = ohci->ir_context_list; list = ohci->ir_context_list;
if (header_size > 0) callback = handle_ir_dualbuffer_packet;
callback = handle_ir_dualbuffer_packet;
else
callback = handle_ir_bufferfill_packet;
} }
/* FIXME: We need a fallback for pre 1.1 OHCI. */
if (callback == handle_ir_dualbuffer_packet && if (callback == handle_ir_dualbuffer_packet &&
ohci->version < OHCI_VERSION_1_1) ohci->version < OHCI_VERSION_1_1)
return ERR_PTR(-EINVAL); return ERR_PTR(-EINVAL);
...@@ -1367,7 +1361,7 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle) ...@@ -1367,7 +1361,7 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle)
{ {
struct iso_context *ctx = container_of(base, struct iso_context, base); struct iso_context *ctx = container_of(base, struct iso_context, base);
struct fw_ohci *ohci = ctx->context.ohci; struct fw_ohci *ohci = ctx->context.ohci;
u32 cycle_match = 0, mode; u32 cycle_match = 0;
int index; int index;
if (ctx->base.type == FW_ISO_CONTEXT_TRANSMIT) { if (ctx->base.type == FW_ISO_CONTEXT_TRANSMIT) {
...@@ -1382,16 +1376,14 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle) ...@@ -1382,16 +1376,14 @@ static int ohci_start_iso(struct fw_iso_context *base, s32 cycle)
} else { } else {
index = ctx - ohci->ir_context_list; index = ctx - ohci->ir_context_list;
if (ctx->base.header_size > 0)
mode = IR_CONTEXT_DUAL_BUFFER_MODE;
else
mode = IR_CONTEXT_BUFFER_FILL;
reg_write(ohci, OHCI1394_IsoRecvIntEventClear, 1 << index); reg_write(ohci, OHCI1394_IsoRecvIntEventClear, 1 << index);
reg_write(ohci, OHCI1394_IsoRecvIntMaskSet, 1 << index); reg_write(ohci, OHCI1394_IsoRecvIntMaskSet, 1 << index);
reg_write(ohci, context_match(ctx->context.regs), reg_write(ohci, context_match(ctx->context.regs),
(ctx->base.tags << 28) | (ctx->base.tags << 28) |
(ctx->base.sync << 8) | ctx->base.channel); (ctx->base.sync << 8) | ctx->base.channel);
context_run(&ctx->context, mode); context_run(&ctx->context,
IR_CONTEXT_DUAL_BUFFER_MODE |
IR_CONTEXT_ISOCH_HEADER);
} }
return 0; return 0;
...@@ -1537,26 +1529,6 @@ ohci_queue_iso_transmit(struct fw_iso_context *base, ...@@ -1537,26 +1529,6 @@ ohci_queue_iso_transmit(struct fw_iso_context *base,
return 0; return 0;
} }
static int
setup_wait_descriptor(struct context *ctx)
{
struct descriptor *d;
dma_addr_t d_bus;
d = context_get_descriptors(ctx, 1, &d_bus);
if (d == NULL)
return -ENOMEM;
d->control = cpu_to_le16(descriptor_input_more |
descriptor_status |
descriptor_branch_always |
descriptor_wait);
context_append(ctx, d, 1, 0);
return 0;
}
static int static int
ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base, ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
struct fw_iso_packet *packet, struct fw_iso_packet *packet,
...@@ -1569,25 +1541,39 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base, ...@@ -1569,25 +1541,39 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
struct fw_iso_packet *p; struct fw_iso_packet *p;
dma_addr_t d_bus, page_bus; dma_addr_t d_bus, page_bus;
u32 z, header_z, length, rest; u32 z, header_z, length, rest;
int page, offset; int page, offset, packet_count, header_size;
/* FIXME: Cycle lost behavior should be configurable: lose /* FIXME: Cycle lost behavior should be configurable: lose
* packet, retransmit or terminate.. */ * packet, retransmit or terminate.. */
if (packet->skip && setup_wait_descriptor(&ctx->context) < 0) if (packet->skip) {
return -ENOMEM; d = context_get_descriptors(&ctx->context, 2, &d_bus);
if (d == NULL)
return -ENOMEM;
db = (struct db_descriptor *) d;
db->control = cpu_to_le16(descriptor_status |
descriptor_branch_always |
descriptor_wait);
db->first_size = cpu_to_le16(ctx->base.header_size + 4);
context_append(&ctx->context, d, 2, 0);
}
p = packet; p = packet;
z = 2; z = 2;
/* The OHCI controller puts the status word in the header
* buffer too, so we need 4 extra bytes per packet. */
packet_count = p->header_length / ctx->base.header_size;
header_size = packet_count * (ctx->base.header_size + 4);
/* Get header size in number of descriptors. */ /* Get header size in number of descriptors. */
header_z = DIV_ROUND_UP(p->header_length, sizeof *d); header_z = DIV_ROUND_UP(header_size, sizeof *d);
page = payload >> PAGE_SHIFT; page = payload >> PAGE_SHIFT;
offset = payload & ~PAGE_MASK; offset = payload & ~PAGE_MASK;
rest = p->payload_length; rest = p->payload_length;
/* FIXME: OHCI 1.0 doesn't support dual buffer receive */ /* FIXME: OHCI 1.0 doesn't support dual buffer receive */
/* FIXME: handle descriptor_wait */
/* FIXME: make packet-per-buffer/dual-buffer a context option */ /* FIXME: make packet-per-buffer/dual-buffer a context option */
while (rest > 0) { while (rest > 0) {
d = context_get_descriptors(&ctx->context, d = context_get_descriptors(&ctx->context,
...@@ -1598,8 +1584,8 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base, ...@@ -1598,8 +1584,8 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
db = (struct db_descriptor *) d; db = (struct db_descriptor *) d;
db->control = cpu_to_le16(descriptor_status | db->control = cpu_to_le16(descriptor_status |
descriptor_branch_always); descriptor_branch_always);
db->first_size = cpu_to_le16(ctx->base.header_size); db->first_size = cpu_to_le16(ctx->base.header_size + 4);
db->first_req_count = cpu_to_le16(p->header_length); db->first_req_count = cpu_to_le16(header_size);
db->first_res_count = db->first_req_count; db->first_res_count = db->first_req_count;
db->first_buffer = cpu_to_le32(d_bus + sizeof *db); db->first_buffer = cpu_to_le32(d_bus + sizeof *db);
...@@ -1625,57 +1611,6 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base, ...@@ -1625,57 +1611,6 @@ ohci_queue_iso_receive_dualbuffer(struct fw_iso_context *base,
return 0; return 0;
} }
static int
ohci_queue_iso_receive_bufferfill(struct fw_iso_context *base,
struct fw_iso_packet *packet,
struct fw_iso_buffer *buffer,
unsigned long payload)
{
struct iso_context *ctx = container_of(base, struct iso_context, base);
struct descriptor *d = NULL;
dma_addr_t d_bus, page_bus;
u32 length, rest;
int page, offset;
page = payload >> PAGE_SHIFT;
offset = payload & ~PAGE_MASK;
rest = packet->payload_length;
if (packet->skip && setup_wait_descriptor(&ctx->context) < 0)
return -ENOMEM;
while (rest > 0) {
d = context_get_descriptors(&ctx->context, 1, &d_bus);
if (d == NULL)
return -ENOMEM;
d->control = cpu_to_le16(descriptor_input_more |
descriptor_status |
descriptor_branch_always);
if (offset + rest < PAGE_SIZE)
length = rest;
else
length = PAGE_SIZE - offset;
page_bus = page_private(buffer->pages[page]);
d->data_address = cpu_to_le32(page_bus + offset);
d->req_count = cpu_to_le16(length);
d->res_count = cpu_to_le16(length);
if (packet->interrupt && length == rest)
d->control |= cpu_to_le16(descriptor_irq_always);
context_append(&ctx->context, d, 1, 0);
offset = (offset + length) & ~PAGE_MASK;
rest -= length;
page++;
}
return 0;
}
static int static int
ohci_queue_iso(struct fw_iso_context *base, ohci_queue_iso(struct fw_iso_context *base,
struct fw_iso_packet *packet, struct fw_iso_packet *packet,
...@@ -1686,9 +1621,6 @@ ohci_queue_iso(struct fw_iso_context *base, ...@@ -1686,9 +1621,6 @@ ohci_queue_iso(struct fw_iso_context *base,
if (base->type == FW_ISO_CONTEXT_TRANSMIT) if (base->type == FW_ISO_CONTEXT_TRANSMIT)
return ohci_queue_iso_transmit(base, packet, buffer, payload); return ohci_queue_iso_transmit(base, packet, buffer, payload);
else if (base->header_size == 0)
return ohci_queue_iso_receive_bufferfill(base, packet,
buffer, payload);
else if (ctx->context.ohci->version >= OHCI_VERSION_1_1) else if (ctx->context.ohci->version >= OHCI_VERSION_1_1)
return ohci_queue_iso_receive_dualbuffer(base, packet, return ohci_queue_iso_receive_dualbuffer(base, packet,
buffer, payload); buffer, payload);
......
Markdown is supported
0%
or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment