rpms/kernel/devel kernel.spec, 1.418, 1.419 linux-2.6-firewire-git-pending.patch, 1.3, 1.4 linux-2.6-firewire-git-update.patch, 1.5, 1.6
Jarod Wilson (jwilson)
fedora-extras-commits at redhat.com
Tue Feb 12 16:42:35 UTC 2008
Author: jwilson
Update of /cvs/pkgs/rpms/kernel/devel
In directory cvs-int.fedora.redhat.com:/tmp/cvs-serv1581
Modified Files:
kernel.spec linux-2.6-firewire-git-pending.patch
linux-2.6-firewire-git-update.patch
Log Message:
* Tue Feb 12 2008 Jarod Wilson <jwilson at redhat.com>
- Fix up firewire bits to apply to 2.6.25
Index: kernel.spec
===================================================================
RCS file: /cvs/pkgs/rpms/kernel/devel/kernel.spec,v
retrieving revision 1.418
retrieving revision 1.419
diff -u -r1.418 -r1.419
--- kernel.spec 12 Feb 2008 15:55:47 -0000 1.418
+++ kernel.spec 12 Feb 2008 16:41:53 -0000 1.419
@@ -1153,11 +1153,11 @@
ApplyPatch linux-2.6-ext4-stable-queue.patch
# linux1394 git patches
-#ApplyPatch linux-2.6-firewire-git-update.patch
+ApplyPatch linux-2.6-firewire-git-update.patch
C=$(wc -l $RPM_SOURCE_DIR/linux-2.6-firewire-git-pending.patch | awk '{print $1}')
-#if [ "$C" -gt 10 ]; then
-#ApplyPatch linux-2.6-firewire-git-pending.patch
-#fi
+if [ "$C" -gt 10 ]; then
+ApplyPatch linux-2.6-firewire-git-pending.patch
+fi
# usb video
ApplyPatch linux-2.6-uvcvideo.patch
@@ -1741,6 +1741,9 @@
%kernel_variant_files -a /%{image_install_path}/xen*-%{KVERREL} -e /etc/ld.so.conf.d/kernelcap-%{KVERREL}.conf %{with_xen} xen
%changelog
+* Tue Feb 12 2008 Jarod Wilson <jwilson at redhat.com>
+- Fix up firewire bits to apply to 2.6.25
+
* Tue Feb 12 2008 Kyle McMartin <kmcmartin at redhat.com>
- Bump.
linux-2.6-firewire-git-pending.patch:
Index: linux-2.6-firewire-git-pending.patch
===================================================================
RCS file: /cvs/pkgs/rpms/kernel/devel/linux-2.6-firewire-git-pending.patch,v
retrieving revision 1.3
retrieving revision 1.4
diff -u -r1.3 -r1.4
--- linux-2.6-firewire-git-pending.patch 10 Feb 2008 19:34:11 -0000 1.3
+++ linux-2.6-firewire-git-pending.patch 12 Feb 2008 16:41:53 -0000 1.4
@@ -3,102 +3,6 @@
# tree, which we think we're going to want...
#
-If a device is being unplugged while fw-sbp2 had a login or reconnect on
-schedule, it would take about half a minute to shut the fw_unit down:
-
-Jan 27 18:34:54 stein firewire_sbp2: logged in to fw2.0 LUN 0000 (0 retries)
-<unplug>
-Jan 27 18:34:59 stein firewire_sbp2: sbp2_scsi_abort
-Jan 27 18:34:59 stein scsi 25:0:0:0: Device offlined - not ready after error recovery
-Jan 27 18:35:01 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:06 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:12 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:17 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:22 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:27 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:32 stein firewire_sbp2: orb reply timed out, rcode=0x11
-Jan 27 18:35:32 stein firewire_sbp2: failed to login to fw2.0 LUN 0000
-Jan 27 18:35:32 stein firewire_sbp2: released fw2.0
-
-After this patch, typically only a few seconds spent in __scsi_add_device
-remain:
-
-Jan 27 19:05:50 stein firewire_sbp2: logged in to fw2.0 LUN 0000 (0 retries)
-<unplug>
-Jan 27 19:05:56 stein firewire_sbp2: sbp2_scsi_abort
-Jan 27 19:05:56 stein scsi 33:0:0:0: Device offlined - not ready after error recovery
-Jan 27 19:05:56 stein firewire_sbp2: released fw2.0
-
-The benefit of this is negligible on simple setups. But on buses with
-several devices, we should avoid any unnecessary blockade of fw-sbp2's
-workqueue thread.
-
-Signed-off-by: Stefan Richter <stefanr at s5r6.in-berlin.de>
-Signed-off-by: Jarod Wilson <jwilson at redhat.com>
----
- drivers/firewire/fw-sbp2.c | 16 +++++++++++-----
- 1 file changed, 11 insertions(+), 5 deletions(-)
-
-Index: linux/drivers/firewire/fw-sbp2.c
-===================================================================
---- linux.orig/drivers/firewire/fw-sbp2.c
-+++ linux/drivers/firewire/fw-sbp2.c
-@@ -499,6 +499,9 @@ sbp2_send_management_orb(struct sbp2_log
- unsigned int timeout;
- int retval = -ENOMEM;
-
-+ if (function == SBP2_LOGOUT_REQUEST && fw_device_is_shutdown(device))
-+ return 0;
-+
- orb = kzalloc(sizeof(*orb), GFP_ATOMIC);
- if (orb == NULL)
- return -ENOMEM;
-@@ -619,16 +622,13 @@ static void sbp2_release_target(struct k
- struct sbp2_logical_unit *lu, *next;
- struct Scsi_Host *shost =
- container_of((void *)tgt, struct Scsi_Host, hostdata[0]);
-- struct fw_device *device = fw_device(tgt->unit->device.parent);
-
- list_for_each_entry_safe(lu, next, &tgt->lu_list, link) {
- if (lu->sdev)
- scsi_remove_device(lu->sdev);
-
-- if (!fw_device_is_shutdown(device))
-- sbp2_send_management_orb(lu, tgt->node_id,
-- lu->generation, SBP2_LOGOUT_REQUEST,
-- lu->login_id, NULL);
-+ sbp2_send_management_orb(lu, tgt->node_id, lu->generation,
-+ SBP2_LOGOUT_REQUEST, lu->login_id, NULL);
-
- fw_core_remove_address_handler(&lu->address_handler);
- list_del(&lu->link);
-@@ -673,6 +673,9 @@ static void sbp2_login(struct work_struc
- struct sbp2_login_response response;
- int generation, node_id, local_node_id;
-
-+ if (fw_device_is_shutdown(device))
-+ goto out;
-+
- generation = device->generation;
- smp_rmb(); /* node_id must not be older than generation */
- node_id = device->node_id;
-@@ -968,6 +971,9 @@ static void sbp2_reconnect(struct work_s
- struct fw_device *device = fw_device(unit->device.parent);
- int generation, node_id, local_node_id;
-
-+ if (fw_device_is_shutdown(device))
-+ goto out;
-+
- generation = device->generation;
- smp_rmb(); /* node_id must not be older than generation */
- node_id = device->node_id;
-
---
-Stefan Richter
--=====-==--- ---= ==-==
-http://arcgraph.de/sr/
-
-
There is a race between shutdown and creation of devices: fw-core may
attempt to add a device with the same name of an already existing
device. http://bugzilla.kernel.org/show_bug.cgi?id=9828
@@ -937,15 +841,14 @@
===================================================================
--- linux.orig/drivers/firewire/fw-sbp2.c
+++ linux/drivers/firewire/fw-sbp2.c
-@@ -41,6 +41,8 @@
+@@ -41,6 +41,7 @@
#include <linux/stringify.h>
#include <linux/timer.h>
#include <linux/workqueue.h>
+#include <asm/atomic.h>
-+#include <asm/system.h>
+ #include <asm/system.h>
#include <scsi/scsi.h>
- #include <scsi/scsi_cmnd.h>
@@ -139,6 +140,7 @@ struct sbp2_logical_unit {
int generation;
int retries;
@@ -1018,9 +921,9 @@
static void sbp2_release_target(struct kref *kref)
{
struct sbp2_target *tgt = container_of(kref, struct sbp2_target, kref);
-@@ -653,6 +705,12 @@ static void sbp2_release_target(struct k
- struct Scsi_Host *shost =
+@@ -654,6 +706,12 @@ static void sbp2_release_target(struct k
container_of((void *)tgt, struct Scsi_Host, hostdata[0]);
+ struct fw_device *device = fw_device(tgt->unit->device.parent);
+ /*
+ * Make sure that the target is unblocked and won't be blocked anymore
linux-2.6-firewire-git-update.patch:
Index: linux-2.6-firewire-git-update.patch
===================================================================
RCS file: /cvs/pkgs/rpms/kernel/devel/linux-2.6-firewire-git-update.patch,v
retrieving revision 1.5
retrieving revision 1.6
diff -u -r1.5 -r1.6
--- linux-2.6-firewire-git-update.patch 28 Jan 2008 22:37:22 -0000 1.5
+++ linux-2.6-firewire-git-update.patch 12 Feb 2008 16:41:53 -0000 1.6
@@ -1,797 +1,7 @@
-diff -Naurp linux-2.6-git/drivers/firewire/fw-cdev.c firewire-git/drivers/firewire/fw-cdev.c
---- linux-2.6-git/drivers/firewire/fw-cdev.c 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-cdev.c 2008-01-25 12:41:22.000000000 -0500
-@@ -206,12 +206,13 @@ fill_bus_reset_event(struct fw_cdev_even
-
- event->closure = client->bus_reset_closure;
- event->type = FW_CDEV_EVENT_BUS_RESET;
-+ event->generation = client->device->generation;
-+ smp_rmb(); /* node_id must not be older than generation */
- event->node_id = client->device->node_id;
- event->local_node_id = card->local_node->node_id;
- event->bm_node_id = 0; /* FIXME: We don't track the BM. */
- event->irm_node_id = card->irm_node->node_id;
- event->root_node_id = card->root_node->node_id;
-- event->generation = card->generation;
- }
-
- static void
-diff -Naurp linux-2.6-git/drivers/firewire/fw-device.c firewire-git/drivers/firewire/fw-device.c
---- linux-2.6-git/drivers/firewire/fw-device.c 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-device.c 2008-01-25 12:41:22.000000000 -0500
-@@ -182,9 +182,14 @@ static void fw_device_release(struct dev
-
- int fw_device_enable_phys_dma(struct fw_device *device)
- {
-+ int generation = device->generation;
-+
-+ /* device->node_id, accessed below, must not be older than generation */
-+ smp_rmb();
-+
- return device->card->driver->enable_phys_dma(device->card,
- device->node_id,
-- device->generation);
-+ generation);
- }
- EXPORT_SYMBOL(fw_device_enable_phys_dma);
-
-@@ -384,17 +389,21 @@ complete_transaction(struct fw_card *car
- complete(&callback_data->done);
- }
-
--static int read_rom(struct fw_device *device, int index, u32 * data)
-+static int
-+read_rom(struct fw_device *device, int generation, int index, u32 *data)
- {
- struct read_quadlet_callback_data callback_data;
- struct fw_transaction t;
- u64 offset;
-
-+ /* device->node_id, accessed below, must not be older than generation */
-+ smp_rmb();
-+
- init_completion(&callback_data.done);
-
- offset = 0xfffff0000400ULL + index * 4;
- fw_send_request(device->card, &t, TCODE_READ_QUADLET_REQUEST,
-- device->node_id, device->generation, device->max_speed,
-+ device->node_id, generation, device->max_speed,
- offset, NULL, 4, complete_transaction, &callback_data);
-
- wait_for_completion(&callback_data.done);
-@@ -404,7 +413,14 @@ static int read_rom(struct fw_device *de
- return callback_data.rcode;
- }
-
--static int read_bus_info_block(struct fw_device *device)
-+/*
-+ * Read the bus info block, perform a speed probe, and read all of the rest of
-+ * the config ROM. We do all this with a cached bus generation. If the bus
-+ * generation changes under us, read_bus_info_block will fail and get retried.
-+ * It's better to start all over in this case because the node from which we
-+ * are reading the ROM may have changed the ROM during the reset.
-+ */
-+static int read_bus_info_block(struct fw_device *device, int generation)
- {
- static u32 rom[256];
- u32 stack[16], sp, key;
-@@ -414,7 +430,7 @@ static int read_bus_info_block(struct fw
-
- /* First read the bus info block. */
- for (i = 0; i < 5; i++) {
-- if (read_rom(device, i, &rom[i]) != RCODE_COMPLETE)
-+ if (read_rom(device, generation, i, &rom[i]) != RCODE_COMPLETE)
- return -1;
- /*
- * As per IEEE1212 7.2, during power-up, devices can
-@@ -449,7 +465,8 @@ static int read_bus_info_block(struct fw
- device->max_speed = device->card->link_speed;
-
- while (device->max_speed > SCODE_100) {
-- if (read_rom(device, 0, &dummy) == RCODE_COMPLETE)
-+ if (read_rom(device, generation, 0, &dummy) ==
-+ RCODE_COMPLETE)
- break;
- device->max_speed--;
- }
-@@ -482,7 +499,7 @@ static int read_bus_info_block(struct fw
- return -1;
-
- /* Read header quadlet for the block to get the length. */
-- if (read_rom(device, i, &rom[i]) != RCODE_COMPLETE)
-+ if (read_rom(device, generation, i, &rom[i]) != RCODE_COMPLETE)
- return -1;
- end = i + (rom[i] >> 16) + 1;
- i++;
-@@ -501,7 +518,8 @@ static int read_bus_info_block(struct fw
- * it references another block, and push it in that case.
- */
- while (i < end) {
-- if (read_rom(device, i, &rom[i]) != RCODE_COMPLETE)
-+ if (read_rom(device, generation, i, &rom[i]) !=
-+ RCODE_COMPLETE)
- return -1;
- if ((key >> 30) == 3 && (rom[i] >> 30) > 1 &&
- sp < ARRAY_SIZE(stack))
-@@ -648,7 +666,7 @@ static void fw_device_init(struct work_s
- * device.
- */
-
-- if (read_bus_info_block(device) < 0) {
-+ if (read_bus_info_block(device, device->generation) < 0) {
- if (device->config_rom_retries < MAX_RETRIES) {
- device->config_rom_retries++;
- schedule_delayed_work(&device->work, RETRY_DELAY);
-@@ -801,6 +819,7 @@ void fw_node_event(struct fw_card *card,
-
- device = node->data;
- device->node_id = node->node_id;
-+ smp_wmb(); /* update node_id before generation */
- device->generation = card->generation;
- if (atomic_read(&device->state) == FW_DEVICE_RUNNING) {
- PREPARE_DELAYED_WORK(&device->work, fw_device_update);
-diff -Naurp linux-2.6-git/drivers/firewire/fw-device.h firewire-git/drivers/firewire/fw-device.h
---- linux-2.6-git/drivers/firewire/fw-device.h 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-device.h 2008-01-25 12:41:22.000000000 -0500
-@@ -35,6 +35,18 @@ struct fw_attribute_group {
- struct attribute *attrs[11];
- };
-
-+/*
-+ * Note, fw_device.generation always has to be read before fw_device.node_id.
-+ * Use SMP memory barriers to ensure this. Otherwise requests will be sent
-+ * to an outdated node_id if the generation was updated in the meantime due
-+ * to a bus reset.
-+ *
-+ * Likewise, fw-core will take care to update .node_id before .generation so
-+ * that whenever fw_device.generation is current WRT the actual bus generation,
-+ * fw_device.node_id is guaranteed to be current too.
-+ *
-+ * The same applies to fw_device.card->node_id vs. fw_device.generation.
-+ */
- struct fw_device {
- atomic_t state;
- struct fw_node *node;
-diff -Naurp linux-2.6-git/drivers/firewire/fw-ohci.c firewire-git/drivers/firewire/fw-ohci.c
---- linux-2.6-git/drivers/firewire/fw-ohci.c 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-ohci.c 2008-01-23 16:01:36.000000000 -0500
-@@ -98,17 +98,48 @@ struct context;
- typedef int (*descriptor_callback_t)(struct context *ctx,
- struct descriptor *d,
- struct descriptor *last);
-+
-+/*
-+ * A buffer that contains a block of DMA-able coherent memory used for
-+ * storing a portion of a DMA descriptor program.
-+ */
-+struct descriptor_buffer {
-+ struct list_head list;
-+ dma_addr_t buffer_bus;
-+ size_t buffer_size;
-+ size_t used;
-+ struct descriptor buffer[0];
-+};
-+
- struct context {
- struct fw_ohci *ohci;
- u32 regs;
-+ int total_allocation;
-
-- struct descriptor *buffer;
-- dma_addr_t buffer_bus;
-- size_t buffer_size;
-- struct descriptor *head_descriptor;
-- struct descriptor *tail_descriptor;
-- struct descriptor *tail_descriptor_last;
-- struct descriptor *prev_descriptor;
-+ /*
-+ * List of page-sized buffers for storing DMA descriptors.
-+ * Head of list contains buffers in use and tail of list contains
-+ * free buffers.
-+ */
-+ struct list_head buffer_list;
-+
-+ /*
-+ * Pointer to a buffer inside buffer_list that contains the tail
-+ * end of the current DMA program.
-+ */
-+ struct descriptor_buffer *buffer_tail;
-+
-+ /*
-+ * The descriptor containing the branch address of the first
-+ * descriptor that has not yet been filled by the device.
-+ */
-+ struct descriptor *last;
-+
-+ /*
-+ * The last descriptor in the DMA program. It contains the branch
-+ * address that must be updated upon appending a new descriptor.
-+ */
-+ struct descriptor *prev;
-
- descriptor_callback_t callback;
-
-@@ -125,6 +156,7 @@ struct context {
- struct iso_context {
- struct fw_iso_context base;
- struct context context;
-+ int excess_bytes;
- void *header;
- size_t header_length;
- };
-@@ -197,8 +229,6 @@ static inline struct fw_ohci *fw_ohci(st
- #define SELF_ID_BUF_SIZE 0x800
- #define OHCI_TCODE_PHY_PACKET 0x0e
- #define OHCI_VERSION_1_1 0x010010
--#define ISO_BUFFER_SIZE (64 * 1024)
--#define AT_BUFFER_SIZE 4096
-
- static char ohci_driver_name[] = KBUILD_MODNAME;
-
-@@ -455,71 +485,108 @@ find_branch_descriptor(struct descriptor
- static void context_tasklet(unsigned long data)
- {
- struct context *ctx = (struct context *) data;
-- struct fw_ohci *ohci = ctx->ohci;
- struct descriptor *d, *last;
- u32 address;
- int z;
-+ struct descriptor_buffer *desc;
-
-- dma_sync_single_for_cpu(ohci->card.device, ctx->buffer_bus,
-- ctx->buffer_size, DMA_TO_DEVICE);
--
-- d = ctx->tail_descriptor;
-- last = ctx->tail_descriptor_last;
--
-+ desc = list_entry(ctx->buffer_list.next,
-+ struct descriptor_buffer, list);
-+ last = ctx->last;
- while (last->branch_address != 0) {
-+ struct descriptor_buffer *old_desc = desc;
- address = le32_to_cpu(last->branch_address);
- z = address & 0xf;
-- d = ctx->buffer + (address - ctx->buffer_bus) / sizeof(*d);
-+ address &= ~0xf;
-+
-+ /* If the branch address points to a buffer outside of the
-+ * current buffer, advance to the next buffer. */
-+ if (address < desc->buffer_bus ||
-+ address >= desc->buffer_bus + desc->used)
-+ desc = list_entry(desc->list.next,
-+ struct descriptor_buffer, list);
-+ d = desc->buffer + (address - desc->buffer_bus) / sizeof(*d);
- last = find_branch_descriptor(d, z);
-
- if (!ctx->callback(ctx, d, last))
- break;
-
-- ctx->tail_descriptor = d;
-- ctx->tail_descriptor_last = last;
-+ if (old_desc != desc) {
-+ /* If we've advanced to the next buffer, move the
-+ * previous buffer to the free list. */
-+ unsigned long flags;
-+ old_desc->used = 0;
-+ spin_lock_irqsave(&ctx->ohci->lock, flags);
-+ list_move_tail(&old_desc->list, &ctx->buffer_list);
-+ spin_unlock_irqrestore(&ctx->ohci->lock, flags);
-+ }
-+ ctx->last = last;
- }
- }
-
-+/*
-+ * Allocate a new buffer and add it to the list of free buffers for this
-+ * context. Must be called with ohci->lock held.
-+ */
-+static int
-+context_add_buffer(struct context *ctx)
-+{
-+ struct descriptor_buffer *desc;
-+ dma_addr_t bus_addr;
-+ int offset;
-+
-+ /*
-+ * 16MB of descriptors should be far more than enough for any DMA
-+ * program. This will catch run-away userspace or DoS attacks.
-+ */
-+ if (ctx->total_allocation >= 16*1024*1024)
-+ return -ENOMEM;
-+
-+ desc = dma_alloc_coherent(ctx->ohci->card.device, PAGE_SIZE,
-+ &bus_addr, GFP_ATOMIC);
-+ if (!desc)
-+ return -ENOMEM;
-+
-+ offset = (void *)&desc->buffer - (void *)desc;
-+ desc->buffer_size = PAGE_SIZE - offset;
-+ desc->buffer_bus = bus_addr + offset;
-+ desc->used = 0;
-+
-+ list_add_tail(&desc->list, &ctx->buffer_list);
-+ ctx->total_allocation += PAGE_SIZE;
-+
-+ return 0;
-+}
-+
- static int
- context_init(struct context *ctx, struct fw_ohci *ohci,
-- size_t buffer_size, u32 regs,
-- descriptor_callback_t callback)
-+ u32 regs, descriptor_callback_t callback)
- {
- ctx->ohci = ohci;
- ctx->regs = regs;
-- ctx->buffer_size = buffer_size;
-- ctx->buffer = kmalloc(buffer_size, GFP_KERNEL);
-- if (ctx->buffer == NULL)
-+ ctx->total_allocation = 0;
-+
-+ INIT_LIST_HEAD(&ctx->buffer_list);
-+ if (context_add_buffer(ctx) < 0)
- return -ENOMEM;
-
-+ ctx->buffer_tail = list_entry(ctx->buffer_list.next,
-+ struct descriptor_buffer, list);
-+
- tasklet_init(&ctx->tasklet, context_tasklet, (unsigned long)ctx);
- ctx->callback = callback;
-
-- ctx->buffer_bus =
-- dma_map_single(ohci->card.device, ctx->buffer,
-- buffer_size, DMA_TO_DEVICE);
-- if (dma_mapping_error(ctx->buffer_bus)) {
-- kfree(ctx->buffer);
-- return -ENOMEM;
-- }
--
-- ctx->head_descriptor = ctx->buffer;
-- ctx->prev_descriptor = ctx->buffer;
-- ctx->tail_descriptor = ctx->buffer;
-- ctx->tail_descriptor_last = ctx->buffer;
--
- /*
- * We put a dummy descriptor in the buffer that has a NULL
- * branch address and looks like it's been sent. That way we
-- * have a descriptor to append DMA programs to. Also, the
-- * ring buffer invariant is that it always has at least one
-- * element so that head == tail means buffer full.
-+ * have a descriptor to append DMA programs to.
- */
--
-- memset(ctx->head_descriptor, 0, sizeof(*ctx->head_descriptor));
-- ctx->head_descriptor->control = cpu_to_le16(DESCRIPTOR_OUTPUT_LAST);
-- ctx->head_descriptor->transfer_status = cpu_to_le16(0x8011);
-- ctx->head_descriptor++;
-+ memset(ctx->buffer_tail->buffer, 0, sizeof(*ctx->buffer_tail->buffer));
-+ ctx->buffer_tail->buffer->control = cpu_to_le16(DESCRIPTOR_OUTPUT_LAST);
-+ ctx->buffer_tail->buffer->transfer_status = cpu_to_le16(0x8011);
-+ ctx->buffer_tail->used += sizeof(*ctx->buffer_tail->buffer);
-+ ctx->last = ctx->buffer_tail->buffer;
-+ ctx->prev = ctx->buffer_tail->buffer;
-
- return 0;
- }
-@@ -528,35 +595,42 @@ static void
- context_release(struct context *ctx)
- {
- struct fw_card *card = &ctx->ohci->card;
-+ struct descriptor_buffer *desc, *tmp;
-
-- dma_unmap_single(card->device, ctx->buffer_bus,
-- ctx->buffer_size, DMA_TO_DEVICE);
-- kfree(ctx->buffer);
-+ list_for_each_entry_safe(desc, tmp, &ctx->buffer_list, list)
-+ dma_free_coherent(card->device, PAGE_SIZE, desc,
-+ desc->buffer_bus -
-+ ((void *)&desc->buffer - (void *)desc));
- }
-
-+/* Must be called with ohci->lock held */
- static struct descriptor *
- context_get_descriptors(struct context *ctx, int z, dma_addr_t *d_bus)
- {
-- struct descriptor *d, *tail, *end;
-+ struct descriptor *d = NULL;
-+ struct descriptor_buffer *desc = ctx->buffer_tail;
-
-- d = ctx->head_descriptor;
-- tail = ctx->tail_descriptor;
-- end = ctx->buffer + ctx->buffer_size / sizeof(*d);
-+ if (z * sizeof(*d) > desc->buffer_size)
-+ return NULL;
-
-- if (d + z <= tail) {
-- goto has_space;
-- } else if (d > tail && d + z <= end) {
-- goto has_space;
-- } else if (d > tail && ctx->buffer + z <= tail) {
-- d = ctx->buffer;
-- goto has_space;
-+ if (z * sizeof(*d) > desc->buffer_size - desc->used) {
-+ /* No room for the descriptor in this buffer, so advance to the
-+ * next one. */
-+
-+ if (desc->list.next == &ctx->buffer_list) {
-+ /* If there is no free buffer next in the list,
-+ * allocate one. */
-+ if (context_add_buffer(ctx) < 0)
-+ return NULL;
-+ }
-+ desc = list_entry(desc->list.next,
-+ struct descriptor_buffer, list);
-+ ctx->buffer_tail = desc;
- }
-
-- return NULL;
--
-- has_space:
-+ d = desc->buffer + desc->used / sizeof(*d);
- memset(d, 0, z * sizeof(*d));
-- *d_bus = ctx->buffer_bus + (d - ctx->buffer) * sizeof(*d);
-+ *d_bus = desc->buffer_bus + desc->used;
-
- return d;
- }
-@@ -566,7 +640,7 @@ static void context_run(struct context *
- struct fw_ohci *ohci = ctx->ohci;
-
- reg_write(ohci, COMMAND_PTR(ctx->regs),
-- le32_to_cpu(ctx->tail_descriptor_last->branch_address));
-+ le32_to_cpu(ctx->last->branch_address));
- reg_write(ohci, CONTROL_CLEAR(ctx->regs), ~0);
- reg_write(ohci, CONTROL_SET(ctx->regs), CONTEXT_RUN | extra);
- flush_writes(ohci);
-@@ -576,15 +650,13 @@ static void context_append(struct contex
- struct descriptor *d, int z, int extra)
- {
- dma_addr_t d_bus;
-+ struct descriptor_buffer *desc = ctx->buffer_tail;
-
-- d_bus = ctx->buffer_bus + (d - ctx->buffer) * sizeof(*d);
--
-- ctx->head_descriptor = d + z + extra;
-- ctx->prev_descriptor->branch_address = cpu_to_le32(d_bus | z);
-- ctx->prev_descriptor = find_branch_descriptor(d, z);
-+ d_bus = desc->buffer_bus + (d - desc->buffer) * sizeof(*d);
-
-- dma_sync_single_for_device(ctx->ohci->card.device, ctx->buffer_bus,
-- ctx->buffer_size, DMA_TO_DEVICE);
-+ desc->used += (z + extra) * sizeof(*d);
-+ ctx->prev->branch_address = cpu_to_le32(d_bus | z);
-+ ctx->prev = find_branch_descriptor(d, z);
-
- reg_write(ctx->ohci, CONTROL_SET(ctx->regs), CONTEXT_WAKE);
- flush_writes(ctx->ohci);
-@@ -1078,6 +1150,13 @@ static irqreturn_t irq_handler(int irq,
- if (unlikely(event & OHCI1394_postedWriteErr))
- fw_error("PCI posted write error\n");
-
-+ if (unlikely(event & OHCI1394_cycleTooLong)) {
-+ if (printk_ratelimit())
-+ fw_notify("isochronous cycle too long\n");
-+ reg_write(ohci, OHCI1394_LinkControlSet,
-+ OHCI1394_LinkControl_cycleMaster);
-+ }
-+
- if (event & OHCI1394_cycle64Seconds) {
- cycle_time = reg_read(ohci, OHCI1394_IsochronousCycleTimer);
- if ((cycle_time & 0x80000000) == 0)
-@@ -1151,8 +1230,8 @@ static int ohci_enable(struct fw_card *c
- OHCI1394_RQPkt | OHCI1394_RSPkt |
- OHCI1394_reqTxComplete | OHCI1394_respTxComplete |
- OHCI1394_isochRx | OHCI1394_isochTx |
-- OHCI1394_postedWriteErr | OHCI1394_cycle64Seconds |
-- OHCI1394_masterIntEnable);
-+ OHCI1394_postedWriteErr | OHCI1394_cycleTooLong |
-+ OHCI1394_cycle64Seconds | OHCI1394_masterIntEnable);
-
- /* Activate link_on bit and contender bit in our self ID packets.*/
- if (ohci_update_phy_reg(card, 4, 0,
-@@ -1408,9 +1487,13 @@ static int handle_ir_dualbuffer_packet(s
- void *p, *end;
- int i;
-
-- if (db->first_res_count > 0 && db->second_res_count > 0)
-- /* This descriptor isn't done yet, stop iteration. */
-- return 0;
-+ if (db->first_res_count > 0 && db->second_res_count > 0) {
-+ if (ctx->excess_bytes <= le16_to_cpu(db->second_req_count)) {
-+ /* This descriptor isn't done yet, stop iteration. */
-+ return 0;
-+ }
-+ ctx->excess_bytes -= le16_to_cpu(db->second_req_count);
-+ }
-
- header_length = le16_to_cpu(db->first_req_count) -
- le16_to_cpu(db->first_res_count);
-@@ -1429,11 +1512,15 @@ static int handle_ir_dualbuffer_packet(s
- *(u32 *) (ctx->header + i) = __swab32(*(u32 *) (p + 4));
- memcpy(ctx->header + i + 4, p + 8, ctx->base.header_size - 4);
- i += ctx->base.header_size;
-+ ctx->excess_bytes +=
-+ (le32_to_cpu(*(u32 *)(p + 4)) >> 16) & 0xffff;
- p += ctx->base.header_size + 4;
- }
--
- ctx->header_length = i;
-
-+ ctx->excess_bytes -= le16_to_cpu(db->second_req_count) -
-+ le16_to_cpu(db->second_res_count);
-+
- if (le16_to_cpu(db->control) & DESCRIPTOR_IRQ_ALWAYS) {
- ir_header = (__le32 *) (db + 1);
- ctx->base.callback(&ctx->base,
-@@ -1452,24 +1539,24 @@ static int handle_ir_packet_per_buffer(s
- {
- struct iso_context *ctx =
- container_of(context, struct iso_context, context);
-- struct descriptor *pd = d + 1;
-+ struct descriptor *pd;
- __le32 *ir_header;
-- size_t header_length;
-- void *p, *end;
-- int i, z;
-+ void *p;
-+ int i;
-
-- if (pd->res_count == pd->req_count)
-+ for (pd = d; pd <= last; pd++) {
-+ if (pd->transfer_status)
-+ break;
-+ }
-+ if (pd > last)
- /* Descriptor(s) not done yet, stop iteration */
- return 0;
-
-- header_length = le16_to_cpu(d->req_count);
--
- i = ctx->header_length;
-- z = le32_to_cpu(pd->branch_address) & 0xf;
-- p = d + z;
-- end = p + header_length;
-+ p = last + 1;
-
-- while (p < end && i + ctx->base.header_size <= PAGE_SIZE) {
-+ if (ctx->base.header_size > 0 &&
-+ i + ctx->base.header_size <= PAGE_SIZE) {
- /*
- * The iso header is byteswapped to little endian by
- * the controller, but the remaining header quadlets
-@@ -1478,14 +1565,11 @@ static int handle_ir_packet_per_buffer(s
- */
- *(u32 *) (ctx->header + i) = __swab32(*(u32 *) (p + 4));
- memcpy(ctx->header + i + 4, p + 8, ctx->base.header_size - 4);
-- i += ctx->base.header_size;
-- p += ctx->base.header_size + 4;
-+ ctx->header_length += ctx->base.header_size;
- }
-
-- ctx->header_length = i;
--
-- if (le16_to_cpu(pd->control) & DESCRIPTOR_IRQ_ALWAYS) {
-- ir_header = (__le32 *) (d + z);
-+ if (le16_to_cpu(last->control) & DESCRIPTOR_IRQ_ALWAYS) {
-+ ir_header = (__le32 *) p;
- ctx->base.callback(&ctx->base,
- le32_to_cpu(ir_header[0]) & 0xffff,
- ctx->header_length, ctx->header,
-@@ -1493,7 +1577,6 @@ static int handle_ir_packet_per_buffer(s
- ctx->header_length = 0;
- }
-
--
- return 1;
- }
-
-@@ -1559,8 +1642,7 @@ ohci_allocate_iso_context(struct fw_card
- if (ctx->header == NULL)
- goto out;
-
-- retval = context_init(&ctx->context, ohci, ISO_BUFFER_SIZE,
-- regs, callback);
-+ retval = context_init(&ctx->context, ohci, regs, callback);
- if (retval < 0)
- goto out_with_header;
-
-@@ -1775,19 +1857,6 @@ ohci_queue_iso_receive_dualbuffer(struct
- * packet, retransmit or terminate..
- */
-
-- if (packet->skip) {
-- d = context_get_descriptors(&ctx->context, 2, &d_bus);
-- if (d == NULL)
-- return -ENOMEM;
--
-- db = (struct db_descriptor *) d;
-- db->control = cpu_to_le16(DESCRIPTOR_STATUS |
-- DESCRIPTOR_BRANCH_ALWAYS |
-- DESCRIPTOR_WAIT);
-- db->first_size = cpu_to_le16(ctx->base.header_size + 4);
-- context_append(&ctx->context, d, 2, 0);
-- }
--
- p = packet;
- z = 2;
-
-@@ -1815,11 +1884,18 @@ ohci_queue_iso_receive_dualbuffer(struct
- db->control = cpu_to_le16(DESCRIPTOR_STATUS |
- DESCRIPTOR_BRANCH_ALWAYS);
- db->first_size = cpu_to_le16(ctx->base.header_size + 4);
-- db->first_req_count = cpu_to_le16(header_size);
-+ if (p->skip && rest == p->payload_length) {
-+ db->control |= cpu_to_le16(DESCRIPTOR_WAIT);
-+ db->first_req_count = db->first_size;
-+ } else {
-+ db->first_req_count = cpu_to_le16(header_size);
-+ }
- db->first_res_count = db->first_req_count;
- db->first_buffer = cpu_to_le32(d_bus + sizeof(*db));
-
-- if (offset + rest < PAGE_SIZE)
-+ if (p->skip && rest == p->payload_length)
-+ length = 4;
-+ else if (offset + rest < PAGE_SIZE)
- length = rest;
- else
- length = PAGE_SIZE - offset;
-@@ -1835,7 +1911,8 @@ ohci_queue_iso_receive_dualbuffer(struct
- context_append(&ctx->context, d, z, header_z);
- offset = (offset + length) & ~PAGE_MASK;
- rest -= length;
-- page++;
-+ if (offset == 0)
-+ page++;
- }
-
- return 0;
-@@ -1849,67 +1926,70 @@ ohci_queue_iso_receive_packet_per_buffer
- {
- struct iso_context *ctx = container_of(base, struct iso_context, base);
- struct descriptor *d = NULL, *pd = NULL;
-- struct fw_iso_packet *p;
-+ struct fw_iso_packet *p = packet;
- dma_addr_t d_bus, page_bus;
- u32 z, header_z, rest;
-- int i, page, offset, packet_count, header_size;
--
-- if (packet->skip) {
-- d = context_get_descriptors(&ctx->context, 1, &d_bus);
-- if (d == NULL)
-- return -ENOMEM;
--
-- d->control = cpu_to_le16(DESCRIPTOR_STATUS |
-- DESCRIPTOR_INPUT_LAST |
-- DESCRIPTOR_BRANCH_ALWAYS |
-- DESCRIPTOR_WAIT);
-- context_append(&ctx->context, d, 1, 0);
-- }
--
-- /* one descriptor for header, one for payload */
-- /* FIXME: handle cases where we need multiple desc. for payload */
-- z = 2;
-- p = packet;
-+ int i, j, length;
-+ int page, offset, packet_count, header_size, payload_per_buffer;
-
- /*
- * The OHCI controller puts the status word in the
- * buffer too, so we need 4 extra bytes per packet.
- */
- packet_count = p->header_length / ctx->base.header_size;
-- header_size = packet_count * (ctx->base.header_size + 4);
-+ header_size = ctx->base.header_size + 4;
-
- /* Get header size in number of descriptors. */
- header_z = DIV_ROUND_UP(header_size, sizeof(*d));
- page = payload >> PAGE_SHIFT;
- offset = payload & ~PAGE_MASK;
-- rest = p->payload_length;
-+ payload_per_buffer = p->payload_length / packet_count;
-
- for (i = 0; i < packet_count; i++) {
- /* d points to the header descriptor */
-+ z = DIV_ROUND_UP(payload_per_buffer + offset, PAGE_SIZE) + 1;
- d = context_get_descriptors(&ctx->context,
-- z + header_z, &d_bus);
-+ z + header_z, &d_bus);
- if (d == NULL)
- return -ENOMEM;
-
-- d->control = cpu_to_le16(DESCRIPTOR_INPUT_MORE);
-+ d->control = cpu_to_le16(DESCRIPTOR_STATUS |
-+ DESCRIPTOR_INPUT_MORE);
-+ if (p->skip && i == 0)
-+ d->control |= cpu_to_le16(DESCRIPTOR_WAIT);
- d->req_count = cpu_to_le16(header_size);
- d->res_count = d->req_count;
-+ d->transfer_status = 0;
- d->data_address = cpu_to_le32(d_bus + (z * sizeof(*d)));
-
-- /* pd points to the payload descriptor */
-- pd = d + 1;
-+ rest = payload_per_buffer;
-+ for (j = 1; j < z; j++) {
-+ pd = d + j;
-+ pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
-+ DESCRIPTOR_INPUT_MORE);
-+
-+ if (offset + rest < PAGE_SIZE)
-+ length = rest;
-+ else
-+ length = PAGE_SIZE - offset;
-+ pd->req_count = cpu_to_le16(length);
-+ pd->res_count = pd->req_count;
-+ pd->transfer_status = 0;
-+
-+ page_bus = page_private(buffer->pages[page]);
-+ pd->data_address = cpu_to_le32(page_bus + offset);
-+
-+ offset = (offset + length) & ~PAGE_MASK;
-+ rest -= length;
-+ if (offset == 0)
-+ page++;
-+ }
- pd->control = cpu_to_le16(DESCRIPTOR_STATUS |
- DESCRIPTOR_INPUT_LAST |
- DESCRIPTOR_BRANCH_ALWAYS);
-- if (p->interrupt)
-+ if (p->interrupt && i == packet_count - 1)
- pd->control |= cpu_to_le16(DESCRIPTOR_IRQ_ALWAYS);
-
-- pd->req_count = cpu_to_le16(rest);
-- pd->res_count = pd->req_count;
--
-- page_bus = page_private(buffer->pages[page]);
-- pd->data_address = cpu_to_le32(page_bus + offset);
--
- context_append(&ctx->context, d, z, header_z);
- }
-
-@@ -1923,16 +2003,22 @@ ohci_queue_iso(struct fw_iso_context *ba
- unsigned long payload)
- {
- struct iso_context *ctx = container_of(base, struct iso_context, base);
-+ unsigned long flags;
-+ int retval;
-
-+ spin_lock_irqsave(&ctx->context.ohci->lock, flags);
- if (base->type == FW_ISO_CONTEXT_TRANSMIT)
-- return ohci_queue_iso_transmit(base, packet, buffer, payload);
-+ retval = ohci_queue_iso_transmit(base, packet, buffer, payload);
- else if (ctx->context.ohci->version >= OHCI_VERSION_1_1)
-- return ohci_queue_iso_receive_dualbuffer(base, packet,
-+ retval = ohci_queue_iso_receive_dualbuffer(base, packet,
- buffer, payload);
- else
-- return ohci_queue_iso_receive_packet_per_buffer(base, packet,
-+ retval = ohci_queue_iso_receive_packet_per_buffer(base, packet,
- buffer,
- payload);
-+ spin_unlock_irqrestore(&ctx->context.ohci->lock, flags);
-+
-+ return retval;
- }
-
- static const struct fw_card_driver ohci_driver = {
-@@ -2004,10 +2090,10 @@ pci_probe(struct pci_dev *dev, const str
- ar_context_init(&ohci->ar_response_ctx, ohci,
- OHCI1394_AsRspRcvContextControlSet);
-
-- context_init(&ohci->at_request_ctx, ohci, AT_BUFFER_SIZE,
-+ context_init(&ohci->at_request_ctx, ohci,
- OHCI1394_AsReqTrContextControlSet, handle_at_packet);
-
-- context_init(&ohci->at_response_ctx, ohci, AT_BUFFER_SIZE,
-+ context_init(&ohci->at_response_ctx, ohci,
- OHCI1394_AsRspTrContextControlSet, handle_at_packet);
-
- reg_write(ohci, OHCI1394_IsoRecvIntMaskSet, ~0);
diff -Naurp linux-2.6-git/drivers/firewire/fw-sbp2.c firewire-git/drivers/firewire/fw-sbp2.c
---- linux-2.6-git/drivers/firewire/fw-sbp2.c 2008-01-28 10:57:54.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-sbp2.c 2008-01-28 10:58:18.000000000 -0500
-@@ -140,26 +140,32 @@ struct sbp2_logical_unit {
+--- linux-2.6-git/drivers/firewire/fw-sbp2.c 2008-02-12 10:05:41.000000000 -0500
++++ firewire-git/drivers/firewire/fw-sbp2.c 2008-02-12 10:27:23.000000000 -0500
+@@ -141,15 +141,13 @@ struct sbp2_logical_unit {
struct sbp2_target {
struct kref kref;
struct fw_unit *unit;
@@ -804,34 +14,21 @@
-
- unsigned workarounds;
- struct list_head lu_list;
+-
+ unsigned int workarounds;
-+ unsigned int mgt_orb_timeout;
+ unsigned int mgt_orb_timeout;
};
--#define SBP2_MAX_SG_ELEMENT_LENGTH 0xf000
--#define SBP2_MAX_SECTORS 255 /* Max sectors supported */
+@@ -160,7 +158,7 @@ struct sbp2_target {
+ */
+ #define SBP2_MIN_LOGIN_ORB_TIMEOUT 5000U /* Timeout in ms */
+ #define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */
-#define SBP2_ORB_TIMEOUT 2000 /* Timeout in ms */
--
-+/*
-+ * Per section 7.4.8 of the SBP-2 spec, a mgt_ORB_timeout value can be
-+ * provided in the config rom. Most devices do provide a value, which
-+ * we'll use for login management orbs, but with some sane limits.
-+ */
-+#define SBP2_MIN_LOGIN_ORB_TIMEOUT 5000U /* Timeout in ms */
-+#define SBP2_MAX_LOGIN_ORB_TIMEOUT 40000U /* Timeout in ms */
+#define SBP2_ORB_TIMEOUT 2000U /* Timeout in ms */
#define SBP2_ORB_NULL 0x80000000
-+#define SBP2_MAX_SG_ELEMENT_LENGTH 0xf000
-
- #define SBP2_DIRECTION_TO_MEDIA 0x0
- #define SBP2_DIRECTION_FROM_MEDIA 0x1
+ #define SBP2_MAX_SG_ELEMENT_LENGTH 0xf000
- /* Unit directory keys */
-+#define SBP2_CSR_UNIT_CHARACTERISTICS 0x3a
- #define SBP2_CSR_FIRMWARE_REVISION 0x3c
- #define SBP2_CSR_LOGICAL_UNIT_NUMBER 0x14
- #define SBP2_CSR_LOGICAL_UNIT_DIRECTORY 0xd4
-@@ -288,7 +294,7 @@ struct sbp2_command_orb {
+@@ -297,7 +295,7 @@ struct sbp2_command_orb {
static const struct {
u32 firmware_revision;
u32 model;
@@ -840,114 +37,40 @@
} sbp2_workarounds_table[] = {
/* DViCO Momobay CX-1 with TSB42AA9 bridge */ {
.firmware_revision = 0x002800,
-@@ -489,6 +495,7 @@ sbp2_send_management_orb(struct sbp2_log
- {
- struct fw_device *device = fw_device(lu->tgt->unit->device.parent);
- struct sbp2_management_orb *orb;
-+ unsigned int timeout;
+@@ -501,6 +499,9 @@ sbp2_send_management_orb(struct sbp2_log
+ unsigned int timeout;
int retval = -ENOMEM;
++ if (function == SBP2_LOGOUT_REQUEST && fw_device_is_shutdown(device))
++ return 0;
++
orb = kzalloc(sizeof(*orb), GFP_ATOMIC);
-@@ -516,9 +523,13 @@ sbp2_send_management_orb(struct sbp2_log
- orb->request.status_fifo.low = lu->address_handler.offset;
-
- if (function == SBP2_LOGIN_REQUEST) {
-+ /* Ask for 2^2 == 4 seconds reconnect grace period */
- orb->request.misc |=
-- MANAGEMENT_ORB_EXCLUSIVE(sbp2_param_exclusive_login) |
-- MANAGEMENT_ORB_RECONNECT(0);
-+ MANAGEMENT_ORB_RECONNECT(2) |
-+ MANAGEMENT_ORB_EXCLUSIVE(sbp2_param_exclusive_login);
-+ timeout = lu->tgt->mgt_orb_timeout;
-+ } else {
-+ timeout = SBP2_ORB_TIMEOUT;
- }
-
- fw_memcpy_to_be32(&orb->request, &orb->request, sizeof(orb->request));
-@@ -535,8 +546,7 @@ sbp2_send_management_orb(struct sbp2_log
- sbp2_send_orb(&orb->base, lu, node_id, generation,
- lu->tgt->management_agent_address);
-
-- wait_for_completion_timeout(&orb->done,
-- msecs_to_jiffies(SBP2_ORB_TIMEOUT));
-+ wait_for_completion_timeout(&orb->done, msecs_to_jiffies(timeout));
-
- retval = -EIO;
- if (sbp2_cancel_orbs(lu) == 0) {
-@@ -608,13 +618,17 @@ static void sbp2_release_target(struct k
- struct sbp2_logical_unit *lu, *next;
- struct Scsi_Host *shost =
- container_of((void *)tgt, struct Scsi_Host, hostdata[0]);
-+ struct fw_device *device = fw_device(tgt->unit->device.parent);
-
- list_for_each_entry_safe(lu, next, &tgt->lu_list, link) {
+ if (orb == NULL)
+ return -ENOMEM;
+@@ -627,10 +628,8 @@ static void sbp2_release_target(struct k
if (lu->sdev)
scsi_remove_device(lu->sdev);
-- sbp2_send_management_orb(lu, tgt->node_id, lu->generation,
-- SBP2_LOGOUT_REQUEST, lu->login_id, NULL);
-+ if (!fw_device_is_shutdown(device))
-+ sbp2_send_management_orb(lu, tgt->node_id,
-+ lu->generation, SBP2_LOGOUT_REQUEST,
-+ lu->login_id, NULL);
-+
+- if (!fw_device_is_shutdown(device))
+- sbp2_send_management_orb(lu, tgt->node_id,
+- lu->generation, SBP2_LOGOUT_REQUEST,
+- lu->login_id, NULL);
++ sbp2_send_management_orb(lu, tgt->node_id, lu->generation,
++ SBP2_LOGOUT_REQUEST, lu->login_id, NULL);
+
fw_core_remove_address_handler(&lu->address_handler);
list_del(&lu->link);
- kfree(lu);
-@@ -628,6 +642,21 @@ static void sbp2_release_target(struct k
-
- static struct workqueue_struct *sbp2_wq;
-
-+/*
-+ * Always get the target's kref when scheduling work on one its units.
-+ * Each workqueue job is responsible to call sbp2_target_put() upon return.
-+ */
-+static void sbp2_queue_work(struct sbp2_logical_unit *lu, unsigned long delay)
-+{
-+ if (queue_delayed_work(sbp2_wq, &lu->work, delay))
-+ kref_get(&lu->tgt->kref);
-+}
-+
-+static void sbp2_target_put(struct sbp2_target *tgt)
-+{
-+ kref_put(&tgt->kref, sbp2_release_target);
-+}
-+
- static void sbp2_reconnect(struct work_struct *work);
-
- static void sbp2_login(struct work_struct *work)
-@@ -643,22 +672,19 @@ static void sbp2_login(struct work_struc
+@@ -675,6 +674,9 @@ static void sbp2_login(struct work_struc
struct sbp2_login_response response;
int generation, node_id, local_node_id;
-- generation = device->card->generation;
-- node_id = device->node->node_id;
-- local_node_id = device->card->local_node->node_id;
-+ generation = device->generation;
-+ smp_rmb(); /* node_id must not be older than generation */
-+ node_id = device->node_id;
-+ local_node_id = device->card->node_id;
-
- if (sbp2_send_management_orb(lu, node_id, generation,
- SBP2_LOGIN_REQUEST, lu->lun, &response) < 0) {
-- if (lu->retries++ < 5) {
-- if (queue_delayed_work(sbp2_wq, &lu->work,
-- DIV_ROUND_UP(HZ, 5)))
-- kref_get(&lu->tgt->kref);
-- } else {
-+ if (lu->retries++ < 5)
-+ sbp2_queue_work(lu, DIV_ROUND_UP(HZ, 5));
-+ else
- fw_error("failed to login to %s LUN %04x\n",
- unit->device.bus_id, lu->lun);
-- }
-- kref_put(&lu->tgt->kref, sbp2_release_target);
-- return;
++ if (fw_device_is_shutdown(device))
+ goto out;
- }
-
- lu->generation = generation;
-@@ -689,18 +715,47 @@ static void sbp2_login(struct work_struc
++
+ generation = device->generation;
+ smp_rmb(); /* node_id must not be older than generation */
+ node_id = device->node_id;
+@@ -718,17 +720,45 @@ static void sbp2_login(struct work_struc
sdev = __scsi_add_device(shost, 0, 0,
scsilun_to_int(&eight_bytes_lun), lu);
if (IS_ERR(sdev)) {
@@ -982,7 +105,6 @@
scsi_device_put(sdev);
+ goto out;
}
-- kref_put(&lu->tgt->kref, sbp2_release_target);
+
+ out_logout_login:
+ smp_rmb(); /* generation may have changed */
@@ -996,42 +118,10 @@
+ * lu->work already. Reset the work from reconnect to login.
+ */
+ PREPARE_DELAYED_WORK(&lu->work, sbp2_login);
-+ out:
-+ sbp2_target_put(lu->tgt);
+ out:
+ sbp2_target_put(lu->tgt);
}
-
- static int sbp2_add_logical_unit(struct sbp2_target *tgt, int lun_entry)
-@@ -750,6 +805,7 @@ static int sbp2_scan_unit_dir(struct sbp
- {
- struct fw_csr_iterator ci;
- int key, value;
-+ unsigned int timeout;
-
- fw_csr_iterator_init(&ci, directory);
- while (fw_csr_iterator_next(&ci, &key, &value)) {
-@@ -772,6 +828,21 @@ static int sbp2_scan_unit_dir(struct sbp
- *firmware_revision = value;
- break;
-
-+ case SBP2_CSR_UNIT_CHARACTERISTICS:
-+ /* the timeout value is stored in 500ms units */
-+ timeout = ((unsigned int) value >> 8 & 0xff) * 500;
-+ timeout = max(timeout, SBP2_MIN_LOGIN_ORB_TIMEOUT);
-+ tgt->mgt_orb_timeout =
-+ min(timeout, SBP2_MAX_LOGIN_ORB_TIMEOUT);
-+
-+ if (timeout > tgt->mgt_orb_timeout)
-+ fw_notify("%s: config rom contains %ds "
-+ "management ORB timeout, limiting "
-+ "to %ds\n", tgt->unit->device.bus_id,
-+ timeout / 1000,
-+ tgt->mgt_orb_timeout / 1000);
-+ break;
-+
- case SBP2_CSR_LOGICAL_UNIT_NUMBER:
- if (sbp2_add_logical_unit(tgt, value) < 0)
- return -ENOMEM;
-@@ -790,7 +861,7 @@ static void sbp2_init_workarounds(struct
+@@ -836,7 +866,7 @@ static void sbp2_init_workarounds(struct
u32 firmware_revision)
{
int i;
@@ -1040,99 +130,17 @@
if (w)
fw_notify("Please notify linux1394-devel at lists.sourceforge.net "
-@@ -865,18 +936,13 @@ static int sbp2_probe(struct device *dev
-
- get_device(&unit->device);
-
-- /*
-- * We schedule work to do the login so we can easily
-- * reschedule retries. Always get the ref before scheduling
-- * work.
-- */
-+ /* Do the login in a workqueue so we can easily reschedule retries. */
- list_for_each_entry(lu, &tgt->lu_list, link)
-- if (queue_delayed_work(sbp2_wq, &lu->work, 0))
-- kref_get(&tgt->kref);
-+ sbp2_queue_work(lu, 0);
- return 0;
-
- fail_tgt_put:
-- kref_put(&tgt->kref, sbp2_release_target);
-+ sbp2_target_put(tgt);
- return -ENOMEM;
-
- fail_shost_put:
-@@ -889,7 +955,7 @@ static int sbp2_remove(struct device *de
- struct fw_unit *unit = fw_unit(dev);
- struct sbp2_target *tgt = unit->device.driver_data;
-
-- kref_put(&tgt->kref, sbp2_release_target);
-+ sbp2_target_put(tgt);
- return 0;
- }
-
-@@ -901,9 +967,10 @@ static void sbp2_reconnect(struct work_s
+@@ -942,6 +972,9 @@ static void sbp2_reconnect(struct work_s
struct fw_device *device = fw_device(unit->device.parent);
int generation, node_id, local_node_id;
-- generation = device->card->generation;
-- node_id = device->node->node_id;
-- local_node_id = device->card->local_node->node_id;
-+ generation = device->generation;
-+ smp_rmb(); /* node_id must not be older than generation */
-+ node_id = device->node_id;
-+ local_node_id = device->card->node_id;
-
- if (sbp2_send_management_orb(lu, node_id, generation,
- SBP2_RECONNECT_REQUEST,
-@@ -915,10 +982,8 @@ static void sbp2_reconnect(struct work_s
- lu->retries = 0;
- PREPARE_DELAYED_WORK(&lu->work, sbp2_login);
- }
-- if (queue_delayed_work(sbp2_wq, &lu->work, DIV_ROUND_UP(HZ, 5)))
-- kref_get(&lu->tgt->kref);
-- kref_put(&lu->tgt->kref, sbp2_release_target);
-- return;
-+ sbp2_queue_work(lu, DIV_ROUND_UP(HZ, 5));
++ if (fw_device_is_shutdown(device))
+ goto out;
- }
-
- lu->generation = generation;
-@@ -930,8 +995,8 @@ static void sbp2_reconnect(struct work_s
-
- sbp2_agent_reset(lu);
- sbp2_cancel_orbs(lu);
--
-- kref_put(&lu->tgt->kref, sbp2_release_target);
-+ out:
-+ sbp2_target_put(lu->tgt);
- }
-
- static void sbp2_update(struct fw_unit *unit)
-@@ -947,8 +1012,7 @@ static void sbp2_update(struct fw_unit *
- */
- list_for_each_entry(lu, &tgt->lu_list, link) {
- lu->retries = 0;
-- if (queue_delayed_work(sbp2_wq, &lu->work, 0))
-- kref_get(&tgt->kref);
-+ sbp2_queue_work(lu, 0);
- }
- }
-
-@@ -1103,9 +1167,9 @@ sbp2_map_scatterlist(struct sbp2_command
- * elements larger than 65535 bytes, some IOMMUs may merge sg elements
- * during DMA mapping, and Linux currently doesn't prevent this.
- */
-- for (i = 0, j = 0; i < count; i++) {
-- sg_len = sg_dma_len(sg + i);
-- sg_addr = sg_dma_address(sg + i);
-+ for (i = 0, j = 0; i < count; i++, sg = sg_next(sg)) {
-+ sg_len = sg_dma_len(sg);
-+ sg_addr = sg_dma_address(sg);
- while (sg_len) {
- /* FIXME: This won't get us out of the pinch. */
- if (unlikely(j >= ARRAY_SIZE(orb->page_table))) {
-@@ -1158,7 +1222,7 @@ static int sbp2_scsi_queuecommand(struct
++
+ generation = device->generation;
+ smp_rmb(); /* node_id must not be older than generation */
+ node_id = device->node_id;
+@@ -1197,7 +1230,7 @@ static int sbp2_scsi_queuecommand(struct
struct sbp2_logical_unit *lu = cmd->device->hostdata;
struct fw_device *device = fw_device(lu->tgt->unit->device.parent);
struct sbp2_command_orb *orb;
@@ -1141,39 +149,17 @@
int retval = SCSI_MLQUEUE_HOST_BUSY;
/*
-diff -Naurp linux-2.6-git/drivers/firewire/fw-topology.c firewire-git/drivers/firewire/fw-topology.c
---- linux-2.6-git/drivers/firewire/fw-topology.c 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-topology.c 2008-01-25 12:41:22.000000000 -0500
-@@ -518,6 +518,11 @@ fw_core_handle_bus_reset(struct fw_card
- card->bm_retries = 0;
+@@ -1277,11 +1310,8 @@ static int sbp2_scsi_slave_alloc(struct
- card->node_id = node_id;
-+ /*
-+ * Update node_id before generation to prevent anybody from using
-+ * a stale node_id together with a current generation.
-+ */
-+ smp_wmb();
- card->generation = generation;
- card->reset_jiffies = jiffies;
- schedule_delayed_work(&card->work, 0);
-diff -Naurp linux-2.6-git/drivers/firewire/fw-transaction.c firewire-git/drivers/firewire/fw-transaction.c
---- linux-2.6-git/drivers/firewire/fw-transaction.c 2008-01-01 22:50:33.000000000 -0500
-+++ firewire-git/drivers/firewire/fw-transaction.c 2008-01-25 12:41:24.000000000 -0500
-@@ -153,7 +153,7 @@ fw_fill_request(struct fw_packet *packet
- int ext_tcode;
+ sdev->allow_restart = 1;
- if (tcode > 0x10) {
-- ext_tcode = tcode - 0x10;
-+ ext_tcode = tcode & ~0x10;
- tcode = TCODE_LOCK_REQUEST;
- } else
- ext_tcode = 0;
-@@ -650,7 +650,7 @@ fw_core_handle_request(struct fw_card *c
- HEADER_GET_OFFSET_HIGH(p->header[1]) << 32) | p->header[2];
- tcode = HEADER_GET_TCODE(p->header[0]);
- destination = HEADER_GET_DESTINATION(p->header[0]);
-- source = HEADER_GET_SOURCE(p->header[0]);
-+ source = HEADER_GET_SOURCE(p->header[1]);
+- /*
+- * Update the dma alignment (minimum alignment requirements for
+- * start and end of DMA transfers) to be a sector
+- */
+- blk_queue_update_dma_alignment(sdev->request_queue, 511);
++ /* SBP-2 requires quadlet alignment of the data buffers. */
++ blk_queue_update_dma_alignment(sdev->request_queue, 4 - 1);
- spin_lock_irqsave(&address_handler_lock, flags);
- handler = lookup_enclosing_address_handler(&address_handler_list,
+ if (lu->tgt->workarounds & SBP2_WORKAROUND_INQUIRY_36)
+ sdev->inquiry_len = 36;
More information about the fedora-extras-commits
mailing list