GNU Linux-libre 4.19.286-gnu1
[releases.git] / drivers / scsi / aic94xx / aic94xx_task.c
1 /*
2  * Aic94xx SAS/SATA Tasks
3  *
4  * Copyright (C) 2005 Adaptec, Inc.  All rights reserved.
5  * Copyright (C) 2005 Luben Tuikov <luben_tuikov@adaptec.com>
6  *
7  * This file is licensed under GPLv2.
8  *
9  * This file is part of the aic94xx driver.
10  *
11  * The aic94xx driver is free software; you can redistribute it and/or
12  * modify it under the terms of the GNU General Public License as
13  * published by the Free Software Foundation; version 2 of the
14  * License.
15  *
16  * The aic94xx driver is distributed in the hope that it will be useful,
17  * but WITHOUT ANY WARRANTY; without even the implied warranty of
18  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the GNU
19  * General Public License for more details.
20  *
21  * You should have received a copy of the GNU General Public License
22  * along with the aic94xx driver; if not, write to the Free Software
23  * Foundation, Inc., 51 Franklin St, Fifth Floor, Boston, MA  02110-1301  USA
24  *
25  */
26
27 #include <linux/spinlock.h>
28 #include "aic94xx.h"
29 #include "aic94xx_sas.h"
30 #include "aic94xx_hwi.h"
31
32 static void asd_unbuild_ata_ascb(struct asd_ascb *a);
33 static void asd_unbuild_smp_ascb(struct asd_ascb *a);
34 static void asd_unbuild_ssp_ascb(struct asd_ascb *a);
35
36 static void asd_can_dequeue(struct asd_ha_struct *asd_ha, int num)
37 {
38         unsigned long flags;
39
40         spin_lock_irqsave(&asd_ha->seq.pend_q_lock, flags);
41         asd_ha->seq.can_queue += num;
42         spin_unlock_irqrestore(&asd_ha->seq.pend_q_lock, flags);
43 }
44
45 /* PCI_DMA_... to our direction translation.
46  */
47 static const u8 data_dir_flags[] = {
48         [PCI_DMA_BIDIRECTIONAL] = DATA_DIR_BYRECIPIENT, /* UNSPECIFIED */
49         [PCI_DMA_TODEVICE]      = DATA_DIR_OUT, /* OUTBOUND */
50         [PCI_DMA_FROMDEVICE]    = DATA_DIR_IN, /* INBOUND */
51         [PCI_DMA_NONE]          = DATA_DIR_NONE, /* NO TRANSFER */
52 };
53
54 static int asd_map_scatterlist(struct sas_task *task,
55                                struct sg_el *sg_arr,
56                                gfp_t gfp_flags)
57 {
58         struct asd_ascb *ascb = task->lldd_task;
59         struct asd_ha_struct *asd_ha = ascb->ha;
60         struct scatterlist *sc;
61         int num_sg, res;
62
63         if (task->data_dir == PCI_DMA_NONE)
64                 return 0;
65
66         if (task->num_scatter == 0) {
67                 void *p = task->scatter;
68                 dma_addr_t dma = pci_map_single(asd_ha->pcidev, p,
69                                                 task->total_xfer_len,
70                                                 task->data_dir);
71                 if (dma_mapping_error(&asd_ha->pcidev->dev, dma))
72                         return -ENOMEM;
73
74                 sg_arr[0].bus_addr = cpu_to_le64((u64)dma);
75                 sg_arr[0].size = cpu_to_le32(task->total_xfer_len);
76                 sg_arr[0].flags |= ASD_SG_EL_LIST_EOL;
77                 return 0;
78         }
79
80         /* STP tasks come from libata which has already mapped
81          * the SG list */
82         if (sas_protocol_ata(task->task_proto))
83                 num_sg = task->num_scatter;
84         else
85                 num_sg = pci_map_sg(asd_ha->pcidev, task->scatter,
86                                     task->num_scatter, task->data_dir);
87         if (num_sg == 0)
88                 return -ENOMEM;
89
90         if (num_sg > 3) {
91                 int i;
92
93                 ascb->sg_arr = asd_alloc_coherent(asd_ha,
94                                                   num_sg*sizeof(struct sg_el),
95                                                   gfp_flags);
96                 if (!ascb->sg_arr) {
97                         res = -ENOMEM;
98                         goto err_unmap;
99                 }
100                 for_each_sg(task->scatter, sc, num_sg, i) {
101                         struct sg_el *sg =
102                                 &((struct sg_el *)ascb->sg_arr->vaddr)[i];
103                         sg->bus_addr = cpu_to_le64((u64)sg_dma_address(sc));
104                         sg->size = cpu_to_le32((u32)sg_dma_len(sc));
105                         if (i == num_sg-1)
106                                 sg->flags |= ASD_SG_EL_LIST_EOL;
107                 }
108
109                 for_each_sg(task->scatter, sc, 2, i) {
110                         sg_arr[i].bus_addr =
111                                 cpu_to_le64((u64)sg_dma_address(sc));
112                         sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc));
113                 }
114                 sg_arr[1].next_sg_offs = 2 * sizeof(*sg_arr);
115                 sg_arr[1].flags |= ASD_SG_EL_LIST_EOS;
116
117                 memset(&sg_arr[2], 0, sizeof(*sg_arr));
118                 sg_arr[2].bus_addr=cpu_to_le64((u64)ascb->sg_arr->dma_handle);
119         } else {
120                 int i;
121                 for_each_sg(task->scatter, sc, num_sg, i) {
122                         sg_arr[i].bus_addr =
123                                 cpu_to_le64((u64)sg_dma_address(sc));
124                         sg_arr[i].size = cpu_to_le32((u32)sg_dma_len(sc));
125                 }
126                 sg_arr[i-1].flags |= ASD_SG_EL_LIST_EOL;
127         }
128
129         return 0;
130 err_unmap:
131         if (sas_protocol_ata(task->task_proto))
132                 pci_unmap_sg(asd_ha->pcidev, task->scatter, task->num_scatter,
133                              task->data_dir);
134         return res;
135 }
136
137 static void asd_unmap_scatterlist(struct asd_ascb *ascb)
138 {
139         struct asd_ha_struct *asd_ha = ascb->ha;
140         struct sas_task *task = ascb->uldd_task;
141
142         if (task->data_dir == PCI_DMA_NONE)
143                 return;
144
145         if (task->num_scatter == 0) {
146                 dma_addr_t dma = (dma_addr_t)
147                        le64_to_cpu(ascb->scb->ssp_task.sg_element[0].bus_addr);
148                 pci_unmap_single(ascb->ha->pcidev, dma, task->total_xfer_len,
149                                  task->data_dir);
150                 return;
151         }
152
153         asd_free_coherent(asd_ha, ascb->sg_arr);
154         if (task->task_proto != SAS_PROTOCOL_STP)
155                 pci_unmap_sg(asd_ha->pcidev, task->scatter, task->num_scatter,
156                              task->data_dir);
157 }
158
159 /* ---------- Task complete tasklet ---------- */
160
161 static void asd_get_response_tasklet(struct asd_ascb *ascb,
162                                      struct done_list_struct *dl)
163 {
164         struct asd_ha_struct *asd_ha = ascb->ha;
165         struct sas_task *task = ascb->uldd_task;
166         struct task_status_struct *ts = &task->task_status;
167         unsigned long flags;
168         struct tc_resp_sb_struct {
169                 __le16 index_escb;
170                 u8     len_lsb;
171                 u8     flags;
172         } __attribute__ ((packed)) *resp_sb = (void *) dl->status_block;
173
174 /*      int  size   = ((resp_sb->flags & 7) << 8) | resp_sb->len_lsb; */
175         int  edb_id = ((resp_sb->flags & 0x70) >> 4)-1;
176         struct asd_ascb *escb;
177         struct asd_dma_tok *edb;
178         void *r;
179
180         spin_lock_irqsave(&asd_ha->seq.tc_index_lock, flags);
181         escb = asd_tc_index_find(&asd_ha->seq,
182                                  (int)le16_to_cpu(resp_sb->index_escb));
183         spin_unlock_irqrestore(&asd_ha->seq.tc_index_lock, flags);
184
185         if (!escb) {
186                 ASD_DPRINTK("Uh-oh! No escb for this dl?!\n");
187                 return;
188         }
189
190         ts->buf_valid_size = 0;
191         edb = asd_ha->seq.edb_arr[edb_id + escb->edb_index];
192         r = edb->vaddr;
193         if (task->task_proto == SAS_PROTOCOL_SSP) {
194                 struct ssp_response_iu *iu =
195                         r + 16 + sizeof(struct ssp_frame_hdr);
196
197                 ts->residual = le32_to_cpu(*(__le32 *)r);
198
199                 sas_ssp_task_response(&asd_ha->pcidev->dev, task, iu);
200         }  else {
201                 struct ata_task_resp *resp = (void *) &ts->buf[0];
202
203                 ts->residual = le32_to_cpu(*(__le32 *)r);
204
205                 if (SAS_STATUS_BUF_SIZE >= sizeof(*resp)) {
206                         resp->frame_len = le16_to_cpu(*(__le16 *)(r+6));
207                         memcpy(&resp->ending_fis[0], r+16, ATA_RESP_FIS_SIZE);
208                         ts->buf_valid_size = sizeof(*resp);
209                 }
210         }
211
212         asd_invalidate_edb(escb, edb_id);
213 }
214
215 static void asd_task_tasklet_complete(struct asd_ascb *ascb,
216                                       struct done_list_struct *dl)
217 {
218         struct sas_task *task = ascb->uldd_task;
219         struct task_status_struct *ts = &task->task_status;
220         unsigned long flags;
221         u8 opcode = dl->opcode;
222
223         asd_can_dequeue(ascb->ha, 1);
224
225 Again:
226         switch (opcode) {
227         case TC_NO_ERROR:
228                 ts->resp = SAS_TASK_COMPLETE;
229                 ts->stat = SAM_STAT_GOOD;
230                 break;
231         case TC_UNDERRUN:
232                 ts->resp = SAS_TASK_COMPLETE;
233                 ts->stat = SAS_DATA_UNDERRUN;
234                 ts->residual = le32_to_cpu(*(__le32 *)dl->status_block);
235                 break;
236         case TC_OVERRUN:
237                 ts->resp = SAS_TASK_COMPLETE;
238                 ts->stat = SAS_DATA_OVERRUN;
239                 ts->residual = 0;
240                 break;
241         case TC_SSP_RESP:
242         case TC_ATA_RESP:
243                 ts->resp = SAS_TASK_COMPLETE;
244                 ts->stat = SAS_PROTO_RESPONSE;
245                 asd_get_response_tasklet(ascb, dl);
246                 break;
247         case TF_OPEN_REJECT:
248                 ts->resp = SAS_TASK_UNDELIVERED;
249                 ts->stat = SAS_OPEN_REJECT;
250                 if (dl->status_block[1] & 2)
251                         ts->open_rej_reason = 1 + dl->status_block[2];
252                 else if (dl->status_block[1] & 1)
253                         ts->open_rej_reason = (dl->status_block[2] >> 4)+10;
254                 else
255                         ts->open_rej_reason = SAS_OREJ_UNKNOWN;
256                 break;
257         case TF_OPEN_TO:
258                 ts->resp = SAS_TASK_UNDELIVERED;
259                 ts->stat = SAS_OPEN_TO;
260                 break;
261         case TF_PHY_DOWN:
262         case TU_PHY_DOWN:
263                 ts->resp = SAS_TASK_UNDELIVERED;
264                 ts->stat = SAS_PHY_DOWN;
265                 break;
266         case TI_PHY_DOWN:
267                 ts->resp = SAS_TASK_COMPLETE;
268                 ts->stat = SAS_PHY_DOWN;
269                 break;
270         case TI_BREAK:
271         case TI_PROTO_ERR:
272         case TI_NAK:
273         case TI_ACK_NAK_TO:
274         case TF_SMP_XMIT_RCV_ERR:
275         case TC_ATA_R_ERR_RECV:
276                 ts->resp = SAS_TASK_COMPLETE;
277                 ts->stat = SAS_INTERRUPTED;
278                 break;
279         case TF_BREAK:
280         case TU_BREAK:
281         case TU_ACK_NAK_TO:
282         case TF_SMPRSP_TO:
283                 ts->resp = SAS_TASK_UNDELIVERED;
284                 ts->stat = SAS_DEV_NO_RESPONSE;
285                 break;
286         case TF_NAK_RECV:
287                 ts->resp = SAS_TASK_COMPLETE;
288                 ts->stat = SAS_NAK_R_ERR;
289                 break;
290         case TA_I_T_NEXUS_LOSS:
291                 opcode = dl->status_block[0];
292                 goto Again;
293                 break;
294         case TF_INV_CONN_HANDLE:
295                 ts->resp = SAS_TASK_UNDELIVERED;
296                 ts->stat = SAS_DEVICE_UNKNOWN;
297                 break;
298         case TF_REQUESTED_N_PENDING:
299                 ts->resp = SAS_TASK_UNDELIVERED;
300                 ts->stat = SAS_PENDING;
301                 break;
302         case TC_TASK_CLEARED:
303         case TA_ON_REQ:
304                 ts->resp = SAS_TASK_COMPLETE;
305                 ts->stat = SAS_ABORTED_TASK;
306                 break;
307
308         case TF_NO_SMP_CONN:
309         case TF_TMF_NO_CTX:
310         case TF_TMF_NO_TAG:
311         case TF_TMF_TAG_FREE:
312         case TF_TMF_TASK_DONE:
313         case TF_TMF_NO_CONN_HANDLE:
314         case TF_IRTT_TO:
315         case TF_IU_SHORT:
316         case TF_DATA_OFFS_ERR:
317                 ts->resp = SAS_TASK_UNDELIVERED;
318                 ts->stat = SAS_DEV_NO_RESPONSE;
319                 break;
320
321         case TC_LINK_ADM_RESP:
322         case TC_CONTROL_PHY:
323         case TC_RESUME:
324         case TC_PARTIAL_SG_LIST:
325         default:
326                 ASD_DPRINTK("%s: dl opcode: 0x%x?\n", __func__, opcode);
327                 break;
328         }
329
330         switch (task->task_proto) {
331         case SAS_PROTOCOL_SATA:
332         case SAS_PROTOCOL_STP:
333                 asd_unbuild_ata_ascb(ascb);
334                 break;
335         case SAS_PROTOCOL_SMP:
336                 asd_unbuild_smp_ascb(ascb);
337                 break;
338         case SAS_PROTOCOL_SSP:
339                 asd_unbuild_ssp_ascb(ascb);
340         default:
341                 break;
342         }
343
344         spin_lock_irqsave(&task->task_state_lock, flags);
345         task->task_state_flags &= ~SAS_TASK_STATE_PENDING;
346         task->task_state_flags &= ~SAS_TASK_AT_INITIATOR;
347         task->task_state_flags |= SAS_TASK_STATE_DONE;
348         if (unlikely((task->task_state_flags & SAS_TASK_STATE_ABORTED))) {
349                 struct completion *completion = ascb->completion;
350                 spin_unlock_irqrestore(&task->task_state_lock, flags);
351                 ASD_DPRINTK("task 0x%p done with opcode 0x%x resp 0x%x "
352                             "stat 0x%x but aborted by upper layer!\n",
353                             task, opcode, ts->resp, ts->stat);
354                 if (completion)
355                         complete(completion);
356         } else {
357                 spin_unlock_irqrestore(&task->task_state_lock, flags);
358                 task->lldd_task = NULL;
359                 asd_ascb_free(ascb);
360                 mb();
361                 task->task_done(task);
362         }
363 }
364
365 /* ---------- ATA ---------- */
366
367 static int asd_build_ata_ascb(struct asd_ascb *ascb, struct sas_task *task,
368                               gfp_t gfp_flags)
369 {
370         struct domain_device *dev = task->dev;
371         struct scb *scb;
372         u8     flags;
373         int    res = 0;
374
375         scb = ascb->scb;
376
377         if (unlikely(task->ata_task.device_control_reg_update))
378                 scb->header.opcode = CONTROL_ATA_DEV;
379         else if (dev->sata_dev.class == ATA_DEV_ATAPI)
380                 scb->header.opcode = INITIATE_ATAPI_TASK;
381         else
382                 scb->header.opcode = INITIATE_ATA_TASK;
383
384         scb->ata_task.proto_conn_rate = (1 << 5); /* STP */
385         if (dev->port->oob_mode == SAS_OOB_MODE)
386                 scb->ata_task.proto_conn_rate |= dev->linkrate;
387
388         scb->ata_task.total_xfer_len = cpu_to_le32(task->total_xfer_len);
389         scb->ata_task.fis = task->ata_task.fis;
390         if (likely(!task->ata_task.device_control_reg_update))
391                 scb->ata_task.fis.flags |= 0x80; /* C=1: update ATA cmd reg */
392         scb->ata_task.fis.flags &= 0xF0; /* PM_PORT field shall be 0 */
393         if (dev->sata_dev.class == ATA_DEV_ATAPI)
394                 memcpy(scb->ata_task.atapi_packet, task->ata_task.atapi_packet,
395                        16);
396         scb->ata_task.sister_scb = cpu_to_le16(0xFFFF);
397         scb->ata_task.conn_handle = cpu_to_le16(
398                 (u16)(unsigned long)dev->lldd_dev);
399
400         if (likely(!task->ata_task.device_control_reg_update)) {
401                 flags = 0;
402                 if (task->ata_task.dma_xfer)
403                         flags |= DATA_XFER_MODE_DMA;
404                 if (task->ata_task.use_ncq &&
405                     dev->sata_dev.class != ATA_DEV_ATAPI)
406                         flags |= ATA_Q_TYPE_NCQ;
407                 flags |= data_dir_flags[task->data_dir];
408                 scb->ata_task.ata_flags = flags;
409
410                 scb->ata_task.retry_count = task->ata_task.retry_count;
411
412                 flags = 0;
413                 if (task->ata_task.set_affil_pol)
414                         flags |= SET_AFFIL_POLICY;
415                 if (task->ata_task.stp_affil_pol)
416                         flags |= STP_AFFIL_POLICY;
417                 scb->ata_task.flags = flags;
418         }
419         ascb->tasklet_complete = asd_task_tasklet_complete;
420
421         if (likely(!task->ata_task.device_control_reg_update))
422                 res = asd_map_scatterlist(task, scb->ata_task.sg_element,
423                                           gfp_flags);
424
425         return res;
426 }
427
428 static void asd_unbuild_ata_ascb(struct asd_ascb *a)
429 {
430         asd_unmap_scatterlist(a);
431 }
432
433 /* ---------- SMP ---------- */
434
435 static int asd_build_smp_ascb(struct asd_ascb *ascb, struct sas_task *task,
436                               gfp_t gfp_flags)
437 {
438         struct asd_ha_struct *asd_ha = ascb->ha;
439         struct domain_device *dev = task->dev;
440         struct scb *scb;
441
442         pci_map_sg(asd_ha->pcidev, &task->smp_task.smp_req, 1,
443                    PCI_DMA_TODEVICE);
444         pci_map_sg(asd_ha->pcidev, &task->smp_task.smp_resp, 1,
445                    PCI_DMA_FROMDEVICE);
446
447         scb = ascb->scb;
448
449         scb->header.opcode = INITIATE_SMP_TASK;
450
451         scb->smp_task.proto_conn_rate = dev->linkrate;
452
453         scb->smp_task.smp_req.bus_addr =
454                 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_req));
455         scb->smp_task.smp_req.size =
456                 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_req)-4);
457
458         scb->smp_task.smp_resp.bus_addr =
459                 cpu_to_le64((u64)sg_dma_address(&task->smp_task.smp_resp));
460         scb->smp_task.smp_resp.size =
461                 cpu_to_le32((u32)sg_dma_len(&task->smp_task.smp_resp)-4);
462
463         scb->smp_task.sister_scb = cpu_to_le16(0xFFFF);
464         scb->smp_task.conn_handle = cpu_to_le16((u16)
465                                                 (unsigned long)dev->lldd_dev);
466
467         ascb->tasklet_complete = asd_task_tasklet_complete;
468
469         return 0;
470 }
471
472 static void asd_unbuild_smp_ascb(struct asd_ascb *a)
473 {
474         struct sas_task *task = a->uldd_task;
475
476         BUG_ON(!task);
477         pci_unmap_sg(a->ha->pcidev, &task->smp_task.smp_req, 1,
478                      PCI_DMA_TODEVICE);
479         pci_unmap_sg(a->ha->pcidev, &task->smp_task.smp_resp, 1,
480                      PCI_DMA_FROMDEVICE);
481 }
482
483 /* ---------- SSP ---------- */
484
485 static int asd_build_ssp_ascb(struct asd_ascb *ascb, struct sas_task *task,
486                               gfp_t gfp_flags)
487 {
488         struct domain_device *dev = task->dev;
489         struct scb *scb;
490         int    res = 0;
491
492         scb = ascb->scb;
493
494         scb->header.opcode = INITIATE_SSP_TASK;
495
496         scb->ssp_task.proto_conn_rate  = (1 << 4); /* SSP */
497         scb->ssp_task.proto_conn_rate |= dev->linkrate;
498         scb->ssp_task.total_xfer_len = cpu_to_le32(task->total_xfer_len);
499         scb->ssp_task.ssp_frame.frame_type = SSP_DATA;
500         memcpy(scb->ssp_task.ssp_frame.hashed_dest_addr, dev->hashed_sas_addr,
501                HASHED_SAS_ADDR_SIZE);
502         memcpy(scb->ssp_task.ssp_frame.hashed_src_addr,
503                dev->port->ha->hashed_sas_addr, HASHED_SAS_ADDR_SIZE);
504         scb->ssp_task.ssp_frame.tptt = cpu_to_be16(0xFFFF);
505
506         memcpy(scb->ssp_task.ssp_cmd.lun, task->ssp_task.LUN, 8);
507         if (task->ssp_task.enable_first_burst)
508                 scb->ssp_task.ssp_cmd.efb_prio_attr |= EFB_MASK;
509         scb->ssp_task.ssp_cmd.efb_prio_attr |= (task->ssp_task.task_prio << 3);
510         scb->ssp_task.ssp_cmd.efb_prio_attr |= (task->ssp_task.task_attr & 7);
511         memcpy(scb->ssp_task.ssp_cmd.cdb, task->ssp_task.cmd->cmnd,
512                task->ssp_task.cmd->cmd_len);
513
514         scb->ssp_task.sister_scb = cpu_to_le16(0xFFFF);
515         scb->ssp_task.conn_handle = cpu_to_le16(
516                 (u16)(unsigned long)dev->lldd_dev);
517         scb->ssp_task.data_dir = data_dir_flags[task->data_dir];
518         scb->ssp_task.retry_count = scb->ssp_task.retry_count;
519
520         ascb->tasklet_complete = asd_task_tasklet_complete;
521
522         res = asd_map_scatterlist(task, scb->ssp_task.sg_element, gfp_flags);
523
524         return res;
525 }
526
527 static void asd_unbuild_ssp_ascb(struct asd_ascb *a)
528 {
529         asd_unmap_scatterlist(a);
530 }
531
532 /* ---------- Execute Task ---------- */
533
534 static int asd_can_queue(struct asd_ha_struct *asd_ha, int num)
535 {
536         int res = 0;
537         unsigned long flags;
538
539         spin_lock_irqsave(&asd_ha->seq.pend_q_lock, flags);
540         if ((asd_ha->seq.can_queue - num) < 0)
541                 res = -SAS_QUEUE_FULL;
542         else
543                 asd_ha->seq.can_queue -= num;
544         spin_unlock_irqrestore(&asd_ha->seq.pend_q_lock, flags);
545
546         return res;
547 }
548
549 int asd_execute_task(struct sas_task *task, gfp_t gfp_flags)
550 {
551         int res = 0;
552         LIST_HEAD(alist);
553         struct sas_task *t = task;
554         struct asd_ascb *ascb = NULL, *a;
555         struct asd_ha_struct *asd_ha = task->dev->port->ha->lldd_ha;
556         unsigned long flags;
557
558         res = asd_can_queue(asd_ha, 1);
559         if (res)
560                 return res;
561
562         res = 1;
563         ascb = asd_ascb_alloc_list(asd_ha, &res, gfp_flags);
564         if (res) {
565                 res = -ENOMEM;
566                 goto out_err;
567         }
568
569         __list_add(&alist, ascb->list.prev, &ascb->list);
570         list_for_each_entry(a, &alist, list) {
571                 a->uldd_task = t;
572                 t->lldd_task = a;
573                 break;
574         }
575         list_for_each_entry(a, &alist, list) {
576                 t = a->uldd_task;
577                 a->uldd_timer = 1;
578                 if (t->task_proto & SAS_PROTOCOL_STP)
579                         t->task_proto = SAS_PROTOCOL_STP;
580                 switch (t->task_proto) {
581                 case SAS_PROTOCOL_SATA:
582                 case SAS_PROTOCOL_STP:
583                         res = asd_build_ata_ascb(a, t, gfp_flags);
584                         break;
585                 case SAS_PROTOCOL_SMP:
586                         res = asd_build_smp_ascb(a, t, gfp_flags);
587                         break;
588                 case SAS_PROTOCOL_SSP:
589                         res = asd_build_ssp_ascb(a, t, gfp_flags);
590                         break;
591                 default:
592                         asd_printk("unknown sas_task proto: 0x%x\n",
593                                    t->task_proto);
594                         res = -ENOMEM;
595                         break;
596                 }
597                 if (res)
598                         goto out_err_unmap;
599
600                 spin_lock_irqsave(&t->task_state_lock, flags);
601                 t->task_state_flags |= SAS_TASK_AT_INITIATOR;
602                 spin_unlock_irqrestore(&t->task_state_lock, flags);
603         }
604         list_del_init(&alist);
605
606         res = asd_post_ascb_list(asd_ha, ascb, 1);
607         if (unlikely(res)) {
608                 a = NULL;
609                 __list_add(&alist, ascb->list.prev, &ascb->list);
610                 goto out_err_unmap;
611         }
612
613         return 0;
614 out_err_unmap:
615         {
616                 struct asd_ascb *b = a;
617                 list_for_each_entry(a, &alist, list) {
618                         if (a == b)
619                                 break;
620                         t = a->uldd_task;
621                         spin_lock_irqsave(&t->task_state_lock, flags);
622                         t->task_state_flags &= ~SAS_TASK_AT_INITIATOR;
623                         spin_unlock_irqrestore(&t->task_state_lock, flags);
624                         switch (t->task_proto) {
625                         case SAS_PROTOCOL_SATA:
626                         case SAS_PROTOCOL_STP:
627                                 asd_unbuild_ata_ascb(a);
628                                 break;
629                         case SAS_PROTOCOL_SMP:
630                                 asd_unbuild_smp_ascb(a);
631                                 break;
632                         case SAS_PROTOCOL_SSP:
633                                 asd_unbuild_ssp_ascb(a);
634                         default:
635                                 break;
636                         }
637                         t->lldd_task = NULL;
638                 }
639         }
640         list_del_init(&alist);
641 out_err:
642         if (ascb)
643                 asd_ascb_free_list(ascb);
644         asd_can_dequeue(asd_ha, 1);
645         return res;
646 }