Mention branches and keyring.
[releases.git] / efct / efct_hw_queues.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (C) 2021 Broadcom. All Rights Reserved. The term
4  * “Broadcom” refers to Broadcom Inc. and/or its subsidiaries.
5  */
6
7 #include "efct_driver.h"
8 #include "efct_hw.h"
9 #include "efct_unsol.h"
10
11 int
12 efct_hw_init_queues(struct efct_hw *hw)
13 {
14         struct hw_eq *eq = NULL;
15         struct hw_cq *cq = NULL;
16         struct hw_wq *wq = NULL;
17         struct hw_mq *mq = NULL;
18
19         struct hw_eq *eqs[EFCT_HW_MAX_NUM_EQ];
20         struct hw_cq *cqs[EFCT_HW_MAX_NUM_EQ];
21         struct hw_rq *rqs[EFCT_HW_MAX_NUM_EQ];
22         u32 i = 0, j;
23
24         hw->eq_count = 0;
25         hw->cq_count = 0;
26         hw->mq_count = 0;
27         hw->wq_count = 0;
28         hw->rq_count = 0;
29         hw->hw_rq_count = 0;
30         INIT_LIST_HEAD(&hw->eq_list);
31
32         for (i = 0; i < hw->config.n_eq; i++) {
33                 /* Create EQ */
34                 eq = efct_hw_new_eq(hw, EFCT_HW_EQ_DEPTH);
35                 if (!eq) {
36                         efct_hw_queue_teardown(hw);
37                         return -ENOMEM;
38                 }
39
40                 eqs[i] = eq;
41
42                 /* Create one MQ */
43                 if (!i) {
44                         cq = efct_hw_new_cq(eq,
45                                             hw->num_qentries[SLI4_QTYPE_CQ]);
46                         if (!cq) {
47                                 efct_hw_queue_teardown(hw);
48                                 return -ENOMEM;
49                         }
50
51                         mq = efct_hw_new_mq(cq, EFCT_HW_MQ_DEPTH);
52                         if (!mq) {
53                                 efct_hw_queue_teardown(hw);
54                                 return -ENOMEM;
55                         }
56                 }
57
58                 /* Create WQ */
59                 cq = efct_hw_new_cq(eq, hw->num_qentries[SLI4_QTYPE_CQ]);
60                 if (!cq) {
61                         efct_hw_queue_teardown(hw);
62                         return -ENOMEM;
63                 }
64
65                 wq = efct_hw_new_wq(cq, hw->num_qentries[SLI4_QTYPE_WQ]);
66                 if (!wq) {
67                         efct_hw_queue_teardown(hw);
68                         return -ENOMEM;
69                 }
70         }
71
72         /* Create CQ set */
73         if (efct_hw_new_cq_set(eqs, cqs, i, hw->num_qentries[SLI4_QTYPE_CQ])) {
74                 efct_hw_queue_teardown(hw);
75                 return -EIO;
76         }
77
78         /* Create RQ set */
79         if (efct_hw_new_rq_set(cqs, rqs, i, EFCT_HW_RQ_ENTRIES_DEF)) {
80                 efct_hw_queue_teardown(hw);
81                 return -EIO;
82         }
83
84         for (j = 0; j < i ; j++) {
85                 rqs[j]->filter_mask = 0;
86                 rqs[j]->is_mrq = true;
87                 rqs[j]->base_mrq_id = rqs[0]->hdr->id;
88         }
89
90         hw->hw_mrq_count = i;
91
92         return 0;
93 }
94
95 int
96 efct_hw_map_wq_cpu(struct efct_hw *hw)
97 {
98         struct efct *efct = hw->os;
99         u32 cpu = 0, i;
100
101         /* Init cpu_map array */
102         hw->wq_cpu_array = kcalloc(num_possible_cpus(), sizeof(void *),
103                                    GFP_KERNEL);
104         if (!hw->wq_cpu_array)
105                 return -ENOMEM;
106
107         for (i = 0; i < hw->config.n_eq; i++) {
108                 const struct cpumask *maskp;
109
110                 /* Get a CPU mask for all CPUs affinitized to this vector */
111                 maskp = pci_irq_get_affinity(efct->pci, i);
112                 if (!maskp) {
113                         efc_log_debug(efct, "maskp null for vector:%d\n", i);
114                         continue;
115                 }
116
117                 /* Loop through all CPUs associated with vector idx */
118                 for_each_cpu_and(cpu, maskp, cpu_present_mask) {
119                         efc_log_debug(efct, "CPU:%d irq vector:%d\n", cpu, i);
120                         hw->wq_cpu_array[cpu] = hw->hw_wq[i];
121                 }
122         }
123
124         return 0;
125 }
126
127 struct hw_eq *
128 efct_hw_new_eq(struct efct_hw *hw, u32 entry_count)
129 {
130         struct hw_eq *eq = kzalloc(sizeof(*eq), GFP_KERNEL);
131
132         if (!eq)
133                 return NULL;
134
135         eq->type = SLI4_QTYPE_EQ;
136         eq->hw = hw;
137         eq->entry_count = entry_count;
138         eq->instance = hw->eq_count++;
139         eq->queue = &hw->eq[eq->instance];
140         INIT_LIST_HEAD(&eq->cq_list);
141
142         if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_EQ, eq->queue, entry_count,
143                             NULL)) {
144                 efc_log_err(hw->os, "EQ[%d] alloc failure\n", eq->instance);
145                 kfree(eq);
146                 return NULL;
147         }
148
149         sli_eq_modify_delay(&hw->sli, eq->queue, 1, 0, 8);
150         hw->hw_eq[eq->instance] = eq;
151         INIT_LIST_HEAD(&eq->list_entry);
152         list_add_tail(&eq->list_entry, &hw->eq_list);
153         efc_log_debug(hw->os, "create eq[%2d] id %3d len %4d\n", eq->instance,
154                       eq->queue->id, eq->entry_count);
155         return eq;
156 }
157
158 struct hw_cq *
159 efct_hw_new_cq(struct hw_eq *eq, u32 entry_count)
160 {
161         struct efct_hw *hw = eq->hw;
162         struct hw_cq *cq = kzalloc(sizeof(*cq), GFP_KERNEL);
163
164         if (!cq)
165                 return NULL;
166
167         cq->eq = eq;
168         cq->type = SLI4_QTYPE_CQ;
169         cq->instance = eq->hw->cq_count++;
170         cq->entry_count = entry_count;
171         cq->queue = &hw->cq[cq->instance];
172
173         INIT_LIST_HEAD(&cq->q_list);
174
175         if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_CQ, cq->queue,
176                             cq->entry_count, eq->queue)) {
177                 efc_log_err(hw->os, "CQ[%d] allocation failure len=%d\n",
178                             eq->instance, eq->entry_count);
179                 kfree(cq);
180                 return NULL;
181         }
182
183         hw->hw_cq[cq->instance] = cq;
184         INIT_LIST_HEAD(&cq->list_entry);
185         list_add_tail(&cq->list_entry, &eq->cq_list);
186         efc_log_debug(hw->os, "create cq[%2d] id %3d len %4d\n", cq->instance,
187                       cq->queue->id, cq->entry_count);
188         return cq;
189 }
190
191 u32
192 efct_hw_new_cq_set(struct hw_eq *eqs[], struct hw_cq *cqs[],
193                    u32 num_cqs, u32 entry_count)
194 {
195         u32 i;
196         struct efct_hw *hw = eqs[0]->hw;
197         struct sli4 *sli4 = &hw->sli;
198         struct hw_cq *cq = NULL;
199         struct sli4_queue *qs[SLI4_MAX_CQ_SET_COUNT];
200         struct sli4_queue *assefct[SLI4_MAX_CQ_SET_COUNT];
201
202         /* Initialise CQS pointers to NULL */
203         for (i = 0; i < num_cqs; i++)
204                 cqs[i] = NULL;
205
206         for (i = 0; i < num_cqs; i++) {
207                 cq = kzalloc(sizeof(*cq), GFP_KERNEL);
208                 if (!cq)
209                         goto error;
210
211                 cqs[i]          = cq;
212                 cq->eq          = eqs[i];
213                 cq->type        = SLI4_QTYPE_CQ;
214                 cq->instance    = hw->cq_count++;
215                 cq->entry_count = entry_count;
216                 cq->queue       = &hw->cq[cq->instance];
217                 qs[i]           = cq->queue;
218                 assefct[i]       = eqs[i]->queue;
219                 INIT_LIST_HEAD(&cq->q_list);
220         }
221
222         if (sli_cq_alloc_set(sli4, qs, num_cqs, entry_count, assefct)) {
223                 efc_log_err(hw->os, "Failed to create CQ Set.\n");
224                 goto error;
225         }
226
227         for (i = 0; i < num_cqs; i++) {
228                 hw->hw_cq[cqs[i]->instance] = cqs[i];
229                 INIT_LIST_HEAD(&cqs[i]->list_entry);
230                 list_add_tail(&cqs[i]->list_entry, &cqs[i]->eq->cq_list);
231         }
232
233         return 0;
234
235 error:
236         for (i = 0; i < num_cqs; i++) {
237                 kfree(cqs[i]);
238                 cqs[i] = NULL;
239         }
240         return -EIO;
241 }
242
243 struct hw_mq *
244 efct_hw_new_mq(struct hw_cq *cq, u32 entry_count)
245 {
246         struct efct_hw *hw = cq->eq->hw;
247         struct hw_mq *mq = kzalloc(sizeof(*mq), GFP_KERNEL);
248
249         if (!mq)
250                 return NULL;
251
252         mq->cq = cq;
253         mq->type = SLI4_QTYPE_MQ;
254         mq->instance = cq->eq->hw->mq_count++;
255         mq->entry_count = entry_count;
256         mq->entry_size = EFCT_HW_MQ_DEPTH;
257         mq->queue = &hw->mq[mq->instance];
258
259         if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_MQ, mq->queue, mq->entry_size,
260                             cq->queue)) {
261                 efc_log_err(hw->os, "MQ allocation failure\n");
262                 kfree(mq);
263                 return NULL;
264         }
265
266         hw->hw_mq[mq->instance] = mq;
267         INIT_LIST_HEAD(&mq->list_entry);
268         list_add_tail(&mq->list_entry, &cq->q_list);
269         efc_log_debug(hw->os, "create mq[%2d] id %3d len %4d\n", mq->instance,
270                       mq->queue->id, mq->entry_count);
271         return mq;
272 }
273
274 struct hw_wq *
275 efct_hw_new_wq(struct hw_cq *cq, u32 entry_count)
276 {
277         struct efct_hw *hw = cq->eq->hw;
278         struct hw_wq *wq = kzalloc(sizeof(*wq), GFP_KERNEL);
279
280         if (!wq)
281                 return NULL;
282
283         wq->hw = cq->eq->hw;
284         wq->cq = cq;
285         wq->type = SLI4_QTYPE_WQ;
286         wq->instance = cq->eq->hw->wq_count++;
287         wq->entry_count = entry_count;
288         wq->queue = &hw->wq[wq->instance];
289         wq->wqec_set_count = EFCT_HW_WQEC_SET_COUNT;
290         wq->wqec_count = wq->wqec_set_count;
291         wq->free_count = wq->entry_count - 1;
292         INIT_LIST_HEAD(&wq->pending_list);
293
294         if (sli_queue_alloc(&hw->sli, SLI4_QTYPE_WQ, wq->queue,
295                             wq->entry_count, cq->queue)) {
296                 efc_log_err(hw->os, "WQ allocation failure\n");
297                 kfree(wq);
298                 return NULL;
299         }
300
301         hw->hw_wq[wq->instance] = wq;
302         INIT_LIST_HEAD(&wq->list_entry);
303         list_add_tail(&wq->list_entry, &cq->q_list);
304         efc_log_debug(hw->os, "create wq[%2d] id %3d len %4d cls %d\n",
305                       wq->instance, wq->queue->id, wq->entry_count, wq->class);
306         return wq;
307 }
308
309 u32
310 efct_hw_new_rq_set(struct hw_cq *cqs[], struct hw_rq *rqs[],
311                    u32 num_rq_pairs, u32 entry_count)
312 {
313         struct efct_hw *hw = cqs[0]->eq->hw;
314         struct hw_rq *rq = NULL;
315         struct sli4_queue *qs[SLI4_MAX_RQ_SET_COUNT * 2] = { NULL };
316         u32 i, q_count, size;
317
318         /* Initialise RQS pointers */
319         for (i = 0; i < num_rq_pairs; i++)
320                 rqs[i] = NULL;
321
322         /*
323          * Allocate an RQ object SET, where each element in set
324          * encapsulates 2 SLI queues (for rq pair)
325          */
326         for (i = 0, q_count = 0; i < num_rq_pairs; i++, q_count += 2) {
327                 rq = kzalloc(sizeof(*rq), GFP_KERNEL);
328                 if (!rq)
329                         goto error;
330
331                 rqs[i] = rq;
332                 rq->instance = hw->hw_rq_count++;
333                 rq->cq = cqs[i];
334                 rq->type = SLI4_QTYPE_RQ;
335                 rq->entry_count = entry_count;
336
337                 /* Header RQ */
338                 rq->hdr = &hw->rq[hw->rq_count];
339                 rq->hdr_entry_size = EFCT_HW_RQ_HEADER_SIZE;
340                 hw->hw_rq_lookup[hw->rq_count] = rq->instance;
341                 hw->rq_count++;
342                 qs[q_count] = rq->hdr;
343
344                 /* Data RQ */
345                 rq->data = &hw->rq[hw->rq_count];
346                 rq->data_entry_size = hw->config.rq_default_buffer_size;
347                 hw->hw_rq_lookup[hw->rq_count] = rq->instance;
348                 hw->rq_count++;
349                 qs[q_count + 1] = rq->data;
350
351                 rq->rq_tracker = NULL;
352         }
353
354         if (sli_fc_rq_set_alloc(&hw->sli, num_rq_pairs, qs,
355                                 cqs[0]->queue->id,
356                             rqs[0]->entry_count,
357                             rqs[0]->hdr_entry_size,
358                             rqs[0]->data_entry_size)) {
359                 efc_log_err(hw->os, "RQ Set alloc failure for base CQ=%d\n",
360                             cqs[0]->queue->id);
361                 goto error;
362         }
363
364         for (i = 0; i < num_rq_pairs; i++) {
365                 hw->hw_rq[rqs[i]->instance] = rqs[i];
366                 INIT_LIST_HEAD(&rqs[i]->list_entry);
367                 list_add_tail(&rqs[i]->list_entry, &cqs[i]->q_list);
368                 size = sizeof(struct efc_hw_sequence *) * rqs[i]->entry_count;
369                 rqs[i]->rq_tracker = kzalloc(size, GFP_KERNEL);
370                 if (!rqs[i]->rq_tracker)
371                         goto error;
372         }
373
374         return 0;
375
376 error:
377         for (i = 0; i < num_rq_pairs; i++) {
378                 if (rqs[i]) {
379                         kfree(rqs[i]->rq_tracker);
380                         kfree(rqs[i]);
381                 }
382         }
383
384         return -EIO;
385 }
386
387 void
388 efct_hw_del_eq(struct hw_eq *eq)
389 {
390         struct hw_cq *cq;
391         struct hw_cq *cq_next;
392
393         if (!eq)
394                 return;
395
396         list_for_each_entry_safe(cq, cq_next, &eq->cq_list, list_entry)
397                 efct_hw_del_cq(cq);
398         list_del(&eq->list_entry);
399         eq->hw->hw_eq[eq->instance] = NULL;
400         kfree(eq);
401 }
402
403 void
404 efct_hw_del_cq(struct hw_cq *cq)
405 {
406         struct hw_q *q;
407         struct hw_q *q_next;
408
409         if (!cq)
410                 return;
411
412         list_for_each_entry_safe(q, q_next, &cq->q_list, list_entry) {
413                 switch (q->type) {
414                 case SLI4_QTYPE_MQ:
415                         efct_hw_del_mq((struct hw_mq *)q);
416                         break;
417                 case SLI4_QTYPE_WQ:
418                         efct_hw_del_wq((struct hw_wq *)q);
419                         break;
420                 case SLI4_QTYPE_RQ:
421                         efct_hw_del_rq((struct hw_rq *)q);
422                         break;
423                 default:
424                         break;
425                 }
426         }
427         list_del(&cq->list_entry);
428         cq->eq->hw->hw_cq[cq->instance] = NULL;
429         kfree(cq);
430 }
431
432 void
433 efct_hw_del_mq(struct hw_mq *mq)
434 {
435         if (!mq)
436                 return;
437
438         list_del(&mq->list_entry);
439         mq->cq->eq->hw->hw_mq[mq->instance] = NULL;
440         kfree(mq);
441 }
442
443 void
444 efct_hw_del_wq(struct hw_wq *wq)
445 {
446         if (!wq)
447                 return;
448
449         list_del(&wq->list_entry);
450         wq->cq->eq->hw->hw_wq[wq->instance] = NULL;
451         kfree(wq);
452 }
453
454 void
455 efct_hw_del_rq(struct hw_rq *rq)
456 {
457         struct efct_hw *hw = NULL;
458
459         if (!rq)
460                 return;
461         /* Free RQ tracker */
462         kfree(rq->rq_tracker);
463         rq->rq_tracker = NULL;
464         list_del(&rq->list_entry);
465         hw = rq->cq->eq->hw;
466         hw->hw_rq[rq->instance] = NULL;
467         kfree(rq);
468 }
469
470 void
471 efct_hw_queue_teardown(struct efct_hw *hw)
472 {
473         struct hw_eq *eq;
474         struct hw_eq *eq_next;
475
476         if (!hw->eq_list.next)
477                 return;
478
479         list_for_each_entry_safe(eq, eq_next, &hw->eq_list, list_entry)
480                 efct_hw_del_eq(eq);
481 }
482
483 static inline int
484 efct_hw_rqpair_find(struct efct_hw *hw, u16 rq_id)
485 {
486         return efct_hw_queue_hash_find(hw->rq_hash, rq_id);
487 }
488
489 static struct efc_hw_sequence *
490 efct_hw_rqpair_get(struct efct_hw *hw, u16 rqindex, u16 bufindex)
491 {
492         struct sli4_queue *rq_hdr = &hw->rq[rqindex];
493         struct efc_hw_sequence *seq = NULL;
494         struct hw_rq *rq = hw->hw_rq[hw->hw_rq_lookup[rqindex]];
495         unsigned long flags = 0;
496
497         if (bufindex >= rq_hdr->length) {
498                 efc_log_err(hw->os,
499                             "RQidx %d bufidx %d exceed ring len %d for id %d\n",
500                             rqindex, bufindex, rq_hdr->length, rq_hdr->id);
501                 return NULL;
502         }
503
504         /* rq_hdr lock also covers rqindex+1 queue */
505         spin_lock_irqsave(&rq_hdr->lock, flags);
506
507         seq = rq->rq_tracker[bufindex];
508         rq->rq_tracker[bufindex] = NULL;
509
510         if (!seq) {
511                 efc_log_err(hw->os,
512                             "RQbuf NULL, rqidx %d, bufidx %d, cur q idx = %d\n",
513                             rqindex, bufindex, rq_hdr->index);
514         }
515
516         spin_unlock_irqrestore(&rq_hdr->lock, flags);
517         return seq;
518 }
519
520 int
521 efct_hw_rqpair_process_rq(struct efct_hw *hw, struct hw_cq *cq,
522                           u8 *cqe)
523 {
524         u16 rq_id;
525         u32 index;
526         int rqindex;
527         int rq_status;
528         u32 h_len;
529         u32 p_len;
530         struct efc_hw_sequence *seq;
531         struct hw_rq *rq;
532
533         rq_status = sli_fc_rqe_rqid_and_index(&hw->sli, cqe,
534                                               &rq_id, &index);
535         if (rq_status != 0) {
536                 switch (rq_status) {
537                 case SLI4_FC_ASYNC_RQ_BUF_LEN_EXCEEDED:
538                 case SLI4_FC_ASYNC_RQ_DMA_FAILURE:
539                         /* just get RQ buffer then return to chip */
540                         rqindex = efct_hw_rqpair_find(hw, rq_id);
541                         if (rqindex < 0) {
542                                 efc_log_debug(hw->os,
543                                               "status=%#x: lookup fail id=%#x\n",
544                                              rq_status, rq_id);
545                                 break;
546                         }
547
548                         /* get RQ buffer */
549                         seq = efct_hw_rqpair_get(hw, rqindex, index);
550
551                         /* return to chip */
552                         if (efct_hw_rqpair_sequence_free(hw, seq)) {
553                                 efc_log_debug(hw->os,
554                                               "status=%#x,fail rtrn buf to RQ\n",
555                                              rq_status);
556                                 break;
557                         }
558                         break;
559                 case SLI4_FC_ASYNC_RQ_INSUFF_BUF_NEEDED:
560                 case SLI4_FC_ASYNC_RQ_INSUFF_BUF_FRM_DISC:
561                         /*
562                          * since RQ buffers were not consumed, cannot return
563                          * them to chip
564                          */
565                         efc_log_debug(hw->os, "Warning: RCQE status=%#x,\n",
566                                       rq_status);
567                         fallthrough;
568                 default:
569                         break;
570                 }
571                 return -EIO;
572         }
573
574         rqindex = efct_hw_rqpair_find(hw, rq_id);
575         if (rqindex < 0) {
576                 efc_log_debug(hw->os, "Error: rq_id lookup failed for id=%#x\n",
577                               rq_id);
578                 return -EIO;
579         }
580
581         rq = hw->hw_rq[hw->hw_rq_lookup[rqindex]];
582         rq->use_count++;
583
584         seq = efct_hw_rqpair_get(hw, rqindex, index);
585         if (WARN_ON(!seq))
586                 return -EIO;
587
588         seq->hw = hw;
589
590         sli_fc_rqe_length(&hw->sli, cqe, &h_len, &p_len);
591         seq->header->dma.len = h_len;
592         seq->payload->dma.len = p_len;
593         seq->fcfi = sli_fc_rqe_fcfi(&hw->sli, cqe);
594         seq->hw_priv = cq->eq;
595
596         efct_unsolicited_cb(hw->os, seq);
597
598         return 0;
599 }
600
601 static int
602 efct_hw_rqpair_put(struct efct_hw *hw, struct efc_hw_sequence *seq)
603 {
604         struct sli4_queue *rq_hdr = &hw->rq[seq->header->rqindex];
605         struct sli4_queue *rq_payload = &hw->rq[seq->payload->rqindex];
606         u32 hw_rq_index = hw->hw_rq_lookup[seq->header->rqindex];
607         struct hw_rq *rq = hw->hw_rq[hw_rq_index];
608         u32 phys_hdr[2];
609         u32 phys_payload[2];
610         int qindex_hdr;
611         int qindex_payload;
612         unsigned long flags = 0;
613
614         /* Update the RQ verification lookup tables */
615         phys_hdr[0] = upper_32_bits(seq->header->dma.phys);
616         phys_hdr[1] = lower_32_bits(seq->header->dma.phys);
617         phys_payload[0] = upper_32_bits(seq->payload->dma.phys);
618         phys_payload[1] = lower_32_bits(seq->payload->dma.phys);
619
620         /* rq_hdr lock also covers payload / header->rqindex+1 queue */
621         spin_lock_irqsave(&rq_hdr->lock, flags);
622
623         /*
624          * Note: The header must be posted last for buffer pair mode because
625          *       posting on the header queue posts the payload queue as well.
626          *       We do not ring the payload queue independently in RQ pair mode.
627          */
628         qindex_payload = sli_rq_write(&hw->sli, rq_payload,
629                                       (void *)phys_payload);
630         qindex_hdr = sli_rq_write(&hw->sli, rq_hdr, (void *)phys_hdr);
631         if (qindex_hdr < 0 ||
632             qindex_payload < 0) {
633                 efc_log_err(hw->os, "RQ_ID=%#x write failed\n", rq_hdr->id);
634                 spin_unlock_irqrestore(&rq_hdr->lock, flags);
635                 return -EIO;
636         }
637
638         /* ensure the indexes are the same */
639         WARN_ON(qindex_hdr != qindex_payload);
640
641         /* Update the lookup table */
642         if (!rq->rq_tracker[qindex_hdr]) {
643                 rq->rq_tracker[qindex_hdr] = seq;
644         } else {
645                 efc_log_debug(hw->os,
646                               "expected rq_tracker[%d][%d] buffer to be NULL\n",
647                               hw_rq_index, qindex_hdr);
648         }
649
650         spin_unlock_irqrestore(&rq_hdr->lock, flags);
651         return 0;
652 }
653
654 int
655 efct_hw_rqpair_sequence_free(struct efct_hw *hw, struct efc_hw_sequence *seq)
656 {
657         int rc = 0;
658
659         /*
660          * Post the data buffer first. Because in RQ pair mode, ringing the
661          * doorbell of the header ring will post the data buffer as well.
662          */
663         if (efct_hw_rqpair_put(hw, seq)) {
664                 efc_log_err(hw->os, "error writing buffers\n");
665                 return -EIO;
666         }
667
668         return rc;
669 }
670
671 int
672 efct_efc_hw_sequence_free(struct efc *efc, struct efc_hw_sequence *seq)
673 {
674         struct efct *efct = efc->base;
675
676         return efct_hw_rqpair_sequence_free(&efct->hw, seq);
677 }