bfdb7af19e496285bdeff3ada8a302b71e746865
[cascardo/linux.git] / drivers / net / ethernet / mellanox / mlx4 / resource_tracker.c
1 /*
2  * Copyright (c) 2004, 2005 Topspin Communications.  All rights reserved.
3  * Copyright (c) 2005, 2006, 2007, 2008 Mellanox Technologies.
4  * All rights reserved.
5  * Copyright (c) 2005, 2006, 2007 Cisco Systems, Inc.  All rights reserved.
6  *
7  * This software is available to you under a choice of one of two
8  * licenses.  You may choose to be licensed under the terms of the GNU
9  * General Public License (GPL) Version 2, available from the file
10  * COPYING in the main directory of this source tree, or the
11  * OpenIB.org BSD license below:
12  *
13  *     Redistribution and use in source and binary forms, with or
14  *     without modification, are permitted provided that the following
15  *     conditions are met:
16  *
17  *      - Redistributions of source code must retain the above
18  *        copyright notice, this list of conditions and the following
19  *        disclaimer.
20  *
21  *      - Redistributions in binary form must reproduce the above
22  *        copyright notice, this list of conditions and the following
23  *        disclaimer in the documentation and/or other materials
24  *        provided with the distribution.
25  *
26  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
27  * EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
28  * MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
29  * NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
30  * BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
31  * ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
32  * CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
33  * SOFTWARE.
34  */
35
36 #include <linux/sched.h>
37 #include <linux/pci.h>
38 #include <linux/errno.h>
39 #include <linux/kernel.h>
40 #include <linux/io.h>
41 #include <linux/slab.h>
42 #include <linux/mlx4/cmd.h>
43 #include <linux/mlx4/qp.h>
44
45 #include "mlx4.h"
46 #include "fw.h"
47
48 #define MLX4_MAC_VALID          (1ull << 63)
49 #define MLX4_MAC_MASK           0x7fffffffffffffffULL
50 #define ETH_ALEN                6
51
52 struct mac_res {
53         struct list_head list;
54         u64 mac;
55         u8 port;
56 };
57
58 struct res_common {
59         struct list_head        list;
60         u32                     res_id;
61         int                     owner;
62         int                     state;
63         int                     from_state;
64         int                     to_state;
65         int                     removing;
66 };
67
68 enum {
69         RES_ANY_BUSY = 1
70 };
71
72 struct res_gid {
73         struct list_head        list;
74         u8                      gid[16];
75         enum mlx4_protocol      prot;
76         enum mlx4_steer_type    steer;
77 };
78
79 enum res_qp_states {
80         RES_QP_BUSY = RES_ANY_BUSY,
81
82         /* QP number was allocated */
83         RES_QP_RESERVED,
84
85         /* ICM memory for QP context was mapped */
86         RES_QP_MAPPED,
87
88         /* QP is in hw ownership */
89         RES_QP_HW
90 };
91
92 static inline const char *qp_states_str(enum res_qp_states state)
93 {
94         switch (state) {
95         case RES_QP_BUSY: return "RES_QP_BUSY";
96         case RES_QP_RESERVED: return "RES_QP_RESERVED";
97         case RES_QP_MAPPED: return "RES_QP_MAPPED";
98         case RES_QP_HW: return "RES_QP_HW";
99         default: return "Unknown";
100         }
101 }
102
103 struct res_qp {
104         struct res_common       com;
105         struct res_mtt         *mtt;
106         struct res_cq          *rcq;
107         struct res_cq          *scq;
108         struct res_srq         *srq;
109         struct list_head        mcg_list;
110         spinlock_t              mcg_spl;
111         int                     local_qpn;
112 };
113
114 enum res_mtt_states {
115         RES_MTT_BUSY = RES_ANY_BUSY,
116         RES_MTT_ALLOCATED,
117 };
118
119 static inline const char *mtt_states_str(enum res_mtt_states state)
120 {
121         switch (state) {
122         case RES_MTT_BUSY: return "RES_MTT_BUSY";
123         case RES_MTT_ALLOCATED: return "RES_MTT_ALLOCATED";
124         default: return "Unknown";
125         }
126 }
127
128 struct res_mtt {
129         struct res_common       com;
130         int                     order;
131         atomic_t                ref_count;
132 };
133
134 enum res_mpt_states {
135         RES_MPT_BUSY = RES_ANY_BUSY,
136         RES_MPT_RESERVED,
137         RES_MPT_MAPPED,
138         RES_MPT_HW,
139 };
140
141 struct res_mpt {
142         struct res_common       com;
143         struct res_mtt         *mtt;
144         int                     key;
145 };
146
147 enum res_eq_states {
148         RES_EQ_BUSY = RES_ANY_BUSY,
149         RES_EQ_RESERVED,
150         RES_EQ_HW,
151 };
152
153 struct res_eq {
154         struct res_common       com;
155         struct res_mtt         *mtt;
156 };
157
158 enum res_cq_states {
159         RES_CQ_BUSY = RES_ANY_BUSY,
160         RES_CQ_ALLOCATED,
161         RES_CQ_HW,
162 };
163
164 struct res_cq {
165         struct res_common       com;
166         struct res_mtt         *mtt;
167         atomic_t                ref_count;
168 };
169
170 enum res_srq_states {
171         RES_SRQ_BUSY = RES_ANY_BUSY,
172         RES_SRQ_ALLOCATED,
173         RES_SRQ_HW,
174 };
175
176 static inline const char *srq_states_str(enum res_srq_states state)
177 {
178         switch (state) {
179         case RES_SRQ_BUSY: return "RES_SRQ_BUSY";
180         case RES_SRQ_ALLOCATED: return "RES_SRQ_ALLOCATED";
181         case RES_SRQ_HW: return "RES_SRQ_HW";
182         default: return "Unknown";
183         }
184 }
185
186 struct res_srq {
187         struct res_common       com;
188         struct res_mtt         *mtt;
189         struct res_cq          *cq;
190         atomic_t                ref_count;
191 };
192
193 enum res_counter_states {
194         RES_COUNTER_BUSY = RES_ANY_BUSY,
195         RES_COUNTER_ALLOCATED,
196 };
197
198 static inline const char *counter_states_str(enum res_counter_states state)
199 {
200         switch (state) {
201         case RES_COUNTER_BUSY: return "RES_COUNTER_BUSY";
202         case RES_COUNTER_ALLOCATED: return "RES_COUNTER_ALLOCATED";
203         default: return "Unknown";
204         }
205 }
206
207 struct res_counter {
208         struct res_common       com;
209         int                     port;
210 };
211
212 /* For Debug uses */
213 static const char *ResourceType(enum mlx4_resource rt)
214 {
215         switch (rt) {
216         case RES_QP: return "RES_QP";
217         case RES_CQ: return "RES_CQ";
218         case RES_SRQ: return "RES_SRQ";
219         case RES_MPT: return "RES_MPT";
220         case RES_MTT: return "RES_MTT";
221         case RES_MAC: return  "RES_MAC";
222         case RES_EQ: return "RES_EQ";
223         case RES_COUNTER: return "RES_COUNTER";
224         default: return "Unknown resource type !!!";
225         };
226 }
227
228 int mlx4_init_resource_tracker(struct mlx4_dev *dev)
229 {
230         struct mlx4_priv *priv = mlx4_priv(dev);
231         int i;
232         int t;
233
234         priv->mfunc.master.res_tracker.slave_list =
235                 kzalloc(dev->num_slaves * sizeof(struct slave_list),
236                         GFP_KERNEL);
237         if (!priv->mfunc.master.res_tracker.slave_list)
238                 return -ENOMEM;
239
240         for (i = 0 ; i < dev->num_slaves; i++) {
241                 for (t = 0; t < MLX4_NUM_OF_RESOURCE_TYPE; ++t)
242                         INIT_LIST_HEAD(&priv->mfunc.master.res_tracker.
243                                        slave_list[i].res_list[t]);
244                 mutex_init(&priv->mfunc.master.res_tracker.slave_list[i].mutex);
245         }
246
247         mlx4_dbg(dev, "Started init_resource_tracker: %ld slaves\n",
248                  dev->num_slaves);
249         for (i = 0 ; i < MLX4_NUM_OF_RESOURCE_TYPE; i++)
250                 INIT_RADIX_TREE(&priv->mfunc.master.res_tracker.res_tree[i],
251                                 GFP_ATOMIC|__GFP_NOWARN);
252
253         spin_lock_init(&priv->mfunc.master.res_tracker.lock);
254         return 0 ;
255 }
256
257 void mlx4_free_resource_tracker(struct mlx4_dev *dev)
258 {
259         struct mlx4_priv *priv = mlx4_priv(dev);
260         int i;
261
262         if (priv->mfunc.master.res_tracker.slave_list) {
263                 for (i = 0 ; i < dev->num_slaves; i++)
264                         mlx4_delete_all_resources_for_slave(dev, i);
265
266                 kfree(priv->mfunc.master.res_tracker.slave_list);
267         }
268 }
269
270 static void update_ud_gid(struct mlx4_dev *dev,
271                           struct mlx4_qp_context *qp_ctx, u8 slave)
272 {
273         u32 ts = (be32_to_cpu(qp_ctx->flags) >> 16) & 0xff;
274
275         if (MLX4_QP_ST_UD == ts)
276                 qp_ctx->pri_path.mgid_index = 0x80 | slave;
277
278         mlx4_dbg(dev, "slave %d, new gid index: 0x%x ",
279                 slave, qp_ctx->pri_path.mgid_index);
280 }
281
282 static int mpt_mask(struct mlx4_dev *dev)
283 {
284         return dev->caps.num_mpts - 1;
285 }
286
287 static void *find_res(struct mlx4_dev *dev, int res_id,
288                       enum mlx4_resource type)
289 {
290         struct mlx4_priv *priv = mlx4_priv(dev);
291
292         return radix_tree_lookup(&priv->mfunc.master.res_tracker.res_tree[type],
293                                  res_id);
294 }
295
296 static int get_res(struct mlx4_dev *dev, int slave, int res_id,
297                    enum mlx4_resource type,
298                    void *res)
299 {
300         struct res_common *r;
301         int err = 0;
302
303         spin_lock_irq(mlx4_tlock(dev));
304         r = find_res(dev, res_id, type);
305         if (!r) {
306                 err = -ENONET;
307                 goto exit;
308         }
309
310         if (r->state == RES_ANY_BUSY) {
311                 err = -EBUSY;
312                 goto exit;
313         }
314
315         if (r->owner != slave) {
316                 err = -EPERM;
317                 goto exit;
318         }
319
320         r->from_state = r->state;
321         r->state = RES_ANY_BUSY;
322         mlx4_dbg(dev, "res %s id 0x%x to busy\n",
323                  ResourceType(type), r->res_id);
324
325         if (res)
326                 *((struct res_common **)res) = r;
327
328 exit:
329         spin_unlock_irq(mlx4_tlock(dev));
330         return err;
331 }
332
333 int mlx4_get_slave_from_resource_id(struct mlx4_dev *dev,
334                                     enum mlx4_resource type,
335                                     int res_id, int *slave)
336 {
337
338         struct res_common *r;
339         int err = -ENOENT;
340         int id = res_id;
341
342         if (type == RES_QP)
343                 id &= 0x7fffff;
344         spin_lock(mlx4_tlock(dev));
345
346         r = find_res(dev, id, type);
347         if (r) {
348                 *slave = r->owner;
349                 err = 0;
350         }
351         spin_unlock(mlx4_tlock(dev));
352
353         return err;
354 }
355
356 static void put_res(struct mlx4_dev *dev, int slave, int res_id,
357                     enum mlx4_resource type)
358 {
359         struct res_common *r;
360
361         spin_lock_irq(mlx4_tlock(dev));
362         r = find_res(dev, res_id, type);
363         if (r)
364                 r->state = r->from_state;
365         spin_unlock_irq(mlx4_tlock(dev));
366 }
367
368 static struct res_common *alloc_qp_tr(int id)
369 {
370         struct res_qp *ret;
371
372         ret = kzalloc(sizeof *ret, GFP_KERNEL);
373         if (!ret)
374                 return NULL;
375
376         ret->com.res_id = id;
377         ret->com.state = RES_QP_RESERVED;
378         ret->local_qpn = id;
379         INIT_LIST_HEAD(&ret->mcg_list);
380         spin_lock_init(&ret->mcg_spl);
381
382         return &ret->com;
383 }
384
385 static struct res_common *alloc_mtt_tr(int id, int order)
386 {
387         struct res_mtt *ret;
388
389         ret = kzalloc(sizeof *ret, GFP_KERNEL);
390         if (!ret)
391                 return NULL;
392
393         ret->com.res_id = id;
394         ret->order = order;
395         ret->com.state = RES_MTT_ALLOCATED;
396         atomic_set(&ret->ref_count, 0);
397
398         return &ret->com;
399 }
400
401 static struct res_common *alloc_mpt_tr(int id, int key)
402 {
403         struct res_mpt *ret;
404
405         ret = kzalloc(sizeof *ret, GFP_KERNEL);
406         if (!ret)
407                 return NULL;
408
409         ret->com.res_id = id;
410         ret->com.state = RES_MPT_RESERVED;
411         ret->key = key;
412
413         return &ret->com;
414 }
415
416 static struct res_common *alloc_eq_tr(int id)
417 {
418         struct res_eq *ret;
419
420         ret = kzalloc(sizeof *ret, GFP_KERNEL);
421         if (!ret)
422                 return NULL;
423
424         ret->com.res_id = id;
425         ret->com.state = RES_EQ_RESERVED;
426
427         return &ret->com;
428 }
429
430 static struct res_common *alloc_cq_tr(int id)
431 {
432         struct res_cq *ret;
433
434         ret = kzalloc(sizeof *ret, GFP_KERNEL);
435         if (!ret)
436                 return NULL;
437
438         ret->com.res_id = id;
439         ret->com.state = RES_CQ_ALLOCATED;
440         atomic_set(&ret->ref_count, 0);
441
442         return &ret->com;
443 }
444
445 static struct res_common *alloc_srq_tr(int id)
446 {
447         struct res_srq *ret;
448
449         ret = kzalloc(sizeof *ret, GFP_KERNEL);
450         if (!ret)
451                 return NULL;
452
453         ret->com.res_id = id;
454         ret->com.state = RES_SRQ_ALLOCATED;
455         atomic_set(&ret->ref_count, 0);
456
457         return &ret->com;
458 }
459
460 static struct res_common *alloc_counter_tr(int id)
461 {
462         struct res_counter *ret;
463
464         ret = kzalloc(sizeof *ret, GFP_KERNEL);
465         if (!ret)
466                 return NULL;
467
468         ret->com.res_id = id;
469         ret->com.state = RES_COUNTER_ALLOCATED;
470
471         return &ret->com;
472 }
473
474 static struct res_common *alloc_tr(int id, enum mlx4_resource type, int slave,
475                                    int extra)
476 {
477         struct res_common *ret;
478
479         switch (type) {
480         case RES_QP:
481                 ret = alloc_qp_tr(id);
482                 break;
483         case RES_MPT:
484                 ret = alloc_mpt_tr(id, extra);
485                 break;
486         case RES_MTT:
487                 ret = alloc_mtt_tr(id, extra);
488                 break;
489         case RES_EQ:
490                 ret = alloc_eq_tr(id);
491                 break;
492         case RES_CQ:
493                 ret = alloc_cq_tr(id);
494                 break;
495         case RES_SRQ:
496                 ret = alloc_srq_tr(id);
497                 break;
498         case RES_MAC:
499                 printk(KERN_ERR "implementation missing\n");
500                 return NULL;
501         case RES_COUNTER:
502                 ret = alloc_counter_tr(id);
503                 break;
504
505         default:
506                 return NULL;
507         }
508         if (ret)
509                 ret->owner = slave;
510
511         return ret;
512 }
513
514 static int add_res_range(struct mlx4_dev *dev, int slave, int base, int count,
515                          enum mlx4_resource type, int extra)
516 {
517         int i;
518         int err;
519         struct mlx4_priv *priv = mlx4_priv(dev);
520         struct res_common **res_arr;
521         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
522         struct radix_tree_root *root = &tracker->res_tree[type];
523
524         res_arr = kzalloc(count * sizeof *res_arr, GFP_KERNEL);
525         if (!res_arr)
526                 return -ENOMEM;
527
528         for (i = 0; i < count; ++i) {
529                 res_arr[i] = alloc_tr(base + i, type, slave, extra);
530                 if (!res_arr[i]) {
531                         for (--i; i >= 0; --i)
532                                 kfree(res_arr[i]);
533
534                         kfree(res_arr);
535                         return -ENOMEM;
536                 }
537         }
538
539         spin_lock_irq(mlx4_tlock(dev));
540         for (i = 0; i < count; ++i) {
541                 if (find_res(dev, base + i, type)) {
542                         err = -EEXIST;
543                         goto undo;
544                 }
545                 err = radix_tree_insert(root, base + i, res_arr[i]);
546                 if (err)
547                         goto undo;
548                 list_add_tail(&res_arr[i]->list,
549                               &tracker->slave_list[slave].res_list[type]);
550         }
551         spin_unlock_irq(mlx4_tlock(dev));
552         kfree(res_arr);
553
554         return 0;
555
556 undo:
557         for (--i; i >= base; --i)
558                 radix_tree_delete(&tracker->res_tree[type], i);
559
560         spin_unlock_irq(mlx4_tlock(dev));
561
562         for (i = 0; i < count; ++i)
563                 kfree(res_arr[i]);
564
565         kfree(res_arr);
566
567         return err;
568 }
569
570 static int remove_qp_ok(struct res_qp *res)
571 {
572         if (res->com.state == RES_QP_BUSY)
573                 return -EBUSY;
574         else if (res->com.state != RES_QP_RESERVED)
575                 return -EPERM;
576
577         return 0;
578 }
579
580 static int remove_mtt_ok(struct res_mtt *res, int order)
581 {
582         if (res->com.state == RES_MTT_BUSY ||
583             atomic_read(&res->ref_count)) {
584                 printk(KERN_DEBUG "%s-%d: state %s, ref_count %d\n",
585                        __func__, __LINE__,
586                        mtt_states_str(res->com.state),
587                        atomic_read(&res->ref_count));
588                 return -EBUSY;
589         } else if (res->com.state != RES_MTT_ALLOCATED)
590                 return -EPERM;
591         else if (res->order != order)
592                 return -EINVAL;
593
594         return 0;
595 }
596
597 static int remove_mpt_ok(struct res_mpt *res)
598 {
599         if (res->com.state == RES_MPT_BUSY)
600                 return -EBUSY;
601         else if (res->com.state != RES_MPT_RESERVED)
602                 return -EPERM;
603
604         return 0;
605 }
606
607 static int remove_eq_ok(struct res_eq *res)
608 {
609         if (res->com.state == RES_MPT_BUSY)
610                 return -EBUSY;
611         else if (res->com.state != RES_MPT_RESERVED)
612                 return -EPERM;
613
614         return 0;
615 }
616
617 static int remove_counter_ok(struct res_counter *res)
618 {
619         if (res->com.state == RES_COUNTER_BUSY)
620                 return -EBUSY;
621         else if (res->com.state != RES_COUNTER_ALLOCATED)
622                 return -EPERM;
623
624         return 0;
625 }
626
627 static int remove_cq_ok(struct res_cq *res)
628 {
629         if (res->com.state == RES_CQ_BUSY)
630                 return -EBUSY;
631         else if (res->com.state != RES_CQ_ALLOCATED)
632                 return -EPERM;
633
634         return 0;
635 }
636
637 static int remove_srq_ok(struct res_srq *res)
638 {
639         if (res->com.state == RES_SRQ_BUSY)
640                 return -EBUSY;
641         else if (res->com.state != RES_SRQ_ALLOCATED)
642                 return -EPERM;
643
644         return 0;
645 }
646
647 static int remove_ok(struct res_common *res, enum mlx4_resource type, int extra)
648 {
649         switch (type) {
650         case RES_QP:
651                 return remove_qp_ok((struct res_qp *)res);
652         case RES_CQ:
653                 return remove_cq_ok((struct res_cq *)res);
654         case RES_SRQ:
655                 return remove_srq_ok((struct res_srq *)res);
656         case RES_MPT:
657                 return remove_mpt_ok((struct res_mpt *)res);
658         case RES_MTT:
659                 return remove_mtt_ok((struct res_mtt *)res, extra);
660         case RES_MAC:
661                 return -ENOSYS;
662         case RES_EQ:
663                 return remove_eq_ok((struct res_eq *)res);
664         case RES_COUNTER:
665                 return remove_counter_ok((struct res_counter *)res);
666         default:
667                 return -EINVAL;
668         }
669 }
670
671 static int rem_res_range(struct mlx4_dev *dev, int slave, int base, int count,
672                          enum mlx4_resource type, int extra)
673 {
674         int i;
675         int err;
676         struct mlx4_priv *priv = mlx4_priv(dev);
677         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
678         struct res_common *r;
679
680         spin_lock_irq(mlx4_tlock(dev));
681         for (i = base; i < base + count; ++i) {
682                 r = radix_tree_lookup(&tracker->res_tree[type], i);
683                 if (!r) {
684                         err = -ENOENT;
685                         goto out;
686                 }
687                 if (r->owner != slave) {
688                         err = -EPERM;
689                         goto out;
690                 }
691                 err = remove_ok(r, type, extra);
692                 if (err)
693                         goto out;
694         }
695
696         for (i = base; i < base + count; ++i) {
697                 r = radix_tree_lookup(&tracker->res_tree[type], i);
698                 radix_tree_delete(&tracker->res_tree[type], i);
699                 list_del(&r->list);
700                 kfree(r);
701         }
702         err = 0;
703
704 out:
705         spin_unlock_irq(mlx4_tlock(dev));
706
707         return err;
708 }
709
710 static int qp_res_start_move_to(struct mlx4_dev *dev, int slave, int qpn,
711                                 enum res_qp_states state, struct res_qp **qp,
712                                 int alloc)
713 {
714         struct mlx4_priv *priv = mlx4_priv(dev);
715         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
716         struct res_qp *r;
717         int err = 0;
718
719         spin_lock_irq(mlx4_tlock(dev));
720         r = radix_tree_lookup(&tracker->res_tree[RES_QP], qpn);
721         if (!r)
722                 err = -ENOENT;
723         else if (r->com.owner != slave)
724                 err = -EPERM;
725         else {
726                 switch (state) {
727                 case RES_QP_BUSY:
728                         mlx4_dbg(dev, "%s: failed RES_QP, 0x%x\n",
729                                  __func__, r->com.res_id);
730                         err = -EBUSY;
731                         break;
732
733                 case RES_QP_RESERVED:
734                         if (r->com.state == RES_QP_MAPPED && !alloc)
735                                 break;
736
737                         mlx4_dbg(dev, "failed RES_QP, 0x%x\n", r->com.res_id);
738                         err = -EINVAL;
739                         break;
740
741                 case RES_QP_MAPPED:
742                         if ((r->com.state == RES_QP_RESERVED && alloc) ||
743                             r->com.state == RES_QP_HW)
744                                 break;
745                         else {
746                                 mlx4_dbg(dev, "failed RES_QP, 0x%x\n",
747                                           r->com.res_id);
748                                 err = -EINVAL;
749                         }
750
751                         break;
752
753                 case RES_QP_HW:
754                         if (r->com.state != RES_QP_MAPPED)
755                                 err = -EINVAL;
756                         break;
757                 default:
758                         err = -EINVAL;
759                 }
760
761                 if (!err) {
762                         r->com.from_state = r->com.state;
763                         r->com.to_state = state;
764                         r->com.state = RES_QP_BUSY;
765                         if (qp)
766                                 *qp = (struct res_qp *)r;
767                 }
768         }
769
770         spin_unlock_irq(mlx4_tlock(dev));
771
772         return err;
773 }
774
775 static int mr_res_start_move_to(struct mlx4_dev *dev, int slave, int index,
776                                 enum res_mpt_states state, struct res_mpt **mpt)
777 {
778         struct mlx4_priv *priv = mlx4_priv(dev);
779         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
780         struct res_mpt *r;
781         int err = 0;
782
783         spin_lock_irq(mlx4_tlock(dev));
784         r = radix_tree_lookup(&tracker->res_tree[RES_MPT], index);
785         if (!r)
786                 err = -ENOENT;
787         else if (r->com.owner != slave)
788                 err = -EPERM;
789         else {
790                 switch (state) {
791                 case RES_MPT_BUSY:
792                         err = -EINVAL;
793                         break;
794
795                 case RES_MPT_RESERVED:
796                         if (r->com.state != RES_MPT_MAPPED)
797                                 err = -EINVAL;
798                         break;
799
800                 case RES_MPT_MAPPED:
801                         if (r->com.state != RES_MPT_RESERVED &&
802                             r->com.state != RES_MPT_HW)
803                                 err = -EINVAL;
804                         break;
805
806                 case RES_MPT_HW:
807                         if (r->com.state != RES_MPT_MAPPED)
808                                 err = -EINVAL;
809                         break;
810                 default:
811                         err = -EINVAL;
812                 }
813
814                 if (!err) {
815                         r->com.from_state = r->com.state;
816                         r->com.to_state = state;
817                         r->com.state = RES_MPT_BUSY;
818                         if (mpt)
819                                 *mpt = (struct res_mpt *)r;
820                 }
821         }
822
823         spin_unlock_irq(mlx4_tlock(dev));
824
825         return err;
826 }
827
828 static int eq_res_start_move_to(struct mlx4_dev *dev, int slave, int index,
829                                 enum res_eq_states state, struct res_eq **eq)
830 {
831         struct mlx4_priv *priv = mlx4_priv(dev);
832         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
833         struct res_eq *r;
834         int err = 0;
835
836         spin_lock_irq(mlx4_tlock(dev));
837         r = radix_tree_lookup(&tracker->res_tree[RES_EQ], index);
838         if (!r)
839                 err = -ENOENT;
840         else if (r->com.owner != slave)
841                 err = -EPERM;
842         else {
843                 switch (state) {
844                 case RES_EQ_BUSY:
845                         err = -EINVAL;
846                         break;
847
848                 case RES_EQ_RESERVED:
849                         if (r->com.state != RES_EQ_HW)
850                                 err = -EINVAL;
851                         break;
852
853                 case RES_EQ_HW:
854                         if (r->com.state != RES_EQ_RESERVED)
855                                 err = -EINVAL;
856                         break;
857
858                 default:
859                         err = -EINVAL;
860                 }
861
862                 if (!err) {
863                         r->com.from_state = r->com.state;
864                         r->com.to_state = state;
865                         r->com.state = RES_EQ_BUSY;
866                         if (eq)
867                                 *eq = r;
868                 }
869         }
870
871         spin_unlock_irq(mlx4_tlock(dev));
872
873         return err;
874 }
875
876 static int cq_res_start_move_to(struct mlx4_dev *dev, int slave, int cqn,
877                                 enum res_cq_states state, struct res_cq **cq)
878 {
879         struct mlx4_priv *priv = mlx4_priv(dev);
880         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
881         struct res_cq *r;
882         int err;
883
884         spin_lock_irq(mlx4_tlock(dev));
885         r = radix_tree_lookup(&tracker->res_tree[RES_CQ], cqn);
886         if (!r)
887                 err = -ENOENT;
888         else if (r->com.owner != slave)
889                 err = -EPERM;
890         else {
891                 switch (state) {
892                 case RES_CQ_BUSY:
893                         err = -EBUSY;
894                         break;
895
896                 case RES_CQ_ALLOCATED:
897                         if (r->com.state != RES_CQ_HW)
898                                 err = -EINVAL;
899                         else if (atomic_read(&r->ref_count))
900                                 err = -EBUSY;
901                         else
902                                 err = 0;
903                         break;
904
905                 case RES_CQ_HW:
906                         if (r->com.state != RES_CQ_ALLOCATED)
907                                 err = -EINVAL;
908                         else
909                                 err = 0;
910                         break;
911
912                 default:
913                         err = -EINVAL;
914                 }
915
916                 if (!err) {
917                         r->com.from_state = r->com.state;
918                         r->com.to_state = state;
919                         r->com.state = RES_CQ_BUSY;
920                         if (cq)
921                                 *cq = r;
922                 }
923         }
924
925         spin_unlock_irq(mlx4_tlock(dev));
926
927         return err;
928 }
929
930 static int srq_res_start_move_to(struct mlx4_dev *dev, int slave, int index,
931                                  enum res_cq_states state, struct res_srq **srq)
932 {
933         struct mlx4_priv *priv = mlx4_priv(dev);
934         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
935         struct res_srq *r;
936         int err = 0;
937
938         spin_lock_irq(mlx4_tlock(dev));
939         r = radix_tree_lookup(&tracker->res_tree[RES_SRQ], index);
940         if (!r)
941                 err = -ENOENT;
942         else if (r->com.owner != slave)
943                 err = -EPERM;
944         else {
945                 switch (state) {
946                 case RES_SRQ_BUSY:
947                         err = -EINVAL;
948                         break;
949
950                 case RES_SRQ_ALLOCATED:
951                         if (r->com.state != RES_SRQ_HW)
952                                 err = -EINVAL;
953                         else if (atomic_read(&r->ref_count))
954                                 err = -EBUSY;
955                         break;
956
957                 case RES_SRQ_HW:
958                         if (r->com.state != RES_SRQ_ALLOCATED)
959                                 err = -EINVAL;
960                         break;
961
962                 default:
963                         err = -EINVAL;
964                 }
965
966                 if (!err) {
967                         r->com.from_state = r->com.state;
968                         r->com.to_state = state;
969                         r->com.state = RES_SRQ_BUSY;
970                         if (srq)
971                                 *srq = r;
972                 }
973         }
974
975         spin_unlock_irq(mlx4_tlock(dev));
976
977         return err;
978 }
979
980 static void res_abort_move(struct mlx4_dev *dev, int slave,
981                            enum mlx4_resource type, int id)
982 {
983         struct mlx4_priv *priv = mlx4_priv(dev);
984         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
985         struct res_common *r;
986
987         spin_lock_irq(mlx4_tlock(dev));
988         r = radix_tree_lookup(&tracker->res_tree[type], id);
989         if (r && (r->owner == slave))
990                 r->state = r->from_state;
991         spin_unlock_irq(mlx4_tlock(dev));
992 }
993
994 static void res_end_move(struct mlx4_dev *dev, int slave,
995                          enum mlx4_resource type, int id)
996 {
997         struct mlx4_priv *priv = mlx4_priv(dev);
998         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
999         struct res_common *r;
1000
1001         spin_lock_irq(mlx4_tlock(dev));
1002         r = radix_tree_lookup(&tracker->res_tree[type], id);
1003         if (r && (r->owner == slave))
1004                 r->state = r->to_state;
1005         spin_unlock_irq(mlx4_tlock(dev));
1006 }
1007
1008 static int valid_reserved(struct mlx4_dev *dev, int slave, int qpn)
1009 {
1010         return mlx4_is_qp_reserved(dev, qpn);
1011 }
1012
1013 static int qp_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1014                         u64 in_param, u64 *out_param)
1015 {
1016         int err;
1017         int count;
1018         int align;
1019         int base;
1020         int qpn;
1021
1022         switch (op) {
1023         case RES_OP_RESERVE:
1024                 count = get_param_l(&in_param);
1025                 align = get_param_h(&in_param);
1026                 err = __mlx4_qp_reserve_range(dev, count, align, &base);
1027                 if (err)
1028                         return err;
1029
1030                 err = add_res_range(dev, slave, base, count, RES_QP, 0);
1031                 if (err) {
1032                         __mlx4_qp_release_range(dev, base, count);
1033                         return err;
1034                 }
1035                 set_param_l(out_param, base);
1036                 break;
1037         case RES_OP_MAP_ICM:
1038                 qpn = get_param_l(&in_param) & 0x7fffff;
1039                 if (valid_reserved(dev, slave, qpn)) {
1040                         err = add_res_range(dev, slave, qpn, 1, RES_QP, 0);
1041                         if (err)
1042                                 return err;
1043                 }
1044
1045                 err = qp_res_start_move_to(dev, slave, qpn, RES_QP_MAPPED,
1046                                            NULL, 1);
1047                 if (err)
1048                         return err;
1049
1050                 if (!valid_reserved(dev, slave, qpn)) {
1051                         err = __mlx4_qp_alloc_icm(dev, qpn);
1052                         if (err) {
1053                                 res_abort_move(dev, slave, RES_QP, qpn);
1054                                 return err;
1055                         }
1056                 }
1057
1058                 res_end_move(dev, slave, RES_QP, qpn);
1059                 break;
1060
1061         default:
1062                 err = -EINVAL;
1063                 break;
1064         }
1065         return err;
1066 }
1067
1068 static int mtt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1069                          u64 in_param, u64 *out_param)
1070 {
1071         int err = -EINVAL;
1072         int base;
1073         int order;
1074
1075         if (op != RES_OP_RESERVE_AND_MAP)
1076                 return err;
1077
1078         order = get_param_l(&in_param);
1079         base = __mlx4_alloc_mtt_range(dev, order);
1080         if (base == -1)
1081                 return -ENOMEM;
1082
1083         err = add_res_range(dev, slave, base, 1, RES_MTT, order);
1084         if (err)
1085                 __mlx4_free_mtt_range(dev, base, order);
1086         else
1087                 set_param_l(out_param, base);
1088
1089         return err;
1090 }
1091
1092 static int mpt_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1093                          u64 in_param, u64 *out_param)
1094 {
1095         int err = -EINVAL;
1096         int index;
1097         int id;
1098         struct res_mpt *mpt;
1099
1100         switch (op) {
1101         case RES_OP_RESERVE:
1102                 index = __mlx4_mr_reserve(dev);
1103                 if (index == -1)
1104                         break;
1105                 id = index & mpt_mask(dev);
1106
1107                 err = add_res_range(dev, slave, id, 1, RES_MPT, index);
1108                 if (err) {
1109                         __mlx4_mr_release(dev, index);
1110                         break;
1111                 }
1112                 set_param_l(out_param, index);
1113                 break;
1114         case RES_OP_MAP_ICM:
1115                 index = get_param_l(&in_param);
1116                 id = index & mpt_mask(dev);
1117                 err = mr_res_start_move_to(dev, slave, id,
1118                                            RES_MPT_MAPPED, &mpt);
1119                 if (err)
1120                         return err;
1121
1122                 err = __mlx4_mr_alloc_icm(dev, mpt->key);
1123                 if (err) {
1124                         res_abort_move(dev, slave, RES_MPT, id);
1125                         return err;
1126                 }
1127
1128                 res_end_move(dev, slave, RES_MPT, id);
1129                 break;
1130         }
1131         return err;
1132 }
1133
1134 static int cq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1135                         u64 in_param, u64 *out_param)
1136 {
1137         int cqn;
1138         int err;
1139
1140         switch (op) {
1141         case RES_OP_RESERVE_AND_MAP:
1142                 err = __mlx4_cq_alloc_icm(dev, &cqn);
1143                 if (err)
1144                         break;
1145
1146                 err = add_res_range(dev, slave, cqn, 1, RES_CQ, 0);
1147                 if (err) {
1148                         __mlx4_cq_free_icm(dev, cqn);
1149                         break;
1150                 }
1151
1152                 set_param_l(out_param, cqn);
1153                 break;
1154
1155         default:
1156                 err = -EINVAL;
1157         }
1158
1159         return err;
1160 }
1161
1162 static int srq_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1163                          u64 in_param, u64 *out_param)
1164 {
1165         int srqn;
1166         int err;
1167
1168         switch (op) {
1169         case RES_OP_RESERVE_AND_MAP:
1170                 err = __mlx4_srq_alloc_icm(dev, &srqn);
1171                 if (err)
1172                         break;
1173
1174                 err = add_res_range(dev, slave, srqn, 1, RES_SRQ, 0);
1175                 if (err) {
1176                         __mlx4_srq_free_icm(dev, srqn);
1177                         break;
1178                 }
1179
1180                 set_param_l(out_param, srqn);
1181                 break;
1182
1183         default:
1184                 err = -EINVAL;
1185         }
1186
1187         return err;
1188 }
1189
1190 static int mac_add_to_slave(struct mlx4_dev *dev, int slave, u64 mac, int port)
1191 {
1192         struct mlx4_priv *priv = mlx4_priv(dev);
1193         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1194         struct mac_res *res;
1195
1196         res = kzalloc(sizeof *res, GFP_KERNEL);
1197         if (!res)
1198                 return -ENOMEM;
1199         res->mac = mac;
1200         res->port = (u8) port;
1201         list_add_tail(&res->list,
1202                       &tracker->slave_list[slave].res_list[RES_MAC]);
1203         return 0;
1204 }
1205
1206 static void mac_del_from_slave(struct mlx4_dev *dev, int slave, u64 mac,
1207                                int port)
1208 {
1209         struct mlx4_priv *priv = mlx4_priv(dev);
1210         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1211         struct list_head *mac_list =
1212                 &tracker->slave_list[slave].res_list[RES_MAC];
1213         struct mac_res *res, *tmp;
1214
1215         list_for_each_entry_safe(res, tmp, mac_list, list) {
1216                 if (res->mac == mac && res->port == (u8) port) {
1217                         list_del(&res->list);
1218                         kfree(res);
1219                         break;
1220                 }
1221         }
1222 }
1223
1224 static void rem_slave_macs(struct mlx4_dev *dev, int slave)
1225 {
1226         struct mlx4_priv *priv = mlx4_priv(dev);
1227         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1228         struct list_head *mac_list =
1229                 &tracker->slave_list[slave].res_list[RES_MAC];
1230         struct mac_res *res, *tmp;
1231
1232         list_for_each_entry_safe(res, tmp, mac_list, list) {
1233                 list_del(&res->list);
1234                 __mlx4_unregister_mac(dev, res->port, res->mac);
1235                 kfree(res);
1236         }
1237 }
1238
1239 static int mac_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1240                          u64 in_param, u64 *out_param)
1241 {
1242         int err = -EINVAL;
1243         int port;
1244         u64 mac;
1245
1246         if (op != RES_OP_RESERVE_AND_MAP)
1247                 return err;
1248
1249         port = get_param_l(out_param);
1250         mac = in_param;
1251
1252         err = __mlx4_register_mac(dev, port, mac);
1253         if (err >= 0) {
1254                 set_param_l(out_param, err);
1255                 err = 0;
1256         }
1257
1258         if (!err) {
1259                 err = mac_add_to_slave(dev, slave, mac, port);
1260                 if (err)
1261                         __mlx4_unregister_mac(dev, port, mac);
1262         }
1263         return err;
1264 }
1265
1266 static int vlan_alloc_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1267                          u64 in_param, u64 *out_param)
1268 {
1269         return 0;
1270 }
1271
1272 int mlx4_ALLOC_RES_wrapper(struct mlx4_dev *dev, int slave,
1273                            struct mlx4_vhcr *vhcr,
1274                            struct mlx4_cmd_mailbox *inbox,
1275                            struct mlx4_cmd_mailbox *outbox,
1276                            struct mlx4_cmd_info *cmd)
1277 {
1278         int err;
1279         int alop = vhcr->op_modifier;
1280
1281         switch (vhcr->in_modifier) {
1282         case RES_QP:
1283                 err = qp_alloc_res(dev, slave, vhcr->op_modifier, alop,
1284                                    vhcr->in_param, &vhcr->out_param);
1285                 break;
1286
1287         case RES_MTT:
1288                 err = mtt_alloc_res(dev, slave, vhcr->op_modifier, alop,
1289                                     vhcr->in_param, &vhcr->out_param);
1290                 break;
1291
1292         case RES_MPT:
1293                 err = mpt_alloc_res(dev, slave, vhcr->op_modifier, alop,
1294                                     vhcr->in_param, &vhcr->out_param);
1295                 break;
1296
1297         case RES_CQ:
1298                 err = cq_alloc_res(dev, slave, vhcr->op_modifier, alop,
1299                                    vhcr->in_param, &vhcr->out_param);
1300                 break;
1301
1302         case RES_SRQ:
1303                 err = srq_alloc_res(dev, slave, vhcr->op_modifier, alop,
1304                                     vhcr->in_param, &vhcr->out_param);
1305                 break;
1306
1307         case RES_MAC:
1308                 err = mac_alloc_res(dev, slave, vhcr->op_modifier, alop,
1309                                     vhcr->in_param, &vhcr->out_param);
1310                 break;
1311
1312         case RES_VLAN:
1313                 err = vlan_alloc_res(dev, slave, vhcr->op_modifier, alop,
1314                                     vhcr->in_param, &vhcr->out_param);
1315                 break;
1316
1317         default:
1318                 err = -EINVAL;
1319                 break;
1320         }
1321
1322         return err;
1323 }
1324
1325 static int qp_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1326                        u64 in_param)
1327 {
1328         int err;
1329         int count;
1330         int base;
1331         int qpn;
1332
1333         switch (op) {
1334         case RES_OP_RESERVE:
1335                 base = get_param_l(&in_param) & 0x7fffff;
1336                 count = get_param_h(&in_param);
1337                 err = rem_res_range(dev, slave, base, count, RES_QP, 0);
1338                 if (err)
1339                         break;
1340                 __mlx4_qp_release_range(dev, base, count);
1341                 break;
1342         case RES_OP_MAP_ICM:
1343                 qpn = get_param_l(&in_param) & 0x7fffff;
1344                 err = qp_res_start_move_to(dev, slave, qpn, RES_QP_RESERVED,
1345                                            NULL, 0);
1346                 if (err)
1347                         return err;
1348
1349                 if (!valid_reserved(dev, slave, qpn))
1350                         __mlx4_qp_free_icm(dev, qpn);
1351
1352                 res_end_move(dev, slave, RES_QP, qpn);
1353
1354                 if (valid_reserved(dev, slave, qpn))
1355                         err = rem_res_range(dev, slave, qpn, 1, RES_QP, 0);
1356                 break;
1357         default:
1358                 err = -EINVAL;
1359                 break;
1360         }
1361         return err;
1362 }
1363
1364 static int mtt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1365                         u64 in_param, u64 *out_param)
1366 {
1367         int err = -EINVAL;
1368         int base;
1369         int order;
1370
1371         if (op != RES_OP_RESERVE_AND_MAP)
1372                 return err;
1373
1374         base = get_param_l(&in_param);
1375         order = get_param_h(&in_param);
1376         err = rem_res_range(dev, slave, base, 1, RES_MTT, order);
1377         if (!err)
1378                 __mlx4_free_mtt_range(dev, base, order);
1379         return err;
1380 }
1381
1382 static int mpt_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1383                         u64 in_param)
1384 {
1385         int err = -EINVAL;
1386         int index;
1387         int id;
1388         struct res_mpt *mpt;
1389
1390         switch (op) {
1391         case RES_OP_RESERVE:
1392                 index = get_param_l(&in_param);
1393                 id = index & mpt_mask(dev);
1394                 err = get_res(dev, slave, id, RES_MPT, &mpt);
1395                 if (err)
1396                         break;
1397                 index = mpt->key;
1398                 put_res(dev, slave, id, RES_MPT);
1399
1400                 err = rem_res_range(dev, slave, id, 1, RES_MPT, 0);
1401                 if (err)
1402                         break;
1403                 __mlx4_mr_release(dev, index);
1404                 break;
1405         case RES_OP_MAP_ICM:
1406                         index = get_param_l(&in_param);
1407                         id = index & mpt_mask(dev);
1408                         err = mr_res_start_move_to(dev, slave, id,
1409                                                    RES_MPT_RESERVED, &mpt);
1410                         if (err)
1411                                 return err;
1412
1413                         __mlx4_mr_free_icm(dev, mpt->key);
1414                         res_end_move(dev, slave, RES_MPT, id);
1415                         return err;
1416                 break;
1417         default:
1418                 err = -EINVAL;
1419                 break;
1420         }
1421         return err;
1422 }
1423
1424 static int cq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1425                        u64 in_param, u64 *out_param)
1426 {
1427         int cqn;
1428         int err;
1429
1430         switch (op) {
1431         case RES_OP_RESERVE_AND_MAP:
1432                 cqn = get_param_l(&in_param);
1433                 err = rem_res_range(dev, slave, cqn, 1, RES_CQ, 0);
1434                 if (err)
1435                         break;
1436
1437                 __mlx4_cq_free_icm(dev, cqn);
1438                 break;
1439
1440         default:
1441                 err = -EINVAL;
1442                 break;
1443         }
1444
1445         return err;
1446 }
1447
1448 static int srq_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1449                         u64 in_param, u64 *out_param)
1450 {
1451         int srqn;
1452         int err;
1453
1454         switch (op) {
1455         case RES_OP_RESERVE_AND_MAP:
1456                 srqn = get_param_l(&in_param);
1457                 err = rem_res_range(dev, slave, srqn, 1, RES_SRQ, 0);
1458                 if (err)
1459                         break;
1460
1461                 __mlx4_srq_free_icm(dev, srqn);
1462                 break;
1463
1464         default:
1465                 err = -EINVAL;
1466                 break;
1467         }
1468
1469         return err;
1470 }
1471
1472 static int mac_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1473                             u64 in_param, u64 *out_param)
1474 {
1475         int port;
1476         int err = 0;
1477
1478         switch (op) {
1479         case RES_OP_RESERVE_AND_MAP:
1480                 port = get_param_l(out_param);
1481                 mac_del_from_slave(dev, slave, in_param, port);
1482                 __mlx4_unregister_mac(dev, port, in_param);
1483                 break;
1484         default:
1485                 err = -EINVAL;
1486                 break;
1487         }
1488
1489         return err;
1490
1491 }
1492
1493 static int vlan_free_res(struct mlx4_dev *dev, int slave, int op, int cmd,
1494                             u64 in_param, u64 *out_param)
1495 {
1496         return 0;
1497 }
1498
1499 int mlx4_FREE_RES_wrapper(struct mlx4_dev *dev, int slave,
1500                           struct mlx4_vhcr *vhcr,
1501                           struct mlx4_cmd_mailbox *inbox,
1502                           struct mlx4_cmd_mailbox *outbox,
1503                           struct mlx4_cmd_info *cmd)
1504 {
1505         int err = -EINVAL;
1506         int alop = vhcr->op_modifier;
1507
1508         switch (vhcr->in_modifier) {
1509         case RES_QP:
1510                 err = qp_free_res(dev, slave, vhcr->op_modifier, alop,
1511                                   vhcr->in_param);
1512                 break;
1513
1514         case RES_MTT:
1515                 err = mtt_free_res(dev, slave, vhcr->op_modifier, alop,
1516                                    vhcr->in_param, &vhcr->out_param);
1517                 break;
1518
1519         case RES_MPT:
1520                 err = mpt_free_res(dev, slave, vhcr->op_modifier, alop,
1521                                    vhcr->in_param);
1522                 break;
1523
1524         case RES_CQ:
1525                 err = cq_free_res(dev, slave, vhcr->op_modifier, alop,
1526                                   vhcr->in_param, &vhcr->out_param);
1527                 break;
1528
1529         case RES_SRQ:
1530                 err = srq_free_res(dev, slave, vhcr->op_modifier, alop,
1531                                    vhcr->in_param, &vhcr->out_param);
1532                 break;
1533
1534         case RES_MAC:
1535                 err = mac_free_res(dev, slave, vhcr->op_modifier, alop,
1536                                    vhcr->in_param, &vhcr->out_param);
1537                 break;
1538
1539         case RES_VLAN:
1540                 err = vlan_free_res(dev, slave, vhcr->op_modifier, alop,
1541                                    vhcr->in_param, &vhcr->out_param);
1542                 break;
1543
1544         default:
1545                 break;
1546         }
1547         return err;
1548 }
1549
1550 /* ugly but other choices are uglier */
1551 static int mr_phys_mpt(struct mlx4_mpt_entry *mpt)
1552 {
1553         return (be32_to_cpu(mpt->flags) >> 9) & 1;
1554 }
1555
1556 static int mr_get_mtt_addr(struct mlx4_mpt_entry *mpt)
1557 {
1558         return (int)be64_to_cpu(mpt->mtt_addr) & 0xfffffff8;
1559 }
1560
1561 static int mr_get_mtt_size(struct mlx4_mpt_entry *mpt)
1562 {
1563         return be32_to_cpu(mpt->mtt_sz);
1564 }
1565
1566 static int qp_get_mtt_addr(struct mlx4_qp_context *qpc)
1567 {
1568         return be32_to_cpu(qpc->mtt_base_addr_l) & 0xfffffff8;
1569 }
1570
1571 static int srq_get_mtt_addr(struct mlx4_srq_context *srqc)
1572 {
1573         return be32_to_cpu(srqc->mtt_base_addr_l) & 0xfffffff8;
1574 }
1575
1576 static int qp_get_mtt_size(struct mlx4_qp_context *qpc)
1577 {
1578         int page_shift = (qpc->log_page_size & 0x3f) + 12;
1579         int log_sq_size = (qpc->sq_size_stride >> 3) & 0xf;
1580         int log_sq_sride = qpc->sq_size_stride & 7;
1581         int log_rq_size = (qpc->rq_size_stride >> 3) & 0xf;
1582         int log_rq_stride = qpc->rq_size_stride & 7;
1583         int srq = (be32_to_cpu(qpc->srqn) >> 24) & 1;
1584         int rss = (be32_to_cpu(qpc->flags) >> 13) & 1;
1585         int xrc = (be32_to_cpu(qpc->local_qpn) >> 23) & 1;
1586         int sq_size;
1587         int rq_size;
1588         int total_pages;
1589         int total_mem;
1590         int page_offset = (be32_to_cpu(qpc->params2) >> 6) & 0x3f;
1591
1592         sq_size = 1 << (log_sq_size + log_sq_sride + 4);
1593         rq_size = (srq|rss|xrc) ? 0 : (1 << (log_rq_size + log_rq_stride + 4));
1594         total_mem = sq_size + rq_size;
1595         total_pages =
1596                 roundup_pow_of_two((total_mem + (page_offset << 6)) >>
1597                                    page_shift);
1598
1599         return total_pages;
1600 }
1601
1602 static int check_mtt_range(struct mlx4_dev *dev, int slave, int start,
1603                            int size, struct res_mtt *mtt)
1604 {
1605         int res_start = mtt->com.res_id;
1606         int res_size = (1 << mtt->order);
1607
1608         if (start < res_start || start + size > res_start + res_size)
1609                 return -EPERM;
1610         return 0;
1611 }
1612
1613 int mlx4_SW2HW_MPT_wrapper(struct mlx4_dev *dev, int slave,
1614                            struct mlx4_vhcr *vhcr,
1615                            struct mlx4_cmd_mailbox *inbox,
1616                            struct mlx4_cmd_mailbox *outbox,
1617                            struct mlx4_cmd_info *cmd)
1618 {
1619         int err;
1620         int index = vhcr->in_modifier;
1621         struct res_mtt *mtt;
1622         struct res_mpt *mpt;
1623         int mtt_base = mr_get_mtt_addr(inbox->buf) / dev->caps.mtt_entry_sz;
1624         int phys;
1625         int id;
1626
1627         id = index & mpt_mask(dev);
1628         err = mr_res_start_move_to(dev, slave, id, RES_MPT_HW, &mpt);
1629         if (err)
1630                 return err;
1631
1632         phys = mr_phys_mpt(inbox->buf);
1633         if (!phys) {
1634                 err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
1635                 if (err)
1636                         goto ex_abort;
1637
1638                 err = check_mtt_range(dev, slave, mtt_base,
1639                                       mr_get_mtt_size(inbox->buf), mtt);
1640                 if (err)
1641                         goto ex_put;
1642
1643                 mpt->mtt = mtt;
1644         }
1645
1646         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1647         if (err)
1648                 goto ex_put;
1649
1650         if (!phys) {
1651                 atomic_inc(&mtt->ref_count);
1652                 put_res(dev, slave, mtt->com.res_id, RES_MTT);
1653         }
1654
1655         res_end_move(dev, slave, RES_MPT, id);
1656         return 0;
1657
1658 ex_put:
1659         if (!phys)
1660                 put_res(dev, slave, mtt->com.res_id, RES_MTT);
1661 ex_abort:
1662         res_abort_move(dev, slave, RES_MPT, id);
1663
1664         return err;
1665 }
1666
1667 int mlx4_HW2SW_MPT_wrapper(struct mlx4_dev *dev, int slave,
1668                            struct mlx4_vhcr *vhcr,
1669                            struct mlx4_cmd_mailbox *inbox,
1670                            struct mlx4_cmd_mailbox *outbox,
1671                            struct mlx4_cmd_info *cmd)
1672 {
1673         int err;
1674         int index = vhcr->in_modifier;
1675         struct res_mpt *mpt;
1676         int id;
1677
1678         id = index & mpt_mask(dev);
1679         err = mr_res_start_move_to(dev, slave, id, RES_MPT_MAPPED, &mpt);
1680         if (err)
1681                 return err;
1682
1683         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1684         if (err)
1685                 goto ex_abort;
1686
1687         if (mpt->mtt)
1688                 atomic_dec(&mpt->mtt->ref_count);
1689
1690         res_end_move(dev, slave, RES_MPT, id);
1691         return 0;
1692
1693 ex_abort:
1694         res_abort_move(dev, slave, RES_MPT, id);
1695
1696         return err;
1697 }
1698
1699 int mlx4_QUERY_MPT_wrapper(struct mlx4_dev *dev, int slave,
1700                            struct mlx4_vhcr *vhcr,
1701                            struct mlx4_cmd_mailbox *inbox,
1702                            struct mlx4_cmd_mailbox *outbox,
1703                            struct mlx4_cmd_info *cmd)
1704 {
1705         int err;
1706         int index = vhcr->in_modifier;
1707         struct res_mpt *mpt;
1708         int id;
1709
1710         id = index & mpt_mask(dev);
1711         err = get_res(dev, slave, id, RES_MPT, &mpt);
1712         if (err)
1713                 return err;
1714
1715         if (mpt->com.from_state != RES_MPT_HW) {
1716                 err = -EBUSY;
1717                 goto out;
1718         }
1719
1720         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1721
1722 out:
1723         put_res(dev, slave, id, RES_MPT);
1724         return err;
1725 }
1726
1727 static int qp_get_rcqn(struct mlx4_qp_context *qpc)
1728 {
1729         return be32_to_cpu(qpc->cqn_recv) & 0xffffff;
1730 }
1731
1732 static int qp_get_scqn(struct mlx4_qp_context *qpc)
1733 {
1734         return be32_to_cpu(qpc->cqn_send) & 0xffffff;
1735 }
1736
1737 static u32 qp_get_srqn(struct mlx4_qp_context *qpc)
1738 {
1739         return be32_to_cpu(qpc->srqn) & 0x1ffffff;
1740 }
1741
1742 int mlx4_RST2INIT_QP_wrapper(struct mlx4_dev *dev, int slave,
1743                              struct mlx4_vhcr *vhcr,
1744                              struct mlx4_cmd_mailbox *inbox,
1745                              struct mlx4_cmd_mailbox *outbox,
1746                              struct mlx4_cmd_info *cmd)
1747 {
1748         int err;
1749         int qpn = vhcr->in_modifier & 0x7fffff;
1750         struct res_mtt *mtt;
1751         struct res_qp *qp;
1752         struct mlx4_qp_context *qpc = inbox->buf + 8;
1753         int mtt_base = qp_get_mtt_addr(qpc) / dev->caps.mtt_entry_sz;
1754         int mtt_size = qp_get_mtt_size(qpc);
1755         struct res_cq *rcq;
1756         struct res_cq *scq;
1757         int rcqn = qp_get_rcqn(qpc);
1758         int scqn = qp_get_scqn(qpc);
1759         u32 srqn = qp_get_srqn(qpc) & 0xffffff;
1760         int use_srq = (qp_get_srqn(qpc) >> 24) & 1;
1761         struct res_srq *srq;
1762         int local_qpn = be32_to_cpu(qpc->local_qpn) & 0xffffff;
1763
1764         err = qp_res_start_move_to(dev, slave, qpn, RES_QP_HW, &qp, 0);
1765         if (err)
1766                 return err;
1767         qp->local_qpn = local_qpn;
1768
1769         err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
1770         if (err)
1771                 goto ex_abort;
1772
1773         err = check_mtt_range(dev, slave, mtt_base, mtt_size, mtt);
1774         if (err)
1775                 goto ex_put_mtt;
1776
1777         err = get_res(dev, slave, rcqn, RES_CQ, &rcq);
1778         if (err)
1779                 goto ex_put_mtt;
1780
1781         if (scqn != rcqn) {
1782                 err = get_res(dev, slave, scqn, RES_CQ, &scq);
1783                 if (err)
1784                         goto ex_put_rcq;
1785         } else
1786                 scq = rcq;
1787
1788         if (use_srq) {
1789                 err = get_res(dev, slave, srqn, RES_SRQ, &srq);
1790                 if (err)
1791                         goto ex_put_scq;
1792         }
1793
1794         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1795         if (err)
1796                 goto ex_put_srq;
1797         atomic_inc(&mtt->ref_count);
1798         qp->mtt = mtt;
1799         atomic_inc(&rcq->ref_count);
1800         qp->rcq = rcq;
1801         atomic_inc(&scq->ref_count);
1802         qp->scq = scq;
1803
1804         if (scqn != rcqn)
1805                 put_res(dev, slave, scqn, RES_CQ);
1806
1807         if (use_srq) {
1808                 atomic_inc(&srq->ref_count);
1809                 put_res(dev, slave, srqn, RES_SRQ);
1810                 qp->srq = srq;
1811         }
1812         put_res(dev, slave, rcqn, RES_CQ);
1813         put_res(dev, slave, mtt_base, RES_MTT);
1814         res_end_move(dev, slave, RES_QP, qpn);
1815
1816         return 0;
1817
1818 ex_put_srq:
1819         if (use_srq)
1820                 put_res(dev, slave, srqn, RES_SRQ);
1821 ex_put_scq:
1822         if (scqn != rcqn)
1823                 put_res(dev, slave, scqn, RES_CQ);
1824 ex_put_rcq:
1825         put_res(dev, slave, rcqn, RES_CQ);
1826 ex_put_mtt:
1827         put_res(dev, slave, mtt_base, RES_MTT);
1828 ex_abort:
1829         res_abort_move(dev, slave, RES_QP, qpn);
1830
1831         return err;
1832 }
1833
1834 static int eq_get_mtt_addr(struct mlx4_eq_context *eqc)
1835 {
1836         return be32_to_cpu(eqc->mtt_base_addr_l) & 0xfffffff8;
1837 }
1838
1839 static int eq_get_mtt_size(struct mlx4_eq_context *eqc)
1840 {
1841         int log_eq_size = eqc->log_eq_size & 0x1f;
1842         int page_shift = (eqc->log_page_size & 0x3f) + 12;
1843
1844         if (log_eq_size + 5 < page_shift)
1845                 return 1;
1846
1847         return 1 << (log_eq_size + 5 - page_shift);
1848 }
1849
1850 static int cq_get_mtt_addr(struct mlx4_cq_context *cqc)
1851 {
1852         return be32_to_cpu(cqc->mtt_base_addr_l) & 0xfffffff8;
1853 }
1854
1855 static int cq_get_mtt_size(struct mlx4_cq_context *cqc)
1856 {
1857         int log_cq_size = (be32_to_cpu(cqc->logsize_usrpage) >> 24) & 0x1f;
1858         int page_shift = (cqc->log_page_size & 0x3f) + 12;
1859
1860         if (log_cq_size + 5 < page_shift)
1861                 return 1;
1862
1863         return 1 << (log_cq_size + 5 - page_shift);
1864 }
1865
1866 int mlx4_SW2HW_EQ_wrapper(struct mlx4_dev *dev, int slave,
1867                           struct mlx4_vhcr *vhcr,
1868                           struct mlx4_cmd_mailbox *inbox,
1869                           struct mlx4_cmd_mailbox *outbox,
1870                           struct mlx4_cmd_info *cmd)
1871 {
1872         int err;
1873         int eqn = vhcr->in_modifier;
1874         int res_id = (slave << 8) | eqn;
1875         struct mlx4_eq_context *eqc = inbox->buf;
1876         int mtt_base = eq_get_mtt_addr(eqc) / dev->caps.mtt_entry_sz;
1877         int mtt_size = eq_get_mtt_size(eqc);
1878         struct res_eq *eq;
1879         struct res_mtt *mtt;
1880
1881         err = add_res_range(dev, slave, res_id, 1, RES_EQ, 0);
1882         if (err)
1883                 return err;
1884         err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_HW, &eq);
1885         if (err)
1886                 goto out_add;
1887
1888         err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
1889         if (err)
1890                 goto out_move;
1891
1892         err = check_mtt_range(dev, slave, mtt_base, mtt_size, mtt);
1893         if (err)
1894                 goto out_put;
1895
1896         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1897         if (err)
1898                 goto out_put;
1899
1900         atomic_inc(&mtt->ref_count);
1901         eq->mtt = mtt;
1902         put_res(dev, slave, mtt->com.res_id, RES_MTT);
1903         res_end_move(dev, slave, RES_EQ, res_id);
1904         return 0;
1905
1906 out_put:
1907         put_res(dev, slave, mtt->com.res_id, RES_MTT);
1908 out_move:
1909         res_abort_move(dev, slave, RES_EQ, res_id);
1910 out_add:
1911         rem_res_range(dev, slave, res_id, 1, RES_EQ, 0);
1912         return err;
1913 }
1914
1915 static int get_containing_mtt(struct mlx4_dev *dev, int slave, int start,
1916                               int len, struct res_mtt **res)
1917 {
1918         struct mlx4_priv *priv = mlx4_priv(dev);
1919         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
1920         struct res_mtt *mtt;
1921         int err = -EINVAL;
1922
1923         spin_lock_irq(mlx4_tlock(dev));
1924         list_for_each_entry(mtt, &tracker->slave_list[slave].res_list[RES_MTT],
1925                             com.list) {
1926                 if (!check_mtt_range(dev, slave, start, len, mtt)) {
1927                         *res = mtt;
1928                         mtt->com.from_state = mtt->com.state;
1929                         mtt->com.state = RES_MTT_BUSY;
1930                         err = 0;
1931                         break;
1932                 }
1933         }
1934         spin_unlock_irq(mlx4_tlock(dev));
1935
1936         return err;
1937 }
1938
1939 int mlx4_WRITE_MTT_wrapper(struct mlx4_dev *dev, int slave,
1940                            struct mlx4_vhcr *vhcr,
1941                            struct mlx4_cmd_mailbox *inbox,
1942                            struct mlx4_cmd_mailbox *outbox,
1943                            struct mlx4_cmd_info *cmd)
1944 {
1945         struct mlx4_mtt mtt;
1946         __be64 *page_list = inbox->buf;
1947         u64 *pg_list = (u64 *)page_list;
1948         int i;
1949         struct res_mtt *rmtt = NULL;
1950         int start = be64_to_cpu(page_list[0]);
1951         int npages = vhcr->in_modifier;
1952         int err;
1953
1954         err = get_containing_mtt(dev, slave, start, npages, &rmtt);
1955         if (err)
1956                 return err;
1957
1958         /* Call the SW implementation of write_mtt:
1959          * - Prepare a dummy mtt struct
1960          * - Translate inbox contents to simple addresses in host endianess */
1961         mtt.offset = 0;  /* TBD this is broken but I don't handle it since
1962                             we don't really use it */
1963         mtt.order = 0;
1964         mtt.page_shift = 0;
1965         for (i = 0; i < npages; ++i)
1966                 pg_list[i + 2] = (be64_to_cpu(page_list[i + 2]) & ~1ULL);
1967
1968         err = __mlx4_write_mtt(dev, &mtt, be64_to_cpu(page_list[0]), npages,
1969                                ((u64 *)page_list + 2));
1970
1971         if (rmtt)
1972                 put_res(dev, slave, rmtt->com.res_id, RES_MTT);
1973
1974         return err;
1975 }
1976
1977 int mlx4_HW2SW_EQ_wrapper(struct mlx4_dev *dev, int slave,
1978                           struct mlx4_vhcr *vhcr,
1979                           struct mlx4_cmd_mailbox *inbox,
1980                           struct mlx4_cmd_mailbox *outbox,
1981                           struct mlx4_cmd_info *cmd)
1982 {
1983         int eqn = vhcr->in_modifier;
1984         int res_id = eqn | (slave << 8);
1985         struct res_eq *eq;
1986         int err;
1987
1988         err = eq_res_start_move_to(dev, slave, res_id, RES_EQ_RESERVED, &eq);
1989         if (err)
1990                 return err;
1991
1992         err = get_res(dev, slave, eq->mtt->com.res_id, RES_MTT, NULL);
1993         if (err)
1994                 goto ex_abort;
1995
1996         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
1997         if (err)
1998                 goto ex_put;
1999
2000         atomic_dec(&eq->mtt->ref_count);
2001         put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
2002         res_end_move(dev, slave, RES_EQ, res_id);
2003         rem_res_range(dev, slave, res_id, 1, RES_EQ, 0);
2004
2005         return 0;
2006
2007 ex_put:
2008         put_res(dev, slave, eq->mtt->com.res_id, RES_MTT);
2009 ex_abort:
2010         res_abort_move(dev, slave, RES_EQ, res_id);
2011
2012         return err;
2013 }
2014
2015 int mlx4_GEN_EQE(struct mlx4_dev *dev, int slave, struct mlx4_eqe *eqe)
2016 {
2017         struct mlx4_priv *priv = mlx4_priv(dev);
2018         struct mlx4_slave_event_eq_info *event_eq;
2019         struct mlx4_cmd_mailbox *mailbox;
2020         u32 in_modifier = 0;
2021         int err;
2022         int res_id;
2023         struct res_eq *req;
2024
2025         if (!priv->mfunc.master.slave_state)
2026                 return -EINVAL;
2027
2028         event_eq = &priv->mfunc.master.slave_state[slave].event_eq[eqe->type];
2029
2030         /* Create the event only if the slave is registered */
2031         if (event_eq->eqn < 0)
2032                 return 0;
2033
2034         mutex_lock(&priv->mfunc.master.gen_eqe_mutex[slave]);
2035         res_id = (slave << 8) | event_eq->eqn;
2036         err = get_res(dev, slave, res_id, RES_EQ, &req);
2037         if (err)
2038                 goto unlock;
2039
2040         if (req->com.from_state != RES_EQ_HW) {
2041                 err = -EINVAL;
2042                 goto put;
2043         }
2044
2045         mailbox = mlx4_alloc_cmd_mailbox(dev);
2046         if (IS_ERR(mailbox)) {
2047                 err = PTR_ERR(mailbox);
2048                 goto put;
2049         }
2050
2051         if (eqe->type == MLX4_EVENT_TYPE_CMD) {
2052                 ++event_eq->token;
2053                 eqe->event.cmd.token = cpu_to_be16(event_eq->token);
2054         }
2055
2056         memcpy(mailbox->buf, (u8 *) eqe, 28);
2057
2058         in_modifier = (slave & 0xff) | ((event_eq->eqn & 0xff) << 16);
2059
2060         err = mlx4_cmd(dev, mailbox->dma, in_modifier, 0,
2061                        MLX4_CMD_GEN_EQE, MLX4_CMD_TIME_CLASS_B,
2062                        MLX4_CMD_NATIVE);
2063
2064         put_res(dev, slave, res_id, RES_EQ);
2065         mutex_unlock(&priv->mfunc.master.gen_eqe_mutex[slave]);
2066         mlx4_free_cmd_mailbox(dev, mailbox);
2067         return err;
2068
2069 put:
2070         put_res(dev, slave, res_id, RES_EQ);
2071
2072 unlock:
2073         mutex_unlock(&priv->mfunc.master.gen_eqe_mutex[slave]);
2074         return err;
2075 }
2076
2077 int mlx4_QUERY_EQ_wrapper(struct mlx4_dev *dev, int slave,
2078                           struct mlx4_vhcr *vhcr,
2079                           struct mlx4_cmd_mailbox *inbox,
2080                           struct mlx4_cmd_mailbox *outbox,
2081                           struct mlx4_cmd_info *cmd)
2082 {
2083         int eqn = vhcr->in_modifier;
2084         int res_id = eqn | (slave << 8);
2085         struct res_eq *eq;
2086         int err;
2087
2088         err = get_res(dev, slave, res_id, RES_EQ, &eq);
2089         if (err)
2090                 return err;
2091
2092         if (eq->com.from_state != RES_EQ_HW) {
2093                 err = -EINVAL;
2094                 goto ex_put;
2095         }
2096
2097         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2098
2099 ex_put:
2100         put_res(dev, slave, res_id, RES_EQ);
2101         return err;
2102 }
2103
2104 int mlx4_SW2HW_CQ_wrapper(struct mlx4_dev *dev, int slave,
2105                           struct mlx4_vhcr *vhcr,
2106                           struct mlx4_cmd_mailbox *inbox,
2107                           struct mlx4_cmd_mailbox *outbox,
2108                           struct mlx4_cmd_info *cmd)
2109 {
2110         int err;
2111         int cqn = vhcr->in_modifier;
2112         struct mlx4_cq_context *cqc = inbox->buf;
2113         int mtt_base = cq_get_mtt_addr(cqc) / dev->caps.mtt_entry_sz;
2114         struct res_cq *cq;
2115         struct res_mtt *mtt;
2116
2117         err = cq_res_start_move_to(dev, slave, cqn, RES_CQ_HW, &cq);
2118         if (err)
2119                 return err;
2120         err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
2121         if (err)
2122                 goto out_move;
2123         err = check_mtt_range(dev, slave, mtt_base, cq_get_mtt_size(cqc), mtt);
2124         if (err)
2125                 goto out_put;
2126         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2127         if (err)
2128                 goto out_put;
2129         atomic_inc(&mtt->ref_count);
2130         cq->mtt = mtt;
2131         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2132         res_end_move(dev, slave, RES_CQ, cqn);
2133         return 0;
2134
2135 out_put:
2136         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2137 out_move:
2138         res_abort_move(dev, slave, RES_CQ, cqn);
2139         return err;
2140 }
2141
2142 int mlx4_HW2SW_CQ_wrapper(struct mlx4_dev *dev, int slave,
2143                           struct mlx4_vhcr *vhcr,
2144                           struct mlx4_cmd_mailbox *inbox,
2145                           struct mlx4_cmd_mailbox *outbox,
2146                           struct mlx4_cmd_info *cmd)
2147 {
2148         int err;
2149         int cqn = vhcr->in_modifier;
2150         struct res_cq *cq;
2151
2152         err = cq_res_start_move_to(dev, slave, cqn, RES_CQ_ALLOCATED, &cq);
2153         if (err)
2154                 return err;
2155         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2156         if (err)
2157                 goto out_move;
2158         atomic_dec(&cq->mtt->ref_count);
2159         res_end_move(dev, slave, RES_CQ, cqn);
2160         return 0;
2161
2162 out_move:
2163         res_abort_move(dev, slave, RES_CQ, cqn);
2164         return err;
2165 }
2166
2167 int mlx4_QUERY_CQ_wrapper(struct mlx4_dev *dev, int slave,
2168                           struct mlx4_vhcr *vhcr,
2169                           struct mlx4_cmd_mailbox *inbox,
2170                           struct mlx4_cmd_mailbox *outbox,
2171                           struct mlx4_cmd_info *cmd)
2172 {
2173         int cqn = vhcr->in_modifier;
2174         struct res_cq *cq;
2175         int err;
2176
2177         err = get_res(dev, slave, cqn, RES_CQ, &cq);
2178         if (err)
2179                 return err;
2180
2181         if (cq->com.from_state != RES_CQ_HW)
2182                 goto ex_put;
2183
2184         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2185 ex_put:
2186         put_res(dev, slave, cqn, RES_CQ);
2187
2188         return err;
2189 }
2190
2191 static int handle_resize(struct mlx4_dev *dev, int slave,
2192                          struct mlx4_vhcr *vhcr,
2193                          struct mlx4_cmd_mailbox *inbox,
2194                          struct mlx4_cmd_mailbox *outbox,
2195                          struct mlx4_cmd_info *cmd,
2196                          struct res_cq *cq)
2197 {
2198         int err;
2199         struct res_mtt *orig_mtt;
2200         struct res_mtt *mtt;
2201         struct mlx4_cq_context *cqc = inbox->buf;
2202         int mtt_base = cq_get_mtt_addr(cqc) / dev->caps.mtt_entry_sz;
2203
2204         err = get_res(dev, slave, cq->mtt->com.res_id, RES_MTT, &orig_mtt);
2205         if (err)
2206                 return err;
2207
2208         if (orig_mtt != cq->mtt) {
2209                 err = -EINVAL;
2210                 goto ex_put;
2211         }
2212
2213         err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
2214         if (err)
2215                 goto ex_put;
2216
2217         err = check_mtt_range(dev, slave, mtt_base, cq_get_mtt_size(cqc), mtt);
2218         if (err)
2219                 goto ex_put1;
2220         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2221         if (err)
2222                 goto ex_put1;
2223         atomic_dec(&orig_mtt->ref_count);
2224         put_res(dev, slave, orig_mtt->com.res_id, RES_MTT);
2225         atomic_inc(&mtt->ref_count);
2226         cq->mtt = mtt;
2227         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2228         return 0;
2229
2230 ex_put1:
2231         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2232 ex_put:
2233         put_res(dev, slave, orig_mtt->com.res_id, RES_MTT);
2234
2235         return err;
2236
2237 }
2238
2239 int mlx4_MODIFY_CQ_wrapper(struct mlx4_dev *dev, int slave,
2240                            struct mlx4_vhcr *vhcr,
2241                            struct mlx4_cmd_mailbox *inbox,
2242                            struct mlx4_cmd_mailbox *outbox,
2243                            struct mlx4_cmd_info *cmd)
2244 {
2245         int cqn = vhcr->in_modifier;
2246         struct res_cq *cq;
2247         int err;
2248
2249         err = get_res(dev, slave, cqn, RES_CQ, &cq);
2250         if (err)
2251                 return err;
2252
2253         if (cq->com.from_state != RES_CQ_HW)
2254                 goto ex_put;
2255
2256         if (vhcr->op_modifier == 0) {
2257                 err = handle_resize(dev, slave, vhcr, inbox, outbox, cmd, cq);
2258                 if (err)
2259                         goto ex_put;
2260         }
2261
2262         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2263 ex_put:
2264         put_res(dev, slave, cqn, RES_CQ);
2265
2266         return err;
2267 }
2268
2269 static int srq_get_mtt_size(struct mlx4_srq_context *srqc)
2270 {
2271         int log_srq_size = (be32_to_cpu(srqc->state_logsize_srqn) >> 24) & 0xf;
2272         int log_rq_stride = srqc->logstride & 7;
2273         int page_shift = (srqc->log_page_size & 0x3f) + 12;
2274
2275         if (log_srq_size + log_rq_stride + 4 < page_shift)
2276                 return 1;
2277
2278         return 1 << (log_srq_size + log_rq_stride + 4 - page_shift);
2279 }
2280
2281 int mlx4_SW2HW_SRQ_wrapper(struct mlx4_dev *dev, int slave,
2282                            struct mlx4_vhcr *vhcr,
2283                            struct mlx4_cmd_mailbox *inbox,
2284                            struct mlx4_cmd_mailbox *outbox,
2285                            struct mlx4_cmd_info *cmd)
2286 {
2287         int err;
2288         int srqn = vhcr->in_modifier;
2289         struct res_mtt *mtt;
2290         struct res_srq *srq;
2291         struct mlx4_srq_context *srqc = inbox->buf;
2292         int mtt_base = srq_get_mtt_addr(srqc) / dev->caps.mtt_entry_sz;
2293
2294         if (srqn != (be32_to_cpu(srqc->state_logsize_srqn) & 0xffffff))
2295                 return -EINVAL;
2296
2297         err = srq_res_start_move_to(dev, slave, srqn, RES_SRQ_HW, &srq);
2298         if (err)
2299                 return err;
2300         err = get_res(dev, slave, mtt_base, RES_MTT, &mtt);
2301         if (err)
2302                 goto ex_abort;
2303         err = check_mtt_range(dev, slave, mtt_base, srq_get_mtt_size(srqc),
2304                               mtt);
2305         if (err)
2306                 goto ex_put_mtt;
2307
2308         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2309         if (err)
2310                 goto ex_put_mtt;
2311
2312         atomic_inc(&mtt->ref_count);
2313         srq->mtt = mtt;
2314         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2315         res_end_move(dev, slave, RES_SRQ, srqn);
2316         return 0;
2317
2318 ex_put_mtt:
2319         put_res(dev, slave, mtt->com.res_id, RES_MTT);
2320 ex_abort:
2321         res_abort_move(dev, slave, RES_SRQ, srqn);
2322
2323         return err;
2324 }
2325
2326 int mlx4_HW2SW_SRQ_wrapper(struct mlx4_dev *dev, int slave,
2327                            struct mlx4_vhcr *vhcr,
2328                            struct mlx4_cmd_mailbox *inbox,
2329                            struct mlx4_cmd_mailbox *outbox,
2330                            struct mlx4_cmd_info *cmd)
2331 {
2332         int err;
2333         int srqn = vhcr->in_modifier;
2334         struct res_srq *srq;
2335
2336         err = srq_res_start_move_to(dev, slave, srqn, RES_SRQ_ALLOCATED, &srq);
2337         if (err)
2338                 return err;
2339         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2340         if (err)
2341                 goto ex_abort;
2342         atomic_dec(&srq->mtt->ref_count);
2343         if (srq->cq)
2344                 atomic_dec(&srq->cq->ref_count);
2345         res_end_move(dev, slave, RES_SRQ, srqn);
2346
2347         return 0;
2348
2349 ex_abort:
2350         res_abort_move(dev, slave, RES_SRQ, srqn);
2351
2352         return err;
2353 }
2354
2355 int mlx4_QUERY_SRQ_wrapper(struct mlx4_dev *dev, int slave,
2356                            struct mlx4_vhcr *vhcr,
2357                            struct mlx4_cmd_mailbox *inbox,
2358                            struct mlx4_cmd_mailbox *outbox,
2359                            struct mlx4_cmd_info *cmd)
2360 {
2361         int err;
2362         int srqn = vhcr->in_modifier;
2363         struct res_srq *srq;
2364
2365         err = get_res(dev, slave, srqn, RES_SRQ, &srq);
2366         if (err)
2367                 return err;
2368         if (srq->com.from_state != RES_SRQ_HW) {
2369                 err = -EBUSY;
2370                 goto out;
2371         }
2372         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2373 out:
2374         put_res(dev, slave, srqn, RES_SRQ);
2375         return err;
2376 }
2377
2378 int mlx4_ARM_SRQ_wrapper(struct mlx4_dev *dev, int slave,
2379                          struct mlx4_vhcr *vhcr,
2380                          struct mlx4_cmd_mailbox *inbox,
2381                          struct mlx4_cmd_mailbox *outbox,
2382                          struct mlx4_cmd_info *cmd)
2383 {
2384         int err;
2385         int srqn = vhcr->in_modifier;
2386         struct res_srq *srq;
2387
2388         err = get_res(dev, slave, srqn, RES_SRQ, &srq);
2389         if (err)
2390                 return err;
2391
2392         if (srq->com.from_state != RES_SRQ_HW) {
2393                 err = -EBUSY;
2394                 goto out;
2395         }
2396
2397         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2398 out:
2399         put_res(dev, slave, srqn, RES_SRQ);
2400         return err;
2401 }
2402
2403 int mlx4_GEN_QP_wrapper(struct mlx4_dev *dev, int slave,
2404                         struct mlx4_vhcr *vhcr,
2405                         struct mlx4_cmd_mailbox *inbox,
2406                         struct mlx4_cmd_mailbox *outbox,
2407                         struct mlx4_cmd_info *cmd)
2408 {
2409         int err;
2410         int qpn = vhcr->in_modifier & 0x7fffff;
2411         struct res_qp *qp;
2412
2413         err = get_res(dev, slave, qpn, RES_QP, &qp);
2414         if (err)
2415                 return err;
2416         if (qp->com.from_state != RES_QP_HW) {
2417                 err = -EBUSY;
2418                 goto out;
2419         }
2420
2421         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2422 out:
2423         put_res(dev, slave, qpn, RES_QP);
2424         return err;
2425 }
2426
2427 int mlx4_INIT2RTR_QP_wrapper(struct mlx4_dev *dev, int slave,
2428                              struct mlx4_vhcr *vhcr,
2429                              struct mlx4_cmd_mailbox *inbox,
2430                              struct mlx4_cmd_mailbox *outbox,
2431                              struct mlx4_cmd_info *cmd)
2432 {
2433         struct mlx4_qp_context *qpc = inbox->buf + 8;
2434
2435         update_ud_gid(dev, qpc, (u8)slave);
2436
2437         return mlx4_GEN_QP_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2438 }
2439
2440 int mlx4_2RST_QP_wrapper(struct mlx4_dev *dev, int slave,
2441                          struct mlx4_vhcr *vhcr,
2442                          struct mlx4_cmd_mailbox *inbox,
2443                          struct mlx4_cmd_mailbox *outbox,
2444                          struct mlx4_cmd_info *cmd)
2445 {
2446         int err;
2447         int qpn = vhcr->in_modifier & 0x7fffff;
2448         struct res_qp *qp;
2449
2450         err = qp_res_start_move_to(dev, slave, qpn, RES_QP_MAPPED, &qp, 0);
2451         if (err)
2452                 return err;
2453         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2454         if (err)
2455                 goto ex_abort;
2456
2457         atomic_dec(&qp->mtt->ref_count);
2458         atomic_dec(&qp->rcq->ref_count);
2459         atomic_dec(&qp->scq->ref_count);
2460         if (qp->srq)
2461                 atomic_dec(&qp->srq->ref_count);
2462         res_end_move(dev, slave, RES_QP, qpn);
2463         return 0;
2464
2465 ex_abort:
2466         res_abort_move(dev, slave, RES_QP, qpn);
2467
2468         return err;
2469 }
2470
2471 static struct res_gid *find_gid(struct mlx4_dev *dev, int slave,
2472                                 struct res_qp *rqp, u8 *gid)
2473 {
2474         struct res_gid *res;
2475
2476         list_for_each_entry(res, &rqp->mcg_list, list) {
2477                 if (!memcmp(res->gid, gid, 16))
2478                         return res;
2479         }
2480         return NULL;
2481 }
2482
2483 static int add_mcg_res(struct mlx4_dev *dev, int slave, struct res_qp *rqp,
2484                        u8 *gid, enum mlx4_protocol prot,
2485                        enum mlx4_steer_type steer)
2486 {
2487         struct res_gid *res;
2488         int err;
2489
2490         res = kzalloc(sizeof *res, GFP_KERNEL);
2491         if (!res)
2492                 return -ENOMEM;
2493
2494         spin_lock_irq(&rqp->mcg_spl);
2495         if (find_gid(dev, slave, rqp, gid)) {
2496                 kfree(res);
2497                 err = -EEXIST;
2498         } else {
2499                 memcpy(res->gid, gid, 16);
2500                 res->prot = prot;
2501                 res->steer = steer;
2502                 list_add_tail(&res->list, &rqp->mcg_list);
2503                 err = 0;
2504         }
2505         spin_unlock_irq(&rqp->mcg_spl);
2506
2507         return err;
2508 }
2509
2510 static int rem_mcg_res(struct mlx4_dev *dev, int slave, struct res_qp *rqp,
2511                        u8 *gid, enum mlx4_protocol prot,
2512                        enum mlx4_steer_type steer)
2513 {
2514         struct res_gid *res;
2515         int err;
2516
2517         spin_lock_irq(&rqp->mcg_spl);
2518         res = find_gid(dev, slave, rqp, gid);
2519         if (!res || res->prot != prot || res->steer != steer)
2520                 err = -EINVAL;
2521         else {
2522                 list_del(&res->list);
2523                 kfree(res);
2524                 err = 0;
2525         }
2526         spin_unlock_irq(&rqp->mcg_spl);
2527
2528         return err;
2529 }
2530
2531 int mlx4_QP_ATTACH_wrapper(struct mlx4_dev *dev, int slave,
2532                                struct mlx4_vhcr *vhcr,
2533                                struct mlx4_cmd_mailbox *inbox,
2534                                struct mlx4_cmd_mailbox *outbox,
2535                                struct mlx4_cmd_info *cmd)
2536 {
2537         struct mlx4_qp qp; /* dummy for calling attach/detach */
2538         u8 *gid = inbox->buf;
2539         enum mlx4_protocol prot = (vhcr->in_modifier >> 28) & 0x7;
2540         int err, err1;
2541         int qpn;
2542         struct res_qp *rqp;
2543         int attach = vhcr->op_modifier;
2544         int block_loopback = vhcr->in_modifier >> 31;
2545         u8 steer_type_mask = 2;
2546         enum mlx4_steer_type type = (gid[7] & steer_type_mask) >> 1;
2547
2548         qpn = vhcr->in_modifier & 0xffffff;
2549         err = get_res(dev, slave, qpn, RES_QP, &rqp);
2550         if (err)
2551                 return err;
2552
2553         qp.qpn = qpn;
2554         if (attach) {
2555                 err = add_mcg_res(dev, slave, rqp, gid, prot, type);
2556                 if (err)
2557                         goto ex_put;
2558
2559                 err = mlx4_qp_attach_common(dev, &qp, gid,
2560                                             block_loopback, prot, type);
2561                 if (err)
2562                         goto ex_rem;
2563         } else {
2564                 err = rem_mcg_res(dev, slave, rqp, gid, prot, type);
2565                 if (err)
2566                         goto ex_put;
2567                 err = mlx4_qp_detach_common(dev, &qp, gid, prot, type);
2568         }
2569
2570         put_res(dev, slave, qpn, RES_QP);
2571         return 0;
2572
2573 ex_rem:
2574         /* ignore error return below, already in error */
2575         err1 = rem_mcg_res(dev, slave, rqp, gid, prot, type);
2576 ex_put:
2577         put_res(dev, slave, qpn, RES_QP);
2578
2579         return err;
2580 }
2581
2582 enum {
2583         BUSY_MAX_RETRIES = 10
2584 };
2585
2586 int mlx4_QUERY_IF_STAT_wrapper(struct mlx4_dev *dev, int slave,
2587                                struct mlx4_vhcr *vhcr,
2588                                struct mlx4_cmd_mailbox *inbox,
2589                                struct mlx4_cmd_mailbox *outbox,
2590                                struct mlx4_cmd_info *cmd)
2591 {
2592         int err;
2593         int index = vhcr->in_modifier & 0xffff;
2594
2595         err = get_res(dev, slave, index, RES_COUNTER, NULL);
2596         if (err)
2597                 return err;
2598
2599         err = mlx4_DMA_wrapper(dev, slave, vhcr, inbox, outbox, cmd);
2600         put_res(dev, slave, index, RES_COUNTER);
2601         return err;
2602 }
2603
2604 static void detach_qp(struct mlx4_dev *dev, int slave, struct res_qp *rqp)
2605 {
2606         struct res_gid *rgid;
2607         struct res_gid *tmp;
2608         int err;
2609         struct mlx4_qp qp; /* dummy for calling attach/detach */
2610
2611         list_for_each_entry_safe(rgid, tmp, &rqp->mcg_list, list) {
2612                 qp.qpn = rqp->local_qpn;
2613                 err = mlx4_qp_detach_common(dev, &qp, rgid->gid, rgid->prot,
2614                                             rgid->steer);
2615                 list_del(&rgid->list);
2616                 kfree(rgid);
2617         }
2618 }
2619
2620 static int _move_all_busy(struct mlx4_dev *dev, int slave,
2621                           enum mlx4_resource type, int print)
2622 {
2623         struct mlx4_priv *priv = mlx4_priv(dev);
2624         struct mlx4_resource_tracker *tracker =
2625                 &priv->mfunc.master.res_tracker;
2626         struct list_head *rlist = &tracker->slave_list[slave].res_list[type];
2627         struct res_common *r;
2628         struct res_common *tmp;
2629         int busy;
2630
2631         busy = 0;
2632         spin_lock_irq(mlx4_tlock(dev));
2633         list_for_each_entry_safe(r, tmp, rlist, list) {
2634                 if (r->owner == slave) {
2635                         if (!r->removing) {
2636                                 if (r->state == RES_ANY_BUSY) {
2637                                         if (print)
2638                                                 mlx4_dbg(dev,
2639                                                          "%s id 0x%x is busy\n",
2640                                                           ResourceType(type),
2641                                                           r->res_id);
2642                                         ++busy;
2643                                 } else {
2644                                         r->from_state = r->state;
2645                                         r->state = RES_ANY_BUSY;
2646                                         r->removing = 1;
2647                                 }
2648                         }
2649                 }
2650         }
2651         spin_unlock_irq(mlx4_tlock(dev));
2652
2653         return busy;
2654 }
2655
2656 static int move_all_busy(struct mlx4_dev *dev, int slave,
2657                          enum mlx4_resource type)
2658 {
2659         unsigned long begin;
2660         int busy;
2661
2662         begin = jiffies;
2663         do {
2664                 busy = _move_all_busy(dev, slave, type, 0);
2665                 if (time_after(jiffies, begin + 5 * HZ))
2666                         break;
2667                 if (busy)
2668                         cond_resched();
2669         } while (busy);
2670
2671         if (busy)
2672                 busy = _move_all_busy(dev, slave, type, 1);
2673
2674         return busy;
2675 }
2676 static void rem_slave_qps(struct mlx4_dev *dev, int slave)
2677 {
2678         struct mlx4_priv *priv = mlx4_priv(dev);
2679         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
2680         struct list_head *qp_list =
2681                 &tracker->slave_list[slave].res_list[RES_QP];
2682         struct res_qp *qp;
2683         struct res_qp *tmp;
2684         int state;
2685         u64 in_param;
2686         int qpn;
2687         int err;
2688
2689         err = move_all_busy(dev, slave, RES_QP);
2690         if (err)
2691                 mlx4_warn(dev, "rem_slave_qps: Could not move all qps to busy"
2692                           "for slave %d\n", slave);
2693
2694         spin_lock_irq(mlx4_tlock(dev));
2695         list_for_each_entry_safe(qp, tmp, qp_list, com.list) {
2696                 spin_unlock_irq(mlx4_tlock(dev));
2697                 if (qp->com.owner == slave) {
2698                         qpn = qp->com.res_id;
2699                         detach_qp(dev, slave, qp);
2700                         state = qp->com.from_state;
2701                         while (state != 0) {
2702                                 switch (state) {
2703                                 case RES_QP_RESERVED:
2704                                         spin_lock_irq(mlx4_tlock(dev));
2705                                         radix_tree_delete(&tracker->res_tree[RES_QP],
2706                                                           qp->com.res_id);
2707                                         list_del(&qp->com.list);
2708                                         spin_unlock_irq(mlx4_tlock(dev));
2709                                         kfree(qp);
2710                                         state = 0;
2711                                         break;
2712                                 case RES_QP_MAPPED:
2713                                         if (!valid_reserved(dev, slave, qpn))
2714                                                 __mlx4_qp_free_icm(dev, qpn);
2715                                         state = RES_QP_RESERVED;
2716                                         break;
2717                                 case RES_QP_HW:
2718                                         in_param = slave;
2719                                         err = mlx4_cmd(dev, in_param,
2720                                                        qp->local_qpn, 2,
2721                                                        MLX4_CMD_2RST_QP,
2722                                                        MLX4_CMD_TIME_CLASS_A,
2723                                                        MLX4_CMD_NATIVE);
2724                                         if (err)
2725                                                 mlx4_dbg(dev, "rem_slave_qps: failed"
2726                                                          " to move slave %d qpn %d to"
2727                                                          " reset\n", slave,
2728                                                          qp->local_qpn);
2729                                         atomic_dec(&qp->rcq->ref_count);
2730                                         atomic_dec(&qp->scq->ref_count);
2731                                         atomic_dec(&qp->mtt->ref_count);
2732                                         if (qp->srq)
2733                                                 atomic_dec(&qp->srq->ref_count);
2734                                         state = RES_QP_MAPPED;
2735                                         break;
2736                                 default:
2737                                         state = 0;
2738                                 }
2739                         }
2740                 }
2741                 spin_lock_irq(mlx4_tlock(dev));
2742         }
2743         spin_unlock_irq(mlx4_tlock(dev));
2744 }
2745
2746 static void rem_slave_srqs(struct mlx4_dev *dev, int slave)
2747 {
2748         struct mlx4_priv *priv = mlx4_priv(dev);
2749         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
2750         struct list_head *srq_list =
2751                 &tracker->slave_list[slave].res_list[RES_SRQ];
2752         struct res_srq *srq;
2753         struct res_srq *tmp;
2754         int state;
2755         u64 in_param;
2756         LIST_HEAD(tlist);
2757         int srqn;
2758         int err;
2759
2760         err = move_all_busy(dev, slave, RES_SRQ);
2761         if (err)
2762                 mlx4_warn(dev, "rem_slave_srqs: Could not move all srqs to "
2763                           "busy for slave %d\n", slave);
2764
2765         spin_lock_irq(mlx4_tlock(dev));
2766         list_for_each_entry_safe(srq, tmp, srq_list, com.list) {
2767                 spin_unlock_irq(mlx4_tlock(dev));
2768                 if (srq->com.owner == slave) {
2769                         srqn = srq->com.res_id;
2770                         state = srq->com.from_state;
2771                         while (state != 0) {
2772                                 switch (state) {
2773                                 case RES_SRQ_ALLOCATED:
2774                                         __mlx4_srq_free_icm(dev, srqn);
2775                                         spin_lock_irq(mlx4_tlock(dev));
2776                                         radix_tree_delete(&tracker->res_tree[RES_SRQ],
2777                                                           srqn);
2778                                         list_del(&srq->com.list);
2779                                         spin_unlock_irq(mlx4_tlock(dev));
2780                                         kfree(srq);
2781                                         state = 0;
2782                                         break;
2783
2784                                 case RES_SRQ_HW:
2785                                         in_param = slave;
2786                                         err = mlx4_cmd(dev, in_param, srqn, 1,
2787                                                        MLX4_CMD_HW2SW_SRQ,
2788                                                        MLX4_CMD_TIME_CLASS_A,
2789                                                        MLX4_CMD_NATIVE);
2790                                         if (err)
2791                                                 mlx4_dbg(dev, "rem_slave_srqs: failed"
2792                                                          " to move slave %d srq %d to"
2793                                                          " SW ownership\n",
2794                                                          slave, srqn);
2795
2796                                         atomic_dec(&srq->mtt->ref_count);
2797                                         if (srq->cq)
2798                                                 atomic_dec(&srq->cq->ref_count);
2799                                         state = RES_SRQ_ALLOCATED;
2800                                         break;
2801
2802                                 default:
2803                                         state = 0;
2804                                 }
2805                         }
2806                 }
2807                 spin_lock_irq(mlx4_tlock(dev));
2808         }
2809         spin_unlock_irq(mlx4_tlock(dev));
2810 }
2811
2812 static void rem_slave_cqs(struct mlx4_dev *dev, int slave)
2813 {
2814         struct mlx4_priv *priv = mlx4_priv(dev);
2815         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
2816         struct list_head *cq_list =
2817                 &tracker->slave_list[slave].res_list[RES_CQ];
2818         struct res_cq *cq;
2819         struct res_cq *tmp;
2820         int state;
2821         u64 in_param;
2822         LIST_HEAD(tlist);
2823         int cqn;
2824         int err;
2825
2826         err = move_all_busy(dev, slave, RES_CQ);
2827         if (err)
2828                 mlx4_warn(dev, "rem_slave_cqs: Could not move all cqs to "
2829                           "busy for slave %d\n", slave);
2830
2831         spin_lock_irq(mlx4_tlock(dev));
2832         list_for_each_entry_safe(cq, tmp, cq_list, com.list) {
2833                 spin_unlock_irq(mlx4_tlock(dev));
2834                 if (cq->com.owner == slave && !atomic_read(&cq->ref_count)) {
2835                         cqn = cq->com.res_id;
2836                         state = cq->com.from_state;
2837                         while (state != 0) {
2838                                 switch (state) {
2839                                 case RES_CQ_ALLOCATED:
2840                                         __mlx4_cq_free_icm(dev, cqn);
2841                                         spin_lock_irq(mlx4_tlock(dev));
2842                                         radix_tree_delete(&tracker->res_tree[RES_CQ],
2843                                                           cqn);
2844                                         list_del(&cq->com.list);
2845                                         spin_unlock_irq(mlx4_tlock(dev));
2846                                         kfree(cq);
2847                                         state = 0;
2848                                         break;
2849
2850                                 case RES_CQ_HW:
2851                                         in_param = slave;
2852                                         err = mlx4_cmd(dev, in_param, cqn, 1,
2853                                                        MLX4_CMD_HW2SW_CQ,
2854                                                        MLX4_CMD_TIME_CLASS_A,
2855                                                        MLX4_CMD_NATIVE);
2856                                         if (err)
2857                                                 mlx4_dbg(dev, "rem_slave_cqs: failed"
2858                                                          " to move slave %d cq %d to"
2859                                                          " SW ownership\n",
2860                                                          slave, cqn);
2861                                         atomic_dec(&cq->mtt->ref_count);
2862                                         state = RES_CQ_ALLOCATED;
2863                                         break;
2864
2865                                 default:
2866                                         state = 0;
2867                                 }
2868                         }
2869                 }
2870                 spin_lock_irq(mlx4_tlock(dev));
2871         }
2872         spin_unlock_irq(mlx4_tlock(dev));
2873 }
2874
2875 static void rem_slave_mrs(struct mlx4_dev *dev, int slave)
2876 {
2877         struct mlx4_priv *priv = mlx4_priv(dev);
2878         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
2879         struct list_head *mpt_list =
2880                 &tracker->slave_list[slave].res_list[RES_MPT];
2881         struct res_mpt *mpt;
2882         struct res_mpt *tmp;
2883         int state;
2884         u64 in_param;
2885         LIST_HEAD(tlist);
2886         int mptn;
2887         int err;
2888
2889         err = move_all_busy(dev, slave, RES_MPT);
2890         if (err)
2891                 mlx4_warn(dev, "rem_slave_mrs: Could not move all mpts to "
2892                           "busy for slave %d\n", slave);
2893
2894         spin_lock_irq(mlx4_tlock(dev));
2895         list_for_each_entry_safe(mpt, tmp, mpt_list, com.list) {
2896                 spin_unlock_irq(mlx4_tlock(dev));
2897                 if (mpt->com.owner == slave) {
2898                         mptn = mpt->com.res_id;
2899                         state = mpt->com.from_state;
2900                         while (state != 0) {
2901                                 switch (state) {
2902                                 case RES_MPT_RESERVED:
2903                                         __mlx4_mr_release(dev, mpt->key);
2904                                         spin_lock_irq(mlx4_tlock(dev));
2905                                         radix_tree_delete(&tracker->res_tree[RES_MPT],
2906                                                           mptn);
2907                                         list_del(&mpt->com.list);
2908                                         spin_unlock_irq(mlx4_tlock(dev));
2909                                         kfree(mpt);
2910                                         state = 0;
2911                                         break;
2912
2913                                 case RES_MPT_MAPPED:
2914                                         __mlx4_mr_free_icm(dev, mpt->key);
2915                                         state = RES_MPT_RESERVED;
2916                                         break;
2917
2918                                 case RES_MPT_HW:
2919                                         in_param = slave;
2920                                         err = mlx4_cmd(dev, in_param, mptn, 0,
2921                                                      MLX4_CMD_HW2SW_MPT,
2922                                                      MLX4_CMD_TIME_CLASS_A,
2923                                                      MLX4_CMD_NATIVE);
2924                                         if (err)
2925                                                 mlx4_dbg(dev, "rem_slave_mrs: failed"
2926                                                          " to move slave %d mpt %d to"
2927                                                          " SW ownership\n",
2928                                                          slave, mptn);
2929                                         if (mpt->mtt)
2930                                                 atomic_dec(&mpt->mtt->ref_count);
2931                                         state = RES_MPT_MAPPED;
2932                                         break;
2933                                 default:
2934                                         state = 0;
2935                                 }
2936                         }
2937                 }
2938                 spin_lock_irq(mlx4_tlock(dev));
2939         }
2940         spin_unlock_irq(mlx4_tlock(dev));
2941 }
2942
2943 static void rem_slave_mtts(struct mlx4_dev *dev, int slave)
2944 {
2945         struct mlx4_priv *priv = mlx4_priv(dev);
2946         struct mlx4_resource_tracker *tracker =
2947                 &priv->mfunc.master.res_tracker;
2948         struct list_head *mtt_list =
2949                 &tracker->slave_list[slave].res_list[RES_MTT];
2950         struct res_mtt *mtt;
2951         struct res_mtt *tmp;
2952         int state;
2953         LIST_HEAD(tlist);
2954         int base;
2955         int err;
2956
2957         err = move_all_busy(dev, slave, RES_MTT);
2958         if (err)
2959                 mlx4_warn(dev, "rem_slave_mtts: Could not move all mtts to "
2960                           "busy for slave %d\n", slave);
2961
2962         spin_lock_irq(mlx4_tlock(dev));
2963         list_for_each_entry_safe(mtt, tmp, mtt_list, com.list) {
2964                 spin_unlock_irq(mlx4_tlock(dev));
2965                 if (mtt->com.owner == slave) {
2966                         base = mtt->com.res_id;
2967                         state = mtt->com.from_state;
2968                         while (state != 0) {
2969                                 switch (state) {
2970                                 case RES_MTT_ALLOCATED:
2971                                         __mlx4_free_mtt_range(dev, base,
2972                                                               mtt->order);
2973                                         spin_lock_irq(mlx4_tlock(dev));
2974                                         radix_tree_delete(&tracker->res_tree[RES_MTT],
2975                                                           base);
2976                                         list_del(&mtt->com.list);
2977                                         spin_unlock_irq(mlx4_tlock(dev));
2978                                         kfree(mtt);
2979                                         state = 0;
2980                                         break;
2981
2982                                 default:
2983                                         state = 0;
2984                                 }
2985                         }
2986                 }
2987                 spin_lock_irq(mlx4_tlock(dev));
2988         }
2989         spin_unlock_irq(mlx4_tlock(dev));
2990 }
2991
2992 static void rem_slave_eqs(struct mlx4_dev *dev, int slave)
2993 {
2994         struct mlx4_priv *priv = mlx4_priv(dev);
2995         struct mlx4_resource_tracker *tracker = &priv->mfunc.master.res_tracker;
2996         struct list_head *eq_list =
2997                 &tracker->slave_list[slave].res_list[RES_EQ];
2998         struct res_eq *eq;
2999         struct res_eq *tmp;
3000         int err;
3001         int state;
3002         LIST_HEAD(tlist);
3003         int eqn;
3004         struct mlx4_cmd_mailbox *mailbox;
3005
3006         err = move_all_busy(dev, slave, RES_EQ);
3007         if (err)
3008                 mlx4_warn(dev, "rem_slave_eqs: Could not move all eqs to "
3009                           "busy for slave %d\n", slave);
3010
3011         spin_lock_irq(mlx4_tlock(dev));
3012         list_for_each_entry_safe(eq, tmp, eq_list, com.list) {
3013                 spin_unlock_irq(mlx4_tlock(dev));
3014                 if (eq->com.owner == slave) {
3015                         eqn = eq->com.res_id;
3016                         state = eq->com.from_state;
3017                         while (state != 0) {
3018                                 switch (state) {
3019                                 case RES_EQ_RESERVED:
3020                                         spin_lock_irq(mlx4_tlock(dev));
3021                                         radix_tree_delete(&tracker->res_tree[RES_EQ],
3022                                                           eqn);
3023                                         list_del(&eq->com.list);
3024                                         spin_unlock_irq(mlx4_tlock(dev));
3025                                         kfree(eq);
3026                                         state = 0;
3027                                         break;
3028
3029                                 case RES_EQ_HW:
3030                                         mailbox = mlx4_alloc_cmd_mailbox(dev);
3031                                         if (IS_ERR(mailbox)) {
3032                                                 cond_resched();
3033                                                 continue;
3034                                         }
3035                                         err = mlx4_cmd_box(dev, slave, 0,
3036                                                            eqn & 0xff, 0,
3037                                                            MLX4_CMD_HW2SW_EQ,
3038                                                            MLX4_CMD_TIME_CLASS_A,
3039                                                            MLX4_CMD_NATIVE);
3040                                         mlx4_dbg(dev, "rem_slave_eqs: failed"
3041                                                  " to move slave %d eqs %d to"
3042                                                  " SW ownership\n", slave, eqn);
3043                                         mlx4_free_cmd_mailbox(dev, mailbox);
3044                                         if (!err) {
3045                                                 atomic_dec(&eq->mtt->ref_count);
3046                                                 state = RES_EQ_RESERVED;
3047                                         }
3048                                         break;
3049
3050                                 default:
3051                                         state = 0;
3052                                 }
3053                         }
3054                 }
3055                 spin_lock_irq(mlx4_tlock(dev));
3056         }
3057         spin_unlock_irq(mlx4_tlock(dev));
3058 }
3059
3060 void mlx4_delete_all_resources_for_slave(struct mlx4_dev *dev, int slave)
3061 {
3062         struct mlx4_priv *priv = mlx4_priv(dev);
3063
3064         mutex_lock(&priv->mfunc.master.res_tracker.slave_list[slave].mutex);
3065         /*VLAN*/
3066         rem_slave_macs(dev, slave);
3067         rem_slave_qps(dev, slave);
3068         rem_slave_srqs(dev, slave);
3069         rem_slave_cqs(dev, slave);
3070         rem_slave_mrs(dev, slave);
3071         rem_slave_eqs(dev, slave);
3072         rem_slave_mtts(dev, slave);
3073         mutex_unlock(&priv->mfunc.master.res_tracker.slave_list[slave].mutex);
3074 }