]> Pileus Git - ~andy/linux/blob - drivers/scsi/qla2xxx/qla_inline.h
Merge branch 'for-3.14' of git://git.kernel.org/pub/scm/linux/kernel/git/tj/percpu
[~andy/linux] / drivers / scsi / qla2xxx / qla_inline.h
1 /*
2  * QLogic Fibre Channel HBA Driver
3  * Copyright (c)  2003-2013 QLogic Corporation
4  *
5  * See LICENSE.qla2xxx for copyright and licensing details.
6  */
7
8 /**
9  * qla24xx_calc_iocbs() - Determine number of Command Type 3 and
10  * Continuation Type 1 IOCBs to allocate.
11  *
12  * @dsds: number of data segment decriptors needed
13  *
14  * Returns the number of IOCB entries needed to store @dsds.
15  */
16 static inline uint16_t
17 qla24xx_calc_iocbs(scsi_qla_host_t *vha, uint16_t dsds)
18 {
19         uint16_t iocbs;
20
21         iocbs = 1;
22         if (dsds > 1) {
23                 iocbs += (dsds - 1) / 5;
24                 if ((dsds - 1) % 5)
25                         iocbs++;
26         }
27         return iocbs;
28 }
29
30 /*
31  * qla2x00_debounce_register
32  *      Debounce register.
33  *
34  * Input:
35  *      port = register address.
36  *
37  * Returns:
38  *      register value.
39  */
40 static __inline__ uint16_t
41 qla2x00_debounce_register(volatile uint16_t __iomem *addr)
42 {
43         volatile uint16_t first;
44         volatile uint16_t second;
45
46         do {
47                 first = RD_REG_WORD(addr);
48                 barrier();
49                 cpu_relax();
50                 second = RD_REG_WORD(addr);
51         } while (first != second);
52
53         return (first);
54 }
55
56 static inline void
57 qla2x00_poll(struct rsp_que *rsp)
58 {
59         unsigned long flags;
60         struct qla_hw_data *ha = rsp->hw;
61         local_irq_save(flags);
62         if (IS_P3P_TYPE(ha))
63                 qla82xx_poll(0, rsp);
64         else
65                 ha->isp_ops->intr_handler(0, rsp);
66         local_irq_restore(flags);
67 }
68
69 static inline uint8_t *
70 host_to_fcp_swap(uint8_t *fcp, uint32_t bsize)
71 {
72        uint32_t *ifcp = (uint32_t *) fcp;
73        uint32_t *ofcp = (uint32_t *) fcp;
74        uint32_t iter = bsize >> 2;
75
76        for (; iter ; iter--)
77                *ofcp++ = swab32(*ifcp++);
78
79        return fcp;
80 }
81
82 static inline void
83 host_to_adap(uint8_t *src, uint8_t *dst, uint32_t bsize)
84 {
85         uint32_t *isrc = (uint32_t *) src;
86         __le32 *odest = (__le32 *) dst;
87         uint32_t iter = bsize >> 2;
88
89         for (; iter ; iter--)
90                 *odest++ = cpu_to_le32(*isrc++);
91 }
92
93 static inline void
94 qla2x00_set_reserved_loop_ids(struct qla_hw_data *ha)
95 {
96         int i;
97
98         if (IS_FWI2_CAPABLE(ha))
99                 return;
100
101         for (i = 0; i < SNS_FIRST_LOOP_ID; i++)
102                 set_bit(i, ha->loop_id_map);
103         set_bit(MANAGEMENT_SERVER, ha->loop_id_map);
104         set_bit(BROADCAST, ha->loop_id_map);
105 }
106
107 static inline int
108 qla2x00_is_reserved_id(scsi_qla_host_t *vha, uint16_t loop_id)
109 {
110         struct qla_hw_data *ha = vha->hw;
111         if (IS_FWI2_CAPABLE(ha))
112                 return (loop_id > NPH_LAST_HANDLE);
113
114         return ((loop_id > ha->max_loop_id && loop_id < SNS_FIRST_LOOP_ID) ||
115             loop_id == MANAGEMENT_SERVER || loop_id == BROADCAST);
116 }
117
118 static inline void
119 qla2x00_clear_loop_id(fc_port_t *fcport) {
120         struct qla_hw_data *ha = fcport->vha->hw;
121
122         if (fcport->loop_id == FC_NO_LOOP_ID ||
123             qla2x00_is_reserved_id(fcport->vha, fcport->loop_id))
124                 return;
125
126         clear_bit(fcport->loop_id, ha->loop_id_map);
127         fcport->loop_id = FC_NO_LOOP_ID;
128 }
129
130 static inline void
131 qla2x00_clean_dsd_pool(struct qla_hw_data *ha, srb_t *sp)
132 {
133         struct dsd_dma *dsd_ptr, *tdsd_ptr;
134         struct crc_context *ctx;
135
136         ctx = (struct crc_context *)GET_CMD_CTX_SP(sp);
137
138         /* clean up allocated prev pool */
139         list_for_each_entry_safe(dsd_ptr, tdsd_ptr,
140             &ctx->dsd_list, list) {
141                 dma_pool_free(ha->dl_dma_pool, dsd_ptr->dsd_addr,
142                     dsd_ptr->dsd_list_dma);
143                 list_del(&dsd_ptr->list);
144                 kfree(dsd_ptr);
145         }
146         INIT_LIST_HEAD(&ctx->dsd_list);
147 }
148
149 static inline void
150 qla2x00_set_fcport_state(fc_port_t *fcport, int state)
151 {
152         int old_state;
153
154         old_state = atomic_read(&fcport->state);
155         atomic_set(&fcport->state, state);
156
157         /* Don't print state transitions during initial allocation of fcport */
158         if (old_state && old_state != state) {
159                 ql_dbg(ql_dbg_disc, fcport->vha, 0x207d,
160                     "FCPort state transitioned from %s to %s - "
161                     "portid=%02x%02x%02x.\n",
162                     port_state_str[old_state], port_state_str[state],
163                     fcport->d_id.b.domain, fcport->d_id.b.area,
164                     fcport->d_id.b.al_pa);
165         }
166 }
167
168 static inline int
169 qla2x00_hba_err_chk_enabled(srb_t *sp)
170 {
171         /*
172          * Uncomment when corresponding SCSI changes are done.
173          *
174         if (!sp->cmd->prot_chk)
175                 return 0;
176          *
177          */
178         switch (scsi_get_prot_op(GET_CMD_SP(sp))) {
179         case SCSI_PROT_READ_STRIP:
180         case SCSI_PROT_WRITE_INSERT:
181                 if (ql2xenablehba_err_chk >= 1)
182                         return 1;
183                 break;
184         case SCSI_PROT_READ_PASS:
185         case SCSI_PROT_WRITE_PASS:
186                 if (ql2xenablehba_err_chk >= 2)
187                         return 1;
188                 break;
189         case SCSI_PROT_READ_INSERT:
190         case SCSI_PROT_WRITE_STRIP:
191                 return 1;
192         }
193         return 0;
194 }
195
196 static inline int
197 qla2x00_reset_active(scsi_qla_host_t *vha)
198 {
199         scsi_qla_host_t *base_vha = pci_get_drvdata(vha->hw->pdev);
200
201         /* Test appropriate base-vha and vha flags. */
202         return test_bit(ISP_ABORT_NEEDED, &base_vha->dpc_flags) ||
203             test_bit(ABORT_ISP_ACTIVE, &base_vha->dpc_flags) ||
204             test_bit(ISP_ABORT_RETRY, &base_vha->dpc_flags) ||
205             test_bit(ISP_ABORT_NEEDED, &vha->dpc_flags) ||
206             test_bit(ABORT_ISP_ACTIVE, &vha->dpc_flags);
207 }
208
209 static inline srb_t *
210 qla2x00_get_sp(scsi_qla_host_t *vha, fc_port_t *fcport, gfp_t flag)
211 {
212         srb_t *sp = NULL;
213         struct qla_hw_data *ha = vha->hw;
214         uint8_t bail;
215
216         QLA_VHA_MARK_BUSY(vha, bail);
217         if (unlikely(bail))
218                 return NULL;
219
220         sp = mempool_alloc(ha->srb_mempool, flag);
221         if (!sp)
222                 goto done;
223
224         memset(sp, 0, sizeof(*sp));
225         sp->fcport = fcport;
226         sp->iocbs = 1;
227 done:
228         if (!sp)
229                 QLA_VHA_MARK_NOT_BUSY(vha);
230         return sp;
231 }
232
233 static inline void
234 qla2x00_rel_sp(scsi_qla_host_t *vha, srb_t *sp)
235 {
236         mempool_free(sp, vha->hw->srb_mempool);
237         QLA_VHA_MARK_NOT_BUSY(vha);
238 }
239
240 static inline void
241 qla2x00_init_timer(srb_t *sp, unsigned long tmo)
242 {
243         init_timer(&sp->u.iocb_cmd.timer);
244         sp->u.iocb_cmd.timer.expires = jiffies + tmo * HZ;
245         sp->u.iocb_cmd.timer.data = (unsigned long)sp;
246         sp->u.iocb_cmd.timer.function = qla2x00_sp_timeout;
247         add_timer(&sp->u.iocb_cmd.timer);
248         sp->free = qla2x00_sp_free;
249         if ((IS_QLAFX00(sp->fcport->vha->hw)) &&
250             (sp->type == SRB_FXIOCB_DCMD))
251                 init_completion(&sp->u.iocb_cmd.u.fxiocb.fxiocb_comp);
252 }
253
254 static inline int
255 qla2x00_gid_list_size(struct qla_hw_data *ha)
256 {
257         if (IS_QLAFX00(ha))
258                 return sizeof(uint32_t) * 32;
259         else
260                 return sizeof(struct gid_list_info) * ha->max_fibre_devices;
261 }
262
263 static inline void
264 qla2x00_do_host_ramp_up(scsi_qla_host_t *vha)
265 {
266         if (vha->hw->cfg_lun_q_depth >= ql2xmaxqdepth)
267                 return;
268
269         /* Wait at least HOST_QUEUE_RAMPDOWN_INTERVAL before ramping up */
270         if (time_before(jiffies, (vha->hw->host_last_rampdown_time +
271             HOST_QUEUE_RAMPDOWN_INTERVAL)))
272                 return;
273
274         /* Wait at least HOST_QUEUE_RAMPUP_INTERVAL between each ramp up */
275         if (time_before(jiffies, (vha->hw->host_last_rampup_time +
276             HOST_QUEUE_RAMPUP_INTERVAL)))
277                 return;
278
279         set_bit(HOST_RAMP_UP_QUEUE_DEPTH, &vha->dpc_flags);
280 }
281
282 static inline void
283 qla2x00_handle_mbx_completion(struct qla_hw_data *ha, int status)
284 {
285         if (test_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags) &&
286             (status & MBX_INTERRUPT) && ha->flags.mbox_int) {
287                 set_bit(MBX_INTERRUPT, &ha->mbx_cmd_flags);
288                 clear_bit(MBX_INTR_WAIT, &ha->mbx_cmd_flags);
289                 complete(&ha->mbx_intr_comp);
290         }
291 }