1 /* SPDX-License-Identifier: GPL-2.0-or-later */
2
3 #ifndef _UFSHCD_PRIV_H_
4 #define _UFSHCD_PRIV_H_
5
6 #include <linux/pm_runtime.h>
7 #include <ufs/ufshcd.h>
8
ufshcd_is_user_access_allowed(struct ufs_hba * hba)9 static inline bool ufshcd_is_user_access_allowed(struct ufs_hba *hba)
10 {
11 return !hba->shutting_down;
12 }
13
14 void ufshcd_schedule_eh_work(struct ufs_hba *hba);
15
ufshcd_keep_autobkops_enabled_except_suspend(struct ufs_hba * hba)16 static inline bool ufshcd_keep_autobkops_enabled_except_suspend(
17 struct ufs_hba *hba)
18 {
19 return hba->caps & UFSHCD_CAP_KEEP_AUTO_BKOPS_ENABLED_EXCEPT_SUSPEND;
20 }
21
ufshcd_wb_get_query_index(struct ufs_hba * hba)22 static inline u8 ufshcd_wb_get_query_index(struct ufs_hba *hba)
23 {
24 if (hba->dev_info.wb_buffer_type == WB_BUF_MODE_LU_DEDICATED)
25 return hba->dev_info.wb_dedicated_lu;
26 return 0;
27 }
28
ufshcd_is_wb_buf_flush_allowed(struct ufs_hba * hba)29 static inline bool ufshcd_is_wb_buf_flush_allowed(struct ufs_hba *hba)
30 {
31 return ufshcd_is_wb_allowed(hba) &&
32 !(hba->quirks & UFSHCI_QUIRK_SKIP_MANUAL_WB_FLUSH_CTRL);
33 }
34
35 #ifdef CONFIG_SCSI_UFS_HWMON
36 void ufs_hwmon_probe(struct ufs_hba *hba, u8 mask);
37 void ufs_hwmon_remove(struct ufs_hba *hba);
38 void ufs_hwmon_notify_event(struct ufs_hba *hba, u8 ee_mask);
39 #else
ufs_hwmon_probe(struct ufs_hba * hba,u8 mask)40 static inline void ufs_hwmon_probe(struct ufs_hba *hba, u8 mask) {}
ufs_hwmon_remove(struct ufs_hba * hba)41 static inline void ufs_hwmon_remove(struct ufs_hba *hba) {}
ufs_hwmon_notify_event(struct ufs_hba * hba,u8 ee_mask)42 static inline void ufs_hwmon_notify_event(struct ufs_hba *hba, u8 ee_mask) {}
43 #endif
44
45 int ufshcd_query_descriptor_retry(struct ufs_hba *hba,
46 enum query_opcode opcode,
47 enum desc_idn idn, u8 index,
48 u8 selector,
49 u8 *desc_buf, int *buf_len);
50 int ufshcd_read_desc_param(struct ufs_hba *hba,
51 enum desc_idn desc_id,
52 int desc_index,
53 u8 param_offset,
54 u8 *param_read_buf,
55 u8 param_size);
56 int ufshcd_query_attr_retry(struct ufs_hba *hba, enum query_opcode opcode,
57 enum attr_idn idn, u8 index, u8 selector,
58 u32 *attr_val);
59 int ufshcd_query_attr(struct ufs_hba *hba, enum query_opcode opcode,
60 enum attr_idn idn, u8 index, u8 selector, u32 *attr_val);
61 int ufshcd_query_flag(struct ufs_hba *hba, enum query_opcode opcode,
62 enum flag_idn idn, u8 index, bool *flag_res);
63 void ufshcd_auto_hibern8_update(struct ufs_hba *hba, u32 ahit);
64 void ufshcd_compl_one_cqe(struct ufs_hba *hba, int task_tag,
65 struct cq_entry *cqe);
66 int ufshcd_mcq_init(struct ufs_hba *hba);
67 int ufshcd_mcq_decide_queue_depth(struct ufs_hba *hba);
68 int ufshcd_mcq_memory_alloc(struct ufs_hba *hba);
69 void ufshcd_mcq_make_queues_operational(struct ufs_hba *hba);
70 void ufshcd_mcq_config_mac(struct ufs_hba *hba, u32 max_active_cmds);
71 void ufshcd_mcq_select_mcq_mode(struct ufs_hba *hba);
72 u32 ufshcd_mcq_read_cqis(struct ufs_hba *hba, int i);
73 void ufshcd_mcq_write_cqis(struct ufs_hba *hba, u32 val, int i);
74 unsigned long ufshcd_mcq_poll_cqe_nolock(struct ufs_hba *hba,
75 struct ufs_hw_queue *hwq);
76 struct ufs_hw_queue *ufshcd_mcq_req_to_hwq(struct ufs_hba *hba,
77 struct request *req);
78 unsigned long ufshcd_mcq_poll_cqe_lock(struct ufs_hba *hba,
79 struct ufs_hw_queue *hwq);
80
81 #define UFSHCD_MCQ_IO_QUEUE_OFFSET 1
82 #define SD_ASCII_STD true
83 #define SD_RAW false
84 int ufshcd_read_string_desc(struct ufs_hba *hba, u8 desc_index,
85 u8 **buf, bool ascii);
86
87 int ufshcd_hold(struct ufs_hba *hba, bool async);
88 void ufshcd_release(struct ufs_hba *hba);
89
90 int ufshcd_send_uic_cmd(struct ufs_hba *hba, struct uic_command *uic_cmd);
91
92 int ufshcd_exec_raw_upiu_cmd(struct ufs_hba *hba,
93 struct utp_upiu_req *req_upiu,
94 struct utp_upiu_req *rsp_upiu,
95 int msgcode,
96 u8 *desc_buff, int *buff_len,
97 enum query_opcode desc_op);
98
99 int ufshcd_wb_toggle(struct ufs_hba *hba, bool enable);
100
101 /* Wrapper functions for safely calling variant operations */
ufshcd_get_var_name(struct ufs_hba * hba)102 static inline const char *ufshcd_get_var_name(struct ufs_hba *hba)
103 {
104 if (hba->vops)
105 return hba->vops->name;
106 return "";
107 }
108
ufshcd_vops_exit(struct ufs_hba * hba)109 static inline void ufshcd_vops_exit(struct ufs_hba *hba)
110 {
111 if (hba->vops && hba->vops->exit)
112 return hba->vops->exit(hba);
113 }
114
ufshcd_vops_get_ufs_hci_version(struct ufs_hba * hba)115 static inline u32 ufshcd_vops_get_ufs_hci_version(struct ufs_hba *hba)
116 {
117 if (hba->vops && hba->vops->get_ufs_hci_version)
118 return hba->vops->get_ufs_hci_version(hba);
119
120 return ufshcd_readl(hba, REG_UFS_VERSION);
121 }
122
ufshcd_vops_clk_scale_notify(struct ufs_hba * hba,bool up,enum ufs_notify_change_status status)123 static inline int ufshcd_vops_clk_scale_notify(struct ufs_hba *hba,
124 bool up, enum ufs_notify_change_status status)
125 {
126 if (hba->vops && hba->vops->clk_scale_notify)
127 return hba->vops->clk_scale_notify(hba, up, status);
128 return 0;
129 }
130
ufshcd_vops_event_notify(struct ufs_hba * hba,enum ufs_event_type evt,void * data)131 static inline void ufshcd_vops_event_notify(struct ufs_hba *hba,
132 enum ufs_event_type evt,
133 void *data)
134 {
135 if (hba->vops && hba->vops->event_notify)
136 hba->vops->event_notify(hba, evt, data);
137 }
138
ufshcd_vops_setup_clocks(struct ufs_hba * hba,bool on,enum ufs_notify_change_status status)139 static inline int ufshcd_vops_setup_clocks(struct ufs_hba *hba, bool on,
140 enum ufs_notify_change_status status)
141 {
142 if (hba->vops && hba->vops->setup_clocks)
143 return hba->vops->setup_clocks(hba, on, status);
144 return 0;
145 }
146
ufshcd_vops_hce_enable_notify(struct ufs_hba * hba,bool status)147 static inline int ufshcd_vops_hce_enable_notify(struct ufs_hba *hba,
148 bool status)
149 {
150 if (hba->vops && hba->vops->hce_enable_notify)
151 return hba->vops->hce_enable_notify(hba, status);
152
153 return 0;
154 }
ufshcd_vops_link_startup_notify(struct ufs_hba * hba,bool status)155 static inline int ufshcd_vops_link_startup_notify(struct ufs_hba *hba,
156 bool status)
157 {
158 if (hba->vops && hba->vops->link_startup_notify)
159 return hba->vops->link_startup_notify(hba, status);
160
161 return 0;
162 }
163
ufshcd_vops_pwr_change_notify(struct ufs_hba * hba,enum ufs_notify_change_status status,struct ufs_pa_layer_attr * dev_max_params,struct ufs_pa_layer_attr * dev_req_params)164 static inline int ufshcd_vops_pwr_change_notify(struct ufs_hba *hba,
165 enum ufs_notify_change_status status,
166 struct ufs_pa_layer_attr *dev_max_params,
167 struct ufs_pa_layer_attr *dev_req_params)
168 {
169 if (hba->vops && hba->vops->pwr_change_notify)
170 return hba->vops->pwr_change_notify(hba, status,
171 dev_max_params, dev_req_params);
172
173 return -ENOTSUPP;
174 }
175
ufshcd_vops_setup_task_mgmt(struct ufs_hba * hba,int tag,u8 tm_function)176 static inline void ufshcd_vops_setup_task_mgmt(struct ufs_hba *hba,
177 int tag, u8 tm_function)
178 {
179 if (hba->vops && hba->vops->setup_task_mgmt)
180 return hba->vops->setup_task_mgmt(hba, tag, tm_function);
181 }
182
ufshcd_vops_hibern8_notify(struct ufs_hba * hba,enum uic_cmd_dme cmd,enum ufs_notify_change_status status)183 static inline void ufshcd_vops_hibern8_notify(struct ufs_hba *hba,
184 enum uic_cmd_dme cmd,
185 enum ufs_notify_change_status status)
186 {
187 if (hba->vops && hba->vops->hibern8_notify)
188 return hba->vops->hibern8_notify(hba, cmd, status);
189 }
190
ufshcd_vops_apply_dev_quirks(struct ufs_hba * hba)191 static inline int ufshcd_vops_apply_dev_quirks(struct ufs_hba *hba)
192 {
193 if (hba->vops && hba->vops->apply_dev_quirks)
194 return hba->vops->apply_dev_quirks(hba);
195 return 0;
196 }
197
ufshcd_vops_fixup_dev_quirks(struct ufs_hba * hba)198 static inline void ufshcd_vops_fixup_dev_quirks(struct ufs_hba *hba)
199 {
200 if (hba->vops && hba->vops->fixup_dev_quirks)
201 hba->vops->fixup_dev_quirks(hba);
202 }
203
ufshcd_vops_suspend(struct ufs_hba * hba,enum ufs_pm_op op,enum ufs_notify_change_status status)204 static inline int ufshcd_vops_suspend(struct ufs_hba *hba, enum ufs_pm_op op,
205 enum ufs_notify_change_status status)
206 {
207 if (hba->vops && hba->vops->suspend)
208 return hba->vops->suspend(hba, op, status);
209
210 return 0;
211 }
212
ufshcd_vops_resume(struct ufs_hba * hba,enum ufs_pm_op op)213 static inline int ufshcd_vops_resume(struct ufs_hba *hba, enum ufs_pm_op op)
214 {
215 if (hba->vops && hba->vops->resume)
216 return hba->vops->resume(hba, op);
217
218 return 0;
219 }
220
ufshcd_vops_dbg_register_dump(struct ufs_hba * hba)221 static inline void ufshcd_vops_dbg_register_dump(struct ufs_hba *hba)
222 {
223 if (hba->vops && hba->vops->dbg_register_dump)
224 hba->vops->dbg_register_dump(hba);
225 }
226
ufshcd_vops_device_reset(struct ufs_hba * hba)227 static inline int ufshcd_vops_device_reset(struct ufs_hba *hba)
228 {
229 if (hba->vops && hba->vops->device_reset)
230 return hba->vops->device_reset(hba);
231
232 return -EOPNOTSUPP;
233 }
234
ufshcd_vops_config_scaling_param(struct ufs_hba * hba,struct devfreq_dev_profile * p,struct devfreq_simple_ondemand_data * data)235 static inline void ufshcd_vops_config_scaling_param(struct ufs_hba *hba,
236 struct devfreq_dev_profile *p,
237 struct devfreq_simple_ondemand_data *data)
238 {
239 if (hba->vops && hba->vops->config_scaling_param)
240 hba->vops->config_scaling_param(hba, p, data);
241 }
242
ufshcd_vops_reinit_notify(struct ufs_hba * hba)243 static inline void ufshcd_vops_reinit_notify(struct ufs_hba *hba)
244 {
245 if (hba->vops && hba->vops->reinit_notify)
246 hba->vops->reinit_notify(hba);
247 }
248
ufshcd_vops_mcq_config_resource(struct ufs_hba * hba)249 static inline int ufshcd_vops_mcq_config_resource(struct ufs_hba *hba)
250 {
251 if (hba->vops && hba->vops->mcq_config_resource)
252 return hba->vops->mcq_config_resource(hba);
253
254 return -EOPNOTSUPP;
255 }
256
ufshcd_mcq_vops_get_hba_mac(struct ufs_hba * hba)257 static inline int ufshcd_mcq_vops_get_hba_mac(struct ufs_hba *hba)
258 {
259 if (hba->vops && hba->vops->get_hba_mac)
260 return hba->vops->get_hba_mac(hba);
261
262 return -EOPNOTSUPP;
263 }
264
ufshcd_mcq_vops_op_runtime_config(struct ufs_hba * hba)265 static inline int ufshcd_mcq_vops_op_runtime_config(struct ufs_hba *hba)
266 {
267 if (hba->vops && hba->vops->op_runtime_config)
268 return hba->vops->op_runtime_config(hba);
269
270 return -EOPNOTSUPP;
271 }
272
ufshcd_vops_get_outstanding_cqs(struct ufs_hba * hba,unsigned long * ocqs)273 static inline int ufshcd_vops_get_outstanding_cqs(struct ufs_hba *hba,
274 unsigned long *ocqs)
275 {
276 if (hba->vops && hba->vops->get_outstanding_cqs)
277 return hba->vops->get_outstanding_cqs(hba, ocqs);
278
279 return -EOPNOTSUPP;
280 }
281
ufshcd_mcq_vops_config_esi(struct ufs_hba * hba)282 static inline int ufshcd_mcq_vops_config_esi(struct ufs_hba *hba)
283 {
284 if (hba->vops && hba->vops->config_esi)
285 return hba->vops->config_esi(hba);
286
287 return -EOPNOTSUPP;
288 }
289
290 extern const struct ufs_pm_lvl_states ufs_pm_lvl_states[];
291
292 /**
293 * ufshcd_scsi_to_upiu_lun - maps scsi LUN to UPIU LUN
294 * @scsi_lun: scsi LUN id
295 *
296 * Returns UPIU LUN id
297 */
ufshcd_scsi_to_upiu_lun(unsigned int scsi_lun)298 static inline u8 ufshcd_scsi_to_upiu_lun(unsigned int scsi_lun)
299 {
300 if (scsi_is_wlun(scsi_lun))
301 return (scsi_lun & UFS_UPIU_MAX_UNIT_NUM_ID)
302 | UFS_UPIU_WLUN_ID;
303 else
304 return scsi_lun & UFS_UPIU_MAX_UNIT_NUM_ID;
305 }
306
307 int __ufshcd_write_ee_control(struct ufs_hba *hba, u32 ee_ctrl_mask);
308 int ufshcd_write_ee_control(struct ufs_hba *hba);
309 int ufshcd_update_ee_control(struct ufs_hba *hba, u16 *mask,
310 const u16 *other_mask, u16 set, u16 clr);
311
ufshcd_update_ee_drv_mask(struct ufs_hba * hba,u16 set,u16 clr)312 static inline int ufshcd_update_ee_drv_mask(struct ufs_hba *hba,
313 u16 set, u16 clr)
314 {
315 return ufshcd_update_ee_control(hba, &hba->ee_drv_mask,
316 &hba->ee_usr_mask, set, clr);
317 }
318
ufshcd_update_ee_usr_mask(struct ufs_hba * hba,u16 set,u16 clr)319 static inline int ufshcd_update_ee_usr_mask(struct ufs_hba *hba,
320 u16 set, u16 clr)
321 {
322 return ufshcd_update_ee_control(hba, &hba->ee_usr_mask,
323 &hba->ee_drv_mask, set, clr);
324 }
325
ufshcd_rpm_get_sync(struct ufs_hba * hba)326 static inline int ufshcd_rpm_get_sync(struct ufs_hba *hba)
327 {
328 return pm_runtime_get_sync(&hba->ufs_device_wlun->sdev_gendev);
329 }
330
ufshcd_rpm_put_sync(struct ufs_hba * hba)331 static inline int ufshcd_rpm_put_sync(struct ufs_hba *hba)
332 {
333 return pm_runtime_put_sync(&hba->ufs_device_wlun->sdev_gendev);
334 }
335
ufshcd_rpm_get_noresume(struct ufs_hba * hba)336 static inline void ufshcd_rpm_get_noresume(struct ufs_hba *hba)
337 {
338 pm_runtime_get_noresume(&hba->ufs_device_wlun->sdev_gendev);
339 }
340
ufshcd_rpm_resume(struct ufs_hba * hba)341 static inline int ufshcd_rpm_resume(struct ufs_hba *hba)
342 {
343 return pm_runtime_resume(&hba->ufs_device_wlun->sdev_gendev);
344 }
345
ufshcd_rpm_put(struct ufs_hba * hba)346 static inline int ufshcd_rpm_put(struct ufs_hba *hba)
347 {
348 return pm_runtime_put(&hba->ufs_device_wlun->sdev_gendev);
349 }
350
351 /**
352 * ufs_is_valid_unit_desc_lun - checks if the given LUN has a unit descriptor
353 * @dev_info: pointer of instance of struct ufs_dev_info
354 * @lun: LU number to check
355 * @return: true if the lun has a matching unit descriptor, false otherwise
356 */
ufs_is_valid_unit_desc_lun(struct ufs_dev_info * dev_info,u8 lun)357 static inline bool ufs_is_valid_unit_desc_lun(struct ufs_dev_info *dev_info, u8 lun)
358 {
359 if (!dev_info || !dev_info->max_lu_supported) {
360 pr_err("Max General LU supported by UFS isn't initialized\n");
361 return false;
362 }
363 return lun == UFS_UPIU_RPMB_WLUN || (lun < dev_info->max_lu_supported);
364 }
365
ufshcd_inc_sq_tail(struct ufs_hw_queue * q)366 static inline void ufshcd_inc_sq_tail(struct ufs_hw_queue *q)
367 {
368 u32 mask = q->max_entries - 1;
369 u32 val;
370
371 q->sq_tail_slot = (q->sq_tail_slot + 1) & mask;
372 val = q->sq_tail_slot * sizeof(struct utp_transfer_req_desc);
373 writel(val, q->mcq_sq_tail);
374 }
375
ufshcd_mcq_update_cq_tail_slot(struct ufs_hw_queue * q)376 static inline void ufshcd_mcq_update_cq_tail_slot(struct ufs_hw_queue *q)
377 {
378 u32 val = readl(q->mcq_cq_tail);
379
380 q->cq_tail_slot = val / sizeof(struct cq_entry);
381 }
382
ufshcd_mcq_is_cq_empty(struct ufs_hw_queue * q)383 static inline bool ufshcd_mcq_is_cq_empty(struct ufs_hw_queue *q)
384 {
385 return q->cq_head_slot == q->cq_tail_slot;
386 }
387
ufshcd_mcq_inc_cq_head_slot(struct ufs_hw_queue * q)388 static inline void ufshcd_mcq_inc_cq_head_slot(struct ufs_hw_queue *q)
389 {
390 q->cq_head_slot++;
391 if (q->cq_head_slot == q->max_entries)
392 q->cq_head_slot = 0;
393 }
394
ufshcd_mcq_update_cq_head(struct ufs_hw_queue * q)395 static inline void ufshcd_mcq_update_cq_head(struct ufs_hw_queue *q)
396 {
397 writel(q->cq_head_slot * sizeof(struct cq_entry), q->mcq_cq_head);
398 }
399
ufshcd_mcq_cur_cqe(struct ufs_hw_queue * q)400 static inline struct cq_entry *ufshcd_mcq_cur_cqe(struct ufs_hw_queue *q)
401 {
402 struct cq_entry *cqe = q->cqe_base_addr;
403
404 return cqe + q->cq_head_slot;
405 }
406 #endif /* _UFSHCD_PRIV_H_ */
407