Lines Matching refs:cl
341 cb->cl->tx_cb_queued++; in mei_tx_cb_enqueue()
353 if (!WARN_ON(cb->cl->tx_cb_queued == 0)) in mei_tx_cb_dequeue()
354 cb->cl->tx_cb_queued--; in mei_tx_cb_dequeue()
367 static void mei_cl_set_read_by_fp(const struct mei_cl *cl, in mei_cl_set_read_by_fp() argument
372 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_set_read_by_fp()
389 static struct mei_cl_cb *mei_io_cb_init(struct mei_cl *cl, in mei_io_cb_init() argument
401 cb->cl = cl; in mei_io_cb_init()
417 const struct mei_cl *cl) in mei_io_list_flush_cl() argument
422 if (cl == cb->cl) { in mei_io_list_flush_cl()
438 const struct mei_cl *cl, in mei_io_tx_list_free_cl() argument
444 if (cl == cb->cl && (!fp || fp == cb->fp)) in mei_io_tx_list_free_cl()
469 static void mei_cl_free_pending(struct mei_cl *cl) in mei_cl_free_pending() argument
473 cb = list_first_entry_or_null(&cl->rd_pending, struct mei_cl_cb, list); in mei_cl_free_pending()
487 struct mei_cl_cb *mei_cl_alloc_cb(struct mei_cl *cl, size_t length, in mei_cl_alloc_cb() argument
493 cb = mei_io_cb_init(cl, fop_type, fp); in mei_cl_alloc_cb()
522 struct mei_cl_cb *mei_cl_enqueue_ctrl_wr_cb(struct mei_cl *cl, size_t length, in mei_cl_enqueue_ctrl_wr_cb() argument
530 length = max_t(size_t, length, mei_cl_mtu(cl)); in mei_cl_enqueue_ctrl_wr_cb()
532 cb = mei_cl_alloc_cb(cl, length, fop_type, fp); in mei_cl_enqueue_ctrl_wr_cb()
536 list_add_tail(&cb->list, &cl->dev->ctrl_wr_list); in mei_cl_enqueue_ctrl_wr_cb()
549 struct mei_cl_cb *mei_cl_read_cb(struct mei_cl *cl, const struct file *fp) in mei_cl_read_cb() argument
554 spin_lock(&cl->rd_completed_lock); in mei_cl_read_cb()
555 list_for_each_entry(cb, &cl->rd_completed, list) in mei_cl_read_cb()
560 spin_unlock(&cl->rd_completed_lock); in mei_cl_read_cb()
572 int mei_cl_flush_queues(struct mei_cl *cl, const struct file *fp) in mei_cl_flush_queues() argument
576 if (WARN_ON(!cl || !cl->dev)) in mei_cl_flush_queues()
579 dev = cl->dev; in mei_cl_flush_queues()
581 cl_dbg(dev, cl, "remove list entry belonging to cl\n"); in mei_cl_flush_queues()
582 mei_io_tx_list_free_cl(&cl->dev->write_list, cl, fp); in mei_cl_flush_queues()
583 mei_io_tx_list_free_cl(&cl->dev->write_waiting_list, cl, fp); in mei_cl_flush_queues()
586 mei_io_list_flush_cl(&cl->dev->ctrl_wr_list, cl); in mei_cl_flush_queues()
587 mei_io_list_flush_cl(&cl->dev->ctrl_rd_list, cl); in mei_cl_flush_queues()
588 mei_cl_free_pending(cl); in mei_cl_flush_queues()
590 spin_lock(&cl->rd_completed_lock); in mei_cl_flush_queues()
591 mei_io_list_free_fp(&cl->rd_completed, fp); in mei_cl_flush_queues()
592 spin_unlock(&cl->rd_completed_lock); in mei_cl_flush_queues()
603 static void mei_cl_init(struct mei_cl *cl, struct mei_device *dev) in mei_cl_init() argument
605 memset(cl, 0, sizeof(*cl)); in mei_cl_init()
606 init_waitqueue_head(&cl->wait); in mei_cl_init()
607 init_waitqueue_head(&cl->rx_wait); in mei_cl_init()
608 init_waitqueue_head(&cl->tx_wait); in mei_cl_init()
609 init_waitqueue_head(&cl->ev_wait); in mei_cl_init()
610 INIT_LIST_HEAD(&cl->vtag_map); in mei_cl_init()
611 spin_lock_init(&cl->rd_completed_lock); in mei_cl_init()
612 INIT_LIST_HEAD(&cl->rd_completed); in mei_cl_init()
613 INIT_LIST_HEAD(&cl->rd_pending); in mei_cl_init()
614 INIT_LIST_HEAD(&cl->link); in mei_cl_init()
615 cl->writing_state = MEI_IDLE; in mei_cl_init()
616 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_init()
617 cl->dev = dev; in mei_cl_init()
628 struct mei_cl *cl; in mei_cl_allocate() local
630 cl = kmalloc(sizeof(*cl), GFP_KERNEL); in mei_cl_allocate()
631 if (!cl) in mei_cl_allocate()
634 mei_cl_init(cl, dev); in mei_cl_allocate()
636 return cl; in mei_cl_allocate()
648 int mei_cl_link(struct mei_cl *cl) in mei_cl_link() argument
653 if (WARN_ON(!cl || !cl->dev)) in mei_cl_link()
656 dev = cl->dev; in mei_cl_link()
672 cl->host_client_id = id; in mei_cl_link()
673 list_add_tail(&cl->link, &dev->file_list); in mei_cl_link()
677 cl->state = MEI_FILE_INITIALIZING; in mei_cl_link()
679 cl_dbg(dev, cl, "link cl\n"); in mei_cl_link()
690 int mei_cl_unlink(struct mei_cl *cl) in mei_cl_unlink() argument
695 if (!cl) in mei_cl_unlink()
698 if (WARN_ON(!cl->dev)) in mei_cl_unlink()
701 dev = cl->dev; in mei_cl_unlink()
703 cl_dbg(dev, cl, "unlink client"); in mei_cl_unlink()
705 if (cl->state == MEI_FILE_UNINITIALIZED) in mei_cl_unlink()
712 if (cl->host_client_id) in mei_cl_unlink()
713 clear_bit(cl->host_client_id, dev->host_clients_map); in mei_cl_unlink()
715 list_del_init(&cl->link); in mei_cl_unlink()
717 cl->state = MEI_FILE_UNINITIALIZED; in mei_cl_unlink()
718 cl->writing_state = MEI_IDLE; in mei_cl_unlink()
720 WARN_ON(!list_empty(&cl->rd_completed) || in mei_cl_unlink()
721 !list_empty(&cl->rd_pending) || in mei_cl_unlink()
722 !list_empty(&cl->link)); in mei_cl_unlink()
769 static void mei_cl_wake_all(struct mei_cl *cl) in mei_cl_wake_all() argument
771 struct mei_device *dev = cl->dev; in mei_cl_wake_all()
774 if (waitqueue_active(&cl->rx_wait)) { in mei_cl_wake_all()
775 cl_dbg(dev, cl, "Waking up reading client!\n"); in mei_cl_wake_all()
776 wake_up_interruptible(&cl->rx_wait); in mei_cl_wake_all()
779 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_wake_all()
780 cl_dbg(dev, cl, "Waking up writing client!\n"); in mei_cl_wake_all()
781 wake_up_interruptible(&cl->tx_wait); in mei_cl_wake_all()
784 if (waitqueue_active(&cl->ev_wait)) { in mei_cl_wake_all()
785 cl_dbg(dev, cl, "Waking up waiting for event clients!\n"); in mei_cl_wake_all()
786 wake_up_interruptible(&cl->ev_wait); in mei_cl_wake_all()
789 if (waitqueue_active(&cl->wait)) { in mei_cl_wake_all()
790 cl_dbg(dev, cl, "Waking up ctrl write clients!\n"); in mei_cl_wake_all()
791 wake_up(&cl->wait); in mei_cl_wake_all()
801 static void mei_cl_set_disconnected(struct mei_cl *cl) in mei_cl_set_disconnected() argument
803 struct mei_device *dev = cl->dev; in mei_cl_set_disconnected()
805 if (cl->state == MEI_FILE_DISCONNECTED || in mei_cl_set_disconnected()
806 cl->state <= MEI_FILE_INITIALIZING) in mei_cl_set_disconnected()
809 cl->state = MEI_FILE_DISCONNECTED; in mei_cl_set_disconnected()
810 mei_io_tx_list_free_cl(&dev->write_list, cl, NULL); in mei_cl_set_disconnected()
811 mei_io_tx_list_free_cl(&dev->write_waiting_list, cl, NULL); in mei_cl_set_disconnected()
812 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_set_disconnected()
813 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_set_disconnected()
814 mei_cl_wake_all(cl); in mei_cl_set_disconnected()
815 cl->rx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
816 cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
817 cl->timer_count = 0; in mei_cl_set_disconnected()
819 if (!cl->me_cl) in mei_cl_set_disconnected()
822 if (!WARN_ON(cl->me_cl->connect_count == 0)) in mei_cl_set_disconnected()
823 cl->me_cl->connect_count--; in mei_cl_set_disconnected()
825 if (cl->me_cl->connect_count == 0) in mei_cl_set_disconnected()
826 cl->me_cl->tx_flow_ctrl_creds = 0; in mei_cl_set_disconnected()
828 mei_me_cl_put(cl->me_cl); in mei_cl_set_disconnected()
829 cl->me_cl = NULL; in mei_cl_set_disconnected()
832 static int mei_cl_set_connecting(struct mei_cl *cl, struct mei_me_client *me_cl) in mei_cl_set_connecting() argument
845 cl->me_cl = me_cl; in mei_cl_set_connecting()
846 cl->state = MEI_FILE_CONNECTING; in mei_cl_set_connecting()
847 cl->me_cl->connect_count++; in mei_cl_set_connecting()
860 static int mei_cl_send_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_disconnect() argument
865 dev = cl->dev; in mei_cl_send_disconnect()
867 ret = mei_hbm_cl_disconnect_req(dev, cl); in mei_cl_send_disconnect()
868 cl->status = ret; in mei_cl_send_disconnect()
870 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_disconnect()
875 cl->timer_count = dev->timeouts.connect; in mei_cl_send_disconnect()
891 int mei_cl_irq_disconnect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_disconnect() argument
894 struct mei_device *dev = cl->dev; in mei_cl_irq_disconnect()
907 ret = mei_cl_send_disconnect(cl, cb); in mei_cl_irq_disconnect()
922 static int __mei_cl_disconnect(struct mei_cl *cl) in __mei_cl_disconnect() argument
928 dev = cl->dev; in __mei_cl_disconnect()
930 cl->state = MEI_FILE_DISCONNECTING; in __mei_cl_disconnect()
932 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DISCONNECT, NULL); in __mei_cl_disconnect()
939 rets = mei_cl_send_disconnect(cl, cb); in __mei_cl_disconnect()
941 cl_err(dev, cl, "failed to disconnect.\n"); in __mei_cl_disconnect()
947 wait_event_timeout(cl->wait, in __mei_cl_disconnect()
948 cl->state == MEI_FILE_DISCONNECT_REPLY || in __mei_cl_disconnect()
949 cl->state == MEI_FILE_DISCONNECTED, in __mei_cl_disconnect()
953 rets = cl->status; in __mei_cl_disconnect()
954 if (cl->state != MEI_FILE_DISCONNECT_REPLY && in __mei_cl_disconnect()
955 cl->state != MEI_FILE_DISCONNECTED) { in __mei_cl_disconnect()
956 cl_dbg(dev, cl, "timeout on disconnect from FW client.\n"); in __mei_cl_disconnect()
962 mei_cl_set_disconnected(cl); in __mei_cl_disconnect()
964 cl_dbg(dev, cl, "successfully disconnected from FW client.\n"); in __mei_cl_disconnect()
979 int mei_cl_disconnect(struct mei_cl *cl) in mei_cl_disconnect() argument
984 if (WARN_ON(!cl || !cl->dev)) in mei_cl_disconnect()
987 dev = cl->dev; in mei_cl_disconnect()
989 cl_dbg(dev, cl, "disconnecting"); in mei_cl_disconnect()
991 if (!mei_cl_is_connected(cl)) in mei_cl_disconnect()
994 if (mei_cl_is_fixed_address(cl)) { in mei_cl_disconnect()
995 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
1001 cl_dbg(dev, cl, "Device is powering down, don't bother with disconnection\n"); in mei_cl_disconnect()
1002 mei_cl_set_disconnected(cl); in mei_cl_disconnect()
1009 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_disconnect()
1013 rets = __mei_cl_disconnect(cl); in mei_cl_disconnect()
1015 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_disconnect()
1031 static bool mei_cl_is_other_connecting(struct mei_cl *cl) in mei_cl_is_other_connecting() argument
1036 dev = cl->dev; in mei_cl_is_other_connecting()
1040 mei_cl_me_id(cl) == mei_cl_me_id(cb->cl)) in mei_cl_is_other_connecting()
1055 static int mei_cl_send_connect(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_send_connect() argument
1060 dev = cl->dev; in mei_cl_send_connect()
1062 ret = mei_hbm_cl_connect_req(dev, cl); in mei_cl_send_connect()
1063 cl->status = ret; in mei_cl_send_connect()
1065 cl->state = MEI_FILE_DISCONNECT_REPLY; in mei_cl_send_connect()
1070 cl->timer_count = dev->timeouts.connect; in mei_cl_send_connect()
1084 int mei_cl_irq_connect(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_connect() argument
1087 struct mei_device *dev = cl->dev; in mei_cl_irq_connect()
1092 if (mei_cl_is_other_connecting(cl)) in mei_cl_irq_connect()
1103 rets = mei_cl_send_connect(cl, cb); in mei_cl_irq_connect()
1121 int mei_cl_connect(struct mei_cl *cl, struct mei_me_client *me_cl, in mei_cl_connect() argument
1128 if (WARN_ON(!cl || !cl->dev || !me_cl)) in mei_cl_connect()
1131 dev = cl->dev; in mei_cl_connect()
1133 rets = mei_cl_set_connecting(cl, me_cl); in mei_cl_connect()
1137 if (mei_cl_is_fixed_address(cl)) { in mei_cl_connect()
1138 cl->state = MEI_FILE_CONNECTED; in mei_cl_connect()
1146 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_connect()
1150 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_CONNECT, fp); in mei_cl_connect()
1157 if (!mei_cl_is_other_connecting(cl) && mei_hbuf_acquire(dev)) { in mei_cl_connect()
1158 rets = mei_cl_send_connect(cl, cb); in mei_cl_connect()
1164 wait_event_timeout(cl->wait, in mei_cl_connect()
1165 (cl->state == MEI_FILE_CONNECTED || in mei_cl_connect()
1166 cl->state == MEI_FILE_DISCONNECTED || in mei_cl_connect()
1167 cl->state == MEI_FILE_DISCONNECT_REQUIRED || in mei_cl_connect()
1168 cl->state == MEI_FILE_DISCONNECT_REPLY), in mei_cl_connect()
1172 if (!mei_cl_is_connected(cl)) { in mei_cl_connect()
1173 if (cl->state == MEI_FILE_DISCONNECT_REQUIRED) { in mei_cl_connect()
1174 mei_io_list_flush_cl(&dev->ctrl_rd_list, cl); in mei_cl_connect()
1175 mei_io_list_flush_cl(&dev->ctrl_wr_list, cl); in mei_cl_connect()
1179 __mei_cl_disconnect(cl); in mei_cl_connect()
1185 if (!cl->status) in mei_cl_connect()
1186 cl->status = -EFAULT; in mei_cl_connect()
1189 rets = cl->status; in mei_cl_connect()
1191 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_connect()
1198 if (!mei_cl_is_connected(cl)) in mei_cl_connect()
1199 mei_cl_set_disconnected(cl); in mei_cl_connect()
1213 struct mei_cl *cl; in mei_cl_alloc_linked() local
1216 cl = mei_cl_allocate(dev); in mei_cl_alloc_linked()
1217 if (!cl) { in mei_cl_alloc_linked()
1222 ret = mei_cl_link(cl); in mei_cl_alloc_linked()
1226 return cl; in mei_cl_alloc_linked()
1228 kfree(cl); in mei_cl_alloc_linked()
1239 static int mei_cl_tx_flow_ctrl_creds(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds() argument
1241 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds()
1244 if (cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1247 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds()
1250 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds()
1251 if (cl->me_cl->tx_flow_ctrl_creds > 0) in mei_cl_tx_flow_ctrl_creds()
1267 static int mei_cl_tx_flow_ctrl_creds_reduce(struct mei_cl *cl) in mei_cl_tx_flow_ctrl_creds_reduce() argument
1269 if (WARN_ON(!cl || !cl->me_cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1272 if (mei_cl_is_fixed_address(cl)) in mei_cl_tx_flow_ctrl_creds_reduce()
1275 if (mei_cl_is_single_recv_buf(cl)) { in mei_cl_tx_flow_ctrl_creds_reduce()
1276 if (WARN_ON(cl->me_cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1278 cl->me_cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1280 if (WARN_ON(cl->tx_flow_ctrl_creds <= 0)) in mei_cl_tx_flow_ctrl_creds_reduce()
1282 cl->tx_flow_ctrl_creds--; in mei_cl_tx_flow_ctrl_creds_reduce()
1322 const struct file *mei_cl_fp_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_fp_by_vtag() argument
1326 list_for_each_entry(vtag_l, &cl->vtag_map, list) in mei_cl_fp_by_vtag()
1328 if ((cl->cldev && mei_cldev_enabled(cl->cldev)) || in mei_cl_fp_by_vtag()
1341 static void mei_cl_reset_read_by_vtag(const struct mei_cl *cl, u8 vtag) in mei_cl_reset_read_by_vtag() argument
1345 list_for_each_entry(vtag_l, &cl->vtag_map, list) { in mei_cl_reset_read_by_vtag()
1359 static void mei_cl_read_vtag_add_fc(struct mei_cl *cl) in mei_cl_read_vtag_add_fc() argument
1363 list_for_each_entry(cl_vtag, &cl->vtag_map, list) { in mei_cl_read_vtag_add_fc()
1365 if (mei_cl_enqueue_ctrl_wr_cb(cl, in mei_cl_read_vtag_add_fc()
1366 mei_cl_mtu(cl), in mei_cl_read_vtag_add_fc()
1369 cl->rx_flow_ctrl_creds++; in mei_cl_read_vtag_add_fc()
1384 int mei_cl_vt_support_check(const struct mei_cl *cl) in mei_cl_vt_support_check() argument
1386 struct mei_device *dev = cl->dev; in mei_cl_vt_support_check()
1391 if (!cl->me_cl) in mei_cl_vt_support_check()
1394 return cl->me_cl->props.vt_supported ? 0 : -EOPNOTSUPP; in mei_cl_vt_support_check()
1405 void mei_cl_add_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_add_rd_completed() argument
1409 if (!mei_cl_vt_support_check(cl)) { in mei_cl_add_rd_completed()
1410 fp = mei_cl_fp_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1417 mei_cl_reset_read_by_vtag(cl, cb->vtag); in mei_cl_add_rd_completed()
1418 mei_cl_read_vtag_add_fc(cl); in mei_cl_add_rd_completed()
1421 spin_lock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1422 list_add_tail(&cb->list, &cl->rd_completed); in mei_cl_add_rd_completed()
1423 spin_unlock(&cl->rd_completed_lock); in mei_cl_add_rd_completed()
1433 void mei_cl_del_rd_completed(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_del_rd_completed() argument
1435 spin_lock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1437 spin_unlock(&cl->rd_completed_lock); in mei_cl_del_rd_completed()
1479 int mei_cl_irq_notify(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_notify() argument
1482 struct mei_device *dev = cl->dev; in mei_cl_irq_notify()
1497 ret = mei_hbm_cl_notify_req(dev, cl, request); in mei_cl_irq_notify()
1499 cl->status = ret; in mei_cl_irq_notify()
1519 int mei_cl_notify_request(struct mei_cl *cl, in mei_cl_notify_request() argument
1527 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_request()
1530 dev = cl->dev; in mei_cl_notify_request()
1533 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_request()
1537 if (!mei_cl_is_connected(cl)) in mei_cl_notify_request()
1543 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_notify_request()
1548 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, fop_type, fp); in mei_cl_notify_request()
1555 if (mei_hbm_cl_notify_req(dev, cl, request)) { in mei_cl_notify_request()
1563 wait_event_timeout(cl->wait, in mei_cl_notify_request()
1564 cl->notify_en == request || in mei_cl_notify_request()
1565 cl->status || in mei_cl_notify_request()
1566 !mei_cl_is_connected(cl), in mei_cl_notify_request()
1570 if (cl->notify_en != request && !cl->status) in mei_cl_notify_request()
1571 cl->status = -EFAULT; in mei_cl_notify_request()
1573 rets = cl->status; in mei_cl_notify_request()
1576 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_notify_request()
1591 void mei_cl_notify(struct mei_cl *cl) in mei_cl_notify() argument
1595 if (!cl || !cl->dev) in mei_cl_notify()
1598 dev = cl->dev; in mei_cl_notify()
1600 if (!cl->notify_en) in mei_cl_notify()
1603 cl_dbg(dev, cl, "notify event"); in mei_cl_notify()
1604 cl->notify_ev = true; in mei_cl_notify()
1605 if (!mei_cl_bus_notify_event(cl)) in mei_cl_notify()
1606 wake_up_interruptible(&cl->ev_wait); in mei_cl_notify()
1608 if (cl->ev_async) in mei_cl_notify()
1609 kill_fasync(&cl->ev_async, SIGIO, POLL_PRI); in mei_cl_notify()
1624 int mei_cl_notify_get(struct mei_cl *cl, bool block, bool *notify_ev) in mei_cl_notify_get() argument
1631 if (WARN_ON(!cl || !cl->dev)) in mei_cl_notify_get()
1634 dev = cl->dev; in mei_cl_notify_get()
1637 cl_dbg(dev, cl, "notifications not supported\n"); in mei_cl_notify_get()
1641 if (!mei_cl_is_connected(cl)) in mei_cl_notify_get()
1644 if (cl->notify_ev) in mei_cl_notify_get()
1651 rets = wait_event_interruptible(cl->ev_wait, cl->notify_ev); in mei_cl_notify_get()
1658 *notify_ev = cl->notify_ev; in mei_cl_notify_get()
1659 cl->notify_ev = false; in mei_cl_notify_get()
1672 int mei_cl_read_start(struct mei_cl *cl, size_t length, const struct file *fp) in mei_cl_read_start() argument
1678 if (WARN_ON(!cl || !cl->dev)) in mei_cl_read_start()
1681 dev = cl->dev; in mei_cl_read_start()
1683 if (!mei_cl_is_connected(cl)) in mei_cl_read_start()
1686 if (!mei_me_cl_is_active(cl->me_cl)) { in mei_cl_read_start()
1687 cl_err(dev, cl, "no such me client\n"); in mei_cl_read_start()
1691 if (mei_cl_is_fixed_address(cl)) in mei_cl_read_start()
1695 if (cl->rx_flow_ctrl_creds) { in mei_cl_read_start()
1696 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1700 cb = mei_cl_enqueue_ctrl_wr_cb(cl, length, MEI_FOP_READ, fp); in mei_cl_read_start()
1704 mei_cl_set_read_by_fp(cl, fp); in mei_cl_read_start()
1709 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_read_start()
1715 rets = mei_hbm_cl_flow_control_req(dev, cl); in mei_cl_read_start()
1719 list_move_tail(&cb->list, &cl->rd_pending); in mei_cl_read_start()
1721 cl->rx_flow_ctrl_creds++; in mei_cl_read_start()
1724 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_read_start()
1776 is_hbm = cb->cl->me_cl->client_id == 0; in mei_msg_hdr_init()
1777 is_gsc = ((!is_hbm) && cb->cl->dev->hbm_f_gsc_supported && mei_ext_hdr_is_gsc(cb->ext_hdr)); in mei_msg_hdr_init()
1798 mei_hdr->host_addr = mei_cl_host_addr(cb->cl); in mei_msg_hdr_init()
1799 mei_hdr->me_addr = mei_cl_me_id(cb->cl); in mei_msg_hdr_init()
1836 int mei_cl_irq_write(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_write() argument
1853 if (WARN_ON(!cl || !cl->dev)) in mei_cl_irq_write()
1856 dev = cl->dev; in mei_cl_irq_write()
1862 rets = first_chunk ? mei_cl_tx_flow_ctrl_creds(cl) : 1; in mei_cl_irq_write()
1867 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_irq_write()
1927 cl->status = 0; in mei_cl_irq_write()
1928 cl->writing_state = MEI_WRITING; in mei_cl_irq_write()
1932 if (mei_cl_tx_flow_ctrl_creds_reduce(cl)) { in mei_cl_irq_write()
1946 cl->status = rets; in mei_cl_irq_write()
1963 ssize_t mei_cl_write(struct mei_cl *cl, struct mei_cl_cb *cb, unsigned long timeout) in mei_cl_write() argument
1979 if (WARN_ON(!cl || !cl->dev)) in mei_cl_write()
1985 dev = cl->dev; in mei_cl_write()
1990 cl_dbg(dev, cl, "buf_len=%zd\n", buf_len); in mei_cl_write()
1998 cl_err(dev, cl, "rpm: get failed %zd\n", rets); in mei_cl_write()
2003 cl->writing_state = MEI_IDLE; in mei_cl_write()
2006 rets = mei_cl_tx_flow_ctrl_creds(cl); in mei_cl_write()
2020 cl_dbg(dev, cl, "No flow control credentials: not sending.\n"); in mei_cl_write()
2026 cl_dbg(dev, cl, "Cannot acquire the host buffer: not sending.\n"); in mei_cl_write()
2068 rets = mei_cl_tx_flow_ctrl_creds_reduce(cl); in mei_cl_write()
2072 cl->writing_state = MEI_WRITING; in mei_cl_write()
2084 if (blocking && cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2087 rets = wait_event_interruptible_timeout(cl->tx_wait, in mei_cl_write()
2088 cl->writing_state == MEI_WRITE_COMPLETE || in mei_cl_write()
2089 (!mei_cl_is_connected(cl)), in mei_cl_write()
2095 mei_io_tx_list_free_cl(&dev->write_list, cl, NULL); in mei_cl_write()
2096 mei_io_tx_list_free_cl(&dev->write_waiting_list, cl, NULL); in mei_cl_write()
2106 if (cl->writing_state != MEI_WRITE_COMPLETE) { in mei_cl_write()
2114 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_write()
2131 void mei_cl_complete(struct mei_cl *cl, struct mei_cl_cb *cb) in mei_cl_complete() argument
2133 struct mei_device *dev = cl->dev; in mei_cl_complete()
2138 cl->writing_state = MEI_WRITE_COMPLETE; in mei_cl_complete()
2139 if (waitqueue_active(&cl->tx_wait)) { in mei_cl_complete()
2140 wake_up_interruptible(&cl->tx_wait); in mei_cl_complete()
2148 mei_cl_add_rd_completed(cl, cb); in mei_cl_complete()
2149 if (!mei_cl_is_fixed_address(cl) && in mei_cl_complete()
2150 !WARN_ON(!cl->rx_flow_ctrl_creds)) in mei_cl_complete()
2151 cl->rx_flow_ctrl_creds--; in mei_cl_complete()
2152 if (!mei_cl_bus_rx_event(cl)) in mei_cl_complete()
2153 wake_up_interruptible(&cl->rx_wait); in mei_cl_complete()
2162 if (waitqueue_active(&cl->wait)) in mei_cl_complete()
2163 wake_up(&cl->wait); in mei_cl_complete()
2168 mei_cl_set_disconnected(cl); in mei_cl_complete()
2183 struct mei_cl *cl; in mei_cl_all_disconnect() local
2185 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_all_disconnect()
2186 mei_cl_set_disconnected(cl); in mei_cl_all_disconnect()
2192 struct mei_cl *cl; in mei_cl_dma_map_find() local
2194 list_for_each_entry(cl, &dev->file_list, link) in mei_cl_dma_map_find()
2195 if (cl->dma.buffer_id == buffer_id) in mei_cl_dma_map_find()
2196 return cl; in mei_cl_dma_map_find()
2209 int mei_cl_irq_dma_map(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_map() argument
2212 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_map()
2225 ret = mei_hbm_cl_dma_map_req(dev, cl); in mei_cl_irq_dma_map()
2227 cl->status = ret; in mei_cl_irq_dma_map()
2245 int mei_cl_irq_dma_unmap(struct mei_cl *cl, struct mei_cl_cb *cb, in mei_cl_irq_dma_unmap() argument
2248 struct mei_device *dev = cl->dev; in mei_cl_irq_dma_unmap()
2261 ret = mei_hbm_cl_dma_unmap_req(dev, cl); in mei_cl_irq_dma_unmap()
2263 cl->status = ret; in mei_cl_irq_dma_unmap()
2272 static int mei_cl_dma_alloc(struct mei_cl *cl, u8 buf_id, size_t size) in mei_cl_dma_alloc() argument
2274 cl->dma.vaddr = dmam_alloc_coherent(cl->dev->dev, size, in mei_cl_dma_alloc()
2275 &cl->dma.daddr, GFP_KERNEL); in mei_cl_dma_alloc()
2276 if (!cl->dma.vaddr) in mei_cl_dma_alloc()
2279 cl->dma.buffer_id = buf_id; in mei_cl_dma_alloc()
2280 cl->dma.size = size; in mei_cl_dma_alloc()
2285 static void mei_cl_dma_free(struct mei_cl *cl) in mei_cl_dma_free() argument
2287 cl->dma.buffer_id = 0; in mei_cl_dma_free()
2288 dmam_free_coherent(cl->dev->dev, in mei_cl_dma_free()
2289 cl->dma.size, cl->dma.vaddr, cl->dma.daddr); in mei_cl_dma_free()
2290 cl->dma.size = 0; in mei_cl_dma_free()
2291 cl->dma.vaddr = NULL; in mei_cl_dma_free()
2292 cl->dma.daddr = 0; in mei_cl_dma_free()
2312 int mei_cl_dma_alloc_and_map(struct mei_cl *cl, const struct file *fp, in mei_cl_dma_alloc_and_map() argument
2319 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_alloc_and_map()
2322 dev = cl->dev; in mei_cl_dma_alloc_and_map()
2325 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_alloc_and_map()
2332 if (mei_cl_is_connected(cl)) in mei_cl_dma_alloc_and_map()
2335 if (cl->dma_mapped) in mei_cl_dma_alloc_and_map()
2339 cl_dbg(dev, cl, "client dma with id %d is already allocated\n", in mei_cl_dma_alloc_and_map()
2340 cl->dma.buffer_id); in mei_cl_dma_alloc_and_map()
2347 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_alloc_and_map()
2351 rets = mei_cl_dma_alloc(cl, buffer_id, size); in mei_cl_dma_alloc_and_map()
2357 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_MAP, fp); in mei_cl_dma_alloc_and_map()
2364 if (mei_hbm_cl_dma_map_req(dev, cl)) { in mei_cl_dma_alloc_and_map()
2371 cl->status = 0; in mei_cl_dma_alloc_and_map()
2374 wait_event_timeout(cl->wait, in mei_cl_dma_alloc_and_map()
2375 cl->dma_mapped || cl->status, in mei_cl_dma_alloc_and_map()
2379 if (!cl->dma_mapped && !cl->status) in mei_cl_dma_alloc_and_map()
2380 cl->status = -EFAULT; in mei_cl_dma_alloc_and_map()
2382 rets = cl->status; in mei_cl_dma_alloc_and_map()
2386 mei_cl_dma_free(cl); in mei_cl_dma_alloc_and_map()
2388 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_alloc_and_map()
2406 int mei_cl_dma_unmap(struct mei_cl *cl, const struct file *fp) in mei_cl_dma_unmap() argument
2412 if (WARN_ON(!cl || !cl->dev)) in mei_cl_dma_unmap()
2415 dev = cl->dev; in mei_cl_dma_unmap()
2418 cl_dbg(dev, cl, "client dma is not supported\n"); in mei_cl_dma_unmap()
2423 if (mei_cl_is_connected(cl)) in mei_cl_dma_unmap()
2426 if (!cl->dma_mapped) in mei_cl_dma_unmap()
2432 cl_err(dev, cl, "rpm: get failed %d\n", rets); in mei_cl_dma_unmap()
2436 cb = mei_cl_enqueue_ctrl_wr_cb(cl, 0, MEI_FOP_DMA_UNMAP, fp); in mei_cl_dma_unmap()
2443 if (mei_hbm_cl_dma_unmap_req(dev, cl)) { in mei_cl_dma_unmap()
2450 cl->status = 0; in mei_cl_dma_unmap()
2453 wait_event_timeout(cl->wait, in mei_cl_dma_unmap()
2454 !cl->dma_mapped || cl->status, in mei_cl_dma_unmap()
2458 if (cl->dma_mapped && !cl->status) in mei_cl_dma_unmap()
2459 cl->status = -EFAULT; in mei_cl_dma_unmap()
2461 rets = cl->status; in mei_cl_dma_unmap()
2464 mei_cl_dma_free(cl); in mei_cl_dma_unmap()
2466 cl_dbg(dev, cl, "rpm: autosuspend\n"); in mei_cl_dma_unmap()