1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * Copyright (c) 2021 MediaTek Inc.
4  * Author: Yunfei Dong <yunfei.dong@mediatek.com>
5  */
6 
7 #include <linux/freezer.h>
8 #include <linux/interrupt.h>
9 #include <linux/kthread.h>
10 
11 #include "mtk_vcodec_dec_pm.h"
12 #include "mtk_vcodec_drv.h"
13 #include "vdec_msg_queue.h"
14 
15 #define VDEC_MSG_QUEUE_TIMEOUT_MS 1500
16 
17 /* the size used to store lat slice header information */
18 #define VDEC_LAT_SLICE_HEADER_SZ    (640 * SZ_1K)
19 
20 /* the size used to store avc error information */
21 #define VDEC_ERR_MAP_SZ_AVC         (17 * SZ_1K)
22 
23 /* core will read the trans buffer which decoded by lat to decode again.
24  * The trans buffer size of FHD and 4K bitstreams are different.
25  */
vde_msg_queue_get_trans_size(int width,int height)26 static int vde_msg_queue_get_trans_size(int width, int height)
27 {
28 	if (width > 1920 || height > 1088)
29 		return 30 * SZ_1M;
30 	else
31 		return 6 * SZ_1M;
32 }
33 
vdec_msg_queue_init_ctx(struct vdec_msg_queue_ctx * ctx,int hardware_index)34 void vdec_msg_queue_init_ctx(struct vdec_msg_queue_ctx *ctx, int hardware_index)
35 {
36 	init_waitqueue_head(&ctx->ready_to_use);
37 	INIT_LIST_HEAD(&ctx->ready_queue);
38 	spin_lock_init(&ctx->ready_lock);
39 	ctx->ready_num = 0;
40 	ctx->hardware_index = hardware_index;
41 }
42 
vdec_get_buf_list(int hardware_index,struct vdec_lat_buf * buf)43 static struct list_head *vdec_get_buf_list(int hardware_index, struct vdec_lat_buf *buf)
44 {
45 	switch (hardware_index) {
46 	case MTK_VDEC_CORE:
47 		return &buf->core_list;
48 	case MTK_VDEC_LAT0:
49 		return &buf->lat_list;
50 	default:
51 		return NULL;
52 	}
53 }
54 
vdec_msg_queue_qbuf(struct vdec_msg_queue_ctx * msg_ctx,struct vdec_lat_buf * buf)55 int vdec_msg_queue_qbuf(struct vdec_msg_queue_ctx *msg_ctx, struct vdec_lat_buf *buf)
56 {
57 	struct list_head *head;
58 
59 	head = vdec_get_buf_list(msg_ctx->hardware_index, buf);
60 	if (!head) {
61 		mtk_v4l2_err("fail to qbuf: %d", msg_ctx->hardware_index);
62 		return -EINVAL;
63 	}
64 
65 	spin_lock(&msg_ctx->ready_lock);
66 	list_add_tail(head, &msg_ctx->ready_queue);
67 	msg_ctx->ready_num++;
68 
69 	if (msg_ctx->hardware_index != MTK_VDEC_CORE)
70 		wake_up_all(&msg_ctx->ready_to_use);
71 	else
72 		queue_work(buf->ctx->dev->core_workqueue,
73 			   &buf->ctx->msg_queue.core_work);
74 
75 	mtk_v4l2_debug(3, "enqueue buf type: %d addr: 0x%p num: %d",
76 		       msg_ctx->hardware_index, buf, msg_ctx->ready_num);
77 	spin_unlock(&msg_ctx->ready_lock);
78 
79 	return 0;
80 }
81 
vdec_msg_queue_wait_event(struct vdec_msg_queue_ctx * msg_ctx)82 static bool vdec_msg_queue_wait_event(struct vdec_msg_queue_ctx *msg_ctx)
83 {
84 	int ret;
85 
86 	ret = wait_event_timeout(msg_ctx->ready_to_use,
87 				 !list_empty(&msg_ctx->ready_queue),
88 				 msecs_to_jiffies(VDEC_MSG_QUEUE_TIMEOUT_MS));
89 	if (!ret)
90 		return false;
91 
92 	return true;
93 }
94 
vdec_msg_queue_dqbuf(struct vdec_msg_queue_ctx * msg_ctx)95 struct vdec_lat_buf *vdec_msg_queue_dqbuf(struct vdec_msg_queue_ctx *msg_ctx)
96 {
97 	struct vdec_lat_buf *buf;
98 	struct list_head *head;
99 	int ret;
100 
101 	spin_lock(&msg_ctx->ready_lock);
102 	if (list_empty(&msg_ctx->ready_queue)) {
103 		mtk_v4l2_debug(3, "queue is NULL, type:%d num: %d",
104 			       msg_ctx->hardware_index, msg_ctx->ready_num);
105 		spin_unlock(&msg_ctx->ready_lock);
106 
107 		if (msg_ctx->hardware_index == MTK_VDEC_CORE)
108 			return NULL;
109 
110 		ret = vdec_msg_queue_wait_event(msg_ctx);
111 		if (!ret)
112 			return NULL;
113 		spin_lock(&msg_ctx->ready_lock);
114 	}
115 
116 	if (msg_ctx->hardware_index == MTK_VDEC_CORE)
117 		buf = list_first_entry(&msg_ctx->ready_queue,
118 				       struct vdec_lat_buf, core_list);
119 	else
120 		buf = list_first_entry(&msg_ctx->ready_queue,
121 				       struct vdec_lat_buf, lat_list);
122 
123 	head = vdec_get_buf_list(msg_ctx->hardware_index, buf);
124 	if (!head) {
125 		spin_unlock(&msg_ctx->ready_lock);
126 		mtk_v4l2_err("fail to dqbuf: %d", msg_ctx->hardware_index);
127 		return NULL;
128 	}
129 	list_del(head);
130 
131 	msg_ctx->ready_num--;
132 	mtk_v4l2_debug(3, "dqueue buf type:%d addr: 0x%p num: %d",
133 		       msg_ctx->hardware_index, buf, msg_ctx->ready_num);
134 	spin_unlock(&msg_ctx->ready_lock);
135 
136 	return buf;
137 }
138 
vdec_msg_queue_update_ube_rptr(struct vdec_msg_queue * msg_queue,uint64_t ube_rptr)139 void vdec_msg_queue_update_ube_rptr(struct vdec_msg_queue *msg_queue, uint64_t ube_rptr)
140 {
141 	spin_lock(&msg_queue->lat_ctx.ready_lock);
142 	msg_queue->wdma_rptr_addr = ube_rptr;
143 	mtk_v4l2_debug(3, "update ube rprt (0x%llx)", ube_rptr);
144 	spin_unlock(&msg_queue->lat_ctx.ready_lock);
145 }
146 
vdec_msg_queue_update_ube_wptr(struct vdec_msg_queue * msg_queue,uint64_t ube_wptr)147 void vdec_msg_queue_update_ube_wptr(struct vdec_msg_queue *msg_queue, uint64_t ube_wptr)
148 {
149 	spin_lock(&msg_queue->lat_ctx.ready_lock);
150 	msg_queue->wdma_wptr_addr = ube_wptr;
151 	mtk_v4l2_debug(3, "update ube wprt: (0x%llx 0x%llx) offset: 0x%llx",
152 		       msg_queue->wdma_rptr_addr, msg_queue->wdma_wptr_addr,
153 		       ube_wptr);
154 	spin_unlock(&msg_queue->lat_ctx.ready_lock);
155 }
156 
vdec_msg_queue_wait_lat_buf_full(struct vdec_msg_queue * msg_queue)157 bool vdec_msg_queue_wait_lat_buf_full(struct vdec_msg_queue *msg_queue)
158 {
159 	long timeout_jiff;
160 	int ret;
161 
162 	timeout_jiff = msecs_to_jiffies(1000 * (NUM_BUFFER_COUNT + 2));
163 	ret = wait_event_timeout(msg_queue->lat_ctx.ready_to_use,
164 				 msg_queue->lat_ctx.ready_num == NUM_BUFFER_COUNT,
165 				 timeout_jiff);
166 	if (ret) {
167 		mtk_v4l2_debug(3, "success to get lat buf: %d",
168 			       msg_queue->lat_ctx.ready_num);
169 		return true;
170 	}
171 	mtk_v4l2_err("failed with lat buf isn't full: %d",
172 		     msg_queue->lat_ctx.ready_num);
173 	return false;
174 }
175 
vdec_msg_queue_deinit(struct vdec_msg_queue * msg_queue,struct mtk_vcodec_ctx * ctx)176 void vdec_msg_queue_deinit(struct vdec_msg_queue *msg_queue,
177 			   struct mtk_vcodec_ctx *ctx)
178 {
179 	struct vdec_lat_buf *lat_buf;
180 	struct mtk_vcodec_mem *mem;
181 	int i;
182 
183 	mem = &msg_queue->wdma_addr;
184 	if (mem->va)
185 		mtk_vcodec_mem_free(ctx, mem);
186 	for (i = 0; i < NUM_BUFFER_COUNT; i++) {
187 		lat_buf = &msg_queue->lat_buf[i];
188 
189 		mem = &lat_buf->wdma_err_addr;
190 		if (mem->va)
191 			mtk_vcodec_mem_free(ctx, mem);
192 
193 		mem = &lat_buf->slice_bc_addr;
194 		if (mem->va)
195 			mtk_vcodec_mem_free(ctx, mem);
196 
197 		kfree(lat_buf->private_data);
198 	}
199 }
200 
vdec_msg_queue_core_work(struct work_struct * work)201 static void vdec_msg_queue_core_work(struct work_struct *work)
202 {
203 	struct vdec_msg_queue *msg_queue =
204 		container_of(work, struct vdec_msg_queue, core_work);
205 	struct mtk_vcodec_ctx *ctx =
206 		container_of(msg_queue, struct mtk_vcodec_ctx, msg_queue);
207 	struct mtk_vcodec_dev *dev = ctx->dev;
208 	struct vdec_lat_buf *lat_buf;
209 
210 	lat_buf = vdec_msg_queue_dqbuf(&dev->msg_queue_core_ctx);
211 	if (!lat_buf)
212 		return;
213 
214 	ctx = lat_buf->ctx;
215 	mtk_vcodec_dec_enable_hardware(ctx, MTK_VDEC_CORE);
216 	mtk_vcodec_set_curr_ctx(dev, ctx, MTK_VDEC_CORE);
217 
218 	lat_buf->core_decode(lat_buf);
219 
220 	mtk_vcodec_set_curr_ctx(dev, NULL, MTK_VDEC_CORE);
221 	mtk_vcodec_dec_disable_hardware(ctx, MTK_VDEC_CORE);
222 	vdec_msg_queue_qbuf(&ctx->msg_queue.lat_ctx, lat_buf);
223 
224 	if (!list_empty(&dev->msg_queue_core_ctx.ready_queue)) {
225 		mtk_v4l2_debug(3, "re-schedule to decode for core: %d",
226 			       dev->msg_queue_core_ctx.ready_num);
227 		queue_work(dev->core_workqueue, &msg_queue->core_work);
228 	}
229 }
230 
vdec_msg_queue_init(struct vdec_msg_queue * msg_queue,struct mtk_vcodec_ctx * ctx,core_decode_cb_t core_decode,int private_size)231 int vdec_msg_queue_init(struct vdec_msg_queue *msg_queue,
232 			struct mtk_vcodec_ctx *ctx, core_decode_cb_t core_decode,
233 			int private_size)
234 {
235 	struct vdec_lat_buf *lat_buf;
236 	int i, err;
237 
238 	/* already init msg queue */
239 	if (msg_queue->wdma_addr.size)
240 		return 0;
241 
242 	vdec_msg_queue_init_ctx(&msg_queue->lat_ctx, MTK_VDEC_LAT0);
243 	INIT_WORK(&msg_queue->core_work, vdec_msg_queue_core_work);
244 	msg_queue->wdma_addr.size =
245 		vde_msg_queue_get_trans_size(ctx->picinfo.buf_w,
246 					     ctx->picinfo.buf_h);
247 
248 	err = mtk_vcodec_mem_alloc(ctx, &msg_queue->wdma_addr);
249 	if (err) {
250 		mtk_v4l2_err("failed to allocate wdma_addr buf");
251 		return -ENOMEM;
252 	}
253 	msg_queue->wdma_rptr_addr = msg_queue->wdma_addr.dma_addr;
254 	msg_queue->wdma_wptr_addr = msg_queue->wdma_addr.dma_addr;
255 
256 	for (i = 0; i < NUM_BUFFER_COUNT; i++) {
257 		lat_buf = &msg_queue->lat_buf[i];
258 
259 		lat_buf->wdma_err_addr.size = VDEC_ERR_MAP_SZ_AVC;
260 		err = mtk_vcodec_mem_alloc(ctx, &lat_buf->wdma_err_addr);
261 		if (err) {
262 			mtk_v4l2_err("failed to allocate wdma_err_addr buf[%d]", i);
263 			goto mem_alloc_err;
264 		}
265 
266 		lat_buf->slice_bc_addr.size = VDEC_LAT_SLICE_HEADER_SZ;
267 		err = mtk_vcodec_mem_alloc(ctx, &lat_buf->slice_bc_addr);
268 		if (err) {
269 			mtk_v4l2_err("failed to allocate wdma_addr buf[%d]", i);
270 			goto mem_alloc_err;
271 		}
272 
273 		lat_buf->private_data = kzalloc(private_size, GFP_KERNEL);
274 		if (!lat_buf->private_data) {
275 			err = -ENOMEM;
276 			goto mem_alloc_err;
277 		}
278 
279 		lat_buf->ctx = ctx;
280 		lat_buf->core_decode = core_decode;
281 		err = vdec_msg_queue_qbuf(&msg_queue->lat_ctx, lat_buf);
282 		if (err) {
283 			mtk_v4l2_err("failed to qbuf buf[%d]", i);
284 			goto mem_alloc_err;
285 		}
286 	}
287 	return 0;
288 
289 mem_alloc_err:
290 	vdec_msg_queue_deinit(msg_queue, ctx);
291 	return err;
292 }
293