1 // SPDX-License-Identifier: GPL-2.0-only
2 /******************************************************************************
3
4 AudioScience HPI driver
5 Copyright (C) 1997-2014 AudioScience Inc. <support@audioscience.com>
6
7
8 Extended Message Function With Response Caching
9
10 (C) Copyright AudioScience Inc. 2002
11 *****************************************************************************/
12 #define SOURCEFILE_NAME "hpimsgx.c"
13 #include "hpi_internal.h"
14 #include "hpi_version.h"
15 #include "hpimsginit.h"
16 #include "hpicmn.h"
17 #include "hpimsgx.h"
18 #include "hpidebug.h"
19
20 static const struct pci_device_id asihpi_pci_tbl[] = {
21 #include "hpipcida.h"
22 };
23
24 static struct hpios_spinlock msgx_lock;
25
26 static hpi_handler_func *hpi_entry_points[HPI_MAX_ADAPTERS];
27 static int logging_enabled = 1;
28
hpi_lookup_entry_point_function(const struct hpi_pci * pci_info)29 static hpi_handler_func *hpi_lookup_entry_point_function(const struct hpi_pci
30 *pci_info)
31 {
32
33 int i;
34
35 for (i = 0; asihpi_pci_tbl[i].vendor != 0; i++) {
36 if (asihpi_pci_tbl[i].vendor != PCI_ANY_ID
37 && asihpi_pci_tbl[i].vendor !=
38 pci_info->pci_dev->vendor)
39 continue;
40 if (asihpi_pci_tbl[i].device != PCI_ANY_ID
41 && asihpi_pci_tbl[i].device !=
42 pci_info->pci_dev->device)
43 continue;
44 if (asihpi_pci_tbl[i].subvendor != PCI_ANY_ID
45 && asihpi_pci_tbl[i].subvendor !=
46 pci_info->pci_dev->subsystem_vendor)
47 continue;
48 if (asihpi_pci_tbl[i].subdevice != PCI_ANY_ID
49 && asihpi_pci_tbl[i].subdevice !=
50 pci_info->pci_dev->subsystem_device)
51 continue;
52
53 /* HPI_DEBUG_LOG(DEBUG, " %x,%lx\n", i,
54 asihpi_pci_tbl[i].driver_data); */
55 return (hpi_handler_func *) asihpi_pci_tbl[i].driver_data;
56 }
57
58 return NULL;
59 }
60
hw_entry_point(struct hpi_message * phm,struct hpi_response * phr)61 static inline void hw_entry_point(struct hpi_message *phm,
62 struct hpi_response *phr)
63 {
64 if ((phm->adapter_index < HPI_MAX_ADAPTERS)
65 && hpi_entry_points[phm->adapter_index])
66 hpi_entry_points[phm->adapter_index] (phm, phr);
67 else
68 hpi_init_response(phr, phm->object, phm->function,
69 HPI_ERROR_PROCESSING_MESSAGE);
70 }
71
72 static void adapter_open(struct hpi_message *phm, struct hpi_response *phr);
73 static void adapter_close(struct hpi_message *phm, struct hpi_response *phr);
74
75 static void mixer_open(struct hpi_message *phm, struct hpi_response *phr);
76 static void mixer_close(struct hpi_message *phm, struct hpi_response *phr);
77
78 static void outstream_open(struct hpi_message *phm, struct hpi_response *phr,
79 void *h_owner);
80 static void outstream_close(struct hpi_message *phm, struct hpi_response *phr,
81 void *h_owner);
82 static void instream_open(struct hpi_message *phm, struct hpi_response *phr,
83 void *h_owner);
84 static void instream_close(struct hpi_message *phm, struct hpi_response *phr,
85 void *h_owner);
86
87 static void HPIMSGX__reset(u16 adapter_index);
88
89 static u16 HPIMSGX__init(struct hpi_message *phm, struct hpi_response *phr);
90 static void HPIMSGX__cleanup(u16 adapter_index, void *h_owner);
91
92 #ifndef DISABLE_PRAGMA_PACK1
93 #pragma pack(push, 1)
94 #endif
95
96 struct hpi_adapter_response {
97 struct hpi_response_header h;
98 struct hpi_adapter_res a;
99 };
100
101 struct hpi_mixer_response {
102 struct hpi_response_header h;
103 struct hpi_mixer_res m;
104 };
105
106 struct hpi_stream_response {
107 struct hpi_response_header h;
108 struct hpi_stream_res d;
109 };
110
111 struct adapter_info {
112 u16 type;
113 u16 num_instreams;
114 u16 num_outstreams;
115 };
116
117 struct asi_open_state {
118 int open_flag;
119 void *h_owner;
120 };
121
122 #ifndef DISABLE_PRAGMA_PACK1
123 #pragma pack(pop)
124 #endif
125
126 /* Globals */
127 static struct hpi_adapter_response rESP_HPI_ADAPTER_OPEN[HPI_MAX_ADAPTERS];
128
129 static struct hpi_stream_response
130 rESP_HPI_OSTREAM_OPEN[HPI_MAX_ADAPTERS][HPI_MAX_STREAMS];
131
132 static struct hpi_stream_response
133 rESP_HPI_ISTREAM_OPEN[HPI_MAX_ADAPTERS][HPI_MAX_STREAMS];
134
135 static struct hpi_mixer_response rESP_HPI_MIXER_OPEN[HPI_MAX_ADAPTERS];
136
137 static struct adapter_info aDAPTER_INFO[HPI_MAX_ADAPTERS];
138
139 /* use these to keep track of opens from user mode apps/DLLs */
140 static struct asi_open_state
141 outstream_user_open[HPI_MAX_ADAPTERS][HPI_MAX_STREAMS];
142
143 static struct asi_open_state
144 instream_user_open[HPI_MAX_ADAPTERS][HPI_MAX_STREAMS];
145
subsys_message(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)146 static void subsys_message(struct hpi_message *phm, struct hpi_response *phr,
147 void *h_owner)
148 {
149 if (phm->adapter_index != HPI_ADAPTER_INDEX_INVALID)
150 HPI_DEBUG_LOG(WARNING,
151 "suspicious adapter index %d in subsys message 0x%x.\n",
152 phm->adapter_index, phm->function);
153
154 switch (phm->function) {
155 case HPI_SUBSYS_GET_VERSION:
156 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM,
157 HPI_SUBSYS_GET_VERSION, 0);
158 phr->u.s.version = HPI_VER >> 8; /* return major.minor */
159 phr->u.s.data = HPI_VER; /* return major.minor.release */
160 break;
161 case HPI_SUBSYS_OPEN:
162 /*do not propagate the message down the chain */
163 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM, HPI_SUBSYS_OPEN, 0);
164 break;
165 case HPI_SUBSYS_CLOSE:
166 /*do not propagate the message down the chain */
167 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM, HPI_SUBSYS_CLOSE,
168 0);
169 HPIMSGX__cleanup(HPIMSGX_ALLADAPTERS, h_owner);
170 break;
171 case HPI_SUBSYS_DRIVER_LOAD:
172 /* Initialize this module's internal state */
173 hpios_msgxlock_init(&msgx_lock);
174 memset(&hpi_entry_points, 0, sizeof(hpi_entry_points));
175 /* Init subsys_findadapters response to no-adapters */
176 HPIMSGX__reset(HPIMSGX_ALLADAPTERS);
177 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM,
178 HPI_SUBSYS_DRIVER_LOAD, 0);
179 /* individual HPIs dont implement driver load */
180 HPI_COMMON(phm, phr);
181 break;
182 case HPI_SUBSYS_DRIVER_UNLOAD:
183 HPI_COMMON(phm, phr);
184 HPIMSGX__cleanup(HPIMSGX_ALLADAPTERS, h_owner);
185 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM,
186 HPI_SUBSYS_DRIVER_UNLOAD, 0);
187 return;
188
189 case HPI_SUBSYS_GET_NUM_ADAPTERS:
190 case HPI_SUBSYS_GET_ADAPTER:
191 HPI_COMMON(phm, phr);
192 break;
193
194 case HPI_SUBSYS_CREATE_ADAPTER:
195 HPIMSGX__init(phm, phr);
196 break;
197
198 default:
199 /* Must explicitly handle every subsys message in this switch */
200 hpi_init_response(phr, HPI_OBJ_SUBSYSTEM, phm->function,
201 HPI_ERROR_INVALID_FUNC);
202 break;
203 }
204 }
205
adapter_message(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)206 static void adapter_message(struct hpi_message *phm, struct hpi_response *phr,
207 void *h_owner)
208 {
209 switch (phm->function) {
210 case HPI_ADAPTER_OPEN:
211 adapter_open(phm, phr);
212 break;
213 case HPI_ADAPTER_CLOSE:
214 adapter_close(phm, phr);
215 break;
216 case HPI_ADAPTER_DELETE:
217 HPIMSGX__cleanup(phm->adapter_index, h_owner);
218 {
219 struct hpi_message hm;
220 struct hpi_response hr;
221 hpi_init_message_response(&hm, &hr, HPI_OBJ_ADAPTER,
222 HPI_ADAPTER_CLOSE);
223 hm.adapter_index = phm->adapter_index;
224 hw_entry_point(&hm, &hr);
225 }
226 hw_entry_point(phm, phr);
227 break;
228
229 default:
230 hw_entry_point(phm, phr);
231 break;
232 }
233 }
234
mixer_message(struct hpi_message * phm,struct hpi_response * phr)235 static void mixer_message(struct hpi_message *phm, struct hpi_response *phr)
236 {
237 switch (phm->function) {
238 case HPI_MIXER_OPEN:
239 mixer_open(phm, phr);
240 break;
241 case HPI_MIXER_CLOSE:
242 mixer_close(phm, phr);
243 break;
244 default:
245 hw_entry_point(phm, phr);
246 break;
247 }
248 }
249
outstream_message(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)250 static void outstream_message(struct hpi_message *phm,
251 struct hpi_response *phr, void *h_owner)
252 {
253 if (phm->obj_index >= aDAPTER_INFO[phm->adapter_index].num_outstreams) {
254 hpi_init_response(phr, HPI_OBJ_OSTREAM, phm->function,
255 HPI_ERROR_INVALID_OBJ_INDEX);
256 return;
257 }
258
259 switch (phm->function) {
260 case HPI_OSTREAM_OPEN:
261 outstream_open(phm, phr, h_owner);
262 break;
263 case HPI_OSTREAM_CLOSE:
264 outstream_close(phm, phr, h_owner);
265 break;
266 default:
267 hw_entry_point(phm, phr);
268 break;
269 }
270 }
271
instream_message(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)272 static void instream_message(struct hpi_message *phm,
273 struct hpi_response *phr, void *h_owner)
274 {
275 if (phm->obj_index >= aDAPTER_INFO[phm->adapter_index].num_instreams) {
276 hpi_init_response(phr, HPI_OBJ_ISTREAM, phm->function,
277 HPI_ERROR_INVALID_OBJ_INDEX);
278 return;
279 }
280
281 switch (phm->function) {
282 case HPI_ISTREAM_OPEN:
283 instream_open(phm, phr, h_owner);
284 break;
285 case HPI_ISTREAM_CLOSE:
286 instream_close(phm, phr, h_owner);
287 break;
288 default:
289 hw_entry_point(phm, phr);
290 break;
291 }
292 }
293
294 /* NOTE: HPI_Message() must be defined in the driver as a wrapper for
295 * HPI_MessageEx so that functions in hpifunc.c compile.
296 */
hpi_send_recv_ex(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)297 void hpi_send_recv_ex(struct hpi_message *phm, struct hpi_response *phr,
298 void *h_owner)
299 {
300
301 if (logging_enabled)
302 HPI_DEBUG_MESSAGE(DEBUG, phm);
303
304 if (phm->type != HPI_TYPE_REQUEST) {
305 hpi_init_response(phr, phm->object, phm->function,
306 HPI_ERROR_INVALID_TYPE);
307 return;
308 }
309
310 if (phm->adapter_index >= HPI_MAX_ADAPTERS
311 && phm->adapter_index != HPIMSGX_ALLADAPTERS) {
312 hpi_init_response(phr, phm->object, phm->function,
313 HPI_ERROR_BAD_ADAPTER_NUMBER);
314 return;
315 }
316
317 switch (phm->object) {
318 case HPI_OBJ_SUBSYSTEM:
319 subsys_message(phm, phr, h_owner);
320 break;
321
322 case HPI_OBJ_ADAPTER:
323 adapter_message(phm, phr, h_owner);
324 break;
325
326 case HPI_OBJ_MIXER:
327 mixer_message(phm, phr);
328 break;
329
330 case HPI_OBJ_OSTREAM:
331 outstream_message(phm, phr, h_owner);
332 break;
333
334 case HPI_OBJ_ISTREAM:
335 instream_message(phm, phr, h_owner);
336 break;
337
338 default:
339 hw_entry_point(phm, phr);
340 break;
341 }
342
343 if (logging_enabled)
344 HPI_DEBUG_RESPONSE(phr);
345
346 if (phr->error >= HPI_ERROR_DSP_COMMUNICATION) {
347 hpi_debug_level_set(HPI_DEBUG_LEVEL_ERROR);
348 logging_enabled = 0;
349 }
350 }
351
adapter_open(struct hpi_message * phm,struct hpi_response * phr)352 static void adapter_open(struct hpi_message *phm, struct hpi_response *phr)
353 {
354 HPI_DEBUG_LOG(VERBOSE, "adapter_open\n");
355 memcpy(phr, &rESP_HPI_ADAPTER_OPEN[phm->adapter_index],
356 sizeof(rESP_HPI_ADAPTER_OPEN[0]));
357 }
358
adapter_close(struct hpi_message * phm,struct hpi_response * phr)359 static void adapter_close(struct hpi_message *phm, struct hpi_response *phr)
360 {
361 HPI_DEBUG_LOG(VERBOSE, "adapter_close\n");
362 hpi_init_response(phr, HPI_OBJ_ADAPTER, HPI_ADAPTER_CLOSE, 0);
363 }
364
mixer_open(struct hpi_message * phm,struct hpi_response * phr)365 static void mixer_open(struct hpi_message *phm, struct hpi_response *phr)
366 {
367 memcpy(phr, &rESP_HPI_MIXER_OPEN[phm->adapter_index],
368 sizeof(rESP_HPI_MIXER_OPEN[0]));
369 }
370
mixer_close(struct hpi_message * phm,struct hpi_response * phr)371 static void mixer_close(struct hpi_message *phm, struct hpi_response *phr)
372 {
373 hpi_init_response(phr, HPI_OBJ_MIXER, HPI_MIXER_CLOSE, 0);
374 }
375
instream_open(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)376 static void instream_open(struct hpi_message *phm, struct hpi_response *phr,
377 void *h_owner)
378 {
379
380 struct hpi_message hm;
381 struct hpi_response hr;
382
383 hpi_init_response(phr, HPI_OBJ_ISTREAM, HPI_ISTREAM_OPEN, 0);
384
385 hpios_msgxlock_lock(&msgx_lock);
386
387 if (instream_user_open[phm->adapter_index][phm->obj_index].open_flag)
388 phr->error = HPI_ERROR_OBJ_ALREADY_OPEN;
389 else if (rESP_HPI_ISTREAM_OPEN[phm->adapter_index]
390 [phm->obj_index].h.error)
391 memcpy(phr,
392 &rESP_HPI_ISTREAM_OPEN[phm->adapter_index][phm->
393 obj_index],
394 sizeof(rESP_HPI_ISTREAM_OPEN[0][0]));
395 else {
396 instream_user_open[phm->adapter_index][phm->
397 obj_index].open_flag = 1;
398 hpios_msgxlock_unlock(&msgx_lock);
399
400 /* issue a reset */
401 hpi_init_message_response(&hm, &hr, HPI_OBJ_ISTREAM,
402 HPI_ISTREAM_RESET);
403 hm.adapter_index = phm->adapter_index;
404 hm.obj_index = phm->obj_index;
405 hw_entry_point(&hm, &hr);
406
407 hpios_msgxlock_lock(&msgx_lock);
408 if (hr.error) {
409 instream_user_open[phm->adapter_index][phm->
410 obj_index].open_flag = 0;
411 phr->error = hr.error;
412 } else {
413 instream_user_open[phm->adapter_index][phm->
414 obj_index].open_flag = 1;
415 instream_user_open[phm->adapter_index][phm->
416 obj_index].h_owner = h_owner;
417 memcpy(phr,
418 &rESP_HPI_ISTREAM_OPEN[phm->adapter_index]
419 [phm->obj_index],
420 sizeof(rESP_HPI_ISTREAM_OPEN[0][0]));
421 }
422 }
423 hpios_msgxlock_unlock(&msgx_lock);
424 }
425
instream_close(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)426 static void instream_close(struct hpi_message *phm, struct hpi_response *phr,
427 void *h_owner)
428 {
429
430 struct hpi_message hm;
431 struct hpi_response hr;
432
433 hpi_init_response(phr, HPI_OBJ_ISTREAM, HPI_ISTREAM_CLOSE, 0);
434
435 hpios_msgxlock_lock(&msgx_lock);
436 if (h_owner ==
437 instream_user_open[phm->adapter_index][phm->
438 obj_index].h_owner) {
439 /* HPI_DEBUG_LOG(INFO,"closing adapter %d "
440 "instream %d owned by %p\n",
441 phm->wAdapterIndex, phm->wObjIndex, hOwner); */
442 instream_user_open[phm->adapter_index][phm->
443 obj_index].h_owner = NULL;
444 hpios_msgxlock_unlock(&msgx_lock);
445 /* issue a reset */
446 hpi_init_message_response(&hm, &hr, HPI_OBJ_ISTREAM,
447 HPI_ISTREAM_RESET);
448 hm.adapter_index = phm->adapter_index;
449 hm.obj_index = phm->obj_index;
450 hw_entry_point(&hm, &hr);
451 hpios_msgxlock_lock(&msgx_lock);
452 if (hr.error) {
453 instream_user_open[phm->adapter_index][phm->
454 obj_index].h_owner = h_owner;
455 phr->error = hr.error;
456 } else {
457 instream_user_open[phm->adapter_index][phm->
458 obj_index].open_flag = 0;
459 instream_user_open[phm->adapter_index][phm->
460 obj_index].h_owner = NULL;
461 }
462 } else {
463 HPI_DEBUG_LOG(WARNING,
464 "%p trying to close %d instream %d owned by %p\n",
465 h_owner, phm->adapter_index, phm->obj_index,
466 instream_user_open[phm->adapter_index][phm->
467 obj_index].h_owner);
468 phr->error = HPI_ERROR_OBJ_NOT_OPEN;
469 }
470 hpios_msgxlock_unlock(&msgx_lock);
471 }
472
outstream_open(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)473 static void outstream_open(struct hpi_message *phm, struct hpi_response *phr,
474 void *h_owner)
475 {
476
477 struct hpi_message hm;
478 struct hpi_response hr;
479
480 hpi_init_response(phr, HPI_OBJ_OSTREAM, HPI_OSTREAM_OPEN, 0);
481
482 hpios_msgxlock_lock(&msgx_lock);
483
484 if (outstream_user_open[phm->adapter_index][phm->obj_index].open_flag)
485 phr->error = HPI_ERROR_OBJ_ALREADY_OPEN;
486 else if (rESP_HPI_OSTREAM_OPEN[phm->adapter_index]
487 [phm->obj_index].h.error)
488 memcpy(phr,
489 &rESP_HPI_OSTREAM_OPEN[phm->adapter_index][phm->
490 obj_index],
491 sizeof(rESP_HPI_OSTREAM_OPEN[0][0]));
492 else {
493 outstream_user_open[phm->adapter_index][phm->
494 obj_index].open_flag = 1;
495 hpios_msgxlock_unlock(&msgx_lock);
496
497 /* issue a reset */
498 hpi_init_message_response(&hm, &hr, HPI_OBJ_OSTREAM,
499 HPI_OSTREAM_RESET);
500 hm.adapter_index = phm->adapter_index;
501 hm.obj_index = phm->obj_index;
502 hw_entry_point(&hm, &hr);
503
504 hpios_msgxlock_lock(&msgx_lock);
505 if (hr.error) {
506 outstream_user_open[phm->adapter_index][phm->
507 obj_index].open_flag = 0;
508 phr->error = hr.error;
509 } else {
510 outstream_user_open[phm->adapter_index][phm->
511 obj_index].open_flag = 1;
512 outstream_user_open[phm->adapter_index][phm->
513 obj_index].h_owner = h_owner;
514 memcpy(phr,
515 &rESP_HPI_OSTREAM_OPEN[phm->adapter_index]
516 [phm->obj_index],
517 sizeof(rESP_HPI_OSTREAM_OPEN[0][0]));
518 }
519 }
520 hpios_msgxlock_unlock(&msgx_lock);
521 }
522
outstream_close(struct hpi_message * phm,struct hpi_response * phr,void * h_owner)523 static void outstream_close(struct hpi_message *phm, struct hpi_response *phr,
524 void *h_owner)
525 {
526
527 struct hpi_message hm;
528 struct hpi_response hr;
529
530 hpi_init_response(phr, HPI_OBJ_OSTREAM, HPI_OSTREAM_CLOSE, 0);
531
532 hpios_msgxlock_lock(&msgx_lock);
533
534 if (h_owner ==
535 outstream_user_open[phm->adapter_index][phm->
536 obj_index].h_owner) {
537 /* HPI_DEBUG_LOG(INFO,"closing adapter %d "
538 "outstream %d owned by %p\n",
539 phm->wAdapterIndex, phm->wObjIndex, hOwner); */
540 outstream_user_open[phm->adapter_index][phm->
541 obj_index].h_owner = NULL;
542 hpios_msgxlock_unlock(&msgx_lock);
543 /* issue a reset */
544 hpi_init_message_response(&hm, &hr, HPI_OBJ_OSTREAM,
545 HPI_OSTREAM_RESET);
546 hm.adapter_index = phm->adapter_index;
547 hm.obj_index = phm->obj_index;
548 hw_entry_point(&hm, &hr);
549 hpios_msgxlock_lock(&msgx_lock);
550 if (hr.error) {
551 outstream_user_open[phm->adapter_index][phm->
552 obj_index].h_owner = h_owner;
553 phr->error = hr.error;
554 } else {
555 outstream_user_open[phm->adapter_index][phm->
556 obj_index].open_flag = 0;
557 outstream_user_open[phm->adapter_index][phm->
558 obj_index].h_owner = NULL;
559 }
560 } else {
561 HPI_DEBUG_LOG(WARNING,
562 "%p trying to close %d outstream %d owned by %p\n",
563 h_owner, phm->adapter_index, phm->obj_index,
564 outstream_user_open[phm->adapter_index][phm->
565 obj_index].h_owner);
566 phr->error = HPI_ERROR_OBJ_NOT_OPEN;
567 }
568 hpios_msgxlock_unlock(&msgx_lock);
569 }
570
adapter_prepare(u16 adapter)571 static u16 adapter_prepare(u16 adapter)
572 {
573 struct hpi_message hm;
574 struct hpi_response hr;
575
576 /* Open the adapter and streams */
577 u16 i;
578
579 /* call to HPI_ADAPTER_OPEN */
580 hpi_init_message_response(&hm, &hr, HPI_OBJ_ADAPTER,
581 HPI_ADAPTER_OPEN);
582 hm.adapter_index = adapter;
583 hw_entry_point(&hm, &hr);
584 memcpy(&rESP_HPI_ADAPTER_OPEN[adapter], &hr,
585 sizeof(rESP_HPI_ADAPTER_OPEN[0]));
586 if (hr.error)
587 return hr.error;
588
589 /* call to HPI_ADAPTER_GET_INFO */
590 hpi_init_message_response(&hm, &hr, HPI_OBJ_ADAPTER,
591 HPI_ADAPTER_GET_INFO);
592 hm.adapter_index = adapter;
593 hw_entry_point(&hm, &hr);
594 if (hr.error)
595 return hr.error;
596
597 aDAPTER_INFO[adapter].num_outstreams = hr.u.ax.info.num_outstreams;
598 aDAPTER_INFO[adapter].num_instreams = hr.u.ax.info.num_instreams;
599 aDAPTER_INFO[adapter].type = hr.u.ax.info.adapter_type;
600
601 /* call to HPI_OSTREAM_OPEN */
602 for (i = 0; i < aDAPTER_INFO[adapter].num_outstreams; i++) {
603 hpi_init_message_response(&hm, &hr, HPI_OBJ_OSTREAM,
604 HPI_OSTREAM_OPEN);
605 hm.adapter_index = adapter;
606 hm.obj_index = i;
607 hw_entry_point(&hm, &hr);
608 memcpy(&rESP_HPI_OSTREAM_OPEN[adapter][i], &hr,
609 sizeof(rESP_HPI_OSTREAM_OPEN[0][0]));
610 outstream_user_open[adapter][i].open_flag = 0;
611 outstream_user_open[adapter][i].h_owner = NULL;
612 }
613
614 /* call to HPI_ISTREAM_OPEN */
615 for (i = 0; i < aDAPTER_INFO[adapter].num_instreams; i++) {
616 hpi_init_message_response(&hm, &hr, HPI_OBJ_ISTREAM,
617 HPI_ISTREAM_OPEN);
618 hm.adapter_index = adapter;
619 hm.obj_index = i;
620 hw_entry_point(&hm, &hr);
621 memcpy(&rESP_HPI_ISTREAM_OPEN[adapter][i], &hr,
622 sizeof(rESP_HPI_ISTREAM_OPEN[0][0]));
623 instream_user_open[adapter][i].open_flag = 0;
624 instream_user_open[adapter][i].h_owner = NULL;
625 }
626
627 /* call to HPI_MIXER_OPEN */
628 hpi_init_message_response(&hm, &hr, HPI_OBJ_MIXER, HPI_MIXER_OPEN);
629 hm.adapter_index = adapter;
630 hw_entry_point(&hm, &hr);
631 memcpy(&rESP_HPI_MIXER_OPEN[adapter], &hr,
632 sizeof(rESP_HPI_MIXER_OPEN[0]));
633
634 return 0;
635 }
636
HPIMSGX__reset(u16 adapter_index)637 static void HPIMSGX__reset(u16 adapter_index)
638 {
639 int i;
640 u16 adapter;
641 struct hpi_response hr;
642
643 if (adapter_index == HPIMSGX_ALLADAPTERS) {
644 for (adapter = 0; adapter < HPI_MAX_ADAPTERS; adapter++) {
645
646 hpi_init_response(&hr, HPI_OBJ_ADAPTER,
647 HPI_ADAPTER_OPEN, HPI_ERROR_BAD_ADAPTER);
648 memcpy(&rESP_HPI_ADAPTER_OPEN[adapter], &hr,
649 sizeof(rESP_HPI_ADAPTER_OPEN[adapter]));
650
651 hpi_init_response(&hr, HPI_OBJ_MIXER, HPI_MIXER_OPEN,
652 HPI_ERROR_INVALID_OBJ);
653 memcpy(&rESP_HPI_MIXER_OPEN[adapter], &hr,
654 sizeof(rESP_HPI_MIXER_OPEN[adapter]));
655
656 for (i = 0; i < HPI_MAX_STREAMS; i++) {
657 hpi_init_response(&hr, HPI_OBJ_OSTREAM,
658 HPI_OSTREAM_OPEN,
659 HPI_ERROR_INVALID_OBJ);
660 memcpy(&rESP_HPI_OSTREAM_OPEN[adapter][i],
661 &hr,
662 sizeof(rESP_HPI_OSTREAM_OPEN[adapter]
663 [i]));
664 hpi_init_response(&hr, HPI_OBJ_ISTREAM,
665 HPI_ISTREAM_OPEN,
666 HPI_ERROR_INVALID_OBJ);
667 memcpy(&rESP_HPI_ISTREAM_OPEN[adapter][i],
668 &hr,
669 sizeof(rESP_HPI_ISTREAM_OPEN[adapter]
670 [i]));
671 }
672 }
673 } else if (adapter_index < HPI_MAX_ADAPTERS) {
674 rESP_HPI_ADAPTER_OPEN[adapter_index].h.error =
675 HPI_ERROR_BAD_ADAPTER;
676 rESP_HPI_MIXER_OPEN[adapter_index].h.error =
677 HPI_ERROR_INVALID_OBJ;
678 for (i = 0; i < HPI_MAX_STREAMS; i++) {
679 rESP_HPI_OSTREAM_OPEN[adapter_index][i].h.error =
680 HPI_ERROR_INVALID_OBJ;
681 rESP_HPI_ISTREAM_OPEN[adapter_index][i].h.error =
682 HPI_ERROR_INVALID_OBJ;
683 }
684 }
685 }
686
HPIMSGX__init(struct hpi_message * phm,struct hpi_response * phr)687 static u16 HPIMSGX__init(struct hpi_message *phm,
688 /* HPI_SUBSYS_CREATE_ADAPTER structure with */
689 /* resource list or NULL=find all */
690 struct hpi_response *phr
691 /* response from HPI_ADAPTER_GET_INFO */
692 )
693 {
694 hpi_handler_func *entry_point_func;
695 struct hpi_response hr;
696
697 /* Init response here so we can pass in previous adapter list */
698 hpi_init_response(&hr, phm->object, phm->function,
699 HPI_ERROR_INVALID_OBJ);
700
701 entry_point_func =
702 hpi_lookup_entry_point_function(phm->u.s.resource.r.pci);
703
704 if (entry_point_func) {
705 HPI_DEBUG_MESSAGE(DEBUG, phm);
706 entry_point_func(phm, &hr);
707 } else {
708 phr->error = HPI_ERROR_PROCESSING_MESSAGE;
709 return phr->error;
710 }
711 if (hr.error == 0) {
712 /* the adapter was created successfully
713 save the mapping for future use */
714 hpi_entry_points[hr.u.s.adapter_index] = entry_point_func;
715 /* prepare adapter (pre-open streams etc.) */
716 HPI_DEBUG_LOG(DEBUG,
717 "HPI_SUBSYS_CREATE_ADAPTER successful,"
718 " preparing adapter\n");
719 adapter_prepare(hr.u.s.adapter_index);
720 }
721 memcpy(phr, &hr, hr.size);
722 return phr->error;
723 }
724
HPIMSGX__cleanup(u16 adapter_index,void * h_owner)725 static void HPIMSGX__cleanup(u16 adapter_index, void *h_owner)
726 {
727 int i, adapter, adapter_limit;
728
729 if (!h_owner)
730 return;
731
732 if (adapter_index == HPIMSGX_ALLADAPTERS) {
733 adapter = 0;
734 adapter_limit = HPI_MAX_ADAPTERS;
735 } else {
736 adapter = adapter_index;
737 adapter_limit = adapter + 1;
738 }
739
740 for (; adapter < adapter_limit; adapter++) {
741 /* printk(KERN_INFO "Cleanup adapter #%d\n",wAdapter); */
742 for (i = 0; i < HPI_MAX_STREAMS; i++) {
743 if (h_owner ==
744 outstream_user_open[adapter][i].h_owner) {
745 struct hpi_message hm;
746 struct hpi_response hr;
747
748 HPI_DEBUG_LOG(DEBUG,
749 "Close adapter %d ostream %d\n",
750 adapter, i);
751
752 hpi_init_message_response(&hm, &hr,
753 HPI_OBJ_OSTREAM, HPI_OSTREAM_RESET);
754 hm.adapter_index = (u16)adapter;
755 hm.obj_index = (u16)i;
756 hw_entry_point(&hm, &hr);
757
758 hm.function = HPI_OSTREAM_HOSTBUFFER_FREE;
759 hw_entry_point(&hm, &hr);
760
761 hm.function = HPI_OSTREAM_GROUP_RESET;
762 hw_entry_point(&hm, &hr);
763
764 outstream_user_open[adapter][i].open_flag = 0;
765 outstream_user_open[adapter][i].h_owner =
766 NULL;
767 }
768 if (h_owner == instream_user_open[adapter][i].h_owner) {
769 struct hpi_message hm;
770 struct hpi_response hr;
771
772 HPI_DEBUG_LOG(DEBUG,
773 "Close adapter %d istream %d\n",
774 adapter, i);
775
776 hpi_init_message_response(&hm, &hr,
777 HPI_OBJ_ISTREAM, HPI_ISTREAM_RESET);
778 hm.adapter_index = (u16)adapter;
779 hm.obj_index = (u16)i;
780 hw_entry_point(&hm, &hr);
781
782 hm.function = HPI_ISTREAM_HOSTBUFFER_FREE;
783 hw_entry_point(&hm, &hr);
784
785 hm.function = HPI_ISTREAM_GROUP_RESET;
786 hw_entry_point(&hm, &hr);
787
788 instream_user_open[adapter][i].open_flag = 0;
789 instream_user_open[adapter][i].h_owner = NULL;
790 }
791 }
792 }
793 }
794