1 // SPDX-License-Identifier: GPL-2.0-only
2 //
3 // Copyright(c) 2021 Intel Corporation
4 //
5 // Authors: Cezary Rojewski <cezary.rojewski@intel.com>
6 // Amadeusz Slawinski <amadeuszx.slawinski@linux.intel.com>
7 //
8
9 #include <linux/acpi.h>
10 #include <acpi/nhlt.h>
11 #include <sound/pcm_params.h>
12 #include <sound/soc.h>
13 #include "avs.h"
14 #include "control.h"
15 #include "path.h"
16 #include "topology.h"
17
18 /* Must be called with adev->comp_list_mutex held. */
19 static struct avs_tplg *
avs_path_find_tplg(struct avs_dev * adev,const char * name)20 avs_path_find_tplg(struct avs_dev *adev, const char *name)
21 {
22 struct avs_soc_component *acomp;
23
24 list_for_each_entry(acomp, &adev->comp_list, node)
25 if (!strcmp(acomp->tplg->name, name))
26 return acomp->tplg;
27 return NULL;
28 }
29
30 static struct avs_path_module *
avs_path_find_module(struct avs_path_pipeline * ppl,u32 template_id)31 avs_path_find_module(struct avs_path_pipeline *ppl, u32 template_id)
32 {
33 struct avs_path_module *mod;
34
35 list_for_each_entry(mod, &ppl->mod_list, node)
36 if (mod->template->id == template_id)
37 return mod;
38 return NULL;
39 }
40
41 static struct avs_path_pipeline *
avs_path_find_pipeline(struct avs_path * path,u32 template_id)42 avs_path_find_pipeline(struct avs_path *path, u32 template_id)
43 {
44 struct avs_path_pipeline *ppl;
45
46 list_for_each_entry(ppl, &path->ppl_list, node)
47 if (ppl->template->id == template_id)
48 return ppl;
49 return NULL;
50 }
51
52 static struct avs_path *
avs_path_find_path(struct avs_dev * adev,const char * name,u32 template_id)53 avs_path_find_path(struct avs_dev *adev, const char *name, u32 template_id)
54 {
55 struct avs_tplg_path_template *pos, *template = NULL;
56 struct avs_tplg *tplg;
57 struct avs_path *path;
58
59 tplg = avs_path_find_tplg(adev, name);
60 if (!tplg)
61 return NULL;
62
63 list_for_each_entry(pos, &tplg->path_tmpl_list, node) {
64 if (pos->id == template_id) {
65 template = pos;
66 break;
67 }
68 }
69 if (!template)
70 return NULL;
71
72 spin_lock(&adev->path_list_lock);
73 /* Only one variant of given path template may be instantiated at a time. */
74 list_for_each_entry(path, &adev->path_list, node) {
75 if (path->template->owner == template) {
76 spin_unlock(&adev->path_list_lock);
77 return path;
78 }
79 }
80
81 spin_unlock(&adev->path_list_lock);
82 return NULL;
83 }
84
avs_test_hw_params(struct snd_pcm_hw_params * params,struct avs_audio_format * fmt)85 static bool avs_test_hw_params(struct snd_pcm_hw_params *params,
86 struct avs_audio_format *fmt)
87 {
88 return (params_rate(params) == fmt->sampling_freq &&
89 params_channels(params) == fmt->num_channels &&
90 params_physical_width(params) == fmt->bit_depth &&
91 snd_pcm_hw_params_bits(params) == fmt->valid_bit_depth);
92 }
93
94 static struct avs_tplg_path *
avs_path_find_variant(struct avs_dev * adev,struct avs_tplg_path_template * template,struct snd_pcm_hw_params * fe_params,struct snd_pcm_hw_params * be_params)95 avs_path_find_variant(struct avs_dev *adev,
96 struct avs_tplg_path_template *template,
97 struct snd_pcm_hw_params *fe_params,
98 struct snd_pcm_hw_params *be_params)
99 {
100 struct avs_tplg_path *variant;
101
102 list_for_each_entry(variant, &template->path_list, node) {
103 dev_dbg(adev->dev, "check FE rate %d chn %d vbd %d bd %d\n",
104 variant->fe_fmt->sampling_freq, variant->fe_fmt->num_channels,
105 variant->fe_fmt->valid_bit_depth, variant->fe_fmt->bit_depth);
106 dev_dbg(adev->dev, "check BE rate %d chn %d vbd %d bd %d\n",
107 variant->be_fmt->sampling_freq, variant->be_fmt->num_channels,
108 variant->be_fmt->valid_bit_depth, variant->be_fmt->bit_depth);
109
110 if (variant->fe_fmt && avs_test_hw_params(fe_params, variant->fe_fmt) &&
111 variant->be_fmt && avs_test_hw_params(be_params, variant->be_fmt))
112 return variant;
113 }
114
115 return NULL;
116 }
117
avs_condpath_find_variant(struct avs_dev * adev,struct avs_tplg_path_template * template,struct avs_path * source,struct avs_path * sink)118 static struct avs_tplg_path *avs_condpath_find_variant(struct avs_dev *adev,
119 struct avs_tplg_path_template *template,
120 struct avs_path *source,
121 struct avs_path *sink)
122 {
123 struct avs_tplg_path *variant;
124
125 list_for_each_entry(variant, &template->path_list, node) {
126 if (variant->source_path_id == source->template->id &&
127 variant->sink_path_id == sink->template->id)
128 return variant;
129 }
130
131 return NULL;
132 }
133
avs_tplg_path_template_id_equal(struct avs_tplg_path_template_id * id,struct avs_tplg_path_template_id * id2)134 static bool avs_tplg_path_template_id_equal(struct avs_tplg_path_template_id *id,
135 struct avs_tplg_path_template_id *id2)
136 {
137 return id->id == id2->id && !sysfs_streq(id->tplg_name, id2->tplg_name);
138 }
139
avs_condpath_find_match(struct avs_dev * adev,struct avs_tplg_path_template * template,struct avs_path * path,int dir)140 static struct avs_path *avs_condpath_find_match(struct avs_dev *adev,
141 struct avs_tplg_path_template *template,
142 struct avs_path *path, int dir)
143 {
144 struct avs_tplg_path_template_id *id, *id2;
145
146 if (dir) {
147 id = &template->source;
148 id2 = &template->sink;
149 } else {
150 id = &template->sink;
151 id2 = &template->source;
152 }
153
154 /* Check whether this path is either source or sink of condpath template. */
155 if (id->id != path->template->owner->id ||
156 strcmp(id->tplg_name, path->template->owner->owner->name))
157 return NULL;
158
159 /* Unidirectional condpaths are allowed. */
160 if (avs_tplg_path_template_id_equal(id, id2))
161 return path;
162
163 /* Now find the counterpart. */
164 return avs_path_find_path(adev, id2->tplg_name, id2->id);
165 }
166
167 static struct acpi_nhlt_config *
168 avs_nhlt_config_or_default(struct avs_dev *adev, struct avs_tplg_module *t);
169
avs_path_set_constraint(struct avs_dev * adev,struct avs_tplg_path_template * template,struct snd_pcm_hw_constraint_list * rate_list,struct snd_pcm_hw_constraint_list * channels_list,struct snd_pcm_hw_constraint_list * sample_bits_list)170 int avs_path_set_constraint(struct avs_dev *adev, struct avs_tplg_path_template *template,
171 struct snd_pcm_hw_constraint_list *rate_list,
172 struct snd_pcm_hw_constraint_list *channels_list,
173 struct snd_pcm_hw_constraint_list *sample_bits_list)
174 {
175 struct avs_tplg_path *path_template;
176 unsigned int *rlist, *clist, *slist;
177 size_t i;
178
179 i = 0;
180 list_for_each_entry(path_template, &template->path_list, node)
181 i++;
182
183 rlist = kcalloc(i, sizeof(*rlist), GFP_KERNEL);
184 clist = kcalloc(i, sizeof(*clist), GFP_KERNEL);
185 slist = kcalloc(i, sizeof(*slist), GFP_KERNEL);
186 if (!rlist || !clist || !slist)
187 return -ENOMEM;
188
189 i = 0;
190 list_for_each_entry(path_template, &template->path_list, node) {
191 struct avs_tplg_pipeline *pipeline_template;
192
193 list_for_each_entry(pipeline_template, &path_template->ppl_list, node) {
194 struct avs_tplg_module *module_template;
195
196 list_for_each_entry(module_template, &pipeline_template->mod_list, node) {
197 const guid_t *type = &module_template->cfg_ext->type;
198 struct acpi_nhlt_config *blob;
199
200 if (!guid_equal(type, &AVS_COPIER_MOD_UUID) &&
201 !guid_equal(type, &AVS_WOVHOSTM_MOD_UUID))
202 continue;
203
204 switch (module_template->cfg_ext->copier.dma_type) {
205 case AVS_DMA_DMIC_LINK_INPUT:
206 case AVS_DMA_I2S_LINK_OUTPUT:
207 case AVS_DMA_I2S_LINK_INPUT:
208 break;
209 default:
210 continue;
211 }
212
213 if (!module_template->nhlt_config) {
214 blob = avs_nhlt_config_or_default(adev, module_template);
215 if (IS_ERR(blob))
216 continue;
217 }
218
219 rlist[i] = path_template->fe_fmt->sampling_freq;
220 clist[i] = path_template->fe_fmt->num_channels;
221 slist[i] = path_template->fe_fmt->bit_depth;
222 i++;
223 }
224 }
225 }
226
227 if (i) {
228 rate_list->count = i;
229 rate_list->list = rlist;
230 channels_list->count = i;
231 channels_list->list = clist;
232 sample_bits_list->count = i;
233 sample_bits_list->list = slist;
234 } else {
235 kfree(rlist);
236 kfree(clist);
237 kfree(slist);
238 }
239
240 return i;
241 }
242
avs_init_node_id(union avs_connector_node_id * node_id,struct avs_tplg_modcfg_ext * te,u32 dma_id)243 static void avs_init_node_id(union avs_connector_node_id *node_id,
244 struct avs_tplg_modcfg_ext *te, u32 dma_id)
245 {
246 node_id->val = 0;
247 node_id->dma_type = te->copier.dma_type;
248
249 switch (node_id->dma_type) {
250 case AVS_DMA_DMIC_LINK_INPUT:
251 case AVS_DMA_I2S_LINK_OUTPUT:
252 case AVS_DMA_I2S_LINK_INPUT:
253 /* Gateway's virtual index is statically assigned in the topology. */
254 node_id->vindex = te->copier.vindex.val;
255 break;
256
257 case AVS_DMA_HDA_HOST_OUTPUT:
258 case AVS_DMA_HDA_HOST_INPUT:
259 /* Gateway's virtual index is dynamically assigned with DMA ID */
260 node_id->vindex = dma_id;
261 break;
262
263 case AVS_DMA_HDA_LINK_OUTPUT:
264 case AVS_DMA_HDA_LINK_INPUT:
265 node_id->vindex = te->copier.vindex.val | dma_id;
266 break;
267
268 default:
269 *node_id = INVALID_NODE_ID;
270 break;
271 }
272 }
273
274 /* Every BLOB contains at least gateway attributes. */
275 static struct acpi_nhlt_config *default_blob = (struct acpi_nhlt_config *)&(u32[2]) {4};
276
277 static struct acpi_nhlt_config *
avs_nhlt_config_or_default(struct avs_dev * adev,struct avs_tplg_module * t)278 avs_nhlt_config_or_default(struct avs_dev *adev, struct avs_tplg_module *t)
279 {
280 struct acpi_nhlt_format_config *fmtcfg;
281 struct avs_tplg_modcfg_ext *te;
282 struct avs_audio_format *fmt;
283 int link_type, dev_type;
284 int bus_id, dir;
285
286 te = t->cfg_ext;
287
288 switch (te->copier.dma_type) {
289 case AVS_DMA_I2S_LINK_OUTPUT:
290 link_type = ACPI_NHLT_LINKTYPE_SSP;
291 dev_type = ACPI_NHLT_DEVICETYPE_CODEC;
292 bus_id = te->copier.vindex.i2s.instance;
293 dir = SNDRV_PCM_STREAM_PLAYBACK;
294 fmt = te->copier.out_fmt;
295 break;
296
297 case AVS_DMA_I2S_LINK_INPUT:
298 link_type = ACPI_NHLT_LINKTYPE_SSP;
299 dev_type = ACPI_NHLT_DEVICETYPE_CODEC;
300 bus_id = te->copier.vindex.i2s.instance;
301 dir = SNDRV_PCM_STREAM_CAPTURE;
302 fmt = t->in_fmt;
303 break;
304
305 case AVS_DMA_DMIC_LINK_INPUT:
306 link_type = ACPI_NHLT_LINKTYPE_PDM;
307 dev_type = -1; /* ignored */
308 bus_id = 0;
309 dir = SNDRV_PCM_STREAM_CAPTURE;
310 fmt = t->in_fmt;
311 break;
312
313 default:
314 return default_blob;
315 }
316
317 /* Override format selection if necessary. */
318 if (te->copier.blob_fmt)
319 fmt = te->copier.blob_fmt;
320
321 fmtcfg = acpi_nhlt_find_fmtcfg(link_type, dev_type, dir, bus_id,
322 fmt->num_channels, fmt->sampling_freq, fmt->valid_bit_depth,
323 fmt->bit_depth);
324 if (!fmtcfg) {
325 dev_warn(adev->dev, "Endpoint format configuration not found.\n");
326 return ERR_PTR(-ENOENT);
327 }
328
329 if (fmtcfg->config.capabilities_size < default_blob->capabilities_size)
330 return ERR_PTR(-ETOOSMALL);
331 /* The firmware expects the payload to be DWORD-aligned. */
332 if (fmtcfg->config.capabilities_size % sizeof(u32))
333 return ERR_PTR(-EINVAL);
334
335 return &fmtcfg->config;
336 }
337
avs_append_dma_cfg(struct avs_dev * adev,struct avs_copier_gtw_cfg * gtw,struct avs_tplg_module * t,u32 dma_id,size_t * cfg_size)338 static int avs_append_dma_cfg(struct avs_dev *adev, struct avs_copier_gtw_cfg *gtw,
339 struct avs_tplg_module *t, u32 dma_id, size_t *cfg_size)
340 {
341 u32 dma_type = t->cfg_ext->copier.dma_type;
342 struct avs_dma_cfg *dma;
343 struct avs_tlv *tlv;
344 size_t tlv_size;
345
346 if (!avs_platattr_test(adev, ALTHDA))
347 return 0;
348
349 switch (dma_type) {
350 case AVS_DMA_HDA_HOST_OUTPUT:
351 case AVS_DMA_HDA_HOST_INPUT:
352 case AVS_DMA_HDA_LINK_OUTPUT:
353 case AVS_DMA_HDA_LINK_INPUT:
354 return 0;
355 default:
356 break;
357 }
358
359 tlv_size = sizeof(*tlv) + sizeof(*dma);
360 if (*cfg_size + tlv_size > AVS_MAILBOX_SIZE)
361 return -E2BIG;
362
363 /* DMA config is a TLV tailing the existing payload. */
364 tlv = (struct avs_tlv *)>w->config.blob[gtw->config_length];
365 tlv->type = AVS_GTW_DMA_CONFIG_ID;
366 tlv->length = sizeof(*dma);
367
368 dma = (struct avs_dma_cfg *)tlv->value;
369 memset(dma, 0, sizeof(*dma));
370 dma->dma_method = AVS_DMA_METHOD_HDA;
371 dma->pre_allocated = true;
372 dma->dma_channel_id = dma_id;
373 dma->stream_id = dma_id + 1;
374
375 gtw->config_length += tlv_size / sizeof(u32);
376 *cfg_size += tlv_size;
377
378 return 0;
379 }
380
avs_fill_gtw_config(struct avs_dev * adev,struct avs_copier_gtw_cfg * gtw,struct avs_tplg_module * t,u32 dma_id,size_t * cfg_size)381 static int avs_fill_gtw_config(struct avs_dev *adev, struct avs_copier_gtw_cfg *gtw,
382 struct avs_tplg_module *t, u32 dma_id, size_t *cfg_size)
383 {
384 struct acpi_nhlt_config *blob;
385 size_t gtw_size;
386
387 if (t->nhlt_config)
388 blob = t->nhlt_config->blob;
389 else
390 blob = avs_nhlt_config_or_default(adev, t);
391 if (IS_ERR(blob))
392 return PTR_ERR(blob);
393
394 gtw_size = blob->capabilities_size;
395 if (*cfg_size + gtw_size > AVS_MAILBOX_SIZE)
396 return -E2BIG;
397
398 gtw->config_length = gtw_size / sizeof(u32);
399 memcpy(gtw->config.blob, blob->capabilities, blob->capabilities_size);
400 *cfg_size += gtw_size;
401
402 return avs_append_dma_cfg(adev, gtw, t, dma_id, cfg_size);
403 }
404
avs_copier_create(struct avs_dev * adev,struct avs_path_module * mod)405 static int avs_copier_create(struct avs_dev *adev, struct avs_path_module *mod)
406 {
407 struct avs_tplg_module *t = mod->template;
408 struct avs_tplg_modcfg_ext *te;
409 struct avs_copier_cfg *cfg;
410 size_t cfg_size;
411 u32 dma_id;
412 int ret;
413
414 te = t->cfg_ext;
415 cfg = adev->modcfg_buf;
416 dma_id = mod->owner->owner->dma_id;
417 cfg_size = offsetof(struct avs_copier_cfg, gtw_cfg.config);
418
419 ret = avs_fill_gtw_config(adev, &cfg->gtw_cfg, t, dma_id, &cfg_size);
420 if (ret)
421 return ret;
422
423 cfg->base.cpc = t->cfg_base->cpc;
424 cfg->base.ibs = t->cfg_base->ibs;
425 cfg->base.obs = t->cfg_base->obs;
426 cfg->base.is_pages = t->cfg_base->is_pages;
427 cfg->base.audio_fmt = *t->in_fmt;
428 cfg->out_fmt = *te->copier.out_fmt;
429 cfg->feature_mask = te->copier.feature_mask;
430 avs_init_node_id(&cfg->gtw_cfg.node_id, te, dma_id);
431 cfg->gtw_cfg.dma_buffer_size = te->copier.dma_buffer_size;
432 mod->gtw_attrs = cfg->gtw_cfg.config.attrs;
433
434 ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
435 t->domain, cfg, cfg_size, &mod->instance_id);
436 return ret;
437 }
438
avs_whm_create(struct avs_dev * adev,struct avs_path_module * mod)439 static int avs_whm_create(struct avs_dev *adev, struct avs_path_module *mod)
440 {
441 struct avs_tplg_module *t = mod->template;
442 struct avs_tplg_modcfg_ext *te;
443 struct avs_whm_cfg *cfg;
444 size_t cfg_size;
445 u32 dma_id;
446 int ret;
447
448 te = t->cfg_ext;
449 cfg = adev->modcfg_buf;
450 dma_id = mod->owner->owner->dma_id;
451 cfg_size = offsetof(struct avs_whm_cfg, gtw_cfg.config);
452
453 ret = avs_fill_gtw_config(adev, &cfg->gtw_cfg, t, dma_id, &cfg_size);
454 if (ret)
455 return ret;
456
457 cfg->base.cpc = t->cfg_base->cpc;
458 cfg->base.ibs = t->cfg_base->ibs;
459 cfg->base.obs = t->cfg_base->obs;
460 cfg->base.is_pages = t->cfg_base->is_pages;
461 cfg->base.audio_fmt = *t->in_fmt;
462 cfg->ref_fmt = *te->whm.ref_fmt;
463 cfg->out_fmt = *te->whm.out_fmt;
464 cfg->wake_tick_period = te->whm.wake_tick_period;
465 avs_init_node_id(&cfg->gtw_cfg.node_id, te, dma_id);
466 cfg->gtw_cfg.dma_buffer_size = te->whm.dma_buffer_size;
467 mod->gtw_attrs = cfg->gtw_cfg.config.attrs;
468
469 ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
470 t->domain, cfg, cfg_size, &mod->instance_id);
471 return ret;
472 }
473
avs_get_module_control(struct avs_path_module * mod,const char * name)474 static struct soc_mixer_control *avs_get_module_control(struct avs_path_module *mod,
475 const char *name)
476 {
477 struct avs_tplg_module *t = mod->template;
478 struct avs_tplg_path_template *path_tmpl;
479 struct snd_soc_dapm_widget *w;
480 int i;
481
482 path_tmpl = t->owner->owner->owner;
483 w = path_tmpl->w;
484
485 for (i = 0; i < w->num_kcontrols; i++) {
486 struct avs_control_data *ctl_data;
487 struct soc_mixer_control *mc;
488
489 mc = (struct soc_mixer_control *)w->kcontrols[i]->private_value;
490 ctl_data = (struct avs_control_data *)mc->dobj.private;
491 if (ctl_data->id == t->ctl_id && strstr(w->kcontrols[i]->id.name, name))
492 return mc;
493 }
494
495 return NULL;
496 }
497
avs_peakvol_set_volume(struct avs_dev * adev,struct avs_path_module * mod,struct soc_mixer_control * mc,long * input)498 int avs_peakvol_set_volume(struct avs_dev *adev, struct avs_path_module *mod,
499 struct soc_mixer_control *mc, long *input)
500 {
501 struct avs_volume_cfg vols[SND_SOC_TPLG_MAX_CHAN] = {{0}};
502 struct avs_control_data *ctl_data;
503 struct avs_tplg_module *t;
504 int ret, i;
505
506 ctl_data = mc->dobj.private;
507 t = mod->template;
508 if (!input)
509 input = ctl_data->values;
510
511 if (mc->num_channels) {
512 for (i = 0; i < mc->num_channels; i++) {
513 vols[i].channel_id = i;
514 vols[i].target_volume = input[i];
515 vols[i].curve_type = t->cfg_ext->peakvol.curve_type;
516 vols[i].curve_duration = t->cfg_ext->peakvol.curve_duration;
517 }
518
519 ret = avs_ipc_peakvol_set_volumes(adev, mod->module_id, mod->instance_id, vols,
520 mc->num_channels);
521 return AVS_IPC_RET(ret);
522 }
523
524 /* Target all channels if no individual selected. */
525 vols[0].channel_id = AVS_ALL_CHANNELS_MASK;
526 vols[0].target_volume = input[0];
527 vols[0].curve_type = t->cfg_ext->peakvol.curve_type;
528 vols[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
529
530 ret = avs_ipc_peakvol_set_volume(adev, mod->module_id, mod->instance_id, &vols[0]);
531 return AVS_IPC_RET(ret);
532 }
533
avs_peakvol_set_mute(struct avs_dev * adev,struct avs_path_module * mod,struct soc_mixer_control * mc,long * input)534 int avs_peakvol_set_mute(struct avs_dev *adev, struct avs_path_module *mod,
535 struct soc_mixer_control *mc, long *input)
536 {
537 struct avs_mute_cfg mutes[SND_SOC_TPLG_MAX_CHAN] = {{0}};
538 struct avs_control_data *ctl_data;
539 struct avs_tplg_module *t;
540 int ret, i;
541
542 ctl_data = mc->dobj.private;
543 t = mod->template;
544 if (!input)
545 input = ctl_data->values;
546
547 if (mc->num_channels) {
548 for (i = 0; i < mc->num_channels; i++) {
549 mutes[i].channel_id = i;
550 mutes[i].mute = !input[i];
551 mutes[i].curve_type = t->cfg_ext->peakvol.curve_type;
552 mutes[i].curve_duration = t->cfg_ext->peakvol.curve_duration;
553 }
554
555 ret = avs_ipc_peakvol_set_mutes(adev, mod->module_id, mod->instance_id, mutes,
556 mc->num_channels);
557 return AVS_IPC_RET(ret);
558 }
559
560 /* Target all channels if no individual selected. */
561 mutes[0].channel_id = AVS_ALL_CHANNELS_MASK;
562 mutes[0].mute = !input[0];
563 mutes[0].curve_type = t->cfg_ext->peakvol.curve_type;
564 mutes[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
565
566 ret = avs_ipc_peakvol_set_mute(adev, mod->module_id, mod->instance_id, &mutes[0]);
567 return AVS_IPC_RET(ret);
568 }
569
avs_peakvol_create(struct avs_dev * adev,struct avs_path_module * mod)570 static int avs_peakvol_create(struct avs_dev *adev, struct avs_path_module *mod)
571 {
572 struct avs_tplg_module *t = mod->template;
573 struct soc_mixer_control *mc;
574 struct avs_peakvol_cfg *cfg;
575 size_t cfg_size;
576 int ret;
577
578 cfg_size = struct_size(cfg, vols, 1);
579 if (cfg_size > AVS_MAILBOX_SIZE)
580 return -EINVAL;
581
582 cfg = adev->modcfg_buf;
583 memset(cfg, 0, cfg_size);
584 cfg->base.cpc = t->cfg_base->cpc;
585 cfg->base.ibs = t->cfg_base->ibs;
586 cfg->base.obs = t->cfg_base->obs;
587 cfg->base.is_pages = t->cfg_base->is_pages;
588 cfg->base.audio_fmt = *t->in_fmt;
589 cfg->vols[0].channel_id = AVS_ALL_CHANNELS_MASK;
590 cfg->vols[0].target_volume = S32_MAX;
591 cfg->vols[0].curve_type = t->cfg_ext->peakvol.curve_type;
592 cfg->vols[0].curve_duration = t->cfg_ext->peakvol.curve_duration;
593
594 ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id, t->core_id,
595 t->domain, cfg, cfg_size, &mod->instance_id);
596 if (ret)
597 return ret;
598
599 /* Now configure both VOLUME and MUTE parameters. */
600 mc = avs_get_module_control(mod, "Volume");
601 if (mc) {
602 ret = avs_peakvol_set_volume(adev, mod, mc, NULL);
603 if (ret)
604 return ret;
605 }
606
607 mc = avs_get_module_control(mod, "Switch");
608 if (mc)
609 return avs_peakvol_set_mute(adev, mod, mc, NULL);
610 return 0;
611 }
612
avs_updown_mix_create(struct avs_dev * adev,struct avs_path_module * mod)613 static int avs_updown_mix_create(struct avs_dev *adev, struct avs_path_module *mod)
614 {
615 struct avs_tplg_module *t = mod->template;
616 struct avs_updown_mixer_cfg cfg;
617 int i;
618
619 cfg.base.cpc = t->cfg_base->cpc;
620 cfg.base.ibs = t->cfg_base->ibs;
621 cfg.base.obs = t->cfg_base->obs;
622 cfg.base.is_pages = t->cfg_base->is_pages;
623 cfg.base.audio_fmt = *t->in_fmt;
624 cfg.out_channel_config = t->cfg_ext->updown_mix.out_channel_config;
625 cfg.coefficients_select = t->cfg_ext->updown_mix.coefficients_select;
626 for (i = 0; i < AVS_COEFF_CHANNELS_MAX; i++)
627 cfg.coefficients[i] = t->cfg_ext->updown_mix.coefficients[i];
628 cfg.channel_map = t->cfg_ext->updown_mix.channel_map;
629
630 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
631 t->core_id, t->domain, &cfg, sizeof(cfg),
632 &mod->instance_id);
633 }
634
avs_src_create(struct avs_dev * adev,struct avs_path_module * mod)635 static int avs_src_create(struct avs_dev *adev, struct avs_path_module *mod)
636 {
637 struct avs_tplg_module *t = mod->template;
638 struct avs_src_cfg cfg;
639
640 cfg.base.cpc = t->cfg_base->cpc;
641 cfg.base.ibs = t->cfg_base->ibs;
642 cfg.base.obs = t->cfg_base->obs;
643 cfg.base.is_pages = t->cfg_base->is_pages;
644 cfg.base.audio_fmt = *t->in_fmt;
645 cfg.out_freq = t->cfg_ext->src.out_freq;
646
647 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
648 t->core_id, t->domain, &cfg, sizeof(cfg),
649 &mod->instance_id);
650 }
651
avs_asrc_create(struct avs_dev * adev,struct avs_path_module * mod)652 static int avs_asrc_create(struct avs_dev *adev, struct avs_path_module *mod)
653 {
654 struct avs_tplg_module *t = mod->template;
655 struct avs_asrc_cfg cfg;
656
657 memset(&cfg, 0, sizeof(cfg));
658 cfg.base.cpc = t->cfg_base->cpc;
659 cfg.base.ibs = t->cfg_base->ibs;
660 cfg.base.obs = t->cfg_base->obs;
661 cfg.base.is_pages = t->cfg_base->is_pages;
662 cfg.base.audio_fmt = *t->in_fmt;
663 cfg.out_freq = t->cfg_ext->asrc.out_freq;
664 cfg.mode = t->cfg_ext->asrc.mode;
665 cfg.disable_jitter_buffer = t->cfg_ext->asrc.disable_jitter_buffer;
666
667 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
668 t->core_id, t->domain, &cfg, sizeof(cfg),
669 &mod->instance_id);
670 }
671
avs_aec_create(struct avs_dev * adev,struct avs_path_module * mod)672 static int avs_aec_create(struct avs_dev *adev, struct avs_path_module *mod)
673 {
674 struct avs_tplg_module *t = mod->template;
675 struct avs_aec_cfg cfg;
676
677 cfg.base.cpc = t->cfg_base->cpc;
678 cfg.base.ibs = t->cfg_base->ibs;
679 cfg.base.obs = t->cfg_base->obs;
680 cfg.base.is_pages = t->cfg_base->is_pages;
681 cfg.base.audio_fmt = *t->in_fmt;
682 cfg.ref_fmt = *t->cfg_ext->aec.ref_fmt;
683 cfg.out_fmt = *t->cfg_ext->aec.out_fmt;
684 cfg.cpc_lp_mode = t->cfg_ext->aec.cpc_lp_mode;
685
686 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
687 t->core_id, t->domain, &cfg, sizeof(cfg),
688 &mod->instance_id);
689 }
690
avs_mux_create(struct avs_dev * adev,struct avs_path_module * mod)691 static int avs_mux_create(struct avs_dev *adev, struct avs_path_module *mod)
692 {
693 struct avs_tplg_module *t = mod->template;
694 struct avs_mux_cfg cfg;
695
696 cfg.base.cpc = t->cfg_base->cpc;
697 cfg.base.ibs = t->cfg_base->ibs;
698 cfg.base.obs = t->cfg_base->obs;
699 cfg.base.is_pages = t->cfg_base->is_pages;
700 cfg.base.audio_fmt = *t->in_fmt;
701 cfg.ref_fmt = *t->cfg_ext->mux.ref_fmt;
702 cfg.out_fmt = *t->cfg_ext->mux.out_fmt;
703
704 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
705 t->core_id, t->domain, &cfg, sizeof(cfg),
706 &mod->instance_id);
707 }
708
avs_wov_create(struct avs_dev * adev,struct avs_path_module * mod)709 static int avs_wov_create(struct avs_dev *adev, struct avs_path_module *mod)
710 {
711 struct avs_tplg_module *t = mod->template;
712 struct avs_wov_cfg cfg;
713
714 cfg.base.cpc = t->cfg_base->cpc;
715 cfg.base.ibs = t->cfg_base->ibs;
716 cfg.base.obs = t->cfg_base->obs;
717 cfg.base.is_pages = t->cfg_base->is_pages;
718 cfg.base.audio_fmt = *t->in_fmt;
719 cfg.cpc_lp_mode = t->cfg_ext->wov.cpc_lp_mode;
720
721 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
722 t->core_id, t->domain, &cfg, sizeof(cfg),
723 &mod->instance_id);
724 }
725
avs_micsel_create(struct avs_dev * adev,struct avs_path_module * mod)726 static int avs_micsel_create(struct avs_dev *adev, struct avs_path_module *mod)
727 {
728 struct avs_tplg_module *t = mod->template;
729 struct avs_micsel_cfg cfg;
730
731 cfg.base.cpc = t->cfg_base->cpc;
732 cfg.base.ibs = t->cfg_base->ibs;
733 cfg.base.obs = t->cfg_base->obs;
734 cfg.base.is_pages = t->cfg_base->is_pages;
735 cfg.base.audio_fmt = *t->in_fmt;
736 cfg.out_fmt = *t->cfg_ext->micsel.out_fmt;
737
738 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
739 t->core_id, t->domain, &cfg, sizeof(cfg),
740 &mod->instance_id);
741 }
742
avs_modbase_create(struct avs_dev * adev,struct avs_path_module * mod)743 static int avs_modbase_create(struct avs_dev *adev, struct avs_path_module *mod)
744 {
745 struct avs_tplg_module *t = mod->template;
746 struct avs_modcfg_base cfg;
747
748 cfg.cpc = t->cfg_base->cpc;
749 cfg.ibs = t->cfg_base->ibs;
750 cfg.obs = t->cfg_base->obs;
751 cfg.is_pages = t->cfg_base->is_pages;
752 cfg.audio_fmt = *t->in_fmt;
753
754 return avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
755 t->core_id, t->domain, &cfg, sizeof(cfg),
756 &mod->instance_id);
757 }
758
avs_modext_create(struct avs_dev * adev,struct avs_path_module * mod)759 static int avs_modext_create(struct avs_dev *adev, struct avs_path_module *mod)
760 {
761 struct avs_tplg_module *t = mod->template;
762 struct avs_tplg_modcfg_ext *tcfg = t->cfg_ext;
763 struct avs_modcfg_ext *cfg;
764 size_t cfg_size, num_pins;
765 int ret, i;
766
767 num_pins = tcfg->generic.num_input_pins + tcfg->generic.num_output_pins;
768 cfg_size = struct_size(cfg, pin_fmts, num_pins);
769
770 if (cfg_size > AVS_MAILBOX_SIZE)
771 return -EINVAL;
772
773 cfg = adev->modcfg_buf;
774 memset(cfg, 0, cfg_size);
775 cfg->base.cpc = t->cfg_base->cpc;
776 cfg->base.ibs = t->cfg_base->ibs;
777 cfg->base.obs = t->cfg_base->obs;
778 cfg->base.is_pages = t->cfg_base->is_pages;
779 cfg->base.audio_fmt = *t->in_fmt;
780 cfg->num_input_pins = tcfg->generic.num_input_pins;
781 cfg->num_output_pins = tcfg->generic.num_output_pins;
782
783 /* configure pin formats */
784 for (i = 0; i < num_pins; i++) {
785 struct avs_tplg_pin_format *tpin = &tcfg->generic.pin_fmts[i];
786 struct avs_pin_format *pin = &cfg->pin_fmts[i];
787
788 pin->pin_index = tpin->pin_index;
789 pin->iobs = tpin->iobs;
790 pin->audio_fmt = *tpin->fmt;
791 }
792
793 ret = avs_dsp_init_module(adev, mod->module_id, mod->owner->instance_id,
794 t->core_id, t->domain, cfg, cfg_size,
795 &mod->instance_id);
796 return ret;
797 }
798
avs_probe_create(struct avs_dev * adev,struct avs_path_module * mod)799 static int avs_probe_create(struct avs_dev *adev, struct avs_path_module *mod)
800 {
801 dev_err(adev->dev, "Probe module can't be instantiated by topology");
802 return -EINVAL;
803 }
804
805 struct avs_module_create {
806 guid_t *guid;
807 int (*create)(struct avs_dev *adev, struct avs_path_module *mod);
808 };
809
810 static struct avs_module_create avs_module_create[] = {
811 { &AVS_MIXIN_MOD_UUID, avs_modbase_create },
812 { &AVS_MIXOUT_MOD_UUID, avs_modbase_create },
813 { &AVS_KPBUFF_MOD_UUID, avs_modbase_create },
814 { &AVS_COPIER_MOD_UUID, avs_copier_create },
815 { &AVS_PEAKVOL_MOD_UUID, avs_peakvol_create },
816 { &AVS_GAIN_MOD_UUID, avs_peakvol_create },
817 { &AVS_MICSEL_MOD_UUID, avs_micsel_create },
818 { &AVS_MUX_MOD_UUID, avs_mux_create },
819 { &AVS_UPDWMIX_MOD_UUID, avs_updown_mix_create },
820 { &AVS_SRCINTC_MOD_UUID, avs_src_create },
821 { &AVS_AEC_MOD_UUID, avs_aec_create },
822 { &AVS_ASRC_MOD_UUID, avs_asrc_create },
823 { &AVS_INTELWOV_MOD_UUID, avs_wov_create },
824 { &AVS_PROBE_MOD_UUID, avs_probe_create },
825 { &AVS_WOVHOSTM_MOD_UUID, avs_whm_create },
826 };
827
avs_path_module_type_create(struct avs_dev * adev,struct avs_path_module * mod)828 static int avs_path_module_type_create(struct avs_dev *adev, struct avs_path_module *mod)
829 {
830 const guid_t *type = &mod->template->cfg_ext->type;
831
832 for (int i = 0; i < ARRAY_SIZE(avs_module_create); i++)
833 if (guid_equal(type, avs_module_create[i].guid))
834 return avs_module_create[i].create(adev, mod);
835
836 return avs_modext_create(adev, mod);
837 }
838
avs_path_module_send_init_configs(struct avs_dev * adev,struct avs_path_module * mod)839 static int avs_path_module_send_init_configs(struct avs_dev *adev, struct avs_path_module *mod)
840 {
841 struct avs_soc_component *acomp;
842
843 acomp = to_avs_soc_component(mod->template->owner->owner->owner->owner->comp);
844
845 u32 num_ids = mod->template->num_config_ids;
846 u32 *ids = mod->template->config_ids;
847
848 for (int i = 0; i < num_ids; i++) {
849 struct avs_tplg_init_config *config = &acomp->tplg->init_configs[ids[i]];
850 size_t len = config->length;
851 void *data = config->data;
852 u32 param = config->param;
853 int ret;
854
855 ret = avs_ipc_set_large_config(adev, mod->module_id, mod->instance_id,
856 param, data, len);
857 if (ret) {
858 dev_err(adev->dev, "send initial module config failed: %d\n", ret);
859 return AVS_IPC_RET(ret);
860 }
861 }
862
863 return 0;
864 }
865
avs_path_module_free(struct avs_dev * adev,struct avs_path_module * mod)866 static void avs_path_module_free(struct avs_dev *adev, struct avs_path_module *mod)
867 {
868 kfree(mod);
869 }
870
871 static struct avs_path_module *
avs_path_module_create(struct avs_dev * adev,struct avs_path_pipeline * owner,struct avs_tplg_module * template)872 avs_path_module_create(struct avs_dev *adev,
873 struct avs_path_pipeline *owner,
874 struct avs_tplg_module *template)
875 {
876 struct avs_path_module *mod;
877 int module_id, ret;
878
879 module_id = avs_get_module_id(adev, &template->cfg_ext->type);
880 if (module_id < 0)
881 return ERR_PTR(module_id);
882
883 mod = kzalloc_obj(*mod);
884 if (!mod)
885 return ERR_PTR(-ENOMEM);
886
887 mod->template = template;
888 mod->module_id = module_id;
889 mod->owner = owner;
890 INIT_LIST_HEAD(&mod->node);
891
892 ret = avs_path_module_type_create(adev, mod);
893 if (ret) {
894 dev_err(adev->dev, "module-type create failed: %d\n", ret);
895 kfree(mod);
896 return ERR_PTR(ret);
897 }
898
899 ret = avs_path_module_send_init_configs(adev, mod);
900 if (ret) {
901 kfree(mod);
902 return ERR_PTR(ret);
903 }
904
905 return mod;
906 }
907
avs_path_binding_arm(struct avs_dev * adev,struct avs_path_binding * binding)908 static int avs_path_binding_arm(struct avs_dev *adev, struct avs_path_binding *binding)
909 {
910 struct avs_path_module *this_mod, *target_mod;
911 struct avs_path_pipeline *target_ppl;
912 struct avs_path *target_path;
913 struct avs_tplg_binding *t;
914
915 t = binding->template;
916 this_mod = avs_path_find_module(binding->owner,
917 t->mod_id);
918 if (!this_mod) {
919 dev_err(adev->dev, "path mod %d not found\n", t->mod_id);
920 return -EINVAL;
921 }
922
923 /* update with target_tplg_name too */
924 target_path = avs_path_find_path(adev, t->target_tplg_name,
925 t->target_path_tmpl_id);
926 if (!target_path) {
927 dev_err(adev->dev, "target path %s:%d not found\n",
928 t->target_tplg_name, t->target_path_tmpl_id);
929 return -EINVAL;
930 }
931
932 target_ppl = avs_path_find_pipeline(target_path,
933 t->target_ppl_id);
934 if (!target_ppl) {
935 dev_err(adev->dev, "target ppl %d not found\n", t->target_ppl_id);
936 return -EINVAL;
937 }
938
939 target_mod = avs_path_find_module(target_ppl, t->target_mod_id);
940 if (!target_mod) {
941 dev_err(adev->dev, "target mod %d not found\n", t->target_mod_id);
942 return -EINVAL;
943 }
944
945 if (t->is_sink) {
946 binding->sink = this_mod;
947 binding->sink_pin = t->mod_pin;
948 binding->source = target_mod;
949 binding->source_pin = t->target_mod_pin;
950 } else {
951 binding->sink = target_mod;
952 binding->sink_pin = t->target_mod_pin;
953 binding->source = this_mod;
954 binding->source_pin = t->mod_pin;
955 }
956
957 return 0;
958 }
959
avs_path_binding_free(struct avs_dev * adev,struct avs_path_binding * binding)960 static void avs_path_binding_free(struct avs_dev *adev, struct avs_path_binding *binding)
961 {
962 kfree(binding);
963 }
964
avs_path_binding_create(struct avs_dev * adev,struct avs_path_pipeline * owner,struct avs_tplg_binding * t)965 static struct avs_path_binding *avs_path_binding_create(struct avs_dev *adev,
966 struct avs_path_pipeline *owner,
967 struct avs_tplg_binding *t)
968 {
969 struct avs_path_binding *binding;
970
971 binding = kzalloc_obj(*binding);
972 if (!binding)
973 return ERR_PTR(-ENOMEM);
974
975 binding->template = t;
976 binding->owner = owner;
977 INIT_LIST_HEAD(&binding->node);
978
979 return binding;
980 }
981
avs_path_pipeline_arm(struct avs_dev * adev,struct avs_path_pipeline * ppl)982 static int avs_path_pipeline_arm(struct avs_dev *adev,
983 struct avs_path_pipeline *ppl)
984 {
985 struct avs_path_module *mod;
986
987 list_for_each_entry(mod, &ppl->mod_list, node) {
988 struct avs_path_module *source, *sink;
989 int ret;
990
991 /*
992 * Only one module (so it's implicitly last) or it is the last
993 * one, either way we don't have next module to bind it to.
994 */
995 if (mod == list_last_entry(&ppl->mod_list,
996 struct avs_path_module, node))
997 break;
998
999 /* bind current module to next module on list */
1000 source = mod;
1001 sink = list_next_entry(mod, node);
1002
1003 ret = avs_ipc_bind(adev, source->module_id, source->instance_id,
1004 sink->module_id, sink->instance_id, 0, 0);
1005 if (ret)
1006 return AVS_IPC_RET(ret);
1007 }
1008
1009 return 0;
1010 }
1011
avs_path_pipeline_free(struct avs_dev * adev,struct avs_path_pipeline * ppl)1012 static void avs_path_pipeline_free(struct avs_dev *adev,
1013 struct avs_path_pipeline *ppl)
1014 {
1015 struct avs_path_binding *binding, *bsave;
1016 struct avs_path_module *mod, *save;
1017
1018 list_for_each_entry_safe(binding, bsave, &ppl->binding_list, node) {
1019 list_del(&binding->node);
1020 avs_path_binding_free(adev, binding);
1021 }
1022
1023 avs_dsp_delete_pipeline(adev, ppl->instance_id);
1024
1025 /* Unload resources occupied by owned modules */
1026 list_for_each_entry_safe(mod, save, &ppl->mod_list, node) {
1027 avs_dsp_delete_module(adev, mod->module_id, mod->instance_id,
1028 mod->owner->instance_id,
1029 mod->template->core_id);
1030 avs_path_module_free(adev, mod);
1031 }
1032
1033 list_del(&ppl->node);
1034 kfree(ppl);
1035 }
1036
1037 static struct avs_path_pipeline *
avs_path_pipeline_create(struct avs_dev * adev,struct avs_path * owner,struct avs_tplg_pipeline * template)1038 avs_path_pipeline_create(struct avs_dev *adev, struct avs_path *owner,
1039 struct avs_tplg_pipeline *template)
1040 {
1041 struct avs_path_pipeline *ppl;
1042 struct avs_tplg_pplcfg *cfg = template->cfg;
1043 struct avs_tplg_module *tmod;
1044 int ret, i;
1045
1046 ppl = kzalloc_obj(*ppl);
1047 if (!ppl)
1048 return ERR_PTR(-ENOMEM);
1049
1050 ppl->template = template;
1051 ppl->owner = owner;
1052 INIT_LIST_HEAD(&ppl->binding_list);
1053 INIT_LIST_HEAD(&ppl->mod_list);
1054 INIT_LIST_HEAD(&ppl->node);
1055
1056 ret = avs_dsp_create_pipeline(adev, cfg->req_size, cfg->priority,
1057 cfg->lp, cfg->attributes,
1058 &ppl->instance_id);
1059 if (ret) {
1060 dev_err(adev->dev, "error creating pipeline %d\n", ret);
1061 kfree(ppl);
1062 return ERR_PTR(ret);
1063 }
1064
1065 list_for_each_entry(tmod, &template->mod_list, node) {
1066 struct avs_path_module *mod;
1067
1068 mod = avs_path_module_create(adev, ppl, tmod);
1069 if (IS_ERR(mod)) {
1070 ret = PTR_ERR(mod);
1071 dev_err(adev->dev, "error creating module %d\n", ret);
1072 goto init_err;
1073 }
1074
1075 list_add_tail(&mod->node, &ppl->mod_list);
1076 }
1077
1078 for (i = 0; i < template->num_bindings; i++) {
1079 struct avs_path_binding *binding;
1080
1081 binding = avs_path_binding_create(adev, ppl, template->bindings[i]);
1082 if (IS_ERR(binding)) {
1083 ret = PTR_ERR(binding);
1084 dev_err(adev->dev, "error creating binding %d\n", ret);
1085 goto init_err;
1086 }
1087
1088 list_add_tail(&binding->node, &ppl->binding_list);
1089 }
1090
1091 return ppl;
1092
1093 init_err:
1094 avs_path_pipeline_free(adev, ppl);
1095 return ERR_PTR(ret);
1096 }
1097
avs_path_init(struct avs_dev * adev,struct avs_path * path,struct avs_tplg_path * template,u32 dma_id)1098 static int avs_path_init(struct avs_dev *adev, struct avs_path *path,
1099 struct avs_tplg_path *template, u32 dma_id)
1100 {
1101 struct avs_tplg_pipeline *tppl;
1102
1103 path->owner = adev;
1104 path->template = template;
1105 path->dma_id = dma_id;
1106 INIT_LIST_HEAD(&path->ppl_list);
1107 INIT_LIST_HEAD(&path->node);
1108 INIT_LIST_HEAD(&path->source_list);
1109 INIT_LIST_HEAD(&path->sink_list);
1110 INIT_LIST_HEAD(&path->source_node);
1111 INIT_LIST_HEAD(&path->sink_node);
1112
1113 /* create all the pipelines */
1114 list_for_each_entry(tppl, &template->ppl_list, node) {
1115 struct avs_path_pipeline *ppl;
1116
1117 ppl = avs_path_pipeline_create(adev, path, tppl);
1118 if (IS_ERR(ppl))
1119 return PTR_ERR(ppl);
1120
1121 list_add_tail(&ppl->node, &path->ppl_list);
1122 }
1123
1124 spin_lock(&adev->path_list_lock);
1125 list_add_tail(&path->node, &adev->path_list);
1126 spin_unlock(&adev->path_list_lock);
1127
1128 return 0;
1129 }
1130
avs_path_arm(struct avs_dev * adev,struct avs_path * path)1131 static int avs_path_arm(struct avs_dev *adev, struct avs_path *path)
1132 {
1133 struct avs_path_pipeline *ppl;
1134 struct avs_path_binding *binding;
1135 int ret;
1136
1137 list_for_each_entry(ppl, &path->ppl_list, node) {
1138 /*
1139 * Arm all ppl bindings before binding internal modules
1140 * as it costs no IPCs which isn't true for the latter.
1141 */
1142 list_for_each_entry(binding, &ppl->binding_list, node) {
1143 ret = avs_path_binding_arm(adev, binding);
1144 if (ret < 0)
1145 return ret;
1146 }
1147
1148 ret = avs_path_pipeline_arm(adev, ppl);
1149 if (ret < 0)
1150 return ret;
1151 }
1152
1153 return 0;
1154 }
1155
avs_path_free_unlocked(struct avs_path * path)1156 static void avs_path_free_unlocked(struct avs_path *path)
1157 {
1158 struct avs_path_pipeline *ppl, *save;
1159
1160 spin_lock(&path->owner->path_list_lock);
1161 list_del(&path->node);
1162 spin_unlock(&path->owner->path_list_lock);
1163
1164 list_for_each_entry_safe(ppl, save, &path->ppl_list, node)
1165 avs_path_pipeline_free(path->owner, ppl);
1166
1167 kfree(path);
1168 }
1169
avs_path_create_unlocked(struct avs_dev * adev,u32 dma_id,struct avs_tplg_path * template)1170 static struct avs_path *avs_path_create_unlocked(struct avs_dev *adev, u32 dma_id,
1171 struct avs_tplg_path *template)
1172 {
1173 struct avs_path *path;
1174 int ret;
1175
1176 path = kzalloc_obj(*path);
1177 if (!path)
1178 return ERR_PTR(-ENOMEM);
1179
1180 ret = avs_path_init(adev, path, template, dma_id);
1181 if (ret < 0)
1182 goto err;
1183
1184 ret = avs_path_arm(adev, path);
1185 if (ret < 0)
1186 goto err;
1187
1188 path->state = AVS_PPL_STATE_INVALID;
1189 return path;
1190 err:
1191 avs_path_free_unlocked(path);
1192 return ERR_PTR(ret);
1193 }
1194
avs_condpath_free(struct avs_dev * adev,struct avs_path * path)1195 static void avs_condpath_free(struct avs_dev *adev, struct avs_path *path)
1196 {
1197 int ret;
1198
1199 list_del(&path->source_node);
1200 list_del(&path->sink_node);
1201
1202 ret = avs_path_reset(path);
1203 if (ret < 0)
1204 dev_err(adev->dev, "reset condpath failed: %d\n", ret);
1205
1206 ret = avs_path_unbind(path);
1207 if (ret < 0)
1208 dev_err(adev->dev, "unbind condpath failed: %d\n", ret);
1209
1210 avs_path_free_unlocked(path);
1211 }
1212
avs_condpath_create(struct avs_dev * adev,struct avs_tplg_path * template,struct avs_path * source,struct avs_path * sink)1213 static struct avs_path *avs_condpath_create(struct avs_dev *adev,
1214 struct avs_tplg_path *template,
1215 struct avs_path *source,
1216 struct avs_path *sink)
1217 {
1218 struct avs_path *path;
1219 int ret;
1220
1221 path = avs_path_create_unlocked(adev, 0, template);
1222 if (IS_ERR(path))
1223 return path;
1224
1225 ret = avs_path_bind(path);
1226 if (ret)
1227 goto err_bind;
1228
1229 ret = avs_path_reset(path);
1230 if (ret)
1231 goto err_reset;
1232
1233 path->source = source;
1234 path->sink = sink;
1235 list_add_tail(&path->source_node, &source->source_list);
1236 list_add_tail(&path->sink_node, &sink->sink_list);
1237
1238 return path;
1239
1240 err_reset:
1241 avs_path_unbind(path);
1242 err_bind:
1243 avs_path_free_unlocked(path);
1244 return ERR_PTR(ret);
1245 }
1246
avs_condpaths_walk(struct avs_dev * adev,struct avs_path * path,int dir)1247 static int avs_condpaths_walk(struct avs_dev *adev, struct avs_path *path, int dir)
1248 {
1249 struct avs_soc_component *acomp;
1250 struct avs_path *source, *sink;
1251 struct avs_path **other;
1252
1253 if (dir) {
1254 source = path;
1255 other = &sink;
1256 } else {
1257 sink = path;
1258 other = &source;
1259 }
1260
1261 list_for_each_entry(acomp, &adev->comp_list, node) {
1262 for (int i = 0; i < acomp->tplg->num_condpath_tmpls; i++) {
1263 struct avs_tplg_path_template *template;
1264 struct avs_tplg_path *variant;
1265 struct avs_path *cpath;
1266
1267 template = &acomp->tplg->condpath_tmpls[i];
1268
1269 /* Do not create unidirectional condpaths twice. */
1270 if (avs_tplg_path_template_id_equal(&template->source,
1271 &template->sink) && dir)
1272 continue;
1273
1274 *other = avs_condpath_find_match(adev, template, path, dir);
1275 if (!*other)
1276 continue;
1277
1278 variant = avs_condpath_find_variant(adev, template, source, sink);
1279 if (!variant)
1280 continue;
1281
1282 cpath = avs_condpath_create(adev, variant, source, sink);
1283 if (IS_ERR(cpath))
1284 return PTR_ERR(cpath);
1285 }
1286 }
1287
1288 return 0;
1289 }
1290
1291 /* Caller responsible for holding adev->path_mutex. */
avs_condpaths_walk_all(struct avs_dev * adev,struct avs_path * path)1292 static int avs_condpaths_walk_all(struct avs_dev *adev, struct avs_path *path)
1293 {
1294 int ret;
1295
1296 ret = avs_condpaths_walk(adev, path, SNDRV_PCM_STREAM_CAPTURE);
1297 if (ret)
1298 return ret;
1299
1300 return avs_condpaths_walk(adev, path, SNDRV_PCM_STREAM_PLAYBACK);
1301 }
1302
avs_path_free(struct avs_path * path)1303 void avs_path_free(struct avs_path *path)
1304 {
1305 struct avs_path *cpath, *csave;
1306 struct avs_dev *adev = path->owner;
1307
1308 mutex_lock(&adev->path_mutex);
1309
1310 /* Free all condpaths this path spawned. */
1311 list_for_each_entry_safe(cpath, csave, &path->source_list, source_node)
1312 avs_condpath_free(path->owner, cpath);
1313 list_for_each_entry_safe(cpath, csave, &path->sink_list, sink_node)
1314 avs_condpath_free(path->owner, cpath);
1315
1316 avs_path_free_unlocked(path);
1317
1318 mutex_unlock(&adev->path_mutex);
1319 }
1320
avs_path_create(struct avs_dev * adev,u32 dma_id,struct avs_tplg_path_template * template,struct snd_pcm_hw_params * fe_params,struct snd_pcm_hw_params * be_params)1321 struct avs_path *avs_path_create(struct avs_dev *adev, u32 dma_id,
1322 struct avs_tplg_path_template *template,
1323 struct snd_pcm_hw_params *fe_params,
1324 struct snd_pcm_hw_params *be_params)
1325 {
1326 struct avs_tplg_path *variant;
1327 struct avs_path *path;
1328 int ret;
1329
1330 variant = avs_path_find_variant(adev, template, fe_params, be_params);
1331 if (!variant) {
1332 dev_err(adev->dev, "no matching variant found\n");
1333 return ERR_PTR(-ENOENT);
1334 }
1335
1336 /* Serialize path and its components creation. */
1337 mutex_lock(&adev->path_mutex);
1338 /* Satisfy needs of avs_path_find_tplg(). */
1339 mutex_lock(&adev->comp_list_mutex);
1340
1341 path = avs_path_create_unlocked(adev, dma_id, variant);
1342 if (IS_ERR(path))
1343 goto exit;
1344
1345 ret = avs_condpaths_walk_all(adev, path);
1346 if (ret) {
1347 avs_path_free_unlocked(path);
1348 path = ERR_PTR(ret);
1349 }
1350
1351 exit:
1352 mutex_unlock(&adev->comp_list_mutex);
1353 mutex_unlock(&adev->path_mutex);
1354
1355 return path;
1356 }
1357
avs_path_bind_prepare(struct avs_dev * adev,struct avs_path_binding * binding)1358 static int avs_path_bind_prepare(struct avs_dev *adev,
1359 struct avs_path_binding *binding)
1360 {
1361 const struct avs_audio_format *src_fmt, *sink_fmt;
1362 struct avs_tplg_module *tsource = binding->source->template;
1363 struct avs_path_module *source = binding->source;
1364 int ret;
1365
1366 /*
1367 * only copier modules about to be bound
1368 * to output pin other than 0 need preparation
1369 */
1370 if (!binding->source_pin)
1371 return 0;
1372 if (!guid_equal(&tsource->cfg_ext->type, &AVS_COPIER_MOD_UUID))
1373 return 0;
1374
1375 src_fmt = tsource->in_fmt;
1376 sink_fmt = binding->sink->template->in_fmt;
1377
1378 ret = avs_ipc_copier_set_sink_format(adev, source->module_id,
1379 source->instance_id, binding->source_pin,
1380 src_fmt, sink_fmt);
1381 if (ret) {
1382 dev_err(adev->dev, "config copier failed: %d\n", ret);
1383 return AVS_IPC_RET(ret);
1384 }
1385
1386 return 0;
1387 }
1388
avs_path_bind(struct avs_path * path)1389 int avs_path_bind(struct avs_path *path)
1390 {
1391 struct avs_path_pipeline *ppl;
1392 struct avs_dev *adev = path->owner;
1393 int ret;
1394
1395 list_for_each_entry(ppl, &path->ppl_list, node) {
1396 struct avs_path_binding *binding;
1397
1398 list_for_each_entry(binding, &ppl->binding_list, node) {
1399 struct avs_path_module *source, *sink;
1400
1401 source = binding->source;
1402 sink = binding->sink;
1403
1404 ret = avs_path_bind_prepare(adev, binding);
1405 if (ret < 0)
1406 return ret;
1407
1408 ret = avs_ipc_bind(adev, source->module_id,
1409 source->instance_id, sink->module_id,
1410 sink->instance_id, binding->sink_pin,
1411 binding->source_pin);
1412 if (ret) {
1413 dev_err(adev->dev, "bind path failed: %d\n", ret);
1414 return AVS_IPC_RET(ret);
1415 }
1416 }
1417 }
1418
1419 return 0;
1420 }
1421
avs_path_unbind(struct avs_path * path)1422 int avs_path_unbind(struct avs_path *path)
1423 {
1424 struct avs_path_pipeline *ppl;
1425 struct avs_dev *adev = path->owner;
1426 int ret;
1427
1428 list_for_each_entry(ppl, &path->ppl_list, node) {
1429 struct avs_path_binding *binding;
1430
1431 list_for_each_entry(binding, &ppl->binding_list, node) {
1432 struct avs_path_module *source, *sink;
1433
1434 source = binding->source;
1435 sink = binding->sink;
1436
1437 ret = avs_ipc_unbind(adev, source->module_id,
1438 source->instance_id, sink->module_id,
1439 sink->instance_id, binding->sink_pin,
1440 binding->source_pin);
1441 if (ret) {
1442 dev_err(adev->dev, "unbind path failed: %d\n", ret);
1443 return AVS_IPC_RET(ret);
1444 }
1445 }
1446 }
1447
1448 return 0;
1449 }
1450
avs_path_reset(struct avs_path * path)1451 int avs_path_reset(struct avs_path *path)
1452 {
1453 struct avs_path_pipeline *ppl;
1454 struct avs_dev *adev = path->owner;
1455 int ret;
1456
1457 if (path->state == AVS_PPL_STATE_RESET)
1458 return 0;
1459
1460 list_for_each_entry(ppl, &path->ppl_list, node) {
1461 ret = avs_ipc_set_pipeline_state(adev, ppl->instance_id,
1462 AVS_PPL_STATE_RESET);
1463 if (ret) {
1464 dev_err(adev->dev, "reset path failed: %d\n", ret);
1465 path->state = AVS_PPL_STATE_INVALID;
1466 return AVS_IPC_RET(ret);
1467 }
1468 }
1469
1470 path->state = AVS_PPL_STATE_RESET;
1471 return 0;
1472 }
1473
avs_condpath_pause(struct avs_dev * adev,struct avs_path * cpath)1474 static int avs_condpath_pause(struct avs_dev *adev, struct avs_path *cpath)
1475 {
1476 struct avs_path_pipeline *ppl;
1477 int ret;
1478
1479 if (cpath->state == AVS_PPL_STATE_PAUSED)
1480 return 0;
1481
1482 list_for_each_entry_reverse(ppl, &cpath->ppl_list, node) {
1483 ret = avs_ipc_set_pipeline_state(adev, ppl->instance_id, AVS_PPL_STATE_PAUSED);
1484 if (ret) {
1485 dev_err(adev->dev, "pause cpath failed: %d\n", ret);
1486 cpath->state = AVS_PPL_STATE_INVALID;
1487 return AVS_IPC_RET(ret);
1488 }
1489 }
1490
1491 cpath->state = AVS_PPL_STATE_PAUSED;
1492 return 0;
1493 }
1494
avs_condpaths_pause(struct avs_dev * adev,struct avs_path * path)1495 static void avs_condpaths_pause(struct avs_dev *adev, struct avs_path *path)
1496 {
1497 struct avs_path *cpath;
1498
1499 mutex_lock(&adev->path_mutex);
1500
1501 /* If either source or sink stops, so do the attached conditional paths. */
1502 list_for_each_entry(cpath, &path->source_list, source_node)
1503 avs_condpath_pause(adev, cpath);
1504 list_for_each_entry(cpath, &path->sink_list, sink_node)
1505 avs_condpath_pause(adev, cpath);
1506
1507 mutex_unlock(&adev->path_mutex);
1508 }
1509
avs_path_pause(struct avs_path * path)1510 int avs_path_pause(struct avs_path *path)
1511 {
1512 struct avs_path_pipeline *ppl;
1513 struct avs_dev *adev = path->owner;
1514 int ret;
1515
1516 if (path->state == AVS_PPL_STATE_PAUSED)
1517 return 0;
1518
1519 avs_condpaths_pause(adev, path);
1520
1521 list_for_each_entry_reverse(ppl, &path->ppl_list, node) {
1522 ret = avs_ipc_set_pipeline_state(adev, ppl->instance_id,
1523 AVS_PPL_STATE_PAUSED);
1524 if (ret) {
1525 dev_err(adev->dev, "pause path failed: %d\n", ret);
1526 path->state = AVS_PPL_STATE_INVALID;
1527 return AVS_IPC_RET(ret);
1528 }
1529 }
1530
1531 path->state = AVS_PPL_STATE_PAUSED;
1532 return 0;
1533 }
1534
avs_condpath_run(struct avs_dev * adev,struct avs_path * cpath,int trigger)1535 static int avs_condpath_run(struct avs_dev *adev, struct avs_path *cpath, int trigger)
1536 {
1537 struct avs_path_pipeline *ppl;
1538 int ret;
1539
1540 if (cpath->state == AVS_PPL_STATE_RUNNING)
1541 return 0;
1542
1543 list_for_each_entry(ppl, &cpath->ppl_list, node) {
1544 if (ppl->template->cfg->trigger != trigger)
1545 continue;
1546
1547 ret = avs_ipc_set_pipeline_state(adev, ppl->instance_id, AVS_PPL_STATE_RUNNING);
1548 if (ret) {
1549 dev_err(adev->dev, "run cpath failed: %d\n", ret);
1550 cpath->state = AVS_PPL_STATE_INVALID;
1551 return AVS_IPC_RET(ret);
1552 }
1553 }
1554
1555 cpath->state = AVS_PPL_STATE_RUNNING;
1556 return 0;
1557 }
1558
avs_condpaths_run(struct avs_dev * adev,struct avs_path * path,int trigger)1559 static void avs_condpaths_run(struct avs_dev *adev, struct avs_path *path, int trigger)
1560 {
1561 struct avs_path *cpath;
1562
1563 mutex_lock(&adev->path_mutex);
1564
1565 /* Run conditional paths only if source and sink are both running. */
1566 list_for_each_entry(cpath, &path->source_list, source_node)
1567 if (cpath->source->state == AVS_PPL_STATE_RUNNING &&
1568 cpath->sink->state == AVS_PPL_STATE_RUNNING)
1569 avs_condpath_run(adev, cpath, trigger);
1570
1571 list_for_each_entry(cpath, &path->sink_list, sink_node)
1572 if (cpath->source->state == AVS_PPL_STATE_RUNNING &&
1573 cpath->sink->state == AVS_PPL_STATE_RUNNING)
1574 avs_condpath_run(adev, cpath, trigger);
1575
1576 mutex_unlock(&adev->path_mutex);
1577 }
1578
avs_path_run(struct avs_path * path,int trigger)1579 int avs_path_run(struct avs_path *path, int trigger)
1580 {
1581 struct avs_path_pipeline *ppl;
1582 struct avs_dev *adev = path->owner;
1583 int ret;
1584
1585 if (path->state == AVS_PPL_STATE_RUNNING && trigger == AVS_TPLG_TRIGGER_AUTO)
1586 return 0;
1587
1588 list_for_each_entry(ppl, &path->ppl_list, node) {
1589 if (ppl->template->cfg->trigger != trigger)
1590 continue;
1591
1592 ret = avs_ipc_set_pipeline_state(adev, ppl->instance_id,
1593 AVS_PPL_STATE_RUNNING);
1594 if (ret) {
1595 dev_err(adev->dev, "run path failed: %d\n", ret);
1596 path->state = AVS_PPL_STATE_INVALID;
1597 return AVS_IPC_RET(ret);
1598 }
1599 }
1600
1601 path->state = AVS_PPL_STATE_RUNNING;
1602
1603 /* Granular pipeline triggering not intended for conditional paths. */
1604 if (trigger == AVS_TPLG_TRIGGER_AUTO)
1605 avs_condpaths_run(adev, path, trigger);
1606
1607 return 0;
1608 }
1609