xref: /linux/sound/core/pcm_native.c (revision db624e82c55f227b84ac9ebfa3de2f6f5fad666b)
1 // SPDX-License-Identifier: GPL-2.0-or-later
2 /*
3  *  Digital Audio (PCM) abstract layer
4  *  Copyright (c) by Jaroslav Kysela <perex@perex.cz>
5  */
6 
7 #include <linux/compat.h>
8 #include <linux/mm.h>
9 #include <linux/module.h>
10 #include <linux/file.h>
11 #include <linux/slab.h>
12 #include <linux/sched/signal.h>
13 #include <linux/time.h>
14 #include <linux/pm_qos.h>
15 #include <linux/io.h>
16 #include <linux/dma-mapping.h>
17 #include <linux/vmalloc.h>
18 #include <sound/core.h>
19 #include <sound/control.h>
20 #include <sound/info.h>
21 #include <sound/pcm.h>
22 #include <sound/pcm_params.h>
23 #include <sound/timer.h>
24 #include <sound/minors.h>
25 #include <linux/uio.h>
26 #include <linux/delay.h>
27 
28 #include "pcm_local.h"
29 
30 #ifdef CONFIG_SND_DEBUG
31 #define CREATE_TRACE_POINTS
32 #include "pcm_param_trace.h"
33 #else
34 #define trace_hw_mask_param_enabled()		0
35 #define trace_hw_interval_param_enabled()	0
36 #define trace_hw_mask_param(substream, type, index, prev, curr)
37 #define trace_hw_interval_param(substream, type, index, prev, curr)
38 #endif
39 
40 /*
41  *  Compatibility
42  */
43 
44 struct snd_pcm_hw_params_old {
45 	unsigned int flags;
46 	unsigned int masks[SNDRV_PCM_HW_PARAM_SUBFORMAT -
47 			   SNDRV_PCM_HW_PARAM_ACCESS + 1];
48 	struct snd_interval intervals[SNDRV_PCM_HW_PARAM_TICK_TIME -
49 					SNDRV_PCM_HW_PARAM_SAMPLE_BITS + 1];
50 	unsigned int rmask;
51 	unsigned int cmask;
52 	unsigned int info;
53 	unsigned int msbits;
54 	unsigned int rate_num;
55 	unsigned int rate_den;
56 	snd_pcm_uframes_t fifo_size;
57 	unsigned char reserved[64];
58 };
59 
60 #ifdef CONFIG_SND_SUPPORT_OLD_API
61 #define SNDRV_PCM_IOCTL_HW_REFINE_OLD _IOWR('A', 0x10, struct snd_pcm_hw_params_old)
62 #define SNDRV_PCM_IOCTL_HW_PARAMS_OLD _IOWR('A', 0x11, struct snd_pcm_hw_params_old)
63 
64 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream,
65 				      struct snd_pcm_hw_params_old __user * _oparams);
66 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream,
67 				      struct snd_pcm_hw_params_old __user * _oparams);
68 #endif
69 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream);
70 
71 /*
72  *
73  */
74 
75 static DECLARE_RWSEM(snd_pcm_link_rwsem);
76 
77 void snd_pcm_group_init(struct snd_pcm_group *group)
78 {
79 	spin_lock_init(&group->lock);
80 	mutex_init(&group->mutex);
81 	INIT_LIST_HEAD(&group->substreams);
82 	refcount_set(&group->refs, 1);
83 }
84 
85 /* define group lock helpers */
86 #define DEFINE_PCM_GROUP_LOCK(action, mutex_action) \
87 static void snd_pcm_group_ ## action(struct snd_pcm_group *group, bool nonatomic) \
88 { \
89 	if (nonatomic) \
90 		mutex_ ## mutex_action(&group->mutex); \
91 	else \
92 		spin_ ## action(&group->lock); \
93 }
94 
95 DEFINE_PCM_GROUP_LOCK(lock, lock);
96 DEFINE_PCM_GROUP_LOCK(unlock, unlock);
97 DEFINE_PCM_GROUP_LOCK(lock_irq, lock);
98 DEFINE_PCM_GROUP_LOCK(unlock_irq, unlock);
99 
100 /**
101  * snd_pcm_stream_lock - Lock the PCM stream
102  * @substream: PCM substream
103  *
104  * This locks the PCM stream's spinlock or mutex depending on the nonatomic
105  * flag of the given substream.  This also takes the global link rw lock
106  * (or rw sem), too, for avoiding the race with linked streams.
107  */
108 void snd_pcm_stream_lock(struct snd_pcm_substream *substream)
109 {
110 	snd_pcm_group_lock(&substream->self_group, substream->pcm->nonatomic);
111 }
112 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock);
113 
114 /**
115  * snd_pcm_stream_unlock - Unlock the PCM stream
116  * @substream: PCM substream
117  *
118  * This unlocks the PCM stream that has been locked via snd_pcm_stream_lock().
119  */
120 void snd_pcm_stream_unlock(struct snd_pcm_substream *substream)
121 {
122 	snd_pcm_group_unlock(&substream->self_group, substream->pcm->nonatomic);
123 }
124 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock);
125 
126 /**
127  * snd_pcm_stream_lock_irq - Lock the PCM stream
128  * @substream: PCM substream
129  *
130  * This locks the PCM stream like snd_pcm_stream_lock() and disables the local
131  * IRQ (only when nonatomic is false).  In nonatomic case, this is identical
132  * as snd_pcm_stream_lock().
133  */
134 void snd_pcm_stream_lock_irq(struct snd_pcm_substream *substream)
135 {
136 	snd_pcm_group_lock_irq(&substream->self_group,
137 			       substream->pcm->nonatomic);
138 }
139 EXPORT_SYMBOL_GPL(snd_pcm_stream_lock_irq);
140 
141 static void snd_pcm_stream_lock_nested(struct snd_pcm_substream *substream)
142 {
143 	struct snd_pcm_group *group = &substream->self_group;
144 
145 	if (substream->pcm->nonatomic)
146 		mutex_lock_nested(&group->mutex, SINGLE_DEPTH_NESTING);
147 	else
148 		spin_lock_nested(&group->lock, SINGLE_DEPTH_NESTING);
149 }
150 
151 /**
152  * snd_pcm_stream_unlock_irq - Unlock the PCM stream
153  * @substream: PCM substream
154  *
155  * This is a counter-part of snd_pcm_stream_lock_irq().
156  */
157 void snd_pcm_stream_unlock_irq(struct snd_pcm_substream *substream)
158 {
159 	snd_pcm_group_unlock_irq(&substream->self_group,
160 				 substream->pcm->nonatomic);
161 }
162 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irq);
163 
164 unsigned long _snd_pcm_stream_lock_irqsave(struct snd_pcm_substream *substream)
165 {
166 	unsigned long flags = 0;
167 	if (substream->pcm->nonatomic)
168 		mutex_lock(&substream->self_group.mutex);
169 	else
170 		spin_lock_irqsave(&substream->self_group.lock, flags);
171 	return flags;
172 }
173 EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave);
174 
175 unsigned long _snd_pcm_stream_lock_irqsave_nested(struct snd_pcm_substream *substream)
176 {
177 	unsigned long flags = 0;
178 	if (substream->pcm->nonatomic)
179 		mutex_lock_nested(&substream->self_group.mutex,
180 				  SINGLE_DEPTH_NESTING);
181 	else
182 		spin_lock_irqsave_nested(&substream->self_group.lock, flags,
183 					 SINGLE_DEPTH_NESTING);
184 	return flags;
185 }
186 EXPORT_SYMBOL_GPL(_snd_pcm_stream_lock_irqsave_nested);
187 
188 /**
189  * snd_pcm_stream_unlock_irqrestore - Unlock the PCM stream
190  * @substream: PCM substream
191  * @flags: irq flags
192  *
193  * This is a counter-part of snd_pcm_stream_lock_irqsave().
194  */
195 void snd_pcm_stream_unlock_irqrestore(struct snd_pcm_substream *substream,
196 				      unsigned long flags)
197 {
198 	if (substream->pcm->nonatomic)
199 		mutex_unlock(&substream->self_group.mutex);
200 	else
201 		spin_unlock_irqrestore(&substream->self_group.lock, flags);
202 }
203 EXPORT_SYMBOL_GPL(snd_pcm_stream_unlock_irqrestore);
204 
205 /* Run PCM ioctl ops */
206 static int snd_pcm_ops_ioctl(struct snd_pcm_substream *substream,
207 			     unsigned cmd, void *arg)
208 {
209 	if (substream->ops->ioctl)
210 		return substream->ops->ioctl(substream, cmd, arg);
211 	else
212 		return snd_pcm_lib_ioctl(substream, cmd, arg);
213 }
214 
215 int snd_pcm_info(struct snd_pcm_substream *substream, struct snd_pcm_info *info)
216 {
217 	struct snd_pcm *pcm = substream->pcm;
218 	struct snd_pcm_str *pstr = substream->pstr;
219 
220 	memset(info, 0, sizeof(*info));
221 	info->card = pcm->card->number;
222 	info->device = pcm->device;
223 	info->stream = substream->stream;
224 	info->subdevice = substream->number;
225 	strscpy(info->id, pcm->id, sizeof(info->id));
226 	strscpy(info->name, pcm->name, sizeof(info->name));
227 	info->dev_class = pcm->dev_class;
228 	info->dev_subclass = pcm->dev_subclass;
229 	info->subdevices_count = pstr->substream_count;
230 	info->subdevices_avail = pstr->substream_count - pstr->substream_opened;
231 	strscpy(info->subname, substream->name, sizeof(info->subname));
232 
233 	return 0;
234 }
235 
236 int snd_pcm_info_user(struct snd_pcm_substream *substream,
237 		      struct snd_pcm_info __user * _info)
238 {
239 	struct snd_pcm_info *info;
240 	int err;
241 
242 	info = kmalloc(sizeof(*info), GFP_KERNEL);
243 	if (! info)
244 		return -ENOMEM;
245 	err = snd_pcm_info(substream, info);
246 	if (err >= 0) {
247 		if (copy_to_user(_info, info, sizeof(*info)))
248 			err = -EFAULT;
249 	}
250 	kfree(info);
251 	return err;
252 }
253 
254 /* macro for simplified cast */
255 #define PARAM_MASK_BIT(b)	(1U << (__force int)(b))
256 
257 static bool hw_support_mmap(struct snd_pcm_substream *substream)
258 {
259 	struct snd_dma_buffer *dmabuf;
260 
261 	if (!(substream->runtime->hw.info & SNDRV_PCM_INFO_MMAP))
262 		return false;
263 
264 	if (substream->ops->mmap || substream->ops->page)
265 		return true;
266 
267 	dmabuf = snd_pcm_get_dma_buf(substream);
268 	if (!dmabuf)
269 		dmabuf = &substream->dma_buffer;
270 	switch (dmabuf->dev.type) {
271 	case SNDRV_DMA_TYPE_UNKNOWN:
272 		/* we can't know the device, so just assume that the driver does
273 		 * everything right
274 		 */
275 		return true;
276 	case SNDRV_DMA_TYPE_CONTINUOUS:
277 	case SNDRV_DMA_TYPE_VMALLOC:
278 		return true;
279 	default:
280 		return dma_can_mmap(dmabuf->dev.dev);
281 	}
282 }
283 
284 static int constrain_mask_params(struct snd_pcm_substream *substream,
285 				 struct snd_pcm_hw_params *params)
286 {
287 	struct snd_pcm_hw_constraints *constrs =
288 					&substream->runtime->hw_constraints;
289 	struct snd_mask *m;
290 	unsigned int k;
291 	struct snd_mask old_mask __maybe_unused;
292 	int changed;
293 
294 	for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) {
295 		m = hw_param_mask(params, k);
296 		if (snd_mask_empty(m))
297 			return -EINVAL;
298 
299 		/* This parameter is not requested to change by a caller. */
300 		if (!(params->rmask & PARAM_MASK_BIT(k)))
301 			continue;
302 
303 		if (trace_hw_mask_param_enabled())
304 			old_mask = *m;
305 
306 		changed = snd_mask_refine(m, constrs_mask(constrs, k));
307 		if (changed < 0)
308 			return changed;
309 		if (changed == 0)
310 			continue;
311 
312 		/* Set corresponding flag so that the caller gets it. */
313 		trace_hw_mask_param(substream, k, 0, &old_mask, m);
314 		params->cmask |= PARAM_MASK_BIT(k);
315 	}
316 
317 	return 0;
318 }
319 
320 static int constrain_interval_params(struct snd_pcm_substream *substream,
321 				     struct snd_pcm_hw_params *params)
322 {
323 	struct snd_pcm_hw_constraints *constrs =
324 					&substream->runtime->hw_constraints;
325 	struct snd_interval *i;
326 	unsigned int k;
327 	struct snd_interval old_interval __maybe_unused;
328 	int changed;
329 
330 	for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) {
331 		i = hw_param_interval(params, k);
332 		if (snd_interval_empty(i))
333 			return -EINVAL;
334 
335 		/* This parameter is not requested to change by a caller. */
336 		if (!(params->rmask & PARAM_MASK_BIT(k)))
337 			continue;
338 
339 		if (trace_hw_interval_param_enabled())
340 			old_interval = *i;
341 
342 		changed = snd_interval_refine(i, constrs_interval(constrs, k));
343 		if (changed < 0)
344 			return changed;
345 		if (changed == 0)
346 			continue;
347 
348 		/* Set corresponding flag so that the caller gets it. */
349 		trace_hw_interval_param(substream, k, 0, &old_interval, i);
350 		params->cmask |= PARAM_MASK_BIT(k);
351 	}
352 
353 	return 0;
354 }
355 
356 static int constrain_params_by_rules(struct snd_pcm_substream *substream,
357 				     struct snd_pcm_hw_params *params)
358 {
359 	struct snd_pcm_hw_constraints *constrs =
360 					&substream->runtime->hw_constraints;
361 	unsigned int k;
362 	unsigned int *rstamps;
363 	unsigned int vstamps[SNDRV_PCM_HW_PARAM_LAST_INTERVAL + 1];
364 	unsigned int stamp;
365 	struct snd_pcm_hw_rule *r;
366 	unsigned int d;
367 	struct snd_mask old_mask __maybe_unused;
368 	struct snd_interval old_interval __maybe_unused;
369 	bool again;
370 	int changed, err = 0;
371 
372 	/*
373 	 * Each application of rule has own sequence number.
374 	 *
375 	 * Each member of 'rstamps' array represents the sequence number of
376 	 * recent application of corresponding rule.
377 	 */
378 	rstamps = kcalloc(constrs->rules_num, sizeof(unsigned int), GFP_KERNEL);
379 	if (!rstamps)
380 		return -ENOMEM;
381 
382 	/*
383 	 * Each member of 'vstamps' array represents the sequence number of
384 	 * recent application of rule in which corresponding parameters were
385 	 * changed.
386 	 *
387 	 * In initial state, elements corresponding to parameters requested by
388 	 * a caller is 1. For unrequested parameters, corresponding members
389 	 * have 0 so that the parameters are never changed anymore.
390 	 */
391 	for (k = 0; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++)
392 		vstamps[k] = (params->rmask & PARAM_MASK_BIT(k)) ? 1 : 0;
393 
394 	/* Due to the above design, actual sequence number starts at 2. */
395 	stamp = 2;
396 retry:
397 	/* Apply all rules in order. */
398 	again = false;
399 	for (k = 0; k < constrs->rules_num; k++) {
400 		r = &constrs->rules[k];
401 
402 		/*
403 		 * Check condition bits of this rule. When the rule has
404 		 * some condition bits, parameter without the bits is
405 		 * never processed. SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP
406 		 * is an example of the condition bits.
407 		 */
408 		if (r->cond && !(r->cond & params->flags))
409 			continue;
410 
411 		/*
412 		 * The 'deps' array includes maximum four dependencies
413 		 * to SNDRV_PCM_HW_PARAM_XXXs for this rule. The fifth
414 		 * member of this array is a sentinel and should be
415 		 * negative value.
416 		 *
417 		 * This rule should be processed in this time when dependent
418 		 * parameters were changed at former applications of the other
419 		 * rules.
420 		 */
421 		for (d = 0; r->deps[d] >= 0; d++) {
422 			if (vstamps[r->deps[d]] > rstamps[k])
423 				break;
424 		}
425 		if (r->deps[d] < 0)
426 			continue;
427 
428 		if (trace_hw_mask_param_enabled()) {
429 			if (hw_is_mask(r->var))
430 				old_mask = *hw_param_mask(params, r->var);
431 		}
432 		if (trace_hw_interval_param_enabled()) {
433 			if (hw_is_interval(r->var))
434 				old_interval = *hw_param_interval(params, r->var);
435 		}
436 
437 		changed = r->func(params, r);
438 		if (changed < 0) {
439 			err = changed;
440 			goto out;
441 		}
442 
443 		/*
444 		 * When the parameter is changed, notify it to the caller
445 		 * by corresponding returned bit, then preparing for next
446 		 * iteration.
447 		 */
448 		if (changed && r->var >= 0) {
449 			if (hw_is_mask(r->var)) {
450 				trace_hw_mask_param(substream, r->var,
451 					k + 1, &old_mask,
452 					hw_param_mask(params, r->var));
453 			}
454 			if (hw_is_interval(r->var)) {
455 				trace_hw_interval_param(substream, r->var,
456 					k + 1, &old_interval,
457 					hw_param_interval(params, r->var));
458 			}
459 
460 			params->cmask |= PARAM_MASK_BIT(r->var);
461 			vstamps[r->var] = stamp;
462 			again = true;
463 		}
464 
465 		rstamps[k] = stamp++;
466 	}
467 
468 	/* Iterate to evaluate all rules till no parameters are changed. */
469 	if (again)
470 		goto retry;
471 
472  out:
473 	kfree(rstamps);
474 	return err;
475 }
476 
477 static int fixup_unreferenced_params(struct snd_pcm_substream *substream,
478 				     struct snd_pcm_hw_params *params)
479 {
480 	const struct snd_interval *i;
481 	const struct snd_mask *m;
482 	struct snd_mask *m_rw;
483 	int err;
484 
485 	if (!params->msbits) {
486 		i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_SAMPLE_BITS);
487 		if (snd_interval_single(i))
488 			params->msbits = snd_interval_value(i);
489 		m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT);
490 		if (snd_mask_single(m)) {
491 			snd_pcm_format_t format = (__force snd_pcm_format_t)snd_mask_min(m);
492 			params->msbits = snd_pcm_format_width(format);
493 		}
494 	}
495 
496 	if (params->msbits) {
497 		m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT);
498 		if (snd_mask_single(m)) {
499 			snd_pcm_format_t format = (__force snd_pcm_format_t)snd_mask_min(m);
500 
501 			if (snd_pcm_format_linear(format) &&
502 			    snd_pcm_format_width(format) != params->msbits) {
503 				m_rw = hw_param_mask(params, SNDRV_PCM_HW_PARAM_SUBFORMAT);
504 				snd_mask_reset(m_rw,
505 					       (__force unsigned)SNDRV_PCM_SUBFORMAT_MSBITS_MAX);
506 				if (snd_mask_empty(m_rw))
507 					return -EINVAL;
508 			}
509 		}
510 	}
511 
512 	if (!params->rate_den) {
513 		i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_RATE);
514 		if (snd_interval_single(i)) {
515 			params->rate_num = snd_interval_value(i);
516 			params->rate_den = 1;
517 		}
518 	}
519 
520 	if (!params->fifo_size) {
521 		m = hw_param_mask_c(params, SNDRV_PCM_HW_PARAM_FORMAT);
522 		i = hw_param_interval_c(params, SNDRV_PCM_HW_PARAM_CHANNELS);
523 		if (snd_mask_single(m) && snd_interval_single(i)) {
524 			err = snd_pcm_ops_ioctl(substream,
525 						SNDRV_PCM_IOCTL1_FIFO_SIZE,
526 						params);
527 			if (err < 0)
528 				return err;
529 		}
530 	}
531 
532 	if (!params->info) {
533 		params->info = substream->runtime->hw.info;
534 		params->info &= ~(SNDRV_PCM_INFO_FIFO_IN_FRAMES |
535 				  SNDRV_PCM_INFO_DRAIN_TRIGGER);
536 		if (!hw_support_mmap(substream))
537 			params->info &= ~(SNDRV_PCM_INFO_MMAP |
538 					  SNDRV_PCM_INFO_MMAP_VALID);
539 	}
540 
541 	return 0;
542 }
543 
544 int snd_pcm_hw_refine(struct snd_pcm_substream *substream,
545 		      struct snd_pcm_hw_params *params)
546 {
547 	int err;
548 
549 	params->info = 0;
550 	params->fifo_size = 0;
551 	if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_SAMPLE_BITS))
552 		params->msbits = 0;
553 	if (params->rmask & PARAM_MASK_BIT(SNDRV_PCM_HW_PARAM_RATE)) {
554 		params->rate_num = 0;
555 		params->rate_den = 0;
556 	}
557 
558 	err = constrain_mask_params(substream, params);
559 	if (err < 0)
560 		return err;
561 
562 	err = constrain_interval_params(substream, params);
563 	if (err < 0)
564 		return err;
565 
566 	err = constrain_params_by_rules(substream, params);
567 	if (err < 0)
568 		return err;
569 
570 	params->rmask = 0;
571 
572 	return 0;
573 }
574 EXPORT_SYMBOL(snd_pcm_hw_refine);
575 
576 static int snd_pcm_hw_refine_user(struct snd_pcm_substream *substream,
577 				  struct snd_pcm_hw_params __user * _params)
578 {
579 	struct snd_pcm_hw_params *params;
580 	int err;
581 
582 	params = memdup_user(_params, sizeof(*params));
583 	if (IS_ERR(params))
584 		return PTR_ERR(params);
585 
586 	err = snd_pcm_hw_refine(substream, params);
587 	if (err < 0)
588 		goto end;
589 
590 	err = fixup_unreferenced_params(substream, params);
591 	if (err < 0)
592 		goto end;
593 
594 	if (copy_to_user(_params, params, sizeof(*params)))
595 		err = -EFAULT;
596 end:
597 	kfree(params);
598 	return err;
599 }
600 
601 static int period_to_usecs(struct snd_pcm_runtime *runtime)
602 {
603 	int usecs;
604 
605 	if (! runtime->rate)
606 		return -1; /* invalid */
607 
608 	/* take 75% of period time as the deadline */
609 	usecs = (750000 / runtime->rate) * runtime->period_size;
610 	usecs += ((750000 % runtime->rate) * runtime->period_size) /
611 		runtime->rate;
612 
613 	return usecs;
614 }
615 
616 static void snd_pcm_set_state(struct snd_pcm_substream *substream,
617 			      snd_pcm_state_t state)
618 {
619 	snd_pcm_stream_lock_irq(substream);
620 	if (substream->runtime->state != SNDRV_PCM_STATE_DISCONNECTED)
621 		__snd_pcm_set_state(substream->runtime, state);
622 	snd_pcm_stream_unlock_irq(substream);
623 }
624 
625 static inline void snd_pcm_timer_notify(struct snd_pcm_substream *substream,
626 					int event)
627 {
628 #ifdef CONFIG_SND_PCM_TIMER
629 	if (substream->timer)
630 		snd_timer_notify(substream->timer, event,
631 					&substream->runtime->trigger_tstamp);
632 #endif
633 }
634 
635 void snd_pcm_sync_stop(struct snd_pcm_substream *substream, bool sync_irq)
636 {
637 	if (substream->runtime && substream->runtime->stop_operating) {
638 		substream->runtime->stop_operating = false;
639 		if (substream->ops && substream->ops->sync_stop)
640 			substream->ops->sync_stop(substream);
641 		else if (sync_irq && substream->pcm->card->sync_irq > 0)
642 			synchronize_irq(substream->pcm->card->sync_irq);
643 	}
644 }
645 
646 /**
647  * snd_pcm_hw_params_choose - choose a configuration defined by @params
648  * @pcm: PCM instance
649  * @params: the hw_params instance
650  *
651  * Choose one configuration from configuration space defined by @params.
652  * The configuration chosen is that obtained fixing in this order:
653  * first access, first format, first subformat, min channels,
654  * min rate, min period time, max buffer size, min tick time
655  *
656  * Return: Zero if successful, or a negative error code on failure.
657  */
658 static int snd_pcm_hw_params_choose(struct snd_pcm_substream *pcm,
659 				    struct snd_pcm_hw_params *params)
660 {
661 	static const int vars[] = {
662 		SNDRV_PCM_HW_PARAM_ACCESS,
663 		SNDRV_PCM_HW_PARAM_FORMAT,
664 		SNDRV_PCM_HW_PARAM_SUBFORMAT,
665 		SNDRV_PCM_HW_PARAM_CHANNELS,
666 		SNDRV_PCM_HW_PARAM_RATE,
667 		SNDRV_PCM_HW_PARAM_PERIOD_TIME,
668 		SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
669 		SNDRV_PCM_HW_PARAM_TICK_TIME,
670 		-1
671 	};
672 	const int *v;
673 	struct snd_mask old_mask __maybe_unused;
674 	struct snd_interval old_interval __maybe_unused;
675 	int changed;
676 
677 	for (v = vars; *v != -1; v++) {
678 		/* Keep old parameter to trace. */
679 		if (trace_hw_mask_param_enabled()) {
680 			if (hw_is_mask(*v))
681 				old_mask = *hw_param_mask(params, *v);
682 		}
683 		if (trace_hw_interval_param_enabled()) {
684 			if (hw_is_interval(*v))
685 				old_interval = *hw_param_interval(params, *v);
686 		}
687 		if (*v != SNDRV_PCM_HW_PARAM_BUFFER_SIZE)
688 			changed = snd_pcm_hw_param_first(pcm, params, *v, NULL);
689 		else
690 			changed = snd_pcm_hw_param_last(pcm, params, *v, NULL);
691 		if (changed < 0)
692 			return changed;
693 		if (changed == 0)
694 			continue;
695 
696 		/* Trace the changed parameter. */
697 		if (hw_is_mask(*v)) {
698 			trace_hw_mask_param(pcm, *v, 0, &old_mask,
699 					    hw_param_mask(params, *v));
700 		}
701 		if (hw_is_interval(*v)) {
702 			trace_hw_interval_param(pcm, *v, 0, &old_interval,
703 						hw_param_interval(params, *v));
704 		}
705 	}
706 
707 	return 0;
708 }
709 
710 /* acquire buffer_mutex; if it's in r/w operation, return -EBUSY, otherwise
711  * block the further r/w operations
712  */
713 static int snd_pcm_buffer_access_lock(struct snd_pcm_runtime *runtime)
714 {
715 	if (!atomic_dec_unless_positive(&runtime->buffer_accessing))
716 		return -EBUSY;
717 	mutex_lock(&runtime->buffer_mutex);
718 	return 0; /* keep buffer_mutex, unlocked by below */
719 }
720 
721 /* release buffer_mutex and clear r/w access flag */
722 static void snd_pcm_buffer_access_unlock(struct snd_pcm_runtime *runtime)
723 {
724 	mutex_unlock(&runtime->buffer_mutex);
725 	atomic_inc(&runtime->buffer_accessing);
726 }
727 
728 #if IS_ENABLED(CONFIG_SND_PCM_OSS)
729 #define is_oss_stream(substream)	((substream)->oss.oss)
730 #else
731 #define is_oss_stream(substream)	false
732 #endif
733 
734 static int snd_pcm_hw_params(struct snd_pcm_substream *substream,
735 			     struct snd_pcm_hw_params *params)
736 {
737 	struct snd_pcm_runtime *runtime;
738 	int err, usecs;
739 	unsigned int bits;
740 	snd_pcm_uframes_t frames;
741 
742 	if (PCM_RUNTIME_CHECK(substream))
743 		return -ENXIO;
744 	runtime = substream->runtime;
745 	err = snd_pcm_buffer_access_lock(runtime);
746 	if (err < 0)
747 		return err;
748 	snd_pcm_stream_lock_irq(substream);
749 	switch (runtime->state) {
750 	case SNDRV_PCM_STATE_OPEN:
751 	case SNDRV_PCM_STATE_SETUP:
752 	case SNDRV_PCM_STATE_PREPARED:
753 		if (!is_oss_stream(substream) &&
754 		    atomic_read(&substream->mmap_count))
755 			err = -EBADFD;
756 		break;
757 	default:
758 		err = -EBADFD;
759 		break;
760 	}
761 	snd_pcm_stream_unlock_irq(substream);
762 	if (err)
763 		goto unlock;
764 
765 	snd_pcm_sync_stop(substream, true);
766 
767 	params->rmask = ~0U;
768 	err = snd_pcm_hw_refine(substream, params);
769 	if (err < 0)
770 		goto _error;
771 
772 	err = snd_pcm_hw_params_choose(substream, params);
773 	if (err < 0)
774 		goto _error;
775 
776 	err = fixup_unreferenced_params(substream, params);
777 	if (err < 0)
778 		goto _error;
779 
780 	if (substream->managed_buffer_alloc) {
781 		err = snd_pcm_lib_malloc_pages(substream,
782 					       params_buffer_bytes(params));
783 		if (err < 0)
784 			goto _error;
785 		runtime->buffer_changed = err > 0;
786 	}
787 
788 	if (substream->ops->hw_params != NULL) {
789 		err = substream->ops->hw_params(substream, params);
790 		if (err < 0)
791 			goto _error;
792 	}
793 
794 	runtime->access = params_access(params);
795 	runtime->format = params_format(params);
796 	runtime->subformat = params_subformat(params);
797 	runtime->channels = params_channels(params);
798 	runtime->rate = params_rate(params);
799 	runtime->period_size = params_period_size(params);
800 	runtime->periods = params_periods(params);
801 	runtime->buffer_size = params_buffer_size(params);
802 	runtime->info = params->info;
803 	runtime->rate_num = params->rate_num;
804 	runtime->rate_den = params->rate_den;
805 	runtime->no_period_wakeup =
806 			(params->info & SNDRV_PCM_INFO_NO_PERIOD_WAKEUP) &&
807 			(params->flags & SNDRV_PCM_HW_PARAMS_NO_PERIOD_WAKEUP);
808 
809 	bits = snd_pcm_format_physical_width(runtime->format);
810 	runtime->sample_bits = bits;
811 	bits *= runtime->channels;
812 	runtime->frame_bits = bits;
813 	frames = 1;
814 	while (bits % 8 != 0) {
815 		bits *= 2;
816 		frames *= 2;
817 	}
818 	runtime->byte_align = bits / 8;
819 	runtime->min_align = frames;
820 
821 	/* Default sw params */
822 	runtime->tstamp_mode = SNDRV_PCM_TSTAMP_NONE;
823 	runtime->period_step = 1;
824 	runtime->control->avail_min = runtime->period_size;
825 	runtime->start_threshold = 1;
826 	runtime->stop_threshold = runtime->buffer_size;
827 	runtime->silence_threshold = 0;
828 	runtime->silence_size = 0;
829 	runtime->boundary = runtime->buffer_size;
830 	while (runtime->boundary * 2 <= LONG_MAX - runtime->buffer_size)
831 		runtime->boundary *= 2;
832 
833 	/* clear the buffer for avoiding possible kernel info leaks */
834 	if (runtime->dma_area && !substream->ops->copy) {
835 		size_t size = runtime->dma_bytes;
836 
837 		if (runtime->info & SNDRV_PCM_INFO_MMAP)
838 			size = PAGE_ALIGN(size);
839 		memset(runtime->dma_area, 0, size);
840 	}
841 
842 	snd_pcm_timer_resolution_change(substream);
843 	snd_pcm_set_state(substream, SNDRV_PCM_STATE_SETUP);
844 
845 	if (cpu_latency_qos_request_active(&substream->latency_pm_qos_req))
846 		cpu_latency_qos_remove_request(&substream->latency_pm_qos_req);
847 	usecs = period_to_usecs(runtime);
848 	if (usecs >= 0)
849 		cpu_latency_qos_add_request(&substream->latency_pm_qos_req,
850 					    usecs);
851 	err = 0;
852  _error:
853 	if (err) {
854 		/* hardware might be unusable from this time,
855 		 * so we force application to retry to set
856 		 * the correct hardware parameter settings
857 		 */
858 		snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN);
859 		if (substream->ops->hw_free != NULL)
860 			substream->ops->hw_free(substream);
861 		if (substream->managed_buffer_alloc)
862 			snd_pcm_lib_free_pages(substream);
863 	}
864  unlock:
865 	snd_pcm_buffer_access_unlock(runtime);
866 	return err;
867 }
868 
869 static int snd_pcm_hw_params_user(struct snd_pcm_substream *substream,
870 				  struct snd_pcm_hw_params __user * _params)
871 {
872 	struct snd_pcm_hw_params *params;
873 	int err;
874 
875 	params = memdup_user(_params, sizeof(*params));
876 	if (IS_ERR(params))
877 		return PTR_ERR(params);
878 
879 	err = snd_pcm_hw_params(substream, params);
880 	if (err < 0)
881 		goto end;
882 
883 	if (copy_to_user(_params, params, sizeof(*params)))
884 		err = -EFAULT;
885 end:
886 	kfree(params);
887 	return err;
888 }
889 
890 static int do_hw_free(struct snd_pcm_substream *substream)
891 {
892 	int result = 0;
893 
894 	snd_pcm_sync_stop(substream, true);
895 	if (substream->ops->hw_free)
896 		result = substream->ops->hw_free(substream);
897 	if (substream->managed_buffer_alloc)
898 		snd_pcm_lib_free_pages(substream);
899 	return result;
900 }
901 
902 static int snd_pcm_hw_free(struct snd_pcm_substream *substream)
903 {
904 	struct snd_pcm_runtime *runtime;
905 	int result = 0;
906 
907 	if (PCM_RUNTIME_CHECK(substream))
908 		return -ENXIO;
909 	runtime = substream->runtime;
910 	result = snd_pcm_buffer_access_lock(runtime);
911 	if (result < 0)
912 		return result;
913 	snd_pcm_stream_lock_irq(substream);
914 	switch (runtime->state) {
915 	case SNDRV_PCM_STATE_SETUP:
916 	case SNDRV_PCM_STATE_PREPARED:
917 		if (atomic_read(&substream->mmap_count))
918 			result = -EBADFD;
919 		break;
920 	default:
921 		result = -EBADFD;
922 		break;
923 	}
924 	snd_pcm_stream_unlock_irq(substream);
925 	if (result)
926 		goto unlock;
927 	result = do_hw_free(substream);
928 	snd_pcm_set_state(substream, SNDRV_PCM_STATE_OPEN);
929 	cpu_latency_qos_remove_request(&substream->latency_pm_qos_req);
930  unlock:
931 	snd_pcm_buffer_access_unlock(runtime);
932 	return result;
933 }
934 
935 static int snd_pcm_sw_params(struct snd_pcm_substream *substream,
936 			     struct snd_pcm_sw_params *params)
937 {
938 	struct snd_pcm_runtime *runtime;
939 	int err;
940 
941 	if (PCM_RUNTIME_CHECK(substream))
942 		return -ENXIO;
943 	runtime = substream->runtime;
944 	snd_pcm_stream_lock_irq(substream);
945 	if (runtime->state == SNDRV_PCM_STATE_OPEN) {
946 		snd_pcm_stream_unlock_irq(substream);
947 		return -EBADFD;
948 	}
949 	snd_pcm_stream_unlock_irq(substream);
950 
951 	if (params->tstamp_mode < 0 ||
952 	    params->tstamp_mode > SNDRV_PCM_TSTAMP_LAST)
953 		return -EINVAL;
954 	if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12) &&
955 	    params->tstamp_type > SNDRV_PCM_TSTAMP_TYPE_LAST)
956 		return -EINVAL;
957 	if (params->avail_min == 0)
958 		return -EINVAL;
959 	if (params->silence_size >= runtime->boundary) {
960 		if (params->silence_threshold != 0)
961 			return -EINVAL;
962 	} else {
963 		if (params->silence_size > params->silence_threshold)
964 			return -EINVAL;
965 		if (params->silence_threshold > runtime->buffer_size)
966 			return -EINVAL;
967 	}
968 	err = 0;
969 	snd_pcm_stream_lock_irq(substream);
970 	runtime->tstamp_mode = params->tstamp_mode;
971 	if (params->proto >= SNDRV_PROTOCOL_VERSION(2, 0, 12))
972 		runtime->tstamp_type = params->tstamp_type;
973 	runtime->period_step = params->period_step;
974 	runtime->control->avail_min = params->avail_min;
975 	runtime->start_threshold = params->start_threshold;
976 	runtime->stop_threshold = params->stop_threshold;
977 	runtime->silence_threshold = params->silence_threshold;
978 	runtime->silence_size = params->silence_size;
979         params->boundary = runtime->boundary;
980 	if (snd_pcm_running(substream)) {
981 		if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK &&
982 		    runtime->silence_size > 0)
983 			snd_pcm_playback_silence(substream, ULONG_MAX);
984 		err = snd_pcm_update_state(substream, runtime);
985 	}
986 	snd_pcm_stream_unlock_irq(substream);
987 	return err;
988 }
989 
990 static int snd_pcm_sw_params_user(struct snd_pcm_substream *substream,
991 				  struct snd_pcm_sw_params __user * _params)
992 {
993 	struct snd_pcm_sw_params params;
994 	int err;
995 	if (copy_from_user(&params, _params, sizeof(params)))
996 		return -EFAULT;
997 	err = snd_pcm_sw_params(substream, &params);
998 	if (copy_to_user(_params, &params, sizeof(params)))
999 		return -EFAULT;
1000 	return err;
1001 }
1002 
1003 static inline snd_pcm_uframes_t
1004 snd_pcm_calc_delay(struct snd_pcm_substream *substream)
1005 {
1006 	snd_pcm_uframes_t delay;
1007 
1008 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
1009 		delay = snd_pcm_playback_hw_avail(substream->runtime);
1010 	else
1011 		delay = snd_pcm_capture_avail(substream->runtime);
1012 	return delay + substream->runtime->delay;
1013 }
1014 
1015 int snd_pcm_status64(struct snd_pcm_substream *substream,
1016 		     struct snd_pcm_status64 *status)
1017 {
1018 	struct snd_pcm_runtime *runtime = substream->runtime;
1019 
1020 	snd_pcm_stream_lock_irq(substream);
1021 
1022 	snd_pcm_unpack_audio_tstamp_config(status->audio_tstamp_data,
1023 					&runtime->audio_tstamp_config);
1024 
1025 	/* backwards compatible behavior */
1026 	if (runtime->audio_tstamp_config.type_requested ==
1027 		SNDRV_PCM_AUDIO_TSTAMP_TYPE_COMPAT) {
1028 		if (runtime->hw.info & SNDRV_PCM_INFO_HAS_WALL_CLOCK)
1029 			runtime->audio_tstamp_config.type_requested =
1030 				SNDRV_PCM_AUDIO_TSTAMP_TYPE_LINK;
1031 		else
1032 			runtime->audio_tstamp_config.type_requested =
1033 				SNDRV_PCM_AUDIO_TSTAMP_TYPE_DEFAULT;
1034 		runtime->audio_tstamp_report.valid = 0;
1035 	} else
1036 		runtime->audio_tstamp_report.valid = 1;
1037 
1038 	status->state = runtime->state;
1039 	status->suspended_state = runtime->suspended_state;
1040 	if (status->state == SNDRV_PCM_STATE_OPEN)
1041 		goto _end;
1042 	status->trigger_tstamp_sec = runtime->trigger_tstamp.tv_sec;
1043 	status->trigger_tstamp_nsec = runtime->trigger_tstamp.tv_nsec;
1044 	if (snd_pcm_running(substream)) {
1045 		snd_pcm_update_hw_ptr(substream);
1046 		if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) {
1047 			status->tstamp_sec = runtime->status->tstamp.tv_sec;
1048 			status->tstamp_nsec =
1049 				runtime->status->tstamp.tv_nsec;
1050 			status->driver_tstamp_sec =
1051 				runtime->driver_tstamp.tv_sec;
1052 			status->driver_tstamp_nsec =
1053 				runtime->driver_tstamp.tv_nsec;
1054 			status->audio_tstamp_sec =
1055 				runtime->status->audio_tstamp.tv_sec;
1056 			status->audio_tstamp_nsec =
1057 				runtime->status->audio_tstamp.tv_nsec;
1058 			if (runtime->audio_tstamp_report.valid == 1)
1059 				/* backwards compatibility, no report provided in COMPAT mode */
1060 				snd_pcm_pack_audio_tstamp_report(&status->audio_tstamp_data,
1061 								&status->audio_tstamp_accuracy,
1062 								&runtime->audio_tstamp_report);
1063 
1064 			goto _tstamp_end;
1065 		}
1066 	} else {
1067 		/* get tstamp only in fallback mode and only if enabled */
1068 		if (runtime->tstamp_mode == SNDRV_PCM_TSTAMP_ENABLE) {
1069 			struct timespec64 tstamp;
1070 
1071 			snd_pcm_gettime(runtime, &tstamp);
1072 			status->tstamp_sec = tstamp.tv_sec;
1073 			status->tstamp_nsec = tstamp.tv_nsec;
1074 		}
1075 	}
1076  _tstamp_end:
1077 	status->appl_ptr = runtime->control->appl_ptr;
1078 	status->hw_ptr = runtime->status->hw_ptr;
1079 	status->avail = snd_pcm_avail(substream);
1080 	status->delay = snd_pcm_running(substream) ?
1081 		snd_pcm_calc_delay(substream) : 0;
1082 	status->avail_max = runtime->avail_max;
1083 	status->overrange = runtime->overrange;
1084 	runtime->avail_max = 0;
1085 	runtime->overrange = 0;
1086  _end:
1087  	snd_pcm_stream_unlock_irq(substream);
1088 	return 0;
1089 }
1090 
1091 static int snd_pcm_status_user64(struct snd_pcm_substream *substream,
1092 				 struct snd_pcm_status64 __user * _status,
1093 				 bool ext)
1094 {
1095 	struct snd_pcm_status64 status;
1096 	int res;
1097 
1098 	memset(&status, 0, sizeof(status));
1099 	/*
1100 	 * with extension, parameters are read/write,
1101 	 * get audio_tstamp_data from user,
1102 	 * ignore rest of status structure
1103 	 */
1104 	if (ext && get_user(status.audio_tstamp_data,
1105 				(u32 __user *)(&_status->audio_tstamp_data)))
1106 		return -EFAULT;
1107 	res = snd_pcm_status64(substream, &status);
1108 	if (res < 0)
1109 		return res;
1110 	if (copy_to_user(_status, &status, sizeof(status)))
1111 		return -EFAULT;
1112 	return 0;
1113 }
1114 
1115 static int snd_pcm_status_user32(struct snd_pcm_substream *substream,
1116 				 struct snd_pcm_status32 __user * _status,
1117 				 bool ext)
1118 {
1119 	struct snd_pcm_status64 status64;
1120 	struct snd_pcm_status32 status32;
1121 	int res;
1122 
1123 	memset(&status64, 0, sizeof(status64));
1124 	memset(&status32, 0, sizeof(status32));
1125 	/*
1126 	 * with extension, parameters are read/write,
1127 	 * get audio_tstamp_data from user,
1128 	 * ignore rest of status structure
1129 	 */
1130 	if (ext && get_user(status64.audio_tstamp_data,
1131 			    (u32 __user *)(&_status->audio_tstamp_data)))
1132 		return -EFAULT;
1133 	res = snd_pcm_status64(substream, &status64);
1134 	if (res < 0)
1135 		return res;
1136 
1137 	status32 = (struct snd_pcm_status32) {
1138 		.state = status64.state,
1139 		.trigger_tstamp_sec = status64.trigger_tstamp_sec,
1140 		.trigger_tstamp_nsec = status64.trigger_tstamp_nsec,
1141 		.tstamp_sec = status64.tstamp_sec,
1142 		.tstamp_nsec = status64.tstamp_nsec,
1143 		.appl_ptr = status64.appl_ptr,
1144 		.hw_ptr = status64.hw_ptr,
1145 		.delay = status64.delay,
1146 		.avail = status64.avail,
1147 		.avail_max = status64.avail_max,
1148 		.overrange = status64.overrange,
1149 		.suspended_state = status64.suspended_state,
1150 		.audio_tstamp_data = status64.audio_tstamp_data,
1151 		.audio_tstamp_sec = status64.audio_tstamp_sec,
1152 		.audio_tstamp_nsec = status64.audio_tstamp_nsec,
1153 		.driver_tstamp_sec = status64.audio_tstamp_sec,
1154 		.driver_tstamp_nsec = status64.audio_tstamp_nsec,
1155 		.audio_tstamp_accuracy = status64.audio_tstamp_accuracy,
1156 	};
1157 
1158 	if (copy_to_user(_status, &status32, sizeof(status32)))
1159 		return -EFAULT;
1160 
1161 	return 0;
1162 }
1163 
1164 static int snd_pcm_channel_info(struct snd_pcm_substream *substream,
1165 				struct snd_pcm_channel_info * info)
1166 {
1167 	struct snd_pcm_runtime *runtime;
1168 	unsigned int channel;
1169 
1170 	channel = info->channel;
1171 	runtime = substream->runtime;
1172 	snd_pcm_stream_lock_irq(substream);
1173 	if (runtime->state == SNDRV_PCM_STATE_OPEN) {
1174 		snd_pcm_stream_unlock_irq(substream);
1175 		return -EBADFD;
1176 	}
1177 	snd_pcm_stream_unlock_irq(substream);
1178 	if (channel >= runtime->channels)
1179 		return -EINVAL;
1180 	memset(info, 0, sizeof(*info));
1181 	info->channel = channel;
1182 	return snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_CHANNEL_INFO, info);
1183 }
1184 
1185 static int snd_pcm_channel_info_user(struct snd_pcm_substream *substream,
1186 				     struct snd_pcm_channel_info __user * _info)
1187 {
1188 	struct snd_pcm_channel_info info;
1189 	int res;
1190 
1191 	if (copy_from_user(&info, _info, sizeof(info)))
1192 		return -EFAULT;
1193 	res = snd_pcm_channel_info(substream, &info);
1194 	if (res < 0)
1195 		return res;
1196 	if (copy_to_user(_info, &info, sizeof(info)))
1197 		return -EFAULT;
1198 	return 0;
1199 }
1200 
1201 static void snd_pcm_trigger_tstamp(struct snd_pcm_substream *substream)
1202 {
1203 	struct snd_pcm_runtime *runtime = substream->runtime;
1204 	if (runtime->trigger_master == NULL)
1205 		return;
1206 	if (runtime->trigger_master == substream) {
1207 		if (!runtime->trigger_tstamp_latched)
1208 			snd_pcm_gettime(runtime, &runtime->trigger_tstamp);
1209 	} else {
1210 		snd_pcm_trigger_tstamp(runtime->trigger_master);
1211 		runtime->trigger_tstamp = runtime->trigger_master->runtime->trigger_tstamp;
1212 	}
1213 	runtime->trigger_master = NULL;
1214 }
1215 
1216 #define ACTION_ARG_IGNORE	(__force snd_pcm_state_t)0
1217 
1218 struct action_ops {
1219 	int (*pre_action)(struct snd_pcm_substream *substream,
1220 			  snd_pcm_state_t state);
1221 	int (*do_action)(struct snd_pcm_substream *substream,
1222 			 snd_pcm_state_t state);
1223 	void (*undo_action)(struct snd_pcm_substream *substream,
1224 			    snd_pcm_state_t state);
1225 	void (*post_action)(struct snd_pcm_substream *substream,
1226 			    snd_pcm_state_t state);
1227 };
1228 
1229 /*
1230  *  this functions is core for handling of linked stream
1231  *  Note: the stream state might be changed also on failure
1232  *  Note2: call with calling stream lock + link lock
1233  */
1234 static int snd_pcm_action_group(const struct action_ops *ops,
1235 				struct snd_pcm_substream *substream,
1236 				snd_pcm_state_t state,
1237 				bool stream_lock)
1238 {
1239 	struct snd_pcm_substream *s = NULL;
1240 	struct snd_pcm_substream *s1;
1241 	int res = 0, depth = 1;
1242 
1243 	snd_pcm_group_for_each_entry(s, substream) {
1244 		if (s != substream) {
1245 			if (!stream_lock)
1246 				mutex_lock_nested(&s->runtime->buffer_mutex, depth);
1247 			else if (s->pcm->nonatomic)
1248 				mutex_lock_nested(&s->self_group.mutex, depth);
1249 			else
1250 				spin_lock_nested(&s->self_group.lock, depth);
1251 			depth++;
1252 		}
1253 		res = ops->pre_action(s, state);
1254 		if (res < 0)
1255 			goto _unlock;
1256 	}
1257 	snd_pcm_group_for_each_entry(s, substream) {
1258 		res = ops->do_action(s, state);
1259 		if (res < 0) {
1260 			if (ops->undo_action) {
1261 				snd_pcm_group_for_each_entry(s1, substream) {
1262 					if (s1 == s) /* failed stream */
1263 						break;
1264 					ops->undo_action(s1, state);
1265 				}
1266 			}
1267 			s = NULL; /* unlock all */
1268 			goto _unlock;
1269 		}
1270 	}
1271 	snd_pcm_group_for_each_entry(s, substream) {
1272 		ops->post_action(s, state);
1273 	}
1274  _unlock:
1275 	/* unlock streams */
1276 	snd_pcm_group_for_each_entry(s1, substream) {
1277 		if (s1 != substream) {
1278 			if (!stream_lock)
1279 				mutex_unlock(&s1->runtime->buffer_mutex);
1280 			else if (s1->pcm->nonatomic)
1281 				mutex_unlock(&s1->self_group.mutex);
1282 			else
1283 				spin_unlock(&s1->self_group.lock);
1284 		}
1285 		if (s1 == s)	/* end */
1286 			break;
1287 	}
1288 	return res;
1289 }
1290 
1291 /*
1292  *  Note: call with stream lock
1293  */
1294 static int snd_pcm_action_single(const struct action_ops *ops,
1295 				 struct snd_pcm_substream *substream,
1296 				 snd_pcm_state_t state)
1297 {
1298 	int res;
1299 
1300 	res = ops->pre_action(substream, state);
1301 	if (res < 0)
1302 		return res;
1303 	res = ops->do_action(substream, state);
1304 	if (res == 0)
1305 		ops->post_action(substream, state);
1306 	else if (ops->undo_action)
1307 		ops->undo_action(substream, state);
1308 	return res;
1309 }
1310 
1311 static void snd_pcm_group_assign(struct snd_pcm_substream *substream,
1312 				 struct snd_pcm_group *new_group)
1313 {
1314 	substream->group = new_group;
1315 	list_move(&substream->link_list, &new_group->substreams);
1316 }
1317 
1318 /*
1319  * Unref and unlock the group, but keep the stream lock;
1320  * when the group becomes empty and no longer referred, destroy itself
1321  */
1322 static void snd_pcm_group_unref(struct snd_pcm_group *group,
1323 				struct snd_pcm_substream *substream)
1324 {
1325 	bool do_free;
1326 
1327 	if (!group)
1328 		return;
1329 	do_free = refcount_dec_and_test(&group->refs);
1330 	snd_pcm_group_unlock(group, substream->pcm->nonatomic);
1331 	if (do_free)
1332 		kfree(group);
1333 }
1334 
1335 /*
1336  * Lock the group inside a stream lock and reference it;
1337  * return the locked group object, or NULL if not linked
1338  */
1339 static struct snd_pcm_group *
1340 snd_pcm_stream_group_ref(struct snd_pcm_substream *substream)
1341 {
1342 	bool nonatomic = substream->pcm->nonatomic;
1343 	struct snd_pcm_group *group;
1344 	bool trylock;
1345 
1346 	for (;;) {
1347 		if (!snd_pcm_stream_linked(substream))
1348 			return NULL;
1349 		group = substream->group;
1350 		/* block freeing the group object */
1351 		refcount_inc(&group->refs);
1352 
1353 		trylock = nonatomic ? mutex_trylock(&group->mutex) :
1354 			spin_trylock(&group->lock);
1355 		if (trylock)
1356 			break; /* OK */
1357 
1358 		/* re-lock for avoiding ABBA deadlock */
1359 		snd_pcm_stream_unlock(substream);
1360 		snd_pcm_group_lock(group, nonatomic);
1361 		snd_pcm_stream_lock(substream);
1362 
1363 		/* check the group again; the above opens a small race window */
1364 		if (substream->group == group)
1365 			break; /* OK */
1366 		/* group changed, try again */
1367 		snd_pcm_group_unref(group, substream);
1368 	}
1369 	return group;
1370 }
1371 
1372 /*
1373  *  Note: call with stream lock
1374  */
1375 static int snd_pcm_action(const struct action_ops *ops,
1376 			  struct snd_pcm_substream *substream,
1377 			  snd_pcm_state_t state)
1378 {
1379 	struct snd_pcm_group *group;
1380 	int res;
1381 
1382 	group = snd_pcm_stream_group_ref(substream);
1383 	if (group)
1384 		res = snd_pcm_action_group(ops, substream, state, true);
1385 	else
1386 		res = snd_pcm_action_single(ops, substream, state);
1387 	snd_pcm_group_unref(group, substream);
1388 	return res;
1389 }
1390 
1391 /*
1392  *  Note: don't use any locks before
1393  */
1394 static int snd_pcm_action_lock_irq(const struct action_ops *ops,
1395 				   struct snd_pcm_substream *substream,
1396 				   snd_pcm_state_t state)
1397 {
1398 	int res;
1399 
1400 	snd_pcm_stream_lock_irq(substream);
1401 	res = snd_pcm_action(ops, substream, state);
1402 	snd_pcm_stream_unlock_irq(substream);
1403 	return res;
1404 }
1405 
1406 /*
1407  */
1408 static int snd_pcm_action_nonatomic(const struct action_ops *ops,
1409 				    struct snd_pcm_substream *substream,
1410 				    snd_pcm_state_t state)
1411 {
1412 	int res;
1413 
1414 	/* Guarantee the group members won't change during non-atomic action */
1415 	down_read(&snd_pcm_link_rwsem);
1416 	res = snd_pcm_buffer_access_lock(substream->runtime);
1417 	if (res < 0)
1418 		goto unlock;
1419 	if (snd_pcm_stream_linked(substream))
1420 		res = snd_pcm_action_group(ops, substream, state, false);
1421 	else
1422 		res = snd_pcm_action_single(ops, substream, state);
1423 	snd_pcm_buffer_access_unlock(substream->runtime);
1424  unlock:
1425 	up_read(&snd_pcm_link_rwsem);
1426 	return res;
1427 }
1428 
1429 /*
1430  * start callbacks
1431  */
1432 static int snd_pcm_pre_start(struct snd_pcm_substream *substream,
1433 			     snd_pcm_state_t state)
1434 {
1435 	struct snd_pcm_runtime *runtime = substream->runtime;
1436 	if (runtime->state != SNDRV_PCM_STATE_PREPARED)
1437 		return -EBADFD;
1438 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK &&
1439 	    !snd_pcm_playback_data(substream))
1440 		return -EPIPE;
1441 	runtime->trigger_tstamp_latched = false;
1442 	runtime->trigger_master = substream;
1443 	return 0;
1444 }
1445 
1446 static int snd_pcm_do_start(struct snd_pcm_substream *substream,
1447 			    snd_pcm_state_t state)
1448 {
1449 	int err;
1450 
1451 	if (substream->runtime->trigger_master != substream)
1452 		return 0;
1453 	err = substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_START);
1454 	/* XRUN happened during the start */
1455 	if (err == -EPIPE)
1456 		__snd_pcm_set_state(substream->runtime, SNDRV_PCM_STATE_XRUN);
1457 	return err;
1458 }
1459 
1460 static void snd_pcm_undo_start(struct snd_pcm_substream *substream,
1461 			       snd_pcm_state_t state)
1462 {
1463 	if (substream->runtime->trigger_master == substream) {
1464 		substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP);
1465 		substream->runtime->stop_operating = true;
1466 	}
1467 }
1468 
1469 static void snd_pcm_post_start(struct snd_pcm_substream *substream,
1470 			       snd_pcm_state_t state)
1471 {
1472 	struct snd_pcm_runtime *runtime = substream->runtime;
1473 	snd_pcm_trigger_tstamp(substream);
1474 	runtime->hw_ptr_jiffies = jiffies;
1475 	runtime->hw_ptr_buffer_jiffies = (runtime->buffer_size * HZ) /
1476 							    runtime->rate;
1477 	__snd_pcm_set_state(runtime, state);
1478 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK &&
1479 	    runtime->silence_size > 0)
1480 		snd_pcm_playback_silence(substream, ULONG_MAX);
1481 	snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTART);
1482 }
1483 
1484 static const struct action_ops snd_pcm_action_start = {
1485 	.pre_action = snd_pcm_pre_start,
1486 	.do_action = snd_pcm_do_start,
1487 	.undo_action = snd_pcm_undo_start,
1488 	.post_action = snd_pcm_post_start
1489 };
1490 
1491 /**
1492  * snd_pcm_start - start all linked streams
1493  * @substream: the PCM substream instance
1494  *
1495  * Return: Zero if successful, or a negative error code.
1496  * The stream lock must be acquired before calling this function.
1497  */
1498 int snd_pcm_start(struct snd_pcm_substream *substream)
1499 {
1500 	return snd_pcm_action(&snd_pcm_action_start, substream,
1501 			      SNDRV_PCM_STATE_RUNNING);
1502 }
1503 
1504 /* take the stream lock and start the streams */
1505 static int snd_pcm_start_lock_irq(struct snd_pcm_substream *substream)
1506 {
1507 	return snd_pcm_action_lock_irq(&snd_pcm_action_start, substream,
1508 				       SNDRV_PCM_STATE_RUNNING);
1509 }
1510 
1511 /*
1512  * stop callbacks
1513  */
1514 static int snd_pcm_pre_stop(struct snd_pcm_substream *substream,
1515 			    snd_pcm_state_t state)
1516 {
1517 	struct snd_pcm_runtime *runtime = substream->runtime;
1518 	if (runtime->state == SNDRV_PCM_STATE_OPEN)
1519 		return -EBADFD;
1520 	runtime->trigger_master = substream;
1521 	return 0;
1522 }
1523 
1524 static int snd_pcm_do_stop(struct snd_pcm_substream *substream,
1525 			   snd_pcm_state_t state)
1526 {
1527 	if (substream->runtime->trigger_master == substream &&
1528 	    snd_pcm_running(substream)) {
1529 		substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_STOP);
1530 		substream->runtime->stop_operating = true;
1531 	}
1532 	return 0; /* unconditionally stop all substreams */
1533 }
1534 
1535 static void snd_pcm_post_stop(struct snd_pcm_substream *substream,
1536 			      snd_pcm_state_t state)
1537 {
1538 	struct snd_pcm_runtime *runtime = substream->runtime;
1539 	if (runtime->state != state) {
1540 		snd_pcm_trigger_tstamp(substream);
1541 		__snd_pcm_set_state(runtime, state);
1542 		snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSTOP);
1543 	}
1544 	wake_up(&runtime->sleep);
1545 	wake_up(&runtime->tsleep);
1546 }
1547 
1548 static const struct action_ops snd_pcm_action_stop = {
1549 	.pre_action = snd_pcm_pre_stop,
1550 	.do_action = snd_pcm_do_stop,
1551 	.post_action = snd_pcm_post_stop
1552 };
1553 
1554 /**
1555  * snd_pcm_stop - try to stop all running streams in the substream group
1556  * @substream: the PCM substream instance
1557  * @state: PCM state after stopping the stream
1558  *
1559  * The state of each stream is then changed to the given state unconditionally.
1560  *
1561  * Return: Zero if successful, or a negative error code.
1562  */
1563 int snd_pcm_stop(struct snd_pcm_substream *substream, snd_pcm_state_t state)
1564 {
1565 	return snd_pcm_action(&snd_pcm_action_stop, substream, state);
1566 }
1567 EXPORT_SYMBOL(snd_pcm_stop);
1568 
1569 /**
1570  * snd_pcm_drain_done - stop the DMA only when the given stream is playback
1571  * @substream: the PCM substream
1572  *
1573  * After stopping, the state is changed to SETUP.
1574  * Unlike snd_pcm_stop(), this affects only the given stream.
1575  *
1576  * Return: Zero if successful, or a negative error code.
1577  */
1578 int snd_pcm_drain_done(struct snd_pcm_substream *substream)
1579 {
1580 	return snd_pcm_action_single(&snd_pcm_action_stop, substream,
1581 				     SNDRV_PCM_STATE_SETUP);
1582 }
1583 
1584 /**
1585  * snd_pcm_stop_xrun - stop the running streams as XRUN
1586  * @substream: the PCM substream instance
1587  *
1588  * This stops the given running substream (and all linked substreams) as XRUN.
1589  * Unlike snd_pcm_stop(), this function takes the substream lock by itself.
1590  *
1591  * Return: Zero if successful, or a negative error code.
1592  */
1593 int snd_pcm_stop_xrun(struct snd_pcm_substream *substream)
1594 {
1595 	unsigned long flags;
1596 
1597 	snd_pcm_stream_lock_irqsave(substream, flags);
1598 	if (substream->runtime && snd_pcm_running(substream))
1599 		__snd_pcm_xrun(substream);
1600 	snd_pcm_stream_unlock_irqrestore(substream, flags);
1601 	return 0;
1602 }
1603 EXPORT_SYMBOL_GPL(snd_pcm_stop_xrun);
1604 
1605 /*
1606  * pause callbacks: pass boolean (to start pause or resume) as state argument
1607  */
1608 #define pause_pushed(state)	(__force bool)(state)
1609 
1610 static int snd_pcm_pre_pause(struct snd_pcm_substream *substream,
1611 			     snd_pcm_state_t state)
1612 {
1613 	struct snd_pcm_runtime *runtime = substream->runtime;
1614 	if (!(runtime->info & SNDRV_PCM_INFO_PAUSE))
1615 		return -ENOSYS;
1616 	if (pause_pushed(state)) {
1617 		if (runtime->state != SNDRV_PCM_STATE_RUNNING)
1618 			return -EBADFD;
1619 	} else if (runtime->state != SNDRV_PCM_STATE_PAUSED)
1620 		return -EBADFD;
1621 	runtime->trigger_master = substream;
1622 	return 0;
1623 }
1624 
1625 static int snd_pcm_do_pause(struct snd_pcm_substream *substream,
1626 			    snd_pcm_state_t state)
1627 {
1628 	if (substream->runtime->trigger_master != substream)
1629 		return 0;
1630 	/* The jiffies check in snd_pcm_update_hw_ptr*() is done by
1631 	 * a delta between the current jiffies, this gives a large enough
1632 	 * delta, effectively to skip the check once.
1633 	 */
1634 	substream->runtime->hw_ptr_jiffies = jiffies - HZ * 1000;
1635 	return substream->ops->trigger(substream,
1636 				       pause_pushed(state) ?
1637 				       SNDRV_PCM_TRIGGER_PAUSE_PUSH :
1638 				       SNDRV_PCM_TRIGGER_PAUSE_RELEASE);
1639 }
1640 
1641 static void snd_pcm_undo_pause(struct snd_pcm_substream *substream,
1642 			       snd_pcm_state_t state)
1643 {
1644 	if (substream->runtime->trigger_master == substream)
1645 		substream->ops->trigger(substream,
1646 					pause_pushed(state) ?
1647 					SNDRV_PCM_TRIGGER_PAUSE_RELEASE :
1648 					SNDRV_PCM_TRIGGER_PAUSE_PUSH);
1649 }
1650 
1651 static void snd_pcm_post_pause(struct snd_pcm_substream *substream,
1652 			       snd_pcm_state_t state)
1653 {
1654 	struct snd_pcm_runtime *runtime = substream->runtime;
1655 	snd_pcm_trigger_tstamp(substream);
1656 	if (pause_pushed(state)) {
1657 		__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_PAUSED);
1658 		snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MPAUSE);
1659 		wake_up(&runtime->sleep);
1660 		wake_up(&runtime->tsleep);
1661 	} else {
1662 		__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_RUNNING);
1663 		snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MCONTINUE);
1664 	}
1665 }
1666 
1667 static const struct action_ops snd_pcm_action_pause = {
1668 	.pre_action = snd_pcm_pre_pause,
1669 	.do_action = snd_pcm_do_pause,
1670 	.undo_action = snd_pcm_undo_pause,
1671 	.post_action = snd_pcm_post_pause
1672 };
1673 
1674 /*
1675  * Push/release the pause for all linked streams.
1676  */
1677 static int snd_pcm_pause(struct snd_pcm_substream *substream, bool push)
1678 {
1679 	return snd_pcm_action(&snd_pcm_action_pause, substream,
1680 			      (__force snd_pcm_state_t)push);
1681 }
1682 
1683 static int snd_pcm_pause_lock_irq(struct snd_pcm_substream *substream,
1684 				  bool push)
1685 {
1686 	return snd_pcm_action_lock_irq(&snd_pcm_action_pause, substream,
1687 				       (__force snd_pcm_state_t)push);
1688 }
1689 
1690 #ifdef CONFIG_PM
1691 /* suspend callback: state argument ignored */
1692 
1693 static int snd_pcm_pre_suspend(struct snd_pcm_substream *substream,
1694 			       snd_pcm_state_t state)
1695 {
1696 	struct snd_pcm_runtime *runtime = substream->runtime;
1697 	switch (runtime->state) {
1698 	case SNDRV_PCM_STATE_SUSPENDED:
1699 		return -EBUSY;
1700 	/* unresumable PCM state; return -EBUSY for skipping suspend */
1701 	case SNDRV_PCM_STATE_OPEN:
1702 	case SNDRV_PCM_STATE_SETUP:
1703 	case SNDRV_PCM_STATE_DISCONNECTED:
1704 		return -EBUSY;
1705 	}
1706 	runtime->trigger_master = substream;
1707 	return 0;
1708 }
1709 
1710 static int snd_pcm_do_suspend(struct snd_pcm_substream *substream,
1711 			      snd_pcm_state_t state)
1712 {
1713 	struct snd_pcm_runtime *runtime = substream->runtime;
1714 	if (runtime->trigger_master != substream)
1715 		return 0;
1716 	if (! snd_pcm_running(substream))
1717 		return 0;
1718 	substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND);
1719 	runtime->stop_operating = true;
1720 	return 0; /* suspend unconditionally */
1721 }
1722 
1723 static void snd_pcm_post_suspend(struct snd_pcm_substream *substream,
1724 				 snd_pcm_state_t state)
1725 {
1726 	struct snd_pcm_runtime *runtime = substream->runtime;
1727 	snd_pcm_trigger_tstamp(substream);
1728 	runtime->suspended_state = runtime->state;
1729 	runtime->status->suspended_state = runtime->suspended_state;
1730 	__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SUSPENDED);
1731 	snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MSUSPEND);
1732 	wake_up(&runtime->sleep);
1733 	wake_up(&runtime->tsleep);
1734 }
1735 
1736 static const struct action_ops snd_pcm_action_suspend = {
1737 	.pre_action = snd_pcm_pre_suspend,
1738 	.do_action = snd_pcm_do_suspend,
1739 	.post_action = snd_pcm_post_suspend
1740 };
1741 
1742 /*
1743  * snd_pcm_suspend - trigger SUSPEND to all linked streams
1744  * @substream: the PCM substream
1745  *
1746  * After this call, all streams are changed to SUSPENDED state.
1747  *
1748  * Return: Zero if successful, or a negative error code.
1749  */
1750 static int snd_pcm_suspend(struct snd_pcm_substream *substream)
1751 {
1752 	int err;
1753 	unsigned long flags;
1754 
1755 	snd_pcm_stream_lock_irqsave(substream, flags);
1756 	err = snd_pcm_action(&snd_pcm_action_suspend, substream,
1757 			     ACTION_ARG_IGNORE);
1758 	snd_pcm_stream_unlock_irqrestore(substream, flags);
1759 	return err;
1760 }
1761 
1762 /**
1763  * snd_pcm_suspend_all - trigger SUSPEND to all substreams in the given pcm
1764  * @pcm: the PCM instance
1765  *
1766  * After this call, all streams are changed to SUSPENDED state.
1767  *
1768  * Return: Zero if successful (or @pcm is %NULL), or a negative error code.
1769  */
1770 int snd_pcm_suspend_all(struct snd_pcm *pcm)
1771 {
1772 	struct snd_pcm_substream *substream;
1773 	int stream, err = 0;
1774 
1775 	if (! pcm)
1776 		return 0;
1777 
1778 	for_each_pcm_substream(pcm, stream, substream) {
1779 		/* FIXME: the open/close code should lock this as well */
1780 		if (!substream->runtime)
1781 			continue;
1782 
1783 		/*
1784 		 * Skip BE dai link PCM's that are internal and may
1785 		 * not have their substream ops set.
1786 		 */
1787 		if (!substream->ops)
1788 			continue;
1789 
1790 		err = snd_pcm_suspend(substream);
1791 		if (err < 0 && err != -EBUSY)
1792 			return err;
1793 	}
1794 
1795 	for_each_pcm_substream(pcm, stream, substream)
1796 		snd_pcm_sync_stop(substream, false);
1797 
1798 	return 0;
1799 }
1800 EXPORT_SYMBOL(snd_pcm_suspend_all);
1801 
1802 /* resume callbacks: state argument ignored */
1803 
1804 static int snd_pcm_pre_resume(struct snd_pcm_substream *substream,
1805 			      snd_pcm_state_t state)
1806 {
1807 	struct snd_pcm_runtime *runtime = substream->runtime;
1808 	if (!(runtime->info & SNDRV_PCM_INFO_RESUME))
1809 		return -ENOSYS;
1810 	runtime->trigger_master = substream;
1811 	return 0;
1812 }
1813 
1814 static int snd_pcm_do_resume(struct snd_pcm_substream *substream,
1815 			     snd_pcm_state_t state)
1816 {
1817 	struct snd_pcm_runtime *runtime = substream->runtime;
1818 	if (runtime->trigger_master != substream)
1819 		return 0;
1820 	/* DMA not running previously? */
1821 	if (runtime->suspended_state != SNDRV_PCM_STATE_RUNNING &&
1822 	    (runtime->suspended_state != SNDRV_PCM_STATE_DRAINING ||
1823 	     substream->stream != SNDRV_PCM_STREAM_PLAYBACK))
1824 		return 0;
1825 	return substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_RESUME);
1826 }
1827 
1828 static void snd_pcm_undo_resume(struct snd_pcm_substream *substream,
1829 				snd_pcm_state_t state)
1830 {
1831 	if (substream->runtime->trigger_master == substream &&
1832 	    snd_pcm_running(substream))
1833 		substream->ops->trigger(substream, SNDRV_PCM_TRIGGER_SUSPEND);
1834 }
1835 
1836 static void snd_pcm_post_resume(struct snd_pcm_substream *substream,
1837 				snd_pcm_state_t state)
1838 {
1839 	struct snd_pcm_runtime *runtime = substream->runtime;
1840 	snd_pcm_trigger_tstamp(substream);
1841 	__snd_pcm_set_state(runtime, runtime->suspended_state);
1842 	snd_pcm_timer_notify(substream, SNDRV_TIMER_EVENT_MRESUME);
1843 }
1844 
1845 static const struct action_ops snd_pcm_action_resume = {
1846 	.pre_action = snd_pcm_pre_resume,
1847 	.do_action = snd_pcm_do_resume,
1848 	.undo_action = snd_pcm_undo_resume,
1849 	.post_action = snd_pcm_post_resume
1850 };
1851 
1852 static int snd_pcm_resume(struct snd_pcm_substream *substream)
1853 {
1854 	return snd_pcm_action_lock_irq(&snd_pcm_action_resume, substream,
1855 				       ACTION_ARG_IGNORE);
1856 }
1857 
1858 #else
1859 
1860 static int snd_pcm_resume(struct snd_pcm_substream *substream)
1861 {
1862 	return -ENOSYS;
1863 }
1864 
1865 #endif /* CONFIG_PM */
1866 
1867 /*
1868  * xrun ioctl
1869  *
1870  * Change the RUNNING stream(s) to XRUN state.
1871  */
1872 static int snd_pcm_xrun(struct snd_pcm_substream *substream)
1873 {
1874 	struct snd_pcm_runtime *runtime = substream->runtime;
1875 	int result;
1876 
1877 	snd_pcm_stream_lock_irq(substream);
1878 	switch (runtime->state) {
1879 	case SNDRV_PCM_STATE_XRUN:
1880 		result = 0;	/* already there */
1881 		break;
1882 	case SNDRV_PCM_STATE_RUNNING:
1883 		__snd_pcm_xrun(substream);
1884 		result = 0;
1885 		break;
1886 	default:
1887 		result = -EBADFD;
1888 	}
1889 	snd_pcm_stream_unlock_irq(substream);
1890 	return result;
1891 }
1892 
1893 /*
1894  * reset ioctl
1895  */
1896 /* reset callbacks:  state argument ignored */
1897 static int snd_pcm_pre_reset(struct snd_pcm_substream *substream,
1898 			     snd_pcm_state_t state)
1899 {
1900 	struct snd_pcm_runtime *runtime = substream->runtime;
1901 	switch (runtime->state) {
1902 	case SNDRV_PCM_STATE_RUNNING:
1903 	case SNDRV_PCM_STATE_PREPARED:
1904 	case SNDRV_PCM_STATE_PAUSED:
1905 	case SNDRV_PCM_STATE_SUSPENDED:
1906 		return 0;
1907 	default:
1908 		return -EBADFD;
1909 	}
1910 }
1911 
1912 static int snd_pcm_do_reset(struct snd_pcm_substream *substream,
1913 			    snd_pcm_state_t state)
1914 {
1915 	struct snd_pcm_runtime *runtime = substream->runtime;
1916 	int err = snd_pcm_ops_ioctl(substream, SNDRV_PCM_IOCTL1_RESET, NULL);
1917 	if (err < 0)
1918 		return err;
1919 	snd_pcm_stream_lock_irq(substream);
1920 	runtime->hw_ptr_base = 0;
1921 	runtime->hw_ptr_interrupt = runtime->status->hw_ptr -
1922 		runtime->status->hw_ptr % runtime->period_size;
1923 	runtime->silence_start = runtime->status->hw_ptr;
1924 	runtime->silence_filled = 0;
1925 	snd_pcm_stream_unlock_irq(substream);
1926 	return 0;
1927 }
1928 
1929 static void snd_pcm_post_reset(struct snd_pcm_substream *substream,
1930 			       snd_pcm_state_t state)
1931 {
1932 	struct snd_pcm_runtime *runtime = substream->runtime;
1933 	snd_pcm_stream_lock_irq(substream);
1934 	runtime->control->appl_ptr = runtime->status->hw_ptr;
1935 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK &&
1936 	    runtime->silence_size > 0)
1937 		snd_pcm_playback_silence(substream, ULONG_MAX);
1938 	snd_pcm_stream_unlock_irq(substream);
1939 }
1940 
1941 static const struct action_ops snd_pcm_action_reset = {
1942 	.pre_action = snd_pcm_pre_reset,
1943 	.do_action = snd_pcm_do_reset,
1944 	.post_action = snd_pcm_post_reset
1945 };
1946 
1947 static int snd_pcm_reset(struct snd_pcm_substream *substream)
1948 {
1949 	return snd_pcm_action_nonatomic(&snd_pcm_action_reset, substream,
1950 					ACTION_ARG_IGNORE);
1951 }
1952 
1953 /*
1954  * prepare ioctl
1955  */
1956 /* pass f_flags as state argument */
1957 static int snd_pcm_pre_prepare(struct snd_pcm_substream *substream,
1958 			       snd_pcm_state_t state)
1959 {
1960 	struct snd_pcm_runtime *runtime = substream->runtime;
1961 	int f_flags = (__force int)state;
1962 
1963 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
1964 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
1965 		return -EBADFD;
1966 	if (snd_pcm_running(substream))
1967 		return -EBUSY;
1968 	substream->f_flags = f_flags;
1969 	return 0;
1970 }
1971 
1972 static int snd_pcm_do_prepare(struct snd_pcm_substream *substream,
1973 			      snd_pcm_state_t state)
1974 {
1975 	int err;
1976 	snd_pcm_sync_stop(substream, true);
1977 	err = substream->ops->prepare(substream);
1978 	if (err < 0)
1979 		return err;
1980 	return snd_pcm_do_reset(substream, state);
1981 }
1982 
1983 static void snd_pcm_post_prepare(struct snd_pcm_substream *substream,
1984 				 snd_pcm_state_t state)
1985 {
1986 	struct snd_pcm_runtime *runtime = substream->runtime;
1987 	runtime->control->appl_ptr = runtime->status->hw_ptr;
1988 	snd_pcm_set_state(substream, SNDRV_PCM_STATE_PREPARED);
1989 }
1990 
1991 static const struct action_ops snd_pcm_action_prepare = {
1992 	.pre_action = snd_pcm_pre_prepare,
1993 	.do_action = snd_pcm_do_prepare,
1994 	.post_action = snd_pcm_post_prepare
1995 };
1996 
1997 /**
1998  * snd_pcm_prepare - prepare the PCM substream to be triggerable
1999  * @substream: the PCM substream instance
2000  * @file: file to refer f_flags
2001  *
2002  * Return: Zero if successful, or a negative error code.
2003  */
2004 static int snd_pcm_prepare(struct snd_pcm_substream *substream,
2005 			   struct file *file)
2006 {
2007 	int f_flags;
2008 
2009 	if (file)
2010 		f_flags = file->f_flags;
2011 	else
2012 		f_flags = substream->f_flags;
2013 
2014 	snd_pcm_stream_lock_irq(substream);
2015 	switch (substream->runtime->state) {
2016 	case SNDRV_PCM_STATE_PAUSED:
2017 		snd_pcm_pause(substream, false);
2018 		fallthrough;
2019 	case SNDRV_PCM_STATE_SUSPENDED:
2020 		snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP);
2021 		break;
2022 	}
2023 	snd_pcm_stream_unlock_irq(substream);
2024 
2025 	return snd_pcm_action_nonatomic(&snd_pcm_action_prepare,
2026 					substream,
2027 					(__force snd_pcm_state_t)f_flags);
2028 }
2029 
2030 /*
2031  * drain ioctl
2032  */
2033 
2034 /* drain init callbacks: state argument ignored */
2035 static int snd_pcm_pre_drain_init(struct snd_pcm_substream *substream,
2036 				  snd_pcm_state_t state)
2037 {
2038 	struct snd_pcm_runtime *runtime = substream->runtime;
2039 	switch (runtime->state) {
2040 	case SNDRV_PCM_STATE_OPEN:
2041 	case SNDRV_PCM_STATE_DISCONNECTED:
2042 	case SNDRV_PCM_STATE_SUSPENDED:
2043 		return -EBADFD;
2044 	}
2045 	runtime->trigger_master = substream;
2046 	return 0;
2047 }
2048 
2049 static int snd_pcm_do_drain_init(struct snd_pcm_substream *substream,
2050 				 snd_pcm_state_t state)
2051 {
2052 	struct snd_pcm_runtime *runtime = substream->runtime;
2053 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) {
2054 		switch (runtime->state) {
2055 		case SNDRV_PCM_STATE_PREPARED:
2056 			/* start playback stream if possible */
2057 			if (! snd_pcm_playback_empty(substream)) {
2058 				snd_pcm_do_start(substream, SNDRV_PCM_STATE_DRAINING);
2059 				snd_pcm_post_start(substream, SNDRV_PCM_STATE_DRAINING);
2060 			} else {
2061 				__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SETUP);
2062 			}
2063 			break;
2064 		case SNDRV_PCM_STATE_RUNNING:
2065 			__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_DRAINING);
2066 			break;
2067 		case SNDRV_PCM_STATE_XRUN:
2068 			__snd_pcm_set_state(runtime, SNDRV_PCM_STATE_SETUP);
2069 			break;
2070 		default:
2071 			break;
2072 		}
2073 	} else {
2074 		/* stop running stream */
2075 		if (runtime->state == SNDRV_PCM_STATE_RUNNING) {
2076 			snd_pcm_state_t new_state;
2077 
2078 			new_state = snd_pcm_capture_avail(runtime) > 0 ?
2079 				SNDRV_PCM_STATE_DRAINING : SNDRV_PCM_STATE_SETUP;
2080 			snd_pcm_do_stop(substream, new_state);
2081 			snd_pcm_post_stop(substream, new_state);
2082 		}
2083 	}
2084 
2085 	if (runtime->state == SNDRV_PCM_STATE_DRAINING &&
2086 	    runtime->trigger_master == substream &&
2087 	    (runtime->hw.info & SNDRV_PCM_INFO_DRAIN_TRIGGER))
2088 		return substream->ops->trigger(substream,
2089 					       SNDRV_PCM_TRIGGER_DRAIN);
2090 
2091 	return 0;
2092 }
2093 
2094 static void snd_pcm_post_drain_init(struct snd_pcm_substream *substream,
2095 				    snd_pcm_state_t state)
2096 {
2097 }
2098 
2099 static const struct action_ops snd_pcm_action_drain_init = {
2100 	.pre_action = snd_pcm_pre_drain_init,
2101 	.do_action = snd_pcm_do_drain_init,
2102 	.post_action = snd_pcm_post_drain_init
2103 };
2104 
2105 /*
2106  * Drain the stream(s).
2107  * When the substream is linked, sync until the draining of all playback streams
2108  * is finished.
2109  * After this call, all streams are supposed to be either SETUP or DRAINING
2110  * (capture only) state.
2111  */
2112 static int snd_pcm_drain(struct snd_pcm_substream *substream,
2113 			 struct file *file)
2114 {
2115 	struct snd_card *card;
2116 	struct snd_pcm_runtime *runtime;
2117 	struct snd_pcm_substream *s;
2118 	struct snd_pcm_group *group;
2119 	wait_queue_entry_t wait;
2120 	int result = 0;
2121 	int nonblock = 0;
2122 
2123 	card = substream->pcm->card;
2124 	runtime = substream->runtime;
2125 
2126 	if (runtime->state == SNDRV_PCM_STATE_OPEN)
2127 		return -EBADFD;
2128 
2129 	if (file) {
2130 		if (file->f_flags & O_NONBLOCK)
2131 			nonblock = 1;
2132 	} else if (substream->f_flags & O_NONBLOCK)
2133 		nonblock = 1;
2134 
2135 	snd_pcm_stream_lock_irq(substream);
2136 	/* resume pause */
2137 	if (runtime->state == SNDRV_PCM_STATE_PAUSED)
2138 		snd_pcm_pause(substream, false);
2139 
2140 	/* pre-start/stop - all running streams are changed to DRAINING state */
2141 	result = snd_pcm_action(&snd_pcm_action_drain_init, substream,
2142 				ACTION_ARG_IGNORE);
2143 	if (result < 0)
2144 		goto unlock;
2145 	/* in non-blocking, we don't wait in ioctl but let caller poll */
2146 	if (nonblock) {
2147 		result = -EAGAIN;
2148 		goto unlock;
2149 	}
2150 
2151 	for (;;) {
2152 		long tout;
2153 		struct snd_pcm_runtime *to_check;
2154 		if (signal_pending(current)) {
2155 			result = -ERESTARTSYS;
2156 			break;
2157 		}
2158 		/* find a substream to drain */
2159 		to_check = NULL;
2160 		group = snd_pcm_stream_group_ref(substream);
2161 		snd_pcm_group_for_each_entry(s, substream) {
2162 			if (s->stream != SNDRV_PCM_STREAM_PLAYBACK)
2163 				continue;
2164 			runtime = s->runtime;
2165 			if (runtime->state == SNDRV_PCM_STATE_DRAINING) {
2166 				to_check = runtime;
2167 				break;
2168 			}
2169 		}
2170 		snd_pcm_group_unref(group, substream);
2171 		if (!to_check)
2172 			break; /* all drained */
2173 		init_waitqueue_entry(&wait, current);
2174 		set_current_state(TASK_INTERRUPTIBLE);
2175 		add_wait_queue(&to_check->sleep, &wait);
2176 		snd_pcm_stream_unlock_irq(substream);
2177 		if (runtime->no_period_wakeup)
2178 			tout = MAX_SCHEDULE_TIMEOUT;
2179 		else {
2180 			tout = 100;
2181 			if (runtime->rate) {
2182 				long t = runtime->buffer_size * 1100 / runtime->rate;
2183 				tout = max(t, tout);
2184 			}
2185 			tout = msecs_to_jiffies(tout);
2186 		}
2187 		tout = schedule_timeout(tout);
2188 
2189 		snd_pcm_stream_lock_irq(substream);
2190 		group = snd_pcm_stream_group_ref(substream);
2191 		snd_pcm_group_for_each_entry(s, substream) {
2192 			if (s->runtime == to_check) {
2193 				remove_wait_queue(&to_check->sleep, &wait);
2194 				break;
2195 			}
2196 		}
2197 		snd_pcm_group_unref(group, substream);
2198 
2199 		if (card->shutdown) {
2200 			result = -ENODEV;
2201 			break;
2202 		}
2203 		if (tout == 0) {
2204 			if (substream->runtime->state == SNDRV_PCM_STATE_SUSPENDED)
2205 				result = -ESTRPIPE;
2206 			else {
2207 				dev_dbg(substream->pcm->card->dev,
2208 					"playback drain timeout (DMA or IRQ trouble?)\n");
2209 				snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP);
2210 				result = -EIO;
2211 			}
2212 			break;
2213 		}
2214 	}
2215 
2216  unlock:
2217 	snd_pcm_stream_unlock_irq(substream);
2218 
2219 	return result;
2220 }
2221 
2222 /*
2223  * drop ioctl
2224  *
2225  * Immediately put all linked substreams into SETUP state.
2226  */
2227 static int snd_pcm_drop(struct snd_pcm_substream *substream)
2228 {
2229 	struct snd_pcm_runtime *runtime;
2230 	int result = 0;
2231 
2232 	if (PCM_RUNTIME_CHECK(substream))
2233 		return -ENXIO;
2234 	runtime = substream->runtime;
2235 
2236 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
2237 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
2238 		return -EBADFD;
2239 
2240 	snd_pcm_stream_lock_irq(substream);
2241 	/* resume pause */
2242 	if (runtime->state == SNDRV_PCM_STATE_PAUSED)
2243 		snd_pcm_pause(substream, false);
2244 
2245 	snd_pcm_stop(substream, SNDRV_PCM_STATE_SETUP);
2246 	/* runtime->control->appl_ptr = runtime->status->hw_ptr; */
2247 	snd_pcm_stream_unlock_irq(substream);
2248 
2249 	return result;
2250 }
2251 
2252 
2253 static bool is_pcm_file(struct file *file)
2254 {
2255 	struct inode *inode = file_inode(file);
2256 	struct snd_pcm *pcm;
2257 	unsigned int minor;
2258 
2259 	if (!S_ISCHR(inode->i_mode) || imajor(inode) != snd_major)
2260 		return false;
2261 	minor = iminor(inode);
2262 	pcm = snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_PLAYBACK);
2263 	if (!pcm)
2264 		pcm = snd_lookup_minor_data(minor, SNDRV_DEVICE_TYPE_PCM_CAPTURE);
2265 	if (!pcm)
2266 		return false;
2267 	snd_card_unref(pcm->card);
2268 	return true;
2269 }
2270 
2271 /*
2272  * PCM link handling
2273  */
2274 static int snd_pcm_link(struct snd_pcm_substream *substream, int fd)
2275 {
2276 	int res = 0;
2277 	struct snd_pcm_file *pcm_file;
2278 	struct snd_pcm_substream *substream1;
2279 	struct snd_pcm_group *group, *target_group;
2280 	bool nonatomic = substream->pcm->nonatomic;
2281 	struct fd f = fdget(fd);
2282 
2283 	if (!f.file)
2284 		return -EBADFD;
2285 	if (!is_pcm_file(f.file)) {
2286 		res = -EBADFD;
2287 		goto _badf;
2288 	}
2289 	pcm_file = f.file->private_data;
2290 	substream1 = pcm_file->substream;
2291 
2292 	if (substream == substream1) {
2293 		res = -EINVAL;
2294 		goto _badf;
2295 	}
2296 
2297 	group = kzalloc(sizeof(*group), GFP_KERNEL);
2298 	if (!group) {
2299 		res = -ENOMEM;
2300 		goto _nolock;
2301 	}
2302 	snd_pcm_group_init(group);
2303 
2304 	down_write(&snd_pcm_link_rwsem);
2305 	if (substream->runtime->state == SNDRV_PCM_STATE_OPEN ||
2306 	    substream->runtime->state != substream1->runtime->state ||
2307 	    substream->pcm->nonatomic != substream1->pcm->nonatomic) {
2308 		res = -EBADFD;
2309 		goto _end;
2310 	}
2311 	if (snd_pcm_stream_linked(substream1)) {
2312 		res = -EALREADY;
2313 		goto _end;
2314 	}
2315 
2316 	snd_pcm_stream_lock_irq(substream);
2317 	if (!snd_pcm_stream_linked(substream)) {
2318 		snd_pcm_group_assign(substream, group);
2319 		group = NULL; /* assigned, don't free this one below */
2320 	}
2321 	target_group = substream->group;
2322 	snd_pcm_stream_unlock_irq(substream);
2323 
2324 	snd_pcm_group_lock_irq(target_group, nonatomic);
2325 	snd_pcm_stream_lock_nested(substream1);
2326 	snd_pcm_group_assign(substream1, target_group);
2327 	refcount_inc(&target_group->refs);
2328 	snd_pcm_stream_unlock(substream1);
2329 	snd_pcm_group_unlock_irq(target_group, nonatomic);
2330  _end:
2331 	up_write(&snd_pcm_link_rwsem);
2332  _nolock:
2333 	kfree(group);
2334  _badf:
2335 	fdput(f);
2336 	return res;
2337 }
2338 
2339 static void relink_to_local(struct snd_pcm_substream *substream)
2340 {
2341 	snd_pcm_stream_lock_nested(substream);
2342 	snd_pcm_group_assign(substream, &substream->self_group);
2343 	snd_pcm_stream_unlock(substream);
2344 }
2345 
2346 static int snd_pcm_unlink(struct snd_pcm_substream *substream)
2347 {
2348 	struct snd_pcm_group *group;
2349 	bool nonatomic = substream->pcm->nonatomic;
2350 	bool do_free = false;
2351 	int res = 0;
2352 
2353 	down_write(&snd_pcm_link_rwsem);
2354 
2355 	if (!snd_pcm_stream_linked(substream)) {
2356 		res = -EALREADY;
2357 		goto _end;
2358 	}
2359 
2360 	group = substream->group;
2361 	snd_pcm_group_lock_irq(group, nonatomic);
2362 
2363 	relink_to_local(substream);
2364 	refcount_dec(&group->refs);
2365 
2366 	/* detach the last stream, too */
2367 	if (list_is_singular(&group->substreams)) {
2368 		relink_to_local(list_first_entry(&group->substreams,
2369 						 struct snd_pcm_substream,
2370 						 link_list));
2371 		do_free = refcount_dec_and_test(&group->refs);
2372 	}
2373 
2374 	snd_pcm_group_unlock_irq(group, nonatomic);
2375 	if (do_free)
2376 		kfree(group);
2377 
2378        _end:
2379 	up_write(&snd_pcm_link_rwsem);
2380 	return res;
2381 }
2382 
2383 /*
2384  * hw configurator
2385  */
2386 static int snd_pcm_hw_rule_mul(struct snd_pcm_hw_params *params,
2387 			       struct snd_pcm_hw_rule *rule)
2388 {
2389 	struct snd_interval t;
2390 	snd_interval_mul(hw_param_interval_c(params, rule->deps[0]),
2391 		     hw_param_interval_c(params, rule->deps[1]), &t);
2392 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2393 }
2394 
2395 static int snd_pcm_hw_rule_div(struct snd_pcm_hw_params *params,
2396 			       struct snd_pcm_hw_rule *rule)
2397 {
2398 	struct snd_interval t;
2399 	snd_interval_div(hw_param_interval_c(params, rule->deps[0]),
2400 		     hw_param_interval_c(params, rule->deps[1]), &t);
2401 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2402 }
2403 
2404 static int snd_pcm_hw_rule_muldivk(struct snd_pcm_hw_params *params,
2405 				   struct snd_pcm_hw_rule *rule)
2406 {
2407 	struct snd_interval t;
2408 	snd_interval_muldivk(hw_param_interval_c(params, rule->deps[0]),
2409 			 hw_param_interval_c(params, rule->deps[1]),
2410 			 (unsigned long) rule->private, &t);
2411 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2412 }
2413 
2414 static int snd_pcm_hw_rule_mulkdiv(struct snd_pcm_hw_params *params,
2415 				   struct snd_pcm_hw_rule *rule)
2416 {
2417 	struct snd_interval t;
2418 	snd_interval_mulkdiv(hw_param_interval_c(params, rule->deps[0]),
2419 			 (unsigned long) rule->private,
2420 			 hw_param_interval_c(params, rule->deps[1]), &t);
2421 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2422 }
2423 
2424 static int snd_pcm_hw_rule_format(struct snd_pcm_hw_params *params,
2425 				  struct snd_pcm_hw_rule *rule)
2426 {
2427 	snd_pcm_format_t k;
2428 	const struct snd_interval *i =
2429 				hw_param_interval_c(params, rule->deps[0]);
2430 	struct snd_mask m;
2431 	struct snd_mask *mask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT);
2432 	snd_mask_any(&m);
2433 	pcm_for_each_format(k) {
2434 		int bits;
2435 		if (!snd_mask_test_format(mask, k))
2436 			continue;
2437 		bits = snd_pcm_format_physical_width(k);
2438 		if (bits <= 0)
2439 			continue; /* ignore invalid formats */
2440 		if ((unsigned)bits < i->min || (unsigned)bits > i->max)
2441 			snd_mask_reset(&m, (__force unsigned)k);
2442 	}
2443 	return snd_mask_refine(mask, &m);
2444 }
2445 
2446 static int snd_pcm_hw_rule_sample_bits(struct snd_pcm_hw_params *params,
2447 				       struct snd_pcm_hw_rule *rule)
2448 {
2449 	struct snd_interval t;
2450 	snd_pcm_format_t k;
2451 
2452 	t.min = UINT_MAX;
2453 	t.max = 0;
2454 	t.openmin = 0;
2455 	t.openmax = 0;
2456 	pcm_for_each_format(k) {
2457 		int bits;
2458 		if (!snd_mask_test_format(hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT), k))
2459 			continue;
2460 		bits = snd_pcm_format_physical_width(k);
2461 		if (bits <= 0)
2462 			continue; /* ignore invalid formats */
2463 		if (t.min > (unsigned)bits)
2464 			t.min = bits;
2465 		if (t.max < (unsigned)bits)
2466 			t.max = bits;
2467 	}
2468 	t.integer = 1;
2469 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2470 }
2471 
2472 #if SNDRV_PCM_RATE_5512 != 1 << 0 || SNDRV_PCM_RATE_192000 != 1 << 12
2473 #error "Change this table"
2474 #endif
2475 
2476 static const unsigned int rates[] = {
2477 	5512, 8000, 11025, 16000, 22050, 32000, 44100,
2478 	48000, 64000, 88200, 96000, 176400, 192000, 352800, 384000
2479 };
2480 
2481 const struct snd_pcm_hw_constraint_list snd_pcm_known_rates = {
2482 	.count = ARRAY_SIZE(rates),
2483 	.list = rates,
2484 };
2485 
2486 static int snd_pcm_hw_rule_rate(struct snd_pcm_hw_params *params,
2487 				struct snd_pcm_hw_rule *rule)
2488 {
2489 	struct snd_pcm_hardware *hw = rule->private;
2490 	return snd_interval_list(hw_param_interval(params, rule->var),
2491 				 snd_pcm_known_rates.count,
2492 				 snd_pcm_known_rates.list, hw->rates);
2493 }
2494 
2495 static int snd_pcm_hw_rule_buffer_bytes_max(struct snd_pcm_hw_params *params,
2496 					    struct snd_pcm_hw_rule *rule)
2497 {
2498 	struct snd_interval t;
2499 	struct snd_pcm_substream *substream = rule->private;
2500 	t.min = 0;
2501 	t.max = substream->buffer_bytes_max;
2502 	t.openmin = 0;
2503 	t.openmax = 0;
2504 	t.integer = 1;
2505 	return snd_interval_refine(hw_param_interval(params, rule->var), &t);
2506 }
2507 
2508 static int snd_pcm_hw_rule_subformats(struct snd_pcm_hw_params *params,
2509 				      struct snd_pcm_hw_rule *rule)
2510 {
2511 	struct snd_mask *sfmask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_SUBFORMAT);
2512 	struct snd_mask *fmask = hw_param_mask(params, SNDRV_PCM_HW_PARAM_FORMAT);
2513 	u32 *subformats = rule->private;
2514 	snd_pcm_format_t f;
2515 	struct snd_mask m;
2516 
2517 	snd_mask_none(&m);
2518 	/* All PCMs support at least the default STD subformat. */
2519 	snd_mask_set(&m, (__force unsigned)SNDRV_PCM_SUBFORMAT_STD);
2520 
2521 	pcm_for_each_format(f) {
2522 		if (!snd_mask_test(fmask, (__force unsigned)f))
2523 			continue;
2524 
2525 		if (f == SNDRV_PCM_FORMAT_S32_LE && *subformats)
2526 			m.bits[0] |= *subformats;
2527 		else if (snd_pcm_format_linear(f))
2528 			snd_mask_set(&m, (__force unsigned)SNDRV_PCM_SUBFORMAT_MSBITS_MAX);
2529 	}
2530 
2531 	return snd_mask_refine(sfmask, &m);
2532 }
2533 
2534 static int snd_pcm_hw_constraint_subformats(struct snd_pcm_runtime *runtime,
2535 					   unsigned int cond, u32 *subformats)
2536 {
2537 	return snd_pcm_hw_rule_add(runtime, cond, -1,
2538 				   snd_pcm_hw_rule_subformats, (void *)subformats,
2539 				   SNDRV_PCM_HW_PARAM_SUBFORMAT,
2540 				   SNDRV_PCM_HW_PARAM_FORMAT, -1);
2541 }
2542 
2543 static int snd_pcm_hw_constraints_init(struct snd_pcm_substream *substream)
2544 {
2545 	struct snd_pcm_runtime *runtime = substream->runtime;
2546 	struct snd_pcm_hw_constraints *constrs = &runtime->hw_constraints;
2547 	int k, err;
2548 
2549 	for (k = SNDRV_PCM_HW_PARAM_FIRST_MASK; k <= SNDRV_PCM_HW_PARAM_LAST_MASK; k++) {
2550 		snd_mask_any(constrs_mask(constrs, k));
2551 	}
2552 
2553 	for (k = SNDRV_PCM_HW_PARAM_FIRST_INTERVAL; k <= SNDRV_PCM_HW_PARAM_LAST_INTERVAL; k++) {
2554 		snd_interval_any(constrs_interval(constrs, k));
2555 	}
2556 
2557 	snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_CHANNELS));
2558 	snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_SIZE));
2559 	snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_BUFFER_BYTES));
2560 	snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_SAMPLE_BITS));
2561 	snd_interval_setinteger(constrs_interval(constrs, SNDRV_PCM_HW_PARAM_FRAME_BITS));
2562 
2563 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FORMAT,
2564 				   snd_pcm_hw_rule_format, NULL,
2565 				   SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1);
2566 	if (err < 0)
2567 		return err;
2568 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS,
2569 				  snd_pcm_hw_rule_sample_bits, NULL,
2570 				  SNDRV_PCM_HW_PARAM_FORMAT,
2571 				  SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1);
2572 	if (err < 0)
2573 		return err;
2574 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_SAMPLE_BITS,
2575 				  snd_pcm_hw_rule_div, NULL,
2576 				  SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1);
2577 	if (err < 0)
2578 		return err;
2579 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS,
2580 				  snd_pcm_hw_rule_mul, NULL,
2581 				  SNDRV_PCM_HW_PARAM_SAMPLE_BITS, SNDRV_PCM_HW_PARAM_CHANNELS, -1);
2582 	if (err < 0)
2583 		return err;
2584 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS,
2585 				  snd_pcm_hw_rule_mulkdiv, (void*) 8,
2586 				  SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1);
2587 	if (err < 0)
2588 		return err;
2589 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_FRAME_BITS,
2590 				  snd_pcm_hw_rule_mulkdiv, (void*) 8,
2591 				  SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_BUFFER_SIZE, -1);
2592 	if (err < 0)
2593 		return err;
2594 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_CHANNELS,
2595 				  snd_pcm_hw_rule_div, NULL,
2596 				  SNDRV_PCM_HW_PARAM_FRAME_BITS, SNDRV_PCM_HW_PARAM_SAMPLE_BITS, -1);
2597 	if (err < 0)
2598 		return err;
2599 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE,
2600 				  snd_pcm_hw_rule_mulkdiv, (void*) 1000000,
2601 				  SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_TIME, -1);
2602 	if (err < 0)
2603 		return err;
2604 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE,
2605 				  snd_pcm_hw_rule_mulkdiv, (void*) 1000000,
2606 				  SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_BUFFER_TIME, -1);
2607 	if (err < 0)
2608 		return err;
2609 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIODS,
2610 				  snd_pcm_hw_rule_div, NULL,
2611 				  SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIOD_SIZE, -1);
2612 	if (err < 0)
2613 		return err;
2614 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE,
2615 				  snd_pcm_hw_rule_div, NULL,
2616 				  SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1);
2617 	if (err < 0)
2618 		return err;
2619 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE,
2620 				  snd_pcm_hw_rule_mulkdiv, (void*) 8,
2621 				  SNDRV_PCM_HW_PARAM_PERIOD_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1);
2622 	if (err < 0)
2623 		return err;
2624 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_SIZE,
2625 				  snd_pcm_hw_rule_muldivk, (void*) 1000000,
2626 				  SNDRV_PCM_HW_PARAM_PERIOD_TIME, SNDRV_PCM_HW_PARAM_RATE, -1);
2627 	if (err < 0)
2628 		return err;
2629 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
2630 				  snd_pcm_hw_rule_mul, NULL,
2631 				  SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_PERIODS, -1);
2632 	if (err < 0)
2633 		return err;
2634 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
2635 				  snd_pcm_hw_rule_mulkdiv, (void*) 8,
2636 				  SNDRV_PCM_HW_PARAM_BUFFER_BYTES, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1);
2637 	if (err < 0)
2638 		return err;
2639 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_SIZE,
2640 				  snd_pcm_hw_rule_muldivk, (void*) 1000000,
2641 				  SNDRV_PCM_HW_PARAM_BUFFER_TIME, SNDRV_PCM_HW_PARAM_RATE, -1);
2642 	if (err < 0)
2643 		return err;
2644 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES,
2645 				  snd_pcm_hw_rule_muldivk, (void*) 8,
2646 				  SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1);
2647 	if (err < 0)
2648 		return err;
2649 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES,
2650 				  snd_pcm_hw_rule_muldivk, (void*) 8,
2651 				  SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_FRAME_BITS, -1);
2652 	if (err < 0)
2653 		return err;
2654 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_TIME,
2655 				  snd_pcm_hw_rule_mulkdiv, (void*) 1000000,
2656 				  SNDRV_PCM_HW_PARAM_PERIOD_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1);
2657 	if (err < 0)
2658 		return err;
2659 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_TIME,
2660 				  snd_pcm_hw_rule_mulkdiv, (void*) 1000000,
2661 				  SNDRV_PCM_HW_PARAM_BUFFER_SIZE, SNDRV_PCM_HW_PARAM_RATE, -1);
2662 	if (err < 0)
2663 		return err;
2664 	return 0;
2665 }
2666 
2667 static int snd_pcm_hw_constraints_complete(struct snd_pcm_substream *substream)
2668 {
2669 	struct snd_pcm_runtime *runtime = substream->runtime;
2670 	struct snd_pcm_hardware *hw = &runtime->hw;
2671 	int err;
2672 	unsigned int mask = 0;
2673 
2674         if (hw->info & SNDRV_PCM_INFO_INTERLEAVED)
2675 		mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_INTERLEAVED);
2676         if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED)
2677 		mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_RW_NONINTERLEAVED);
2678 	if (hw_support_mmap(substream)) {
2679 		if (hw->info & SNDRV_PCM_INFO_INTERLEAVED)
2680 			mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_INTERLEAVED);
2681 		if (hw->info & SNDRV_PCM_INFO_NONINTERLEAVED)
2682 			mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_NONINTERLEAVED);
2683 		if (hw->info & SNDRV_PCM_INFO_COMPLEX)
2684 			mask |= PARAM_MASK_BIT(SNDRV_PCM_ACCESS_MMAP_COMPLEX);
2685 	}
2686 	err = snd_pcm_hw_constraint_mask(runtime, SNDRV_PCM_HW_PARAM_ACCESS, mask);
2687 	if (err < 0)
2688 		return err;
2689 
2690 	err = snd_pcm_hw_constraint_mask64(runtime, SNDRV_PCM_HW_PARAM_FORMAT, hw->formats);
2691 	if (err < 0)
2692 		return err;
2693 
2694 	err = snd_pcm_hw_constraint_subformats(runtime, 0, &hw->subformats);
2695 	if (err < 0)
2696 		return err;
2697 
2698 	err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_CHANNELS,
2699 					   hw->channels_min, hw->channels_max);
2700 	if (err < 0)
2701 		return err;
2702 
2703 	err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_RATE,
2704 					   hw->rate_min, hw->rate_max);
2705 	if (err < 0)
2706 		return err;
2707 
2708 	err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIOD_BYTES,
2709 					   hw->period_bytes_min, hw->period_bytes_max);
2710 	if (err < 0)
2711 		return err;
2712 
2713 	err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_PERIODS,
2714 					   hw->periods_min, hw->periods_max);
2715 	if (err < 0)
2716 		return err;
2717 
2718 	err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES,
2719 					   hw->period_bytes_min, hw->buffer_bytes_max);
2720 	if (err < 0)
2721 		return err;
2722 
2723 	err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES,
2724 				  snd_pcm_hw_rule_buffer_bytes_max, substream,
2725 				  SNDRV_PCM_HW_PARAM_BUFFER_BYTES, -1);
2726 	if (err < 0)
2727 		return err;
2728 
2729 	/* FIXME: remove */
2730 	if (runtime->dma_bytes) {
2731 		err = snd_pcm_hw_constraint_minmax(runtime, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, 0, runtime->dma_bytes);
2732 		if (err < 0)
2733 			return err;
2734 	}
2735 
2736 	if (!(hw->rates & (SNDRV_PCM_RATE_KNOT | SNDRV_PCM_RATE_CONTINUOUS))) {
2737 		err = snd_pcm_hw_rule_add(runtime, 0, SNDRV_PCM_HW_PARAM_RATE,
2738 					  snd_pcm_hw_rule_rate, hw,
2739 					  SNDRV_PCM_HW_PARAM_RATE, -1);
2740 		if (err < 0)
2741 			return err;
2742 	}
2743 
2744 	/* FIXME: this belong to lowlevel */
2745 	snd_pcm_hw_constraint_integer(runtime, SNDRV_PCM_HW_PARAM_PERIOD_SIZE);
2746 
2747 	return 0;
2748 }
2749 
2750 static void pcm_release_private(struct snd_pcm_substream *substream)
2751 {
2752 	if (snd_pcm_stream_linked(substream))
2753 		snd_pcm_unlink(substream);
2754 }
2755 
2756 void snd_pcm_release_substream(struct snd_pcm_substream *substream)
2757 {
2758 	substream->ref_count--;
2759 	if (substream->ref_count > 0)
2760 		return;
2761 
2762 	snd_pcm_drop(substream);
2763 	if (substream->hw_opened) {
2764 		if (substream->runtime->state != SNDRV_PCM_STATE_OPEN)
2765 			do_hw_free(substream);
2766 		substream->ops->close(substream);
2767 		substream->hw_opened = 0;
2768 	}
2769 	if (cpu_latency_qos_request_active(&substream->latency_pm_qos_req))
2770 		cpu_latency_qos_remove_request(&substream->latency_pm_qos_req);
2771 	if (substream->pcm_release) {
2772 		substream->pcm_release(substream);
2773 		substream->pcm_release = NULL;
2774 	}
2775 	snd_pcm_detach_substream(substream);
2776 }
2777 EXPORT_SYMBOL(snd_pcm_release_substream);
2778 
2779 int snd_pcm_open_substream(struct snd_pcm *pcm, int stream,
2780 			   struct file *file,
2781 			   struct snd_pcm_substream **rsubstream)
2782 {
2783 	struct snd_pcm_substream *substream;
2784 	int err;
2785 
2786 	err = snd_pcm_attach_substream(pcm, stream, file, &substream);
2787 	if (err < 0)
2788 		return err;
2789 	if (substream->ref_count > 1) {
2790 		*rsubstream = substream;
2791 		return 0;
2792 	}
2793 
2794 	err = snd_pcm_hw_constraints_init(substream);
2795 	if (err < 0) {
2796 		pcm_dbg(pcm, "snd_pcm_hw_constraints_init failed\n");
2797 		goto error;
2798 	}
2799 
2800 	err = substream->ops->open(substream);
2801 	if (err < 0)
2802 		goto error;
2803 
2804 	substream->hw_opened = 1;
2805 
2806 	err = snd_pcm_hw_constraints_complete(substream);
2807 	if (err < 0) {
2808 		pcm_dbg(pcm, "snd_pcm_hw_constraints_complete failed\n");
2809 		goto error;
2810 	}
2811 
2812 	/* automatically set EXPLICIT_SYNC flag in the managed mode whenever
2813 	 * the DMA buffer requires it
2814 	 */
2815 	if (substream->managed_buffer_alloc &&
2816 	    substream->dma_buffer.dev.need_sync)
2817 		substream->runtime->hw.info |= SNDRV_PCM_INFO_EXPLICIT_SYNC;
2818 
2819 	*rsubstream = substream;
2820 	return 0;
2821 
2822  error:
2823 	snd_pcm_release_substream(substream);
2824 	return err;
2825 }
2826 EXPORT_SYMBOL(snd_pcm_open_substream);
2827 
2828 static int snd_pcm_open_file(struct file *file,
2829 			     struct snd_pcm *pcm,
2830 			     int stream)
2831 {
2832 	struct snd_pcm_file *pcm_file;
2833 	struct snd_pcm_substream *substream;
2834 	int err;
2835 
2836 	err = snd_pcm_open_substream(pcm, stream, file, &substream);
2837 	if (err < 0)
2838 		return err;
2839 
2840 	pcm_file = kzalloc(sizeof(*pcm_file), GFP_KERNEL);
2841 	if (pcm_file == NULL) {
2842 		snd_pcm_release_substream(substream);
2843 		return -ENOMEM;
2844 	}
2845 	pcm_file->substream = substream;
2846 	if (substream->ref_count == 1)
2847 		substream->pcm_release = pcm_release_private;
2848 	file->private_data = pcm_file;
2849 
2850 	return 0;
2851 }
2852 
2853 static int snd_pcm_playback_open(struct inode *inode, struct file *file)
2854 {
2855 	struct snd_pcm *pcm;
2856 	int err = nonseekable_open(inode, file);
2857 	if (err < 0)
2858 		return err;
2859 	pcm = snd_lookup_minor_data(iminor(inode),
2860 				    SNDRV_DEVICE_TYPE_PCM_PLAYBACK);
2861 	err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_PLAYBACK);
2862 	if (pcm)
2863 		snd_card_unref(pcm->card);
2864 	return err;
2865 }
2866 
2867 static int snd_pcm_capture_open(struct inode *inode, struct file *file)
2868 {
2869 	struct snd_pcm *pcm;
2870 	int err = nonseekable_open(inode, file);
2871 	if (err < 0)
2872 		return err;
2873 	pcm = snd_lookup_minor_data(iminor(inode),
2874 				    SNDRV_DEVICE_TYPE_PCM_CAPTURE);
2875 	err = snd_pcm_open(file, pcm, SNDRV_PCM_STREAM_CAPTURE);
2876 	if (pcm)
2877 		snd_card_unref(pcm->card);
2878 	return err;
2879 }
2880 
2881 static int snd_pcm_open(struct file *file, struct snd_pcm *pcm, int stream)
2882 {
2883 	int err;
2884 	wait_queue_entry_t wait;
2885 
2886 	if (pcm == NULL) {
2887 		err = -ENODEV;
2888 		goto __error1;
2889 	}
2890 	err = snd_card_file_add(pcm->card, file);
2891 	if (err < 0)
2892 		goto __error1;
2893 	if (!try_module_get(pcm->card->module)) {
2894 		err = -EFAULT;
2895 		goto __error2;
2896 	}
2897 	init_waitqueue_entry(&wait, current);
2898 	add_wait_queue(&pcm->open_wait, &wait);
2899 	mutex_lock(&pcm->open_mutex);
2900 	while (1) {
2901 		err = snd_pcm_open_file(file, pcm, stream);
2902 		if (err >= 0)
2903 			break;
2904 		if (err == -EAGAIN) {
2905 			if (file->f_flags & O_NONBLOCK) {
2906 				err = -EBUSY;
2907 				break;
2908 			}
2909 		} else
2910 			break;
2911 		set_current_state(TASK_INTERRUPTIBLE);
2912 		mutex_unlock(&pcm->open_mutex);
2913 		schedule();
2914 		mutex_lock(&pcm->open_mutex);
2915 		if (pcm->card->shutdown) {
2916 			err = -ENODEV;
2917 			break;
2918 		}
2919 		if (signal_pending(current)) {
2920 			err = -ERESTARTSYS;
2921 			break;
2922 		}
2923 	}
2924 	remove_wait_queue(&pcm->open_wait, &wait);
2925 	mutex_unlock(&pcm->open_mutex);
2926 	if (err < 0)
2927 		goto __error;
2928 	return err;
2929 
2930       __error:
2931 	module_put(pcm->card->module);
2932       __error2:
2933       	snd_card_file_remove(pcm->card, file);
2934       __error1:
2935       	return err;
2936 }
2937 
2938 static int snd_pcm_release(struct inode *inode, struct file *file)
2939 {
2940 	struct snd_pcm *pcm;
2941 	struct snd_pcm_substream *substream;
2942 	struct snd_pcm_file *pcm_file;
2943 
2944 	pcm_file = file->private_data;
2945 	substream = pcm_file->substream;
2946 	if (snd_BUG_ON(!substream))
2947 		return -ENXIO;
2948 	pcm = substream->pcm;
2949 
2950 	/* block until the device gets woken up as it may touch the hardware */
2951 	snd_power_wait(pcm->card);
2952 
2953 	mutex_lock(&pcm->open_mutex);
2954 	snd_pcm_release_substream(substream);
2955 	kfree(pcm_file);
2956 	mutex_unlock(&pcm->open_mutex);
2957 	wake_up(&pcm->open_wait);
2958 	module_put(pcm->card->module);
2959 	snd_card_file_remove(pcm->card, file);
2960 	return 0;
2961 }
2962 
2963 /* check and update PCM state; return 0 or a negative error
2964  * call this inside PCM lock
2965  */
2966 static int do_pcm_hwsync(struct snd_pcm_substream *substream)
2967 {
2968 	switch (substream->runtime->state) {
2969 	case SNDRV_PCM_STATE_DRAINING:
2970 		if (substream->stream == SNDRV_PCM_STREAM_CAPTURE)
2971 			return -EBADFD;
2972 		fallthrough;
2973 	case SNDRV_PCM_STATE_RUNNING:
2974 		return snd_pcm_update_hw_ptr(substream);
2975 	case SNDRV_PCM_STATE_PREPARED:
2976 	case SNDRV_PCM_STATE_PAUSED:
2977 		return 0;
2978 	case SNDRV_PCM_STATE_SUSPENDED:
2979 		return -ESTRPIPE;
2980 	case SNDRV_PCM_STATE_XRUN:
2981 		return -EPIPE;
2982 	default:
2983 		return -EBADFD;
2984 	}
2985 }
2986 
2987 /* increase the appl_ptr; returns the processed frames or a negative error */
2988 static snd_pcm_sframes_t forward_appl_ptr(struct snd_pcm_substream *substream,
2989 					  snd_pcm_uframes_t frames,
2990 					   snd_pcm_sframes_t avail)
2991 {
2992 	struct snd_pcm_runtime *runtime = substream->runtime;
2993 	snd_pcm_sframes_t appl_ptr;
2994 	int ret;
2995 
2996 	if (avail <= 0)
2997 		return 0;
2998 	if (frames > (snd_pcm_uframes_t)avail)
2999 		frames = avail;
3000 	appl_ptr = runtime->control->appl_ptr + frames;
3001 	if (appl_ptr >= (snd_pcm_sframes_t)runtime->boundary)
3002 		appl_ptr -= runtime->boundary;
3003 	ret = pcm_lib_apply_appl_ptr(substream, appl_ptr);
3004 	return ret < 0 ? ret : frames;
3005 }
3006 
3007 /* decrease the appl_ptr; returns the processed frames or zero for error */
3008 static snd_pcm_sframes_t rewind_appl_ptr(struct snd_pcm_substream *substream,
3009 					 snd_pcm_uframes_t frames,
3010 					 snd_pcm_sframes_t avail)
3011 {
3012 	struct snd_pcm_runtime *runtime = substream->runtime;
3013 	snd_pcm_sframes_t appl_ptr;
3014 	int ret;
3015 
3016 	if (avail <= 0)
3017 		return 0;
3018 	if (frames > (snd_pcm_uframes_t)avail)
3019 		frames = avail;
3020 	appl_ptr = runtime->control->appl_ptr - frames;
3021 	if (appl_ptr < 0)
3022 		appl_ptr += runtime->boundary;
3023 	ret = pcm_lib_apply_appl_ptr(substream, appl_ptr);
3024 	/* NOTE: we return zero for errors because PulseAudio gets depressed
3025 	 * upon receiving an error from rewind ioctl and stops processing
3026 	 * any longer.  Returning zero means that no rewind is done, so
3027 	 * it's not absolutely wrong to answer like that.
3028 	 */
3029 	return ret < 0 ? 0 : frames;
3030 }
3031 
3032 static snd_pcm_sframes_t snd_pcm_rewind(struct snd_pcm_substream *substream,
3033 					snd_pcm_uframes_t frames)
3034 {
3035 	snd_pcm_sframes_t ret;
3036 
3037 	if (frames == 0)
3038 		return 0;
3039 
3040 	snd_pcm_stream_lock_irq(substream);
3041 	ret = do_pcm_hwsync(substream);
3042 	if (!ret)
3043 		ret = rewind_appl_ptr(substream, frames,
3044 				      snd_pcm_hw_avail(substream));
3045 	snd_pcm_stream_unlock_irq(substream);
3046 	if (ret >= 0)
3047 		snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE);
3048 	return ret;
3049 }
3050 
3051 static snd_pcm_sframes_t snd_pcm_forward(struct snd_pcm_substream *substream,
3052 					 snd_pcm_uframes_t frames)
3053 {
3054 	snd_pcm_sframes_t ret;
3055 
3056 	if (frames == 0)
3057 		return 0;
3058 
3059 	snd_pcm_stream_lock_irq(substream);
3060 	ret = do_pcm_hwsync(substream);
3061 	if (!ret)
3062 		ret = forward_appl_ptr(substream, frames,
3063 				       snd_pcm_avail(substream));
3064 	snd_pcm_stream_unlock_irq(substream);
3065 	if (ret >= 0)
3066 		snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE);
3067 	return ret;
3068 }
3069 
3070 static int snd_pcm_delay(struct snd_pcm_substream *substream,
3071 			 snd_pcm_sframes_t *delay)
3072 {
3073 	int err;
3074 
3075 	snd_pcm_stream_lock_irq(substream);
3076 	err = do_pcm_hwsync(substream);
3077 	if (delay && !err)
3078 		*delay = snd_pcm_calc_delay(substream);
3079 	snd_pcm_stream_unlock_irq(substream);
3080 	snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_CPU);
3081 
3082 	return err;
3083 }
3084 
3085 static inline int snd_pcm_hwsync(struct snd_pcm_substream *substream)
3086 {
3087 	return snd_pcm_delay(substream, NULL);
3088 }
3089 
3090 static int snd_pcm_sync_ptr(struct snd_pcm_substream *substream,
3091 			    struct snd_pcm_sync_ptr __user *_sync_ptr)
3092 {
3093 	struct snd_pcm_runtime *runtime = substream->runtime;
3094 	struct snd_pcm_sync_ptr sync_ptr;
3095 	volatile struct snd_pcm_mmap_status *status;
3096 	volatile struct snd_pcm_mmap_control *control;
3097 	int err;
3098 
3099 	memset(&sync_ptr, 0, sizeof(sync_ptr));
3100 	if (get_user(sync_ptr.flags, (unsigned __user *)&(_sync_ptr->flags)))
3101 		return -EFAULT;
3102 	if (copy_from_user(&sync_ptr.c.control, &(_sync_ptr->c.control), sizeof(struct snd_pcm_mmap_control)))
3103 		return -EFAULT;
3104 	status = runtime->status;
3105 	control = runtime->control;
3106 	if (sync_ptr.flags & SNDRV_PCM_SYNC_PTR_HWSYNC) {
3107 		err = snd_pcm_hwsync(substream);
3108 		if (err < 0)
3109 			return err;
3110 	}
3111 	snd_pcm_stream_lock_irq(substream);
3112 	if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL)) {
3113 		err = pcm_lib_apply_appl_ptr(substream,
3114 					     sync_ptr.c.control.appl_ptr);
3115 		if (err < 0) {
3116 			snd_pcm_stream_unlock_irq(substream);
3117 			return err;
3118 		}
3119 	} else {
3120 		sync_ptr.c.control.appl_ptr = control->appl_ptr;
3121 	}
3122 	if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN))
3123 		control->avail_min = sync_ptr.c.control.avail_min;
3124 	else
3125 		sync_ptr.c.control.avail_min = control->avail_min;
3126 	sync_ptr.s.status.state = status->state;
3127 	sync_ptr.s.status.hw_ptr = status->hw_ptr;
3128 	sync_ptr.s.status.tstamp = status->tstamp;
3129 	sync_ptr.s.status.suspended_state = status->suspended_state;
3130 	sync_ptr.s.status.audio_tstamp = status->audio_tstamp;
3131 	snd_pcm_stream_unlock_irq(substream);
3132 	if (!(sync_ptr.flags & SNDRV_PCM_SYNC_PTR_APPL))
3133 		snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE);
3134 	if (copy_to_user(_sync_ptr, &sync_ptr, sizeof(sync_ptr)))
3135 		return -EFAULT;
3136 	return 0;
3137 }
3138 
3139 struct snd_pcm_mmap_status32 {
3140 	snd_pcm_state_t state;
3141 	s32 pad1;
3142 	u32 hw_ptr;
3143 	s32 tstamp_sec;
3144 	s32 tstamp_nsec;
3145 	snd_pcm_state_t suspended_state;
3146 	s32 audio_tstamp_sec;
3147 	s32 audio_tstamp_nsec;
3148 } __packed;
3149 
3150 struct snd_pcm_mmap_control32 {
3151 	u32 appl_ptr;
3152 	u32 avail_min;
3153 };
3154 
3155 struct snd_pcm_sync_ptr32 {
3156 	u32 flags;
3157 	union {
3158 		struct snd_pcm_mmap_status32 status;
3159 		unsigned char reserved[64];
3160 	} s;
3161 	union {
3162 		struct snd_pcm_mmap_control32 control;
3163 		unsigned char reserved[64];
3164 	} c;
3165 } __packed;
3166 
3167 /* recalcuate the boundary within 32bit */
3168 static snd_pcm_uframes_t recalculate_boundary(struct snd_pcm_runtime *runtime)
3169 {
3170 	snd_pcm_uframes_t boundary;
3171 
3172 	if (! runtime->buffer_size)
3173 		return 0;
3174 	boundary = runtime->buffer_size;
3175 	while (boundary * 2 <= 0x7fffffffUL - runtime->buffer_size)
3176 		boundary *= 2;
3177 	return boundary;
3178 }
3179 
3180 static int snd_pcm_ioctl_sync_ptr_compat(struct snd_pcm_substream *substream,
3181 					 struct snd_pcm_sync_ptr32 __user *src)
3182 {
3183 	struct snd_pcm_runtime *runtime = substream->runtime;
3184 	volatile struct snd_pcm_mmap_status *status;
3185 	volatile struct snd_pcm_mmap_control *control;
3186 	u32 sflags;
3187 	struct snd_pcm_mmap_control scontrol;
3188 	struct snd_pcm_mmap_status sstatus;
3189 	snd_pcm_uframes_t boundary;
3190 	int err;
3191 
3192 	if (snd_BUG_ON(!runtime))
3193 		return -EINVAL;
3194 
3195 	if (get_user(sflags, &src->flags) ||
3196 	    get_user(scontrol.appl_ptr, &src->c.control.appl_ptr) ||
3197 	    get_user(scontrol.avail_min, &src->c.control.avail_min))
3198 		return -EFAULT;
3199 	if (sflags & SNDRV_PCM_SYNC_PTR_HWSYNC) {
3200 		err = snd_pcm_hwsync(substream);
3201 		if (err < 0)
3202 			return err;
3203 	}
3204 	status = runtime->status;
3205 	control = runtime->control;
3206 	boundary = recalculate_boundary(runtime);
3207 	if (! boundary)
3208 		boundary = 0x7fffffff;
3209 	snd_pcm_stream_lock_irq(substream);
3210 	/* FIXME: we should consider the boundary for the sync from app */
3211 	if (!(sflags & SNDRV_PCM_SYNC_PTR_APPL)) {
3212 		err = pcm_lib_apply_appl_ptr(substream,
3213 				scontrol.appl_ptr);
3214 		if (err < 0) {
3215 			snd_pcm_stream_unlock_irq(substream);
3216 			return err;
3217 		}
3218 	} else
3219 		scontrol.appl_ptr = control->appl_ptr % boundary;
3220 	if (!(sflags & SNDRV_PCM_SYNC_PTR_AVAIL_MIN))
3221 		control->avail_min = scontrol.avail_min;
3222 	else
3223 		scontrol.avail_min = control->avail_min;
3224 	sstatus.state = status->state;
3225 	sstatus.hw_ptr = status->hw_ptr % boundary;
3226 	sstatus.tstamp = status->tstamp;
3227 	sstatus.suspended_state = status->suspended_state;
3228 	sstatus.audio_tstamp = status->audio_tstamp;
3229 	snd_pcm_stream_unlock_irq(substream);
3230 	if (!(sflags & SNDRV_PCM_SYNC_PTR_APPL))
3231 		snd_pcm_dma_buffer_sync(substream, SNDRV_DMA_SYNC_DEVICE);
3232 	if (put_user(sstatus.state, &src->s.status.state) ||
3233 	    put_user(sstatus.hw_ptr, &src->s.status.hw_ptr) ||
3234 	    put_user(sstatus.tstamp.tv_sec, &src->s.status.tstamp_sec) ||
3235 	    put_user(sstatus.tstamp.tv_nsec, &src->s.status.tstamp_nsec) ||
3236 	    put_user(sstatus.suspended_state, &src->s.status.suspended_state) ||
3237 	    put_user(sstatus.audio_tstamp.tv_sec, &src->s.status.audio_tstamp_sec) ||
3238 	    put_user(sstatus.audio_tstamp.tv_nsec, &src->s.status.audio_tstamp_nsec) ||
3239 	    put_user(scontrol.appl_ptr, &src->c.control.appl_ptr) ||
3240 	    put_user(scontrol.avail_min, &src->c.control.avail_min))
3241 		return -EFAULT;
3242 
3243 	return 0;
3244 }
3245 #define __SNDRV_PCM_IOCTL_SYNC_PTR32 _IOWR('A', 0x23, struct snd_pcm_sync_ptr32)
3246 
3247 static int snd_pcm_tstamp(struct snd_pcm_substream *substream, int __user *_arg)
3248 {
3249 	struct snd_pcm_runtime *runtime = substream->runtime;
3250 	int arg;
3251 
3252 	if (get_user(arg, _arg))
3253 		return -EFAULT;
3254 	if (arg < 0 || arg > SNDRV_PCM_TSTAMP_TYPE_LAST)
3255 		return -EINVAL;
3256 	runtime->tstamp_type = arg;
3257 	return 0;
3258 }
3259 
3260 static int snd_pcm_xferi_frames_ioctl(struct snd_pcm_substream *substream,
3261 				      struct snd_xferi __user *_xferi)
3262 {
3263 	struct snd_xferi xferi;
3264 	struct snd_pcm_runtime *runtime = substream->runtime;
3265 	snd_pcm_sframes_t result;
3266 
3267 	if (runtime->state == SNDRV_PCM_STATE_OPEN)
3268 		return -EBADFD;
3269 	if (put_user(0, &_xferi->result))
3270 		return -EFAULT;
3271 	if (copy_from_user(&xferi, _xferi, sizeof(xferi)))
3272 		return -EFAULT;
3273 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
3274 		result = snd_pcm_lib_write(substream, xferi.buf, xferi.frames);
3275 	else
3276 		result = snd_pcm_lib_read(substream, xferi.buf, xferi.frames);
3277 	if (put_user(result, &_xferi->result))
3278 		return -EFAULT;
3279 	return result < 0 ? result : 0;
3280 }
3281 
3282 static int snd_pcm_xfern_frames_ioctl(struct snd_pcm_substream *substream,
3283 				      struct snd_xfern __user *_xfern)
3284 {
3285 	struct snd_xfern xfern;
3286 	struct snd_pcm_runtime *runtime = substream->runtime;
3287 	void *bufs;
3288 	snd_pcm_sframes_t result;
3289 
3290 	if (runtime->state == SNDRV_PCM_STATE_OPEN)
3291 		return -EBADFD;
3292 	if (runtime->channels > 128)
3293 		return -EINVAL;
3294 	if (put_user(0, &_xfern->result))
3295 		return -EFAULT;
3296 	if (copy_from_user(&xfern, _xfern, sizeof(xfern)))
3297 		return -EFAULT;
3298 
3299 	bufs = memdup_user(xfern.bufs, sizeof(void *) * runtime->channels);
3300 	if (IS_ERR(bufs))
3301 		return PTR_ERR(bufs);
3302 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
3303 		result = snd_pcm_lib_writev(substream, bufs, xfern.frames);
3304 	else
3305 		result = snd_pcm_lib_readv(substream, bufs, xfern.frames);
3306 	kfree(bufs);
3307 	if (put_user(result, &_xfern->result))
3308 		return -EFAULT;
3309 	return result < 0 ? result : 0;
3310 }
3311 
3312 static int snd_pcm_rewind_ioctl(struct snd_pcm_substream *substream,
3313 				snd_pcm_uframes_t __user *_frames)
3314 {
3315 	snd_pcm_uframes_t frames;
3316 	snd_pcm_sframes_t result;
3317 
3318 	if (get_user(frames, _frames))
3319 		return -EFAULT;
3320 	if (put_user(0, _frames))
3321 		return -EFAULT;
3322 	result = snd_pcm_rewind(substream, frames);
3323 	if (put_user(result, _frames))
3324 		return -EFAULT;
3325 	return result < 0 ? result : 0;
3326 }
3327 
3328 static int snd_pcm_forward_ioctl(struct snd_pcm_substream *substream,
3329 				 snd_pcm_uframes_t __user *_frames)
3330 {
3331 	snd_pcm_uframes_t frames;
3332 	snd_pcm_sframes_t result;
3333 
3334 	if (get_user(frames, _frames))
3335 		return -EFAULT;
3336 	if (put_user(0, _frames))
3337 		return -EFAULT;
3338 	result = snd_pcm_forward(substream, frames);
3339 	if (put_user(result, _frames))
3340 		return -EFAULT;
3341 	return result < 0 ? result : 0;
3342 }
3343 
3344 static int snd_pcm_common_ioctl(struct file *file,
3345 				 struct snd_pcm_substream *substream,
3346 				 unsigned int cmd, void __user *arg)
3347 {
3348 	struct snd_pcm_file *pcm_file = file->private_data;
3349 	int res;
3350 
3351 	if (PCM_RUNTIME_CHECK(substream))
3352 		return -ENXIO;
3353 
3354 	if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3355 		return -EBADFD;
3356 
3357 	res = snd_power_wait(substream->pcm->card);
3358 	if (res < 0)
3359 		return res;
3360 
3361 	switch (cmd) {
3362 	case SNDRV_PCM_IOCTL_PVERSION:
3363 		return put_user(SNDRV_PCM_VERSION, (int __user *)arg) ? -EFAULT : 0;
3364 	case SNDRV_PCM_IOCTL_INFO:
3365 		return snd_pcm_info_user(substream, arg);
3366 	case SNDRV_PCM_IOCTL_TSTAMP:	/* just for compatibility */
3367 		return 0;
3368 	case SNDRV_PCM_IOCTL_TTSTAMP:
3369 		return snd_pcm_tstamp(substream, arg);
3370 	case SNDRV_PCM_IOCTL_USER_PVERSION:
3371 		if (get_user(pcm_file->user_pversion,
3372 			     (unsigned int __user *)arg))
3373 			return -EFAULT;
3374 		return 0;
3375 	case SNDRV_PCM_IOCTL_HW_REFINE:
3376 		return snd_pcm_hw_refine_user(substream, arg);
3377 	case SNDRV_PCM_IOCTL_HW_PARAMS:
3378 		return snd_pcm_hw_params_user(substream, arg);
3379 	case SNDRV_PCM_IOCTL_HW_FREE:
3380 		return snd_pcm_hw_free(substream);
3381 	case SNDRV_PCM_IOCTL_SW_PARAMS:
3382 		return snd_pcm_sw_params_user(substream, arg);
3383 	case SNDRV_PCM_IOCTL_STATUS32:
3384 		return snd_pcm_status_user32(substream, arg, false);
3385 	case SNDRV_PCM_IOCTL_STATUS_EXT32:
3386 		return snd_pcm_status_user32(substream, arg, true);
3387 	case SNDRV_PCM_IOCTL_STATUS64:
3388 		return snd_pcm_status_user64(substream, arg, false);
3389 	case SNDRV_PCM_IOCTL_STATUS_EXT64:
3390 		return snd_pcm_status_user64(substream, arg, true);
3391 	case SNDRV_PCM_IOCTL_CHANNEL_INFO:
3392 		return snd_pcm_channel_info_user(substream, arg);
3393 	case SNDRV_PCM_IOCTL_PREPARE:
3394 		return snd_pcm_prepare(substream, file);
3395 	case SNDRV_PCM_IOCTL_RESET:
3396 		return snd_pcm_reset(substream);
3397 	case SNDRV_PCM_IOCTL_START:
3398 		return snd_pcm_start_lock_irq(substream);
3399 	case SNDRV_PCM_IOCTL_LINK:
3400 		return snd_pcm_link(substream, (int)(unsigned long) arg);
3401 	case SNDRV_PCM_IOCTL_UNLINK:
3402 		return snd_pcm_unlink(substream);
3403 	case SNDRV_PCM_IOCTL_RESUME:
3404 		return snd_pcm_resume(substream);
3405 	case SNDRV_PCM_IOCTL_XRUN:
3406 		return snd_pcm_xrun(substream);
3407 	case SNDRV_PCM_IOCTL_HWSYNC:
3408 		return snd_pcm_hwsync(substream);
3409 	case SNDRV_PCM_IOCTL_DELAY:
3410 	{
3411 		snd_pcm_sframes_t delay = 0;
3412 		snd_pcm_sframes_t __user *res = arg;
3413 		int err;
3414 
3415 		err = snd_pcm_delay(substream, &delay);
3416 		if (err)
3417 			return err;
3418 		if (put_user(delay, res))
3419 			return -EFAULT;
3420 		return 0;
3421 	}
3422 	case __SNDRV_PCM_IOCTL_SYNC_PTR32:
3423 		return snd_pcm_ioctl_sync_ptr_compat(substream, arg);
3424 	case __SNDRV_PCM_IOCTL_SYNC_PTR64:
3425 		return snd_pcm_sync_ptr(substream, arg);
3426 #ifdef CONFIG_SND_SUPPORT_OLD_API
3427 	case SNDRV_PCM_IOCTL_HW_REFINE_OLD:
3428 		return snd_pcm_hw_refine_old_user(substream, arg);
3429 	case SNDRV_PCM_IOCTL_HW_PARAMS_OLD:
3430 		return snd_pcm_hw_params_old_user(substream, arg);
3431 #endif
3432 	case SNDRV_PCM_IOCTL_DRAIN:
3433 		return snd_pcm_drain(substream, file);
3434 	case SNDRV_PCM_IOCTL_DROP:
3435 		return snd_pcm_drop(substream);
3436 	case SNDRV_PCM_IOCTL_PAUSE:
3437 		return snd_pcm_pause_lock_irq(substream, (unsigned long)arg);
3438 	case SNDRV_PCM_IOCTL_WRITEI_FRAMES:
3439 	case SNDRV_PCM_IOCTL_READI_FRAMES:
3440 		return snd_pcm_xferi_frames_ioctl(substream, arg);
3441 	case SNDRV_PCM_IOCTL_WRITEN_FRAMES:
3442 	case SNDRV_PCM_IOCTL_READN_FRAMES:
3443 		return snd_pcm_xfern_frames_ioctl(substream, arg);
3444 	case SNDRV_PCM_IOCTL_REWIND:
3445 		return snd_pcm_rewind_ioctl(substream, arg);
3446 	case SNDRV_PCM_IOCTL_FORWARD:
3447 		return snd_pcm_forward_ioctl(substream, arg);
3448 	}
3449 	pcm_dbg(substream->pcm, "unknown ioctl = 0x%x\n", cmd);
3450 	return -ENOTTY;
3451 }
3452 
3453 static long snd_pcm_ioctl(struct file *file, unsigned int cmd,
3454 			  unsigned long arg)
3455 {
3456 	struct snd_pcm_file *pcm_file;
3457 
3458 	pcm_file = file->private_data;
3459 
3460 	if (((cmd >> 8) & 0xff) != 'A')
3461 		return -ENOTTY;
3462 
3463 	return snd_pcm_common_ioctl(file, pcm_file->substream, cmd,
3464 				     (void __user *)arg);
3465 }
3466 
3467 /**
3468  * snd_pcm_kernel_ioctl - Execute PCM ioctl in the kernel-space
3469  * @substream: PCM substream
3470  * @cmd: IOCTL cmd
3471  * @arg: IOCTL argument
3472  *
3473  * The function is provided primarily for OSS layer and USB gadget drivers,
3474  * and it allows only the limited set of ioctls (hw_params, sw_params,
3475  * prepare, start, drain, drop, forward).
3476  *
3477  * Return: zero if successful, or a negative error code
3478  */
3479 int snd_pcm_kernel_ioctl(struct snd_pcm_substream *substream,
3480 			 unsigned int cmd, void *arg)
3481 {
3482 	snd_pcm_uframes_t *frames = arg;
3483 	snd_pcm_sframes_t result;
3484 
3485 	if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3486 		return -EBADFD;
3487 
3488 	switch (cmd) {
3489 	case SNDRV_PCM_IOCTL_FORWARD:
3490 	{
3491 		/* provided only for OSS; capture-only and no value returned */
3492 		if (substream->stream != SNDRV_PCM_STREAM_CAPTURE)
3493 			return -EINVAL;
3494 		result = snd_pcm_forward(substream, *frames);
3495 		return result < 0 ? result : 0;
3496 	}
3497 	case SNDRV_PCM_IOCTL_HW_PARAMS:
3498 		return snd_pcm_hw_params(substream, arg);
3499 	case SNDRV_PCM_IOCTL_SW_PARAMS:
3500 		return snd_pcm_sw_params(substream, arg);
3501 	case SNDRV_PCM_IOCTL_PREPARE:
3502 		return snd_pcm_prepare(substream, NULL);
3503 	case SNDRV_PCM_IOCTL_START:
3504 		return snd_pcm_start_lock_irq(substream);
3505 	case SNDRV_PCM_IOCTL_DRAIN:
3506 		return snd_pcm_drain(substream, NULL);
3507 	case SNDRV_PCM_IOCTL_DROP:
3508 		return snd_pcm_drop(substream);
3509 	case SNDRV_PCM_IOCTL_DELAY:
3510 		return snd_pcm_delay(substream, frames);
3511 	default:
3512 		return -EINVAL;
3513 	}
3514 }
3515 EXPORT_SYMBOL(snd_pcm_kernel_ioctl);
3516 
3517 static ssize_t snd_pcm_read(struct file *file, char __user *buf, size_t count,
3518 			    loff_t * offset)
3519 {
3520 	struct snd_pcm_file *pcm_file;
3521 	struct snd_pcm_substream *substream;
3522 	struct snd_pcm_runtime *runtime;
3523 	snd_pcm_sframes_t result;
3524 
3525 	pcm_file = file->private_data;
3526 	substream = pcm_file->substream;
3527 	if (PCM_RUNTIME_CHECK(substream))
3528 		return -ENXIO;
3529 	runtime = substream->runtime;
3530 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
3531 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3532 		return -EBADFD;
3533 	if (!frame_aligned(runtime, count))
3534 		return -EINVAL;
3535 	count = bytes_to_frames(runtime, count);
3536 	result = snd_pcm_lib_read(substream, buf, count);
3537 	if (result > 0)
3538 		result = frames_to_bytes(runtime, result);
3539 	return result;
3540 }
3541 
3542 static ssize_t snd_pcm_write(struct file *file, const char __user *buf,
3543 			     size_t count, loff_t * offset)
3544 {
3545 	struct snd_pcm_file *pcm_file;
3546 	struct snd_pcm_substream *substream;
3547 	struct snd_pcm_runtime *runtime;
3548 	snd_pcm_sframes_t result;
3549 
3550 	pcm_file = file->private_data;
3551 	substream = pcm_file->substream;
3552 	if (PCM_RUNTIME_CHECK(substream))
3553 		return -ENXIO;
3554 	runtime = substream->runtime;
3555 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
3556 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3557 		return -EBADFD;
3558 	if (!frame_aligned(runtime, count))
3559 		return -EINVAL;
3560 	count = bytes_to_frames(runtime, count);
3561 	result = snd_pcm_lib_write(substream, buf, count);
3562 	if (result > 0)
3563 		result = frames_to_bytes(runtime, result);
3564 	return result;
3565 }
3566 
3567 static ssize_t snd_pcm_readv(struct kiocb *iocb, struct iov_iter *to)
3568 {
3569 	struct snd_pcm_file *pcm_file;
3570 	struct snd_pcm_substream *substream;
3571 	struct snd_pcm_runtime *runtime;
3572 	snd_pcm_sframes_t result;
3573 	unsigned long i;
3574 	void __user **bufs;
3575 	snd_pcm_uframes_t frames;
3576 	const struct iovec *iov = iter_iov(to);
3577 
3578 	pcm_file = iocb->ki_filp->private_data;
3579 	substream = pcm_file->substream;
3580 	if (PCM_RUNTIME_CHECK(substream))
3581 		return -ENXIO;
3582 	runtime = substream->runtime;
3583 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
3584 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3585 		return -EBADFD;
3586 	if (!user_backed_iter(to))
3587 		return -EINVAL;
3588 	if (to->nr_segs > 1024 || to->nr_segs != runtime->channels)
3589 		return -EINVAL;
3590 	if (!frame_aligned(runtime, iov->iov_len))
3591 		return -EINVAL;
3592 	frames = bytes_to_samples(runtime, iov->iov_len);
3593 	bufs = kmalloc_array(to->nr_segs, sizeof(void *), GFP_KERNEL);
3594 	if (bufs == NULL)
3595 		return -ENOMEM;
3596 	for (i = 0; i < to->nr_segs; ++i) {
3597 		bufs[i] = iov->iov_base;
3598 		iov++;
3599 	}
3600 	result = snd_pcm_lib_readv(substream, bufs, frames);
3601 	if (result > 0)
3602 		result = frames_to_bytes(runtime, result);
3603 	kfree(bufs);
3604 	return result;
3605 }
3606 
3607 static ssize_t snd_pcm_writev(struct kiocb *iocb, struct iov_iter *from)
3608 {
3609 	struct snd_pcm_file *pcm_file;
3610 	struct snd_pcm_substream *substream;
3611 	struct snd_pcm_runtime *runtime;
3612 	snd_pcm_sframes_t result;
3613 	unsigned long i;
3614 	void __user **bufs;
3615 	snd_pcm_uframes_t frames;
3616 	const struct iovec *iov = iter_iov(from);
3617 
3618 	pcm_file = iocb->ki_filp->private_data;
3619 	substream = pcm_file->substream;
3620 	if (PCM_RUNTIME_CHECK(substream))
3621 		return -ENXIO;
3622 	runtime = substream->runtime;
3623 	if (runtime->state == SNDRV_PCM_STATE_OPEN ||
3624 	    runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3625 		return -EBADFD;
3626 	if (!user_backed_iter(from))
3627 		return -EINVAL;
3628 	if (from->nr_segs > 128 || from->nr_segs != runtime->channels ||
3629 	    !frame_aligned(runtime, iov->iov_len))
3630 		return -EINVAL;
3631 	frames = bytes_to_samples(runtime, iov->iov_len);
3632 	bufs = kmalloc_array(from->nr_segs, sizeof(void *), GFP_KERNEL);
3633 	if (bufs == NULL)
3634 		return -ENOMEM;
3635 	for (i = 0; i < from->nr_segs; ++i) {
3636 		bufs[i] = iov->iov_base;
3637 		iov++;
3638 	}
3639 	result = snd_pcm_lib_writev(substream, bufs, frames);
3640 	if (result > 0)
3641 		result = frames_to_bytes(runtime, result);
3642 	kfree(bufs);
3643 	return result;
3644 }
3645 
3646 static __poll_t snd_pcm_poll(struct file *file, poll_table *wait)
3647 {
3648 	struct snd_pcm_file *pcm_file;
3649 	struct snd_pcm_substream *substream;
3650 	struct snd_pcm_runtime *runtime;
3651 	__poll_t mask, ok;
3652 	snd_pcm_uframes_t avail;
3653 
3654 	pcm_file = file->private_data;
3655 
3656 	substream = pcm_file->substream;
3657 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK)
3658 		ok = EPOLLOUT | EPOLLWRNORM;
3659 	else
3660 		ok = EPOLLIN | EPOLLRDNORM;
3661 	if (PCM_RUNTIME_CHECK(substream))
3662 		return ok | EPOLLERR;
3663 
3664 	runtime = substream->runtime;
3665 	if (runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3666 		return ok | EPOLLERR;
3667 
3668 	poll_wait(file, &runtime->sleep, wait);
3669 
3670 	mask = 0;
3671 	snd_pcm_stream_lock_irq(substream);
3672 	avail = snd_pcm_avail(substream);
3673 	switch (runtime->state) {
3674 	case SNDRV_PCM_STATE_RUNNING:
3675 	case SNDRV_PCM_STATE_PREPARED:
3676 	case SNDRV_PCM_STATE_PAUSED:
3677 		if (avail >= runtime->control->avail_min)
3678 			mask = ok;
3679 		break;
3680 	case SNDRV_PCM_STATE_DRAINING:
3681 		if (substream->stream == SNDRV_PCM_STREAM_CAPTURE) {
3682 			mask = ok;
3683 			if (!avail)
3684 				mask |= EPOLLERR;
3685 		}
3686 		break;
3687 	default:
3688 		mask = ok | EPOLLERR;
3689 		break;
3690 	}
3691 	snd_pcm_stream_unlock_irq(substream);
3692 	return mask;
3693 }
3694 
3695 /*
3696  * mmap support
3697  */
3698 
3699 /*
3700  * Only on coherent architectures, we can mmap the status and the control records
3701  * for effcient data transfer.  On others, we have to use HWSYNC ioctl...
3702  */
3703 #if defined(CONFIG_X86) || defined(CONFIG_PPC) || defined(CONFIG_ALPHA)
3704 /*
3705  * mmap status record
3706  */
3707 static vm_fault_t snd_pcm_mmap_status_fault(struct vm_fault *vmf)
3708 {
3709 	struct snd_pcm_substream *substream = vmf->vma->vm_private_data;
3710 	struct snd_pcm_runtime *runtime;
3711 
3712 	if (substream == NULL)
3713 		return VM_FAULT_SIGBUS;
3714 	runtime = substream->runtime;
3715 	vmf->page = virt_to_page(runtime->status);
3716 	get_page(vmf->page);
3717 	return 0;
3718 }
3719 
3720 static const struct vm_operations_struct snd_pcm_vm_ops_status =
3721 {
3722 	.fault =	snd_pcm_mmap_status_fault,
3723 };
3724 
3725 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file,
3726 			       struct vm_area_struct *area)
3727 {
3728 	long size;
3729 	if (!(area->vm_flags & VM_READ))
3730 		return -EINVAL;
3731 	size = area->vm_end - area->vm_start;
3732 	if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_status)))
3733 		return -EINVAL;
3734 	area->vm_ops = &snd_pcm_vm_ops_status;
3735 	area->vm_private_data = substream;
3736 	vm_flags_mod(area, VM_DONTEXPAND | VM_DONTDUMP,
3737 		     VM_WRITE | VM_MAYWRITE);
3738 
3739 	return 0;
3740 }
3741 
3742 /*
3743  * mmap control record
3744  */
3745 static vm_fault_t snd_pcm_mmap_control_fault(struct vm_fault *vmf)
3746 {
3747 	struct snd_pcm_substream *substream = vmf->vma->vm_private_data;
3748 	struct snd_pcm_runtime *runtime;
3749 
3750 	if (substream == NULL)
3751 		return VM_FAULT_SIGBUS;
3752 	runtime = substream->runtime;
3753 	vmf->page = virt_to_page(runtime->control);
3754 	get_page(vmf->page);
3755 	return 0;
3756 }
3757 
3758 static const struct vm_operations_struct snd_pcm_vm_ops_control =
3759 {
3760 	.fault =	snd_pcm_mmap_control_fault,
3761 };
3762 
3763 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file,
3764 				struct vm_area_struct *area)
3765 {
3766 	long size;
3767 	if (!(area->vm_flags & VM_READ))
3768 		return -EINVAL;
3769 	size = area->vm_end - area->vm_start;
3770 	if (size != PAGE_ALIGN(sizeof(struct snd_pcm_mmap_control)))
3771 		return -EINVAL;
3772 	area->vm_ops = &snd_pcm_vm_ops_control;
3773 	area->vm_private_data = substream;
3774 	vm_flags_set(area, VM_DONTEXPAND | VM_DONTDUMP);
3775 	return 0;
3776 }
3777 
3778 static bool pcm_status_mmap_allowed(struct snd_pcm_file *pcm_file)
3779 {
3780 	/* If drivers require the explicit sync (typically for non-coherent
3781 	 * pages), we have to disable the mmap of status and control data
3782 	 * to enforce the control via SYNC_PTR ioctl.
3783 	 */
3784 	if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_EXPLICIT_SYNC)
3785 		return false;
3786 	/* See pcm_control_mmap_allowed() below.
3787 	 * Since older alsa-lib requires both status and control mmaps to be
3788 	 * coupled, we have to disable the status mmap for old alsa-lib, too.
3789 	 */
3790 	if (pcm_file->user_pversion < SNDRV_PROTOCOL_VERSION(2, 0, 14) &&
3791 	    (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR))
3792 		return false;
3793 	return true;
3794 }
3795 
3796 static bool pcm_control_mmap_allowed(struct snd_pcm_file *pcm_file)
3797 {
3798 	if (pcm_file->no_compat_mmap)
3799 		return false;
3800 	/* see above */
3801 	if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_EXPLICIT_SYNC)
3802 		return false;
3803 	/* Disallow the control mmap when SYNC_APPLPTR flag is set;
3804 	 * it enforces the user-space to fall back to snd_pcm_sync_ptr(),
3805 	 * thus it effectively assures the manual update of appl_ptr.
3806 	 */
3807 	if (pcm_file->substream->runtime->hw.info & SNDRV_PCM_INFO_SYNC_APPLPTR)
3808 		return false;
3809 	return true;
3810 }
3811 
3812 #else /* ! coherent mmap */
3813 /*
3814  * don't support mmap for status and control records.
3815  */
3816 #define pcm_status_mmap_allowed(pcm_file)	false
3817 #define pcm_control_mmap_allowed(pcm_file)	false
3818 
3819 static int snd_pcm_mmap_status(struct snd_pcm_substream *substream, struct file *file,
3820 			       struct vm_area_struct *area)
3821 {
3822 	return -ENXIO;
3823 }
3824 static int snd_pcm_mmap_control(struct snd_pcm_substream *substream, struct file *file,
3825 				struct vm_area_struct *area)
3826 {
3827 	return -ENXIO;
3828 }
3829 #endif /* coherent mmap */
3830 
3831 /*
3832  * fault callback for mmapping a RAM page
3833  */
3834 static vm_fault_t snd_pcm_mmap_data_fault(struct vm_fault *vmf)
3835 {
3836 	struct snd_pcm_substream *substream = vmf->vma->vm_private_data;
3837 	struct snd_pcm_runtime *runtime;
3838 	unsigned long offset;
3839 	struct page * page;
3840 	size_t dma_bytes;
3841 
3842 	if (substream == NULL)
3843 		return VM_FAULT_SIGBUS;
3844 	runtime = substream->runtime;
3845 	offset = vmf->pgoff << PAGE_SHIFT;
3846 	dma_bytes = PAGE_ALIGN(runtime->dma_bytes);
3847 	if (offset > dma_bytes - PAGE_SIZE)
3848 		return VM_FAULT_SIGBUS;
3849 	if (substream->ops->page)
3850 		page = substream->ops->page(substream, offset);
3851 	else if (!snd_pcm_get_dma_buf(substream))
3852 		page = virt_to_page(runtime->dma_area + offset);
3853 	else
3854 		page = snd_sgbuf_get_page(snd_pcm_get_dma_buf(substream), offset);
3855 	if (!page)
3856 		return VM_FAULT_SIGBUS;
3857 	get_page(page);
3858 	vmf->page = page;
3859 	return 0;
3860 }
3861 
3862 static const struct vm_operations_struct snd_pcm_vm_ops_data = {
3863 	.open =		snd_pcm_mmap_data_open,
3864 	.close =	snd_pcm_mmap_data_close,
3865 };
3866 
3867 static const struct vm_operations_struct snd_pcm_vm_ops_data_fault = {
3868 	.open =		snd_pcm_mmap_data_open,
3869 	.close =	snd_pcm_mmap_data_close,
3870 	.fault =	snd_pcm_mmap_data_fault,
3871 };
3872 
3873 /*
3874  * mmap the DMA buffer on RAM
3875  */
3876 
3877 /**
3878  * snd_pcm_lib_default_mmap - Default PCM data mmap function
3879  * @substream: PCM substream
3880  * @area: VMA
3881  *
3882  * This is the default mmap handler for PCM data.  When mmap pcm_ops is NULL,
3883  * this function is invoked implicitly.
3884  *
3885  * Return: zero if successful, or a negative error code
3886  */
3887 int snd_pcm_lib_default_mmap(struct snd_pcm_substream *substream,
3888 			     struct vm_area_struct *area)
3889 {
3890 	vm_flags_set(area, VM_DONTEXPAND | VM_DONTDUMP);
3891 	if (!substream->ops->page &&
3892 	    !snd_dma_buffer_mmap(snd_pcm_get_dma_buf(substream), area))
3893 		return 0;
3894 	/* mmap with fault handler */
3895 	area->vm_ops = &snd_pcm_vm_ops_data_fault;
3896 	return 0;
3897 }
3898 EXPORT_SYMBOL_GPL(snd_pcm_lib_default_mmap);
3899 
3900 /*
3901  * mmap the DMA buffer on I/O memory area
3902  */
3903 #if SNDRV_PCM_INFO_MMAP_IOMEM
3904 /**
3905  * snd_pcm_lib_mmap_iomem - Default PCM data mmap function for I/O mem
3906  * @substream: PCM substream
3907  * @area: VMA
3908  *
3909  * When your hardware uses the iomapped pages as the hardware buffer and
3910  * wants to mmap it, pass this function as mmap pcm_ops.  Note that this
3911  * is supposed to work only on limited architectures.
3912  *
3913  * Return: zero if successful, or a negative error code
3914  */
3915 int snd_pcm_lib_mmap_iomem(struct snd_pcm_substream *substream,
3916 			   struct vm_area_struct *area)
3917 {
3918 	struct snd_pcm_runtime *runtime = substream->runtime;
3919 
3920 	area->vm_page_prot = pgprot_noncached(area->vm_page_prot);
3921 	return vm_iomap_memory(area, runtime->dma_addr, runtime->dma_bytes);
3922 }
3923 EXPORT_SYMBOL(snd_pcm_lib_mmap_iomem);
3924 #endif /* SNDRV_PCM_INFO_MMAP */
3925 
3926 /*
3927  * mmap DMA buffer
3928  */
3929 int snd_pcm_mmap_data(struct snd_pcm_substream *substream, struct file *file,
3930 		      struct vm_area_struct *area)
3931 {
3932 	struct snd_pcm_runtime *runtime;
3933 	long size;
3934 	unsigned long offset;
3935 	size_t dma_bytes;
3936 	int err;
3937 
3938 	if (substream->stream == SNDRV_PCM_STREAM_PLAYBACK) {
3939 		if (!(area->vm_flags & (VM_WRITE|VM_READ)))
3940 			return -EINVAL;
3941 	} else {
3942 		if (!(area->vm_flags & VM_READ))
3943 			return -EINVAL;
3944 	}
3945 	runtime = substream->runtime;
3946 	if (runtime->state == SNDRV_PCM_STATE_OPEN)
3947 		return -EBADFD;
3948 	if (!(runtime->info & SNDRV_PCM_INFO_MMAP))
3949 		return -ENXIO;
3950 	if (runtime->access == SNDRV_PCM_ACCESS_RW_INTERLEAVED ||
3951 	    runtime->access == SNDRV_PCM_ACCESS_RW_NONINTERLEAVED)
3952 		return -EINVAL;
3953 	size = area->vm_end - area->vm_start;
3954 	offset = area->vm_pgoff << PAGE_SHIFT;
3955 	dma_bytes = PAGE_ALIGN(runtime->dma_bytes);
3956 	if ((size_t)size > dma_bytes)
3957 		return -EINVAL;
3958 	if (offset > dma_bytes - size)
3959 		return -EINVAL;
3960 
3961 	area->vm_ops = &snd_pcm_vm_ops_data;
3962 	area->vm_private_data = substream;
3963 	if (substream->ops->mmap)
3964 		err = substream->ops->mmap(substream, area);
3965 	else
3966 		err = snd_pcm_lib_default_mmap(substream, area);
3967 	if (!err)
3968 		atomic_inc(&substream->mmap_count);
3969 	return err;
3970 }
3971 EXPORT_SYMBOL(snd_pcm_mmap_data);
3972 
3973 static int snd_pcm_mmap(struct file *file, struct vm_area_struct *area)
3974 {
3975 	struct snd_pcm_file * pcm_file;
3976 	struct snd_pcm_substream *substream;
3977 	unsigned long offset;
3978 
3979 	pcm_file = file->private_data;
3980 	substream = pcm_file->substream;
3981 	if (PCM_RUNTIME_CHECK(substream))
3982 		return -ENXIO;
3983 	if (substream->runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
3984 		return -EBADFD;
3985 
3986 	offset = area->vm_pgoff << PAGE_SHIFT;
3987 	switch (offset) {
3988 	case SNDRV_PCM_MMAP_OFFSET_STATUS_OLD:
3989 		if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT))
3990 			return -ENXIO;
3991 		fallthrough;
3992 	case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW:
3993 		if (!pcm_status_mmap_allowed(pcm_file))
3994 			return -ENXIO;
3995 		return snd_pcm_mmap_status(substream, file, area);
3996 	case SNDRV_PCM_MMAP_OFFSET_CONTROL_OLD:
3997 		if (pcm_file->no_compat_mmap || !IS_ENABLED(CONFIG_64BIT))
3998 			return -ENXIO;
3999 		fallthrough;
4000 	case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW:
4001 		if (!pcm_control_mmap_allowed(pcm_file))
4002 			return -ENXIO;
4003 		return snd_pcm_mmap_control(substream, file, area);
4004 	default:
4005 		return snd_pcm_mmap_data(substream, file, area);
4006 	}
4007 	return 0;
4008 }
4009 
4010 static int snd_pcm_fasync(int fd, struct file * file, int on)
4011 {
4012 	struct snd_pcm_file * pcm_file;
4013 	struct snd_pcm_substream *substream;
4014 	struct snd_pcm_runtime *runtime;
4015 
4016 	pcm_file = file->private_data;
4017 	substream = pcm_file->substream;
4018 	if (PCM_RUNTIME_CHECK(substream))
4019 		return -ENXIO;
4020 	runtime = substream->runtime;
4021 	if (runtime->state == SNDRV_PCM_STATE_DISCONNECTED)
4022 		return -EBADFD;
4023 	return snd_fasync_helper(fd, file, on, &runtime->fasync);
4024 }
4025 
4026 /*
4027  * ioctl32 compat
4028  */
4029 #ifdef CONFIG_COMPAT
4030 #include "pcm_compat.c"
4031 #else
4032 #define snd_pcm_ioctl_compat	NULL
4033 #endif
4034 
4035 /*
4036  *  To be removed helpers to keep binary compatibility
4037  */
4038 
4039 #ifdef CONFIG_SND_SUPPORT_OLD_API
4040 #define __OLD_TO_NEW_MASK(x) ((x&7)|((x&0x07fffff8)<<5))
4041 #define __NEW_TO_OLD_MASK(x) ((x&7)|((x&0xffffff00)>>5))
4042 
4043 static void snd_pcm_hw_convert_from_old_params(struct snd_pcm_hw_params *params,
4044 					       struct snd_pcm_hw_params_old *oparams)
4045 {
4046 	unsigned int i;
4047 
4048 	memset(params, 0, sizeof(*params));
4049 	params->flags = oparams->flags;
4050 	for (i = 0; i < ARRAY_SIZE(oparams->masks); i++)
4051 		params->masks[i].bits[0] = oparams->masks[i];
4052 	memcpy(params->intervals, oparams->intervals, sizeof(oparams->intervals));
4053 	params->rmask = __OLD_TO_NEW_MASK(oparams->rmask);
4054 	params->cmask = __OLD_TO_NEW_MASK(oparams->cmask);
4055 	params->info = oparams->info;
4056 	params->msbits = oparams->msbits;
4057 	params->rate_num = oparams->rate_num;
4058 	params->rate_den = oparams->rate_den;
4059 	params->fifo_size = oparams->fifo_size;
4060 }
4061 
4062 static void snd_pcm_hw_convert_to_old_params(struct snd_pcm_hw_params_old *oparams,
4063 					     struct snd_pcm_hw_params *params)
4064 {
4065 	unsigned int i;
4066 
4067 	memset(oparams, 0, sizeof(*oparams));
4068 	oparams->flags = params->flags;
4069 	for (i = 0; i < ARRAY_SIZE(oparams->masks); i++)
4070 		oparams->masks[i] = params->masks[i].bits[0];
4071 	memcpy(oparams->intervals, params->intervals, sizeof(oparams->intervals));
4072 	oparams->rmask = __NEW_TO_OLD_MASK(params->rmask);
4073 	oparams->cmask = __NEW_TO_OLD_MASK(params->cmask);
4074 	oparams->info = params->info;
4075 	oparams->msbits = params->msbits;
4076 	oparams->rate_num = params->rate_num;
4077 	oparams->rate_den = params->rate_den;
4078 	oparams->fifo_size = params->fifo_size;
4079 }
4080 
4081 static int snd_pcm_hw_refine_old_user(struct snd_pcm_substream *substream,
4082 				      struct snd_pcm_hw_params_old __user * _oparams)
4083 {
4084 	struct snd_pcm_hw_params *params;
4085 	struct snd_pcm_hw_params_old *oparams = NULL;
4086 	int err;
4087 
4088 	params = kmalloc(sizeof(*params), GFP_KERNEL);
4089 	if (!params)
4090 		return -ENOMEM;
4091 
4092 	oparams = memdup_user(_oparams, sizeof(*oparams));
4093 	if (IS_ERR(oparams)) {
4094 		err = PTR_ERR(oparams);
4095 		goto out;
4096 	}
4097 	snd_pcm_hw_convert_from_old_params(params, oparams);
4098 	err = snd_pcm_hw_refine(substream, params);
4099 	if (err < 0)
4100 		goto out_old;
4101 
4102 	err = fixup_unreferenced_params(substream, params);
4103 	if (err < 0)
4104 		goto out_old;
4105 
4106 	snd_pcm_hw_convert_to_old_params(oparams, params);
4107 	if (copy_to_user(_oparams, oparams, sizeof(*oparams)))
4108 		err = -EFAULT;
4109 out_old:
4110 	kfree(oparams);
4111 out:
4112 	kfree(params);
4113 	return err;
4114 }
4115 
4116 static int snd_pcm_hw_params_old_user(struct snd_pcm_substream *substream,
4117 				      struct snd_pcm_hw_params_old __user * _oparams)
4118 {
4119 	struct snd_pcm_hw_params *params;
4120 	struct snd_pcm_hw_params_old *oparams = NULL;
4121 	int err;
4122 
4123 	params = kmalloc(sizeof(*params), GFP_KERNEL);
4124 	if (!params)
4125 		return -ENOMEM;
4126 
4127 	oparams = memdup_user(_oparams, sizeof(*oparams));
4128 	if (IS_ERR(oparams)) {
4129 		err = PTR_ERR(oparams);
4130 		goto out;
4131 	}
4132 
4133 	snd_pcm_hw_convert_from_old_params(params, oparams);
4134 	err = snd_pcm_hw_params(substream, params);
4135 	if (err < 0)
4136 		goto out_old;
4137 
4138 	snd_pcm_hw_convert_to_old_params(oparams, params);
4139 	if (copy_to_user(_oparams, oparams, sizeof(*oparams)))
4140 		err = -EFAULT;
4141 out_old:
4142 	kfree(oparams);
4143 out:
4144 	kfree(params);
4145 	return err;
4146 }
4147 #endif /* CONFIG_SND_SUPPORT_OLD_API */
4148 
4149 #ifndef CONFIG_MMU
4150 static unsigned long snd_pcm_get_unmapped_area(struct file *file,
4151 					       unsigned long addr,
4152 					       unsigned long len,
4153 					       unsigned long pgoff,
4154 					       unsigned long flags)
4155 {
4156 	struct snd_pcm_file *pcm_file = file->private_data;
4157 	struct snd_pcm_substream *substream = pcm_file->substream;
4158 	struct snd_pcm_runtime *runtime = substream->runtime;
4159 	unsigned long offset = pgoff << PAGE_SHIFT;
4160 
4161 	switch (offset) {
4162 	case SNDRV_PCM_MMAP_OFFSET_STATUS_NEW:
4163 		return (unsigned long)runtime->status;
4164 	case SNDRV_PCM_MMAP_OFFSET_CONTROL_NEW:
4165 		return (unsigned long)runtime->control;
4166 	default:
4167 		return (unsigned long)runtime->dma_area + offset;
4168 	}
4169 }
4170 #else
4171 # define snd_pcm_get_unmapped_area NULL
4172 #endif
4173 
4174 /*
4175  *  Register section
4176  */
4177 
4178 const struct file_operations snd_pcm_f_ops[2] = {
4179 	{
4180 		.owner =		THIS_MODULE,
4181 		.write =		snd_pcm_write,
4182 		.write_iter =		snd_pcm_writev,
4183 		.open =			snd_pcm_playback_open,
4184 		.release =		snd_pcm_release,
4185 		.llseek =		no_llseek,
4186 		.poll =			snd_pcm_poll,
4187 		.unlocked_ioctl =	snd_pcm_ioctl,
4188 		.compat_ioctl = 	snd_pcm_ioctl_compat,
4189 		.mmap =			snd_pcm_mmap,
4190 		.fasync =		snd_pcm_fasync,
4191 		.get_unmapped_area =	snd_pcm_get_unmapped_area,
4192 	},
4193 	{
4194 		.owner =		THIS_MODULE,
4195 		.read =			snd_pcm_read,
4196 		.read_iter =		snd_pcm_readv,
4197 		.open =			snd_pcm_capture_open,
4198 		.release =		snd_pcm_release,
4199 		.llseek =		no_llseek,
4200 		.poll =			snd_pcm_poll,
4201 		.unlocked_ioctl =	snd_pcm_ioctl,
4202 		.compat_ioctl = 	snd_pcm_ioctl_compat,
4203 		.mmap =			snd_pcm_mmap,
4204 		.fasync =		snd_pcm_fasync,
4205 		.get_unmapped_area =	snd_pcm_get_unmapped_area,
4206 	}
4207 };
4208