Lines Matching full:ipc

24 	struct avs_ipc *ipc = adev->ipc;
28 if (ipc->in_d0ix == enable)
33 /* Prevent further d0ix attempts on conscious IPC failure. */
35 atomic_inc(&ipc->d0ix_disable_depth);
37 ipc->in_d0ix = false;
41 ipc->in_d0ix = enable;
47 if (atomic_read(&adev->ipc->d0ix_disable_depth))
50 mod_delayed_work(system_power_efficient_wq, &adev->ipc->d0ix_work,
56 struct avs_ipc *ipc = container_of(work, struct avs_ipc, d0ix_work.work);
58 avs_dsp_set_d0ix(to_avs_dev(ipc->dev), true);
63 struct avs_ipc *ipc = adev->ipc;
65 if (!atomic_read(&ipc->d0ix_disable_depth)) {
66 cancel_delayed_work_sync(&ipc->d0ix_work);
75 struct avs_ipc *ipc = adev->ipc;
78 if (atomic_inc_return(&ipc->d0ix_disable_depth) == 1) {
79 cancel_delayed_work_sync(&ipc->d0ix_work);
88 struct avs_ipc *ipc = adev->ipc;
90 if (atomic_dec_and_test(&ipc->d0ix_disable_depth))
91 queue_delayed_work(system_power_efficient_wq, &ipc->d0ix_work,
148 atomic_set(&adev->ipc->recovering, 0);
153 struct avs_ipc *ipc = container_of(work, struct avs_ipc, recovery_work);
155 avs_dsp_recovery(to_avs_dev(ipc->dev));
160 struct avs_ipc *ipc = adev->ipc;
163 ipc->ready = false;
165 if (!atomic_add_unless(&ipc->recovering, 1, 1)) {
173 if (current_work() != &ipc->d0ix_work.work)
174 cancel_delayed_work_sync(&ipc->d0ix_work);
175 ipc->in_d0ix = false;
182 schedule_work(&ipc->recovery_work);
187 struct avs_ipc *ipc = adev->ipc;
195 ipc->rx.header = header;
201 ipc->rx.size = min_t(u32, AVS_MAILBOX_SIZE,
204 memcpy_fromio(ipc->rx.data, avs_uplink_addr(adev), ipc->rx.size);
205 trace_avs_msg_payload(ipc->rx.data, ipc->rx.size);
222 if (!adev->ipc->ready && msg.notify_msg_type != AVS_NOTIFY_FW_READY) {
268 adev->ipc->ready = true;
289 struct avs_ipc *ipc = adev->ipc;
297 spin_lock_irq(&ipc->rx_lock);
299 ipc->rx_completed = true;
300 spin_unlock_irq(&ipc->rx_lock);
305 complete(&ipc->busy_completion);
308 static bool avs_ipc_is_busy(struct avs_ipc *ipc)
310 struct avs_dev *adev = to_avs_dev(ipc->dev);
318 static int avs_ipc_wait_busy_completion(struct avs_ipc *ipc, int timeout)
324 ret = wait_for_completion_timeout(&ipc->busy_completion, msecs_to_jiffies(timeout));
327 if (!ipc->ready)
331 if (!avs_ipc_is_busy(ipc))
337 wait_for_completion_killable(&ipc->busy_completion);
341 spin_lock(&ipc->rx_lock);
342 if (!ipc->rx_completed) {
346 reinit_completion(&ipc->busy_completion);
347 spin_unlock(&ipc->rx_lock);
351 spin_unlock(&ipc->rx_lock);
355 spin_unlock(&ipc->rx_lock);
359 static void avs_ipc_msg_init(struct avs_ipc *ipc, struct avs_ipc_msg *reply)
361 lockdep_assert_held(&ipc->rx_lock);
363 ipc->rx.header = 0;
364 ipc->rx.size = reply ? reply->size : 0;
365 ipc->rx_completed = false;
367 reinit_completion(&ipc->done_completion);
368 reinit_completion(&ipc->busy_completion);
394 struct avs_ipc *ipc = adev->ipc;
397 if (!ipc->ready)
400 mutex_lock(&ipc->msg_mutex);
402 spin_lock(&ipc->rx_lock);
403 avs_ipc_msg_init(ipc, reply);
405 spin_unlock(&ipc->rx_lock);
407 ret = avs_ipc_wait_busy_completion(ipc, timeout);
418 ret = ipc->rx.rsp.status;
420 * If IPC channel is blocked e.g.: due to ongoing recovery,
433 reply->header = ipc->rx.header;
434 reply->size = ipc->rx.size;
435 if (reply->data && ipc->rx.size)
436 memcpy(reply->data, ipc->rx.data, reply->size);
440 mutex_unlock(&ipc->msg_mutex);
481 return avs_dsp_send_msg_timeout(adev, request, reply, adev->ipc->default_timeout_ms, name);
494 return avs_dsp_send_pm_msg_timeout(adev, request, reply, adev->ipc->default_timeout_ms,
501 struct avs_ipc *ipc = adev->ipc;
504 mutex_lock(&ipc->msg_mutex);
506 spin_lock(&ipc->rx_lock);
507 avs_ipc_msg_init(ipc, NULL);
513 spin_unlock(&ipc->rx_lock);
518 ret = wait_for_completion_timeout(&ipc->done_completion, msecs_to_jiffies(timeout));
525 mutex_unlock(&ipc->msg_mutex);
538 return avs_dsp_send_rom_msg_timeout(adev, request, adev->ipc->default_timeout_ms, name);
558 int avs_ipc_init(struct avs_ipc *ipc, struct device *dev)
560 ipc->rx.data = devm_kzalloc(dev, AVS_MAILBOX_SIZE, GFP_KERNEL);
561 if (!ipc->rx.data)
564 ipc->dev = dev;
565 ipc->ready = false;
566 ipc->default_timeout_ms = AVS_IPC_TIMEOUT_MS;
567 INIT_WORK(&ipc->recovery_work, avs_dsp_recovery_work);
568 INIT_DELAYED_WORK(&ipc->d0ix_work, avs_dsp_d0ix_work);
569 init_completion(&ipc->done_completion);
570 init_completion(&ipc->busy_completion);
571 spin_lock_init(&ipc->rx_lock);
572 mutex_init(&ipc->msg_mutex);
577 void avs_ipc_block(struct avs_ipc *ipc)
579 ipc->ready = false;
580 cancel_work_sync(&ipc->recovery_work);
581 cancel_delayed_work_sync(&ipc->d0ix_work);
582 ipc->in_d0ix = false;