Linux Audio

Check our new training course

Real-Time Linux with PREEMPT_RT training

Feb 18-20, 2025
Register
Loading...
v6.13.7
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 *  sst.c - Intel SST Driver for audio engine
  4 *
  5 *  Copyright (C) 2008-14	Intel Corp
  6 *  Authors:	Vinod Koul <vinod.koul@intel.com>
  7 *		Harsha Priya <priya.harsha@intel.com>
  8 *		Dharageswari R <dharageswari.r@intel.com>
  9 *		KP Jeeja <jeeja.kp@intel.com>
 10 *  ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 11 *
 12 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 13 */
 14#include <linux/module.h>
 15#include <linux/fs.h>
 16#include <linux/interrupt.h>
 17#include <linux/io.h>
 18#include <linux/firmware.h>
 19#include <linux/pci.h>
 20#include <linux/pm_runtime.h>
 21#include <linux/pm_qos.h>
 22#include <linux/async.h>
 23#include <linux/acpi.h>
 24#include <linux/sysfs.h>
 25#include <sound/core.h>
 26#include <sound/soc.h>
 27#include <asm/platform_sst_audio.h>
 28#include "../sst-mfld-platform.h"
 29#include "sst.h"
 30
 31MODULE_AUTHOR("Vinod Koul <vinod.koul@intel.com>");
 32MODULE_AUTHOR("Harsha Priya <priya.harsha@intel.com>");
 33MODULE_DESCRIPTION("Intel (R) SST(R) Audio Engine Driver");
 34MODULE_LICENSE("GPL v2");
 35
 36static inline bool sst_is_process_reply(u32 msg_id)
 37{
 38	return ((msg_id & PROCESS_MSG) ? true : false);
 39}
 40
 41static inline bool sst_validate_mailbox_size(unsigned int size)
 42{
 43	return ((size <= SST_MAILBOX_SIZE) ? true : false);
 44}
 45
 46static irqreturn_t intel_sst_interrupt_mrfld(int irq, void *context)
 47{
 48	union interrupt_reg_mrfld isr;
 49	union ipc_header_mrfld header;
 50	union sst_imr_reg_mrfld imr;
 51	struct ipc_post *msg = NULL;
 52	unsigned int size;
 53	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
 54	irqreturn_t retval = IRQ_HANDLED;
 55
 56	/* Interrupt arrived, check src */
 57	isr.full = sst_shim_read64(drv->shim, SST_ISRX);
 58
 59	if (isr.part.done_interrupt) {
 60		/* Clear done bit */
 61		spin_lock(&drv->ipc_spin_lock);
 62		header.full = sst_shim_read64(drv->shim,
 63					drv->ipc_reg.ipcx);
 64		header.p.header_high.part.done = 0;
 65		sst_shim_write64(drv->shim, drv->ipc_reg.ipcx, header.full);
 66
 67		/* write 1 to clear status register */;
 68		isr.part.done_interrupt = 1;
 69		sst_shim_write64(drv->shim, SST_ISRX, isr.full);
 70		spin_unlock(&drv->ipc_spin_lock);
 71
 72		/* we can send more messages to DSP so trigger work */
 73		queue_work(drv->post_msg_wq, &drv->ipc_post_msg_wq);
 74		retval = IRQ_HANDLED;
 75	}
 76
 77	if (isr.part.busy_interrupt) {
 78		/* message from dsp so copy that */
 79		spin_lock(&drv->ipc_spin_lock);
 80		imr.full = sst_shim_read64(drv->shim, SST_IMRX);
 81		imr.part.busy_interrupt = 1;
 82		sst_shim_write64(drv->shim, SST_IMRX, imr.full);
 83		spin_unlock(&drv->ipc_spin_lock);
 84		header.full =  sst_shim_read64(drv->shim, drv->ipc_reg.ipcd);
 85
 86		if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) {
 87			drv->ops->clear_interrupt(drv);
 88			return IRQ_HANDLED;
 89		}
 90
 91		if (header.p.header_high.part.large) {
 92			size = header.p.header_low_payload;
 93			if (sst_validate_mailbox_size(size)) {
 94				memcpy_fromio(msg->mailbox_data,
 95					drv->mailbox + drv->mailbox_recv_offset, size);
 96			} else {
 97				dev_err(drv->dev,
 98					"Mailbox not copied, payload size is: %u\n", size);
 99				header.p.header_low_payload = 0;
100			}
101		}
102
103		msg->mrfld_header = header;
104		msg->is_process_reply =
105			sst_is_process_reply(header.p.header_high.part.msg_id);
106		spin_lock(&drv->rx_msg_lock);
107		list_add_tail(&msg->node, &drv->rx_list);
108		spin_unlock(&drv->rx_msg_lock);
109		drv->ops->clear_interrupt(drv);
110		retval = IRQ_WAKE_THREAD;
111	}
112	return retval;
113}
114
115static irqreturn_t intel_sst_irq_thread_mrfld(int irq, void *context)
116{
117	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
118	struct ipc_post *__msg, *msg;
119	unsigned long irq_flags;
120
121	spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
122	if (list_empty(&drv->rx_list)) {
123		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
124		return IRQ_HANDLED;
125	}
126
127	list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) {
128		list_del(&msg->node);
129		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
130		if (msg->is_process_reply)
131			drv->ops->process_message(msg);
132		else
133			drv->ops->process_reply(drv, msg);
134
135		if (msg->is_large)
136			kfree(msg->mailbox_data);
137		kfree(msg);
138		spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
139	}
140	spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
141	return IRQ_HANDLED;
142}
143
144static int sst_save_dsp_context_v2(struct intel_sst_drv *sst)
145{
146	int ret = 0;
147
148	ret = sst_prepare_and_post_msg(sst, SST_TASK_ID_MEDIA, IPC_CMD,
149			IPC_PREP_D3, PIPE_RSVD, 0, NULL, NULL,
150			true, true, false, true);
151
152	if (ret < 0) {
153		dev_err(sst->dev, "not suspending FW!!, Err: %d\n", ret);
154		return -EIO;
155	}
156
157	return 0;
158}
159
160
161static struct intel_sst_ops mrfld_ops = {
162	.interrupt = intel_sst_interrupt_mrfld,
163	.irq_thread = intel_sst_irq_thread_mrfld,
164	.clear_interrupt = intel_sst_clear_intr_mrfld,
165	.start = sst_start_mrfld,
166	.reset = intel_sst_reset_dsp_mrfld,
167	.post_message = sst_post_message_mrfld,
168	.process_reply = sst_process_reply_mrfld,
169	.save_dsp_context =  sst_save_dsp_context_v2,
170	.alloc_stream = sst_alloc_stream_mrfld,
171	.post_download = sst_post_download_mrfld,
172};
173
174int sst_driver_ops(struct intel_sst_drv *sst)
175{
176
177	switch (sst->dev_id) {
178	case PCI_DEVICE_ID_INTEL_SST_TNG:
179	case PCI_DEVICE_ID_INTEL_SST_BYT:
180	case PCI_DEVICE_ID_INTEL_SST_BSW:
181		sst->tstamp = SST_TIME_STAMP_MRFLD;
182		sst->ops = &mrfld_ops;
183		return 0;
184
185	default:
186		dev_err(sst->dev,
187			"SST Driver capabilities missing for dev_id: %x",
188			sst->dev_id);
189		return -EINVAL;
190	}
191}
192
193void sst_process_pending_msg(struct work_struct *work)
194{
195	struct intel_sst_drv *ctx = container_of(work,
196			struct intel_sst_drv, ipc_post_msg_wq);
197
198	ctx->ops->post_message(ctx, NULL, false);
199}
200
201static int sst_workqueue_init(struct intel_sst_drv *ctx)
202{
203	INIT_LIST_HEAD(&ctx->memcpy_list);
204	INIT_LIST_HEAD(&ctx->rx_list);
205	INIT_LIST_HEAD(&ctx->ipc_dispatch_list);
206	INIT_LIST_HEAD(&ctx->block_list);
207	INIT_WORK(&ctx->ipc_post_msg_wq, sst_process_pending_msg);
208	init_waitqueue_head(&ctx->wait_queue);
209
210	ctx->post_msg_wq =
211		create_singlethread_workqueue("sst_post_msg_wq");
212	if (!ctx->post_msg_wq)
213		return -EBUSY;
214	return 0;
215}
216
217static void sst_init_locks(struct intel_sst_drv *ctx)
218{
219	mutex_init(&ctx->sst_lock);
220	spin_lock_init(&ctx->rx_msg_lock);
221	spin_lock_init(&ctx->ipc_spin_lock);
222	spin_lock_init(&ctx->block_lock);
223}
224
225/*
226 * Driver handles PCI IDs in ACPI - sst_acpi_probe() - and we are using only
227 * device ID part. If real ACPI ID appears, the kstrtouint() returns error, so
228 * we are fine with using unsigned short as dev_id type.
229 */
230int sst_alloc_drv_context(struct intel_sst_drv **ctx,
231		struct device *dev, unsigned short dev_id)
232{
233	*ctx = devm_kzalloc(dev, sizeof(struct intel_sst_drv), GFP_KERNEL);
234	if (!(*ctx))
235		return -ENOMEM;
236
237	(*ctx)->dev = dev;
238	(*ctx)->dev_id = dev_id;
239
240	return 0;
241}
242EXPORT_SYMBOL_GPL(sst_alloc_drv_context);
243
244static ssize_t firmware_version_show(struct device *dev,
245			    struct device_attribute *attr, char *buf)
246{
247	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
248
249	if (ctx->fw_version.type == 0 && ctx->fw_version.major == 0 &&
250	    ctx->fw_version.minor == 0 && ctx->fw_version.build == 0)
251		return sysfs_emit(buf, "FW not yet loaded\n");
252	else
253		return sysfs_emit(buf, "v%02x.%02x.%02x.%02x\n",
254				  ctx->fw_version.type, ctx->fw_version.major,
255				  ctx->fw_version.minor, ctx->fw_version.build);
256
257}
258
259static DEVICE_ATTR_RO(firmware_version);
260
261static const struct attribute *sst_fw_version_attrs[] = {
262	&dev_attr_firmware_version.attr,
263	NULL,
264};
265
266static const struct attribute_group sst_fw_version_attr_group = {
267	.attrs = (struct attribute **)sst_fw_version_attrs,
268};
269
270int sst_context_init(struct intel_sst_drv *ctx)
271{
272	int ret = 0, i;
273
274	if (!ctx->pdata)
275		return -EINVAL;
276
277	if (!ctx->pdata->probe_data)
278		return -EINVAL;
279
280	memcpy(&ctx->info, ctx->pdata->probe_data, sizeof(ctx->info));
281
282	ret = sst_driver_ops(ctx);
283	if (ret != 0)
284		return -EINVAL;
285
286	sst_init_locks(ctx);
287	sst_set_fw_state_locked(ctx, SST_RESET);
288
289	/* pvt_id 0 reserved for async messages */
290	ctx->pvt_id = 1;
291	ctx->stream_cnt = 0;
292	ctx->fw_in_mem = NULL;
293	/* we use memcpy, so set to 0 */
294	ctx->use_dma = 0;
295	ctx->use_lli = 0;
296
297	if (sst_workqueue_init(ctx))
298		return -EINVAL;
299
300	ctx->mailbox_recv_offset = ctx->pdata->ipc_info->mbox_recv_off;
301	ctx->ipc_reg.ipcx = SST_IPCX + ctx->pdata->ipc_info->ipc_offset;
302	ctx->ipc_reg.ipcd = SST_IPCD + ctx->pdata->ipc_info->ipc_offset;
303
304	dev_info(ctx->dev, "Got drv data max stream %d\n",
305				ctx->info.max_streams);
306
307	for (i = 1; i <= ctx->info.max_streams; i++) {
308		struct stream_info *stream = &ctx->streams[i];
309
310		memset(stream, 0, sizeof(*stream));
311		stream->pipe_id = PIPE_RSVD;
312		mutex_init(&stream->lock);
313	}
314
315	/* Register the ISR */
316	ret = devm_request_threaded_irq(ctx->dev, ctx->irq_num, ctx->ops->interrupt,
317					ctx->ops->irq_thread, 0, SST_DRV_NAME,
318					ctx);
319	if (ret)
320		goto do_free_mem;
321
322	dev_dbg(ctx->dev, "Registered IRQ %#x\n", ctx->irq_num);
323
324	/* default intr are unmasked so set this as masked */
325	sst_shim_write64(ctx->shim, SST_IMRX, 0xFFFF0038);
326
327	ctx->qos = devm_kzalloc(ctx->dev,
328		sizeof(struct pm_qos_request), GFP_KERNEL);
329	if (!ctx->qos) {
330		ret = -ENOMEM;
331		goto do_free_mem;
332	}
333	cpu_latency_qos_add_request(ctx->qos, PM_QOS_DEFAULT_VALUE);
334
335	dev_dbg(ctx->dev, "Requesting FW %s now...\n", ctx->firmware_name);
336	ret = request_firmware_nowait(THIS_MODULE, true, ctx->firmware_name,
337				      ctx->dev, GFP_KERNEL, ctx, sst_firmware_load_cb);
338	if (ret) {
339		dev_err(ctx->dev, "Firmware download failed:%d\n", ret);
340		goto do_free_mem;
341	}
342
343	ret = sysfs_create_group(&ctx->dev->kobj,
344				 &sst_fw_version_attr_group);
345	if (ret) {
346		dev_err(ctx->dev,
347			"Unable to create sysfs\n");
348		goto err_sysfs;
349	}
350
351	sst_register(ctx->dev);
352	return 0;
353err_sysfs:
354	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
355
356do_free_mem:
357	destroy_workqueue(ctx->post_msg_wq);
358	return ret;
359}
360EXPORT_SYMBOL_GPL(sst_context_init);
361
362void sst_context_cleanup(struct intel_sst_drv *ctx)
363{
364	pm_runtime_get_noresume(ctx->dev);
365	pm_runtime_disable(ctx->dev);
366	sst_unregister(ctx->dev);
367	sst_set_fw_state_locked(ctx, SST_SHUTDOWN);
368	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
369	destroy_workqueue(ctx->post_msg_wq);
370	cpu_latency_qos_remove_request(ctx->qos);
371	kfree(ctx->fw_sg_list.src);
372	kfree(ctx->fw_sg_list.dst);
373	ctx->fw_sg_list.list_len = 0;
374	kfree(ctx->fw_in_mem);
375	ctx->fw_in_mem = NULL;
376	sst_memcpy_free_resources(ctx);
377}
378EXPORT_SYMBOL_GPL(sst_context_cleanup);
379
380void sst_configure_runtime_pm(struct intel_sst_drv *ctx)
381{
382	pm_runtime_set_autosuspend_delay(ctx->dev, SST_SUSPEND_DELAY);
383	pm_runtime_use_autosuspend(ctx->dev);
384	/*
385	 * For acpi devices, the actual physical device state is
386	 * initially active. So change the state to active before
387	 * enabling the pm
388	 */
389
390	if (!acpi_disabled)
391		pm_runtime_set_active(ctx->dev);
392
393	pm_runtime_enable(ctx->dev);
394
395	if (acpi_disabled)
396		pm_runtime_set_active(ctx->dev);
397	else
398		pm_runtime_put_noidle(ctx->dev);
399}
400EXPORT_SYMBOL_GPL(sst_configure_runtime_pm);
401
402static int intel_sst_runtime_suspend(struct device *dev)
403{
404	int ret = 0;
405	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
406
407	if (ctx->sst_state == SST_RESET) {
408		dev_dbg(dev, "LPE is already in RESET state, No action\n");
409		return 0;
410	}
411	/* save fw context */
412	if (ctx->ops->save_dsp_context(ctx))
413		return -EBUSY;
414
415	/* Move the SST state to Reset */
416	sst_set_fw_state_locked(ctx, SST_RESET);
417
418	synchronize_irq(ctx->irq_num);
419	flush_workqueue(ctx->post_msg_wq);
420
421	ctx->ops->reset(ctx);
422
423	return ret;
424}
425
426static int intel_sst_suspend(struct device *dev)
427{
428	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
429	struct sst_fw_save *fw_save;
430	int i, ret;
431
432	/* check first if we are already in SW reset */
433	if (ctx->sst_state == SST_RESET)
434		return 0;
435
436	/*
437	 * check if any stream is active and running
438	 * they should already by suspend by soc_suspend
439	 */
440	for (i = 1; i <= ctx->info.max_streams; i++) {
441		struct stream_info *stream = &ctx->streams[i];
442
443		if (stream->status == STREAM_RUNNING) {
444			dev_err(dev, "stream %d is running, can't suspend, abort\n", i);
445			return -EBUSY;
446		}
447
448		if (ctx->pdata->streams_lost_on_suspend) {
449			stream->resume_status = stream->status;
450			stream->resume_prev = stream->prev;
451			if (stream->status != STREAM_UN_INIT)
452				sst_free_stream(ctx, i);
453		}
454	}
455	synchronize_irq(ctx->irq_num);
456	flush_workqueue(ctx->post_msg_wq);
457
458	/* Move the SST state to Reset */
459	sst_set_fw_state_locked(ctx, SST_RESET);
460
461	/* tell DSP we are suspending */
462	if (ctx->ops->save_dsp_context(ctx))
463		return -EBUSY;
464
465	/* save the memories */
466	fw_save = kzalloc(sizeof(*fw_save), GFP_KERNEL);
467	if (!fw_save)
468		return -ENOMEM;
469	fw_save->iram = kvzalloc(ctx->iram_end - ctx->iram_base, GFP_KERNEL);
470	if (!fw_save->iram) {
471		ret = -ENOMEM;
472		goto iram;
473	}
474	fw_save->dram = kvzalloc(ctx->dram_end - ctx->dram_base, GFP_KERNEL);
475	if (!fw_save->dram) {
476		ret = -ENOMEM;
477		goto dram;
478	}
479	fw_save->sram = kvzalloc(SST_MAILBOX_SIZE, GFP_KERNEL);
480	if (!fw_save->sram) {
481		ret = -ENOMEM;
482		goto sram;
483	}
484
485	fw_save->ddr = kvzalloc(ctx->ddr_end - ctx->ddr_base, GFP_KERNEL);
486	if (!fw_save->ddr) {
487		ret = -ENOMEM;
488		goto ddr;
489	}
490
491	memcpy32_fromio(fw_save->iram, ctx->iram, ctx->iram_end - ctx->iram_base);
492	memcpy32_fromio(fw_save->dram, ctx->dram, ctx->dram_end - ctx->dram_base);
493	memcpy32_fromio(fw_save->sram, ctx->mailbox, SST_MAILBOX_SIZE);
494	memcpy32_fromio(fw_save->ddr, ctx->ddr, ctx->ddr_end - ctx->ddr_base);
495
496	ctx->fw_save = fw_save;
497	ctx->ops->reset(ctx);
498	return 0;
499ddr:
500	kvfree(fw_save->sram);
501sram:
502	kvfree(fw_save->dram);
503dram:
504	kvfree(fw_save->iram);
505iram:
506	kfree(fw_save);
507	return ret;
508}
509
510static int intel_sst_resume(struct device *dev)
511{
512	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
513	struct sst_fw_save *fw_save = ctx->fw_save;
514	struct sst_block *block;
515	int i, ret = 0;
516
517	if (!fw_save)
518		return 0;
519
520	sst_set_fw_state_locked(ctx, SST_FW_LOADING);
521
522	/* we have to restore the memory saved */
523	ctx->ops->reset(ctx);
524
525	ctx->fw_save = NULL;
526
527	memcpy32_toio(ctx->iram, fw_save->iram, ctx->iram_end - ctx->iram_base);
528	memcpy32_toio(ctx->dram, fw_save->dram, ctx->dram_end - ctx->dram_base);
529	memcpy32_toio(ctx->mailbox, fw_save->sram, SST_MAILBOX_SIZE);
530	memcpy32_toio(ctx->ddr, fw_save->ddr, ctx->ddr_end - ctx->ddr_base);
531
532	kvfree(fw_save->sram);
533	kvfree(fw_save->dram);
534	kvfree(fw_save->iram);
535	kvfree(fw_save->ddr);
536	kfree(fw_save);
537
538	block = sst_create_block(ctx, 0, FW_DWNL_ID);
539	if (block == NULL)
540		return -ENOMEM;
541
542
543	/* start and wait for ack */
544	ctx->ops->start(ctx);
545	ret = sst_wait_timeout(ctx, block);
546	if (ret) {
547		dev_err(ctx->dev, "fw download failed %d\n", ret);
548		/* FW download failed due to timeout */
549		ret = -EBUSY;
550
551	} else {
552		sst_set_fw_state_locked(ctx, SST_FW_RUNNING);
553	}
554
555	if (ctx->pdata->streams_lost_on_suspend) {
556		for (i = 1; i <= ctx->info.max_streams; i++) {
557			struct stream_info *stream = &ctx->streams[i];
558
559			if (stream->resume_status != STREAM_UN_INIT) {
560				dev_dbg(ctx->dev, "Re-allocing stream %d status %d prev %d\n",
561					i, stream->resume_status,
562					stream->resume_prev);
563				sst_realloc_stream(ctx, i);
564				stream->status = stream->resume_status;
565				stream->prev = stream->resume_prev;
566			}
567		}
568	}
569
570	sst_free_block(ctx, block);
571	return ret;
572}
573
574const struct dev_pm_ops intel_sst_pm = {
575	.suspend = intel_sst_suspend,
576	.resume = intel_sst_resume,
577	.runtime_suspend = intel_sst_runtime_suspend,
578};
579EXPORT_SYMBOL_GPL(intel_sst_pm);
v6.2
  1// SPDX-License-Identifier: GPL-2.0-only
  2/*
  3 *  sst.c - Intel SST Driver for audio engine
  4 *
  5 *  Copyright (C) 2008-14	Intel Corp
  6 *  Authors:	Vinod Koul <vinod.koul@intel.com>
  7 *		Harsha Priya <priya.harsha@intel.com>
  8 *		Dharageswari R <dharageswari.r@intel.com>
  9 *		KP Jeeja <jeeja.kp@intel.com>
 10 *  ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 11 *
 12 * ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
 13 */
 14#include <linux/module.h>
 15#include <linux/fs.h>
 16#include <linux/interrupt.h>
 17#include <linux/io.h>
 18#include <linux/firmware.h>
 
 19#include <linux/pm_runtime.h>
 20#include <linux/pm_qos.h>
 21#include <linux/async.h>
 22#include <linux/acpi.h>
 23#include <linux/sysfs.h>
 24#include <sound/core.h>
 25#include <sound/soc.h>
 26#include <asm/platform_sst_audio.h>
 27#include "../sst-mfld-platform.h"
 28#include "sst.h"
 29
 30MODULE_AUTHOR("Vinod Koul <vinod.koul@intel.com>");
 31MODULE_AUTHOR("Harsha Priya <priya.harsha@intel.com>");
 32MODULE_DESCRIPTION("Intel (R) SST(R) Audio Engine Driver");
 33MODULE_LICENSE("GPL v2");
 34
 35static inline bool sst_is_process_reply(u32 msg_id)
 36{
 37	return ((msg_id & PROCESS_MSG) ? true : false);
 38}
 39
 40static inline bool sst_validate_mailbox_size(unsigned int size)
 41{
 42	return ((size <= SST_MAILBOX_SIZE) ? true : false);
 43}
 44
 45static irqreturn_t intel_sst_interrupt_mrfld(int irq, void *context)
 46{
 47	union interrupt_reg_mrfld isr;
 48	union ipc_header_mrfld header;
 49	union sst_imr_reg_mrfld imr;
 50	struct ipc_post *msg = NULL;
 51	unsigned int size;
 52	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
 53	irqreturn_t retval = IRQ_HANDLED;
 54
 55	/* Interrupt arrived, check src */
 56	isr.full = sst_shim_read64(drv->shim, SST_ISRX);
 57
 58	if (isr.part.done_interrupt) {
 59		/* Clear done bit */
 60		spin_lock(&drv->ipc_spin_lock);
 61		header.full = sst_shim_read64(drv->shim,
 62					drv->ipc_reg.ipcx);
 63		header.p.header_high.part.done = 0;
 64		sst_shim_write64(drv->shim, drv->ipc_reg.ipcx, header.full);
 65
 66		/* write 1 to clear status register */;
 67		isr.part.done_interrupt = 1;
 68		sst_shim_write64(drv->shim, SST_ISRX, isr.full);
 69		spin_unlock(&drv->ipc_spin_lock);
 70
 71		/* we can send more messages to DSP so trigger work */
 72		queue_work(drv->post_msg_wq, &drv->ipc_post_msg_wq);
 73		retval = IRQ_HANDLED;
 74	}
 75
 76	if (isr.part.busy_interrupt) {
 77		/* message from dsp so copy that */
 78		spin_lock(&drv->ipc_spin_lock);
 79		imr.full = sst_shim_read64(drv->shim, SST_IMRX);
 80		imr.part.busy_interrupt = 1;
 81		sst_shim_write64(drv->shim, SST_IMRX, imr.full);
 82		spin_unlock(&drv->ipc_spin_lock);
 83		header.full =  sst_shim_read64(drv->shim, drv->ipc_reg.ipcd);
 84
 85		if (sst_create_ipc_msg(&msg, header.p.header_high.part.large)) {
 86			drv->ops->clear_interrupt(drv);
 87			return IRQ_HANDLED;
 88		}
 89
 90		if (header.p.header_high.part.large) {
 91			size = header.p.header_low_payload;
 92			if (sst_validate_mailbox_size(size)) {
 93				memcpy_fromio(msg->mailbox_data,
 94					drv->mailbox + drv->mailbox_recv_offset, size);
 95			} else {
 96				dev_err(drv->dev,
 97					"Mailbox not copied, payload size is: %u\n", size);
 98				header.p.header_low_payload = 0;
 99			}
100		}
101
102		msg->mrfld_header = header;
103		msg->is_process_reply =
104			sst_is_process_reply(header.p.header_high.part.msg_id);
105		spin_lock(&drv->rx_msg_lock);
106		list_add_tail(&msg->node, &drv->rx_list);
107		spin_unlock(&drv->rx_msg_lock);
108		drv->ops->clear_interrupt(drv);
109		retval = IRQ_WAKE_THREAD;
110	}
111	return retval;
112}
113
114static irqreturn_t intel_sst_irq_thread_mrfld(int irq, void *context)
115{
116	struct intel_sst_drv *drv = (struct intel_sst_drv *) context;
117	struct ipc_post *__msg, *msg;
118	unsigned long irq_flags;
119
120	spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
121	if (list_empty(&drv->rx_list)) {
122		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
123		return IRQ_HANDLED;
124	}
125
126	list_for_each_entry_safe(msg, __msg, &drv->rx_list, node) {
127		list_del(&msg->node);
128		spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
129		if (msg->is_process_reply)
130			drv->ops->process_message(msg);
131		else
132			drv->ops->process_reply(drv, msg);
133
134		if (msg->is_large)
135			kfree(msg->mailbox_data);
136		kfree(msg);
137		spin_lock_irqsave(&drv->rx_msg_lock, irq_flags);
138	}
139	spin_unlock_irqrestore(&drv->rx_msg_lock, irq_flags);
140	return IRQ_HANDLED;
141}
142
143static int sst_save_dsp_context_v2(struct intel_sst_drv *sst)
144{
145	int ret = 0;
146
147	ret = sst_prepare_and_post_msg(sst, SST_TASK_ID_MEDIA, IPC_CMD,
148			IPC_PREP_D3, PIPE_RSVD, 0, NULL, NULL,
149			true, true, false, true);
150
151	if (ret < 0) {
152		dev_err(sst->dev, "not suspending FW!!, Err: %d\n", ret);
153		return -EIO;
154	}
155
156	return 0;
157}
158
159
160static struct intel_sst_ops mrfld_ops = {
161	.interrupt = intel_sst_interrupt_mrfld,
162	.irq_thread = intel_sst_irq_thread_mrfld,
163	.clear_interrupt = intel_sst_clear_intr_mrfld,
164	.start = sst_start_mrfld,
165	.reset = intel_sst_reset_dsp_mrfld,
166	.post_message = sst_post_message_mrfld,
167	.process_reply = sst_process_reply_mrfld,
168	.save_dsp_context =  sst_save_dsp_context_v2,
169	.alloc_stream = sst_alloc_stream_mrfld,
170	.post_download = sst_post_download_mrfld,
171};
172
173int sst_driver_ops(struct intel_sst_drv *sst)
174{
175
176	switch (sst->dev_id) {
177	case SST_MRFLD_PCI_ID:
178	case SST_BYT_ACPI_ID:
179	case SST_CHV_ACPI_ID:
180		sst->tstamp = SST_TIME_STAMP_MRFLD;
181		sst->ops = &mrfld_ops;
182		return 0;
183
184	default:
185		dev_err(sst->dev,
186			"SST Driver capabilities missing for dev_id: %x",
187			sst->dev_id);
188		return -EINVAL;
189	}
190}
191
192void sst_process_pending_msg(struct work_struct *work)
193{
194	struct intel_sst_drv *ctx = container_of(work,
195			struct intel_sst_drv, ipc_post_msg_wq);
196
197	ctx->ops->post_message(ctx, NULL, false);
198}
199
200static int sst_workqueue_init(struct intel_sst_drv *ctx)
201{
202	INIT_LIST_HEAD(&ctx->memcpy_list);
203	INIT_LIST_HEAD(&ctx->rx_list);
204	INIT_LIST_HEAD(&ctx->ipc_dispatch_list);
205	INIT_LIST_HEAD(&ctx->block_list);
206	INIT_WORK(&ctx->ipc_post_msg_wq, sst_process_pending_msg);
207	init_waitqueue_head(&ctx->wait_queue);
208
209	ctx->post_msg_wq =
210		create_singlethread_workqueue("sst_post_msg_wq");
211	if (!ctx->post_msg_wq)
212		return -EBUSY;
213	return 0;
214}
215
216static void sst_init_locks(struct intel_sst_drv *ctx)
217{
218	mutex_init(&ctx->sst_lock);
219	spin_lock_init(&ctx->rx_msg_lock);
220	spin_lock_init(&ctx->ipc_spin_lock);
221	spin_lock_init(&ctx->block_lock);
222}
223
 
 
 
 
 
224int sst_alloc_drv_context(struct intel_sst_drv **ctx,
225		struct device *dev, unsigned int dev_id)
226{
227	*ctx = devm_kzalloc(dev, sizeof(struct intel_sst_drv), GFP_KERNEL);
228	if (!(*ctx))
229		return -ENOMEM;
230
231	(*ctx)->dev = dev;
232	(*ctx)->dev_id = dev_id;
233
234	return 0;
235}
236EXPORT_SYMBOL_GPL(sst_alloc_drv_context);
237
238static ssize_t firmware_version_show(struct device *dev,
239			    struct device_attribute *attr, char *buf)
240{
241	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
242
243	if (ctx->fw_version.type == 0 && ctx->fw_version.major == 0 &&
244	    ctx->fw_version.minor == 0 && ctx->fw_version.build == 0)
245		return sysfs_emit(buf, "FW not yet loaded\n");
246	else
247		return sysfs_emit(buf, "v%02x.%02x.%02x.%02x\n",
248				  ctx->fw_version.type, ctx->fw_version.major,
249				  ctx->fw_version.minor, ctx->fw_version.build);
250
251}
252
253static DEVICE_ATTR_RO(firmware_version);
254
255static const struct attribute *sst_fw_version_attrs[] = {
256	&dev_attr_firmware_version.attr,
257	NULL,
258};
259
260static const struct attribute_group sst_fw_version_attr_group = {
261	.attrs = (struct attribute **)sst_fw_version_attrs,
262};
263
264int sst_context_init(struct intel_sst_drv *ctx)
265{
266	int ret = 0, i;
267
268	if (!ctx->pdata)
269		return -EINVAL;
270
271	if (!ctx->pdata->probe_data)
272		return -EINVAL;
273
274	memcpy(&ctx->info, ctx->pdata->probe_data, sizeof(ctx->info));
275
276	ret = sst_driver_ops(ctx);
277	if (ret != 0)
278		return -EINVAL;
279
280	sst_init_locks(ctx);
281	sst_set_fw_state_locked(ctx, SST_RESET);
282
283	/* pvt_id 0 reserved for async messages */
284	ctx->pvt_id = 1;
285	ctx->stream_cnt = 0;
286	ctx->fw_in_mem = NULL;
287	/* we use memcpy, so set to 0 */
288	ctx->use_dma = 0;
289	ctx->use_lli = 0;
290
291	if (sst_workqueue_init(ctx))
292		return -EINVAL;
293
294	ctx->mailbox_recv_offset = ctx->pdata->ipc_info->mbox_recv_off;
295	ctx->ipc_reg.ipcx = SST_IPCX + ctx->pdata->ipc_info->ipc_offset;
296	ctx->ipc_reg.ipcd = SST_IPCD + ctx->pdata->ipc_info->ipc_offset;
297
298	dev_info(ctx->dev, "Got drv data max stream %d\n",
299				ctx->info.max_streams);
300
301	for (i = 1; i <= ctx->info.max_streams; i++) {
302		struct stream_info *stream = &ctx->streams[i];
303
304		memset(stream, 0, sizeof(*stream));
305		stream->pipe_id = PIPE_RSVD;
306		mutex_init(&stream->lock);
307	}
308
309	/* Register the ISR */
310	ret = devm_request_threaded_irq(ctx->dev, ctx->irq_num, ctx->ops->interrupt,
311					ctx->ops->irq_thread, 0, SST_DRV_NAME,
312					ctx);
313	if (ret)
314		goto do_free_mem;
315
316	dev_dbg(ctx->dev, "Registered IRQ %#x\n", ctx->irq_num);
317
318	/* default intr are unmasked so set this as masked */
319	sst_shim_write64(ctx->shim, SST_IMRX, 0xFFFF0038);
320
321	ctx->qos = devm_kzalloc(ctx->dev,
322		sizeof(struct pm_qos_request), GFP_KERNEL);
323	if (!ctx->qos) {
324		ret = -ENOMEM;
325		goto do_free_mem;
326	}
327	cpu_latency_qos_add_request(ctx->qos, PM_QOS_DEFAULT_VALUE);
328
329	dev_dbg(ctx->dev, "Requesting FW %s now...\n", ctx->firmware_name);
330	ret = request_firmware_nowait(THIS_MODULE, true, ctx->firmware_name,
331				      ctx->dev, GFP_KERNEL, ctx, sst_firmware_load_cb);
332	if (ret) {
333		dev_err(ctx->dev, "Firmware download failed:%d\n", ret);
334		goto do_free_mem;
335	}
336
337	ret = sysfs_create_group(&ctx->dev->kobj,
338				 &sst_fw_version_attr_group);
339	if (ret) {
340		dev_err(ctx->dev,
341			"Unable to create sysfs\n");
342		goto err_sysfs;
343	}
344
345	sst_register(ctx->dev);
346	return 0;
347err_sysfs:
348	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
349
350do_free_mem:
351	destroy_workqueue(ctx->post_msg_wq);
352	return ret;
353}
354EXPORT_SYMBOL_GPL(sst_context_init);
355
356void sst_context_cleanup(struct intel_sst_drv *ctx)
357{
358	pm_runtime_get_noresume(ctx->dev);
359	pm_runtime_disable(ctx->dev);
360	sst_unregister(ctx->dev);
361	sst_set_fw_state_locked(ctx, SST_SHUTDOWN);
362	sysfs_remove_group(&ctx->dev->kobj, &sst_fw_version_attr_group);
363	destroy_workqueue(ctx->post_msg_wq);
364	cpu_latency_qos_remove_request(ctx->qos);
365	kfree(ctx->fw_sg_list.src);
366	kfree(ctx->fw_sg_list.dst);
367	ctx->fw_sg_list.list_len = 0;
368	kfree(ctx->fw_in_mem);
369	ctx->fw_in_mem = NULL;
370	sst_memcpy_free_resources(ctx);
371}
372EXPORT_SYMBOL_GPL(sst_context_cleanup);
373
374void sst_configure_runtime_pm(struct intel_sst_drv *ctx)
375{
376	pm_runtime_set_autosuspend_delay(ctx->dev, SST_SUSPEND_DELAY);
377	pm_runtime_use_autosuspend(ctx->dev);
378	/*
379	 * For acpi devices, the actual physical device state is
380	 * initially active. So change the state to active before
381	 * enabling the pm
382	 */
383
384	if (!acpi_disabled)
385		pm_runtime_set_active(ctx->dev);
386
387	pm_runtime_enable(ctx->dev);
388
389	if (acpi_disabled)
390		pm_runtime_set_active(ctx->dev);
391	else
392		pm_runtime_put_noidle(ctx->dev);
393}
394EXPORT_SYMBOL_GPL(sst_configure_runtime_pm);
395
396static int intel_sst_runtime_suspend(struct device *dev)
397{
398	int ret = 0;
399	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
400
401	if (ctx->sst_state == SST_RESET) {
402		dev_dbg(dev, "LPE is already in RESET state, No action\n");
403		return 0;
404	}
405	/* save fw context */
406	if (ctx->ops->save_dsp_context(ctx))
407		return -EBUSY;
408
409	/* Move the SST state to Reset */
410	sst_set_fw_state_locked(ctx, SST_RESET);
411
412	synchronize_irq(ctx->irq_num);
413	flush_workqueue(ctx->post_msg_wq);
414
415	ctx->ops->reset(ctx);
416
417	return ret;
418}
419
420static int intel_sst_suspend(struct device *dev)
421{
422	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
423	struct sst_fw_save *fw_save;
424	int i, ret;
425
426	/* check first if we are already in SW reset */
427	if (ctx->sst_state == SST_RESET)
428		return 0;
429
430	/*
431	 * check if any stream is active and running
432	 * they should already by suspend by soc_suspend
433	 */
434	for (i = 1; i <= ctx->info.max_streams; i++) {
435		struct stream_info *stream = &ctx->streams[i];
436
437		if (stream->status == STREAM_RUNNING) {
438			dev_err(dev, "stream %d is running, can't suspend, abort\n", i);
439			return -EBUSY;
440		}
441
442		if (ctx->pdata->streams_lost_on_suspend) {
443			stream->resume_status = stream->status;
444			stream->resume_prev = stream->prev;
445			if (stream->status != STREAM_UN_INIT)
446				sst_free_stream(ctx, i);
447		}
448	}
449	synchronize_irq(ctx->irq_num);
450	flush_workqueue(ctx->post_msg_wq);
451
452	/* Move the SST state to Reset */
453	sst_set_fw_state_locked(ctx, SST_RESET);
454
455	/* tell DSP we are suspending */
456	if (ctx->ops->save_dsp_context(ctx))
457		return -EBUSY;
458
459	/* save the memories */
460	fw_save = kzalloc(sizeof(*fw_save), GFP_KERNEL);
461	if (!fw_save)
462		return -ENOMEM;
463	fw_save->iram = kvzalloc(ctx->iram_end - ctx->iram_base, GFP_KERNEL);
464	if (!fw_save->iram) {
465		ret = -ENOMEM;
466		goto iram;
467	}
468	fw_save->dram = kvzalloc(ctx->dram_end - ctx->dram_base, GFP_KERNEL);
469	if (!fw_save->dram) {
470		ret = -ENOMEM;
471		goto dram;
472	}
473	fw_save->sram = kvzalloc(SST_MAILBOX_SIZE, GFP_KERNEL);
474	if (!fw_save->sram) {
475		ret = -ENOMEM;
476		goto sram;
477	}
478
479	fw_save->ddr = kvzalloc(ctx->ddr_end - ctx->ddr_base, GFP_KERNEL);
480	if (!fw_save->ddr) {
481		ret = -ENOMEM;
482		goto ddr;
483	}
484
485	memcpy32_fromio(fw_save->iram, ctx->iram, ctx->iram_end - ctx->iram_base);
486	memcpy32_fromio(fw_save->dram, ctx->dram, ctx->dram_end - ctx->dram_base);
487	memcpy32_fromio(fw_save->sram, ctx->mailbox, SST_MAILBOX_SIZE);
488	memcpy32_fromio(fw_save->ddr, ctx->ddr, ctx->ddr_end - ctx->ddr_base);
489
490	ctx->fw_save = fw_save;
491	ctx->ops->reset(ctx);
492	return 0;
493ddr:
494	kvfree(fw_save->sram);
495sram:
496	kvfree(fw_save->dram);
497dram:
498	kvfree(fw_save->iram);
499iram:
500	kfree(fw_save);
501	return ret;
502}
503
504static int intel_sst_resume(struct device *dev)
505{
506	struct intel_sst_drv *ctx = dev_get_drvdata(dev);
507	struct sst_fw_save *fw_save = ctx->fw_save;
508	struct sst_block *block;
509	int i, ret = 0;
510
511	if (!fw_save)
512		return 0;
513
514	sst_set_fw_state_locked(ctx, SST_FW_LOADING);
515
516	/* we have to restore the memory saved */
517	ctx->ops->reset(ctx);
518
519	ctx->fw_save = NULL;
520
521	memcpy32_toio(ctx->iram, fw_save->iram, ctx->iram_end - ctx->iram_base);
522	memcpy32_toio(ctx->dram, fw_save->dram, ctx->dram_end - ctx->dram_base);
523	memcpy32_toio(ctx->mailbox, fw_save->sram, SST_MAILBOX_SIZE);
524	memcpy32_toio(ctx->ddr, fw_save->ddr, ctx->ddr_end - ctx->ddr_base);
525
526	kvfree(fw_save->sram);
527	kvfree(fw_save->dram);
528	kvfree(fw_save->iram);
529	kvfree(fw_save->ddr);
530	kfree(fw_save);
531
532	block = sst_create_block(ctx, 0, FW_DWNL_ID);
533	if (block == NULL)
534		return -ENOMEM;
535
536
537	/* start and wait for ack */
538	ctx->ops->start(ctx);
539	ret = sst_wait_timeout(ctx, block);
540	if (ret) {
541		dev_err(ctx->dev, "fw download failed %d\n", ret);
542		/* FW download failed due to timeout */
543		ret = -EBUSY;
544
545	} else {
546		sst_set_fw_state_locked(ctx, SST_FW_RUNNING);
547	}
548
549	if (ctx->pdata->streams_lost_on_suspend) {
550		for (i = 1; i <= ctx->info.max_streams; i++) {
551			struct stream_info *stream = &ctx->streams[i];
552
553			if (stream->resume_status != STREAM_UN_INIT) {
554				dev_dbg(ctx->dev, "Re-allocing stream %d status %d prev %d\n",
555					i, stream->resume_status,
556					stream->resume_prev);
557				sst_realloc_stream(ctx, i);
558				stream->status = stream->resume_status;
559				stream->prev = stream->resume_prev;
560			}
561		}
562	}
563
564	sst_free_block(ctx, block);
565	return ret;
566}
567
568const struct dev_pm_ops intel_sst_pm = {
569	.suspend = intel_sst_suspend,
570	.resume = intel_sst_resume,
571	.runtime_suspend = intel_sst_runtime_suspend,
572};
573EXPORT_SYMBOL_GPL(intel_sst_pm);