// SPDX-License-Identifier: (GPL-2.0 OR MIT) // // Copyright (c) 2018 BayLibre, SAS. // Author: Jerome Brunet #include #include #include #include #include #include #include #include #include #include "axg-fifo.h" /* * This file implements the platform operations common to the playback and * capture frontend DAI. The logic behind this two types of fifo is very * similar but some difference exist. * These differences the respective DAI drivers */ static struct snd_pcm_hardware axg_fifo_hw = { .info = (SNDRV_PCM_INFO_INTERLEAVED | SNDRV_PCM_INFO_MMAP | SNDRV_PCM_INFO_MMAP_VALID | SNDRV_PCM_INFO_BLOCK_TRANSFER | SNDRV_PCM_INFO_PAUSE), .formats = AXG_FIFO_FORMATS, .rate_min = 5512, .rate_max = 192000, .channels_min = 1, .channels_max = AXG_FIFO_CH_MAX, .period_bytes_min = AXG_FIFO_MIN_DEPTH, .period_bytes_max = UINT_MAX, .periods_min = 2, .periods_max = UINT_MAX, /* No real justification for this */ .buffer_bytes_max = 1 * 1024 * 1024, }; static struct snd_soc_dai *axg_fifo_dai(struct snd_pcm_substream *ss) { struct snd_soc_pcm_runtime *rtd = ss->private_data; return rtd->cpu_dai; } static struct axg_fifo *axg_fifo_data(struct snd_pcm_substream *ss) { struct snd_soc_dai *dai = axg_fifo_dai(ss); return snd_soc_dai_get_drvdata(dai); } static struct device *axg_fifo_dev(struct snd_pcm_substream *ss) { struct snd_soc_dai *dai = axg_fifo_dai(ss); return dai->dev; } static void __dma_enable(struct axg_fifo *fifo, bool enable) { regmap_update_bits(fifo->map, FIFO_CTRL0, CTRL0_DMA_EN, enable ? CTRL0_DMA_EN : 0); } static int axg_fifo_pcm_trigger(struct snd_pcm_substream *ss, int cmd) { struct axg_fifo *fifo = axg_fifo_data(ss); switch (cmd) { case SNDRV_PCM_TRIGGER_START: case SNDRV_PCM_TRIGGER_RESUME: case SNDRV_PCM_TRIGGER_PAUSE_RELEASE: __dma_enable(fifo, true); break; case SNDRV_PCM_TRIGGER_SUSPEND: case SNDRV_PCM_TRIGGER_PAUSE_PUSH: case SNDRV_PCM_TRIGGER_STOP: __dma_enable(fifo, false); break; default: return -EINVAL; } return 0; } static snd_pcm_uframes_t axg_fifo_pcm_pointer(struct snd_pcm_substream *ss) { struct axg_fifo *fifo = axg_fifo_data(ss); struct snd_pcm_runtime *runtime = ss->runtime; unsigned int addr; regmap_read(fifo->map, FIFO_STATUS2, &addr); return bytes_to_frames(runtime, addr - (unsigned int)runtime->dma_addr); } static int axg_fifo_pcm_hw_params(struct snd_pcm_substream *ss, struct snd_pcm_hw_params *params) { struct snd_pcm_runtime *runtime = ss->runtime; struct axg_fifo *fifo = axg_fifo_data(ss); dma_addr_t end_ptr; unsigned int burst_num; int ret; ret = snd_pcm_lib_malloc_pages(ss, params_buffer_bytes(params)); if (ret < 0) return ret; /* Setup dma memory pointers */ end_ptr = runtime->dma_addr + runtime->dma_bytes - AXG_FIFO_BURST; regmap_write(fifo->map, FIFO_START_ADDR, runtime->dma_addr); regmap_write(fifo->map, FIFO_FINISH_ADDR, end_ptr); /* Setup interrupt periodicity */ burst_num = params_period_bytes(params) / AXG_FIFO_BURST; regmap_write(fifo->map, FIFO_INT_ADDR, burst_num); /* Enable block count irq */ regmap_update_bits(fifo->map, FIFO_CTRL0, CTRL0_INT_EN(FIFO_INT_COUNT_REPEAT), CTRL0_INT_EN(FIFO_INT_COUNT_REPEAT)); return 0; } static int axg_fifo_pcm_hw_free(struct snd_pcm_substream *ss) { struct axg_fifo *fifo = axg_fifo_data(ss); /* Disable the block count irq */ regmap_update_bits(fifo->map, FIFO_CTRL0, CTRL0_INT_EN(FIFO_INT_COUNT_REPEAT), 0); return snd_pcm_lib_free_pages(ss); } static void axg_fifo_ack_irq(struct axg_fifo *fifo, u8 mask) { regmap_update_bits(fifo->map, FIFO_CTRL1, CTRL1_INT_CLR(FIFO_INT_MASK), CTRL1_INT_CLR(mask)); /* Clear must also be cleared */ regmap_update_bits(fifo->map, FIFO_CTRL1, CTRL1_INT_CLR(FIFO_INT_MASK), 0); } static irqreturn_t axg_fifo_pcm_irq_block(int irq, void *dev_id) { struct snd_pcm_substream *ss = dev_id; struct axg_fifo *fifo = axg_fifo_data(ss); unsigned int status; regmap_read(fifo->map, FIFO_STATUS1, &status); status = STATUS1_INT_STS(status) & FIFO_INT_MASK; if (status & FIFO_INT_COUNT_REPEAT) snd_pcm_period_elapsed(ss); else dev_dbg(axg_fifo_dev(ss), "unexpected irq - STS 0x%02x\n", status); /* Ack irqs */ axg_fifo_ack_irq(fifo, status); return IRQ_RETVAL(status); } static int axg_fifo_pcm_open(struct snd_pcm_substream *ss) { struct axg_fifo *fifo = axg_fifo_data(ss); struct device *dev = axg_fifo_dev(ss); int ret; snd_soc_set_runtime_hwparams(ss, &axg_fifo_hw); /* * Make sure the buffer and period size are multiple of the FIFO * minimum depth size */ ret = snd_pcm_hw_constraint_step(ss->runtime, 0, SNDRV_PCM_HW_PARAM_BUFFER_BYTES, AXG_FIFO_MIN_DEPTH); if (ret) return ret; ret = snd_pcm_hw_constraint_step(ss->runtime, 0, SNDRV_PCM_HW_PARAM_PERIOD_BYTES, AXG_FIFO_MIN_DEPTH); if (ret) return ret; ret = request_irq(fifo->irq, axg_fifo_pcm_irq_block, 0, dev_name(dev), ss); if (ret) return ret; /* Enable pclk to access registers and clock the fifo ip */ ret = clk_prepare_enable(fifo->pclk); if (ret) goto free_irq; /* Setup status2 so it reports the memory pointer */ regmap_update_bits(fifo->map, FIFO_CTRL1, CTRL1_STATUS2_SEL_MASK, CTRL1_STATUS2_SEL(STATUS2_SEL_DDR_READ)); /* Make sure the dma is initially disabled */ __dma_enable(fifo, false); /* Disable irqs until params are ready */ regmap_update_bits(fifo->map, FIFO_CTRL0, CTRL0_INT_EN(FIFO_INT_MASK), 0); /* Clear any pending interrupt */ axg_fifo_ack_irq(fifo, FIFO_INT_MASK); /* Take memory arbitror out of reset */ ret = reset_control_deassert(fifo->arb); if (ret) goto free_clk; return 0; free_clk: clk_disable_unprepare(fifo->pclk); free_irq: free_irq(fifo->irq, ss); return ret; } static int axg_fifo_pcm_close(struct snd_pcm_substream *ss) { struct axg_fifo *fifo = axg_fifo_data(ss); int ret; /* Put the memory arbitror back in reset */ ret = reset_control_assert(fifo->arb); /* Disable fifo ip and register access */ clk_disable_unprepare(fifo->pclk); /* remove IRQ */ free_irq(fifo->irq, ss); return ret; } const struct snd_pcm_ops axg_fifo_pcm_ops = { .open = axg_fifo_pcm_open, .close = axg_fifo_pcm_close, .ioctl = snd_pcm_lib_ioctl, .hw_params = axg_fifo_pcm_hw_params, .hw_free = axg_fifo_pcm_hw_free, .pointer = axg_fifo_pcm_pointer, .trigger = axg_fifo_pcm_trigger, }; EXPORT_SYMBOL_GPL(axg_fifo_pcm_ops); int axg_fifo_pcm_new(struct snd_soc_pcm_runtime *rtd, unsigned int type) { struct snd_card *card = rtd->card->snd_card; size_t size = axg_fifo_hw.buffer_bytes_max; return snd_pcm_lib_preallocate_pages(rtd->pcm->streams[type].substream, SNDRV_DMA_TYPE_DEV, card->dev, size, size); } EXPORT_SYMBOL_GPL(axg_fifo_pcm_new); static const struct regmap_config axg_fifo_regmap_cfg = { .reg_bits = 32, .val_bits = 32, .reg_stride = 4, .max_register = FIFO_STATUS2, }; int axg_fifo_probe(struct platform_device *pdev) { struct device *dev = &pdev->dev; const struct axg_fifo_match_data *data; struct axg_fifo *fifo; struct resource *res; void __iomem *regs; data = of_device_get_match_data(dev); if (!data) { dev_err(dev, "failed to match device\n"); return -ENODEV; } fifo = devm_kzalloc(dev, sizeof(*fifo), GFP_KERNEL); if (!fifo) return -ENOMEM; platform_set_drvdata(pdev, fifo); res = platform_get_resource(pdev, IORESOURCE_MEM, 0); regs = devm_ioremap_resource(dev, res); if (IS_ERR(regs)) return PTR_ERR(regs); fifo->map = devm_regmap_init_mmio(dev, regs, &axg_fifo_regmap_cfg); if (IS_ERR(fifo->map)) { dev_err(dev, "failed to init regmap: %ld\n", PTR_ERR(fifo->map)); return PTR_ERR(fifo->map); } fifo->pclk = devm_clk_get(dev, NULL); if (IS_ERR(fifo->pclk)) { if (PTR_ERR(fifo->pclk) != -EPROBE_DEFER) dev_err(dev, "failed to get pclk: %ld\n", PTR_ERR(fifo->pclk)); return PTR_ERR(fifo->pclk); } fifo->arb = devm_reset_control_get_exclusive(dev, NULL); if (IS_ERR(fifo->arb)) { if (PTR_ERR(fifo->arb) != -EPROBE_DEFER) dev_err(dev, "failed to get arb reset: %ld\n", PTR_ERR(fifo->arb)); return PTR_ERR(fifo->arb); } fifo->irq = of_irq_get(dev->of_node, 0); if (fifo->irq <= 0) { dev_err(dev, "failed to get irq: %d\n", fifo->irq); return fifo->irq; } return devm_snd_soc_register_component(dev, data->component_drv, data->dai_drv, 1); } EXPORT_SYMBOL_GPL(axg_fifo_probe); MODULE_DESCRIPTION("Amlogic AXG fifo driver"); MODULE_AUTHOR("Jerome Brunet "); MODULE_LICENSE("GPL v2");