mirror of
https://source.denx.de/u-boot/u-boot.git
synced 2025-12-19 16:31:27 +01:00
Currently xilinx_pm_request API supports four u32 payloads. However the legacy SMC format supports five u32 request payloads and extended SMC format supports six u32 request payloads. Add support for the same in xilinx_pm_request API. Also add two dummy arguments to all the callers of xilinx_pm_request. The TF-A always fills seven u32 return payload so add support for the same in xilinx_pm_request API. Signed-off-by: Naman Trivedi <naman.trivedimanojbhai@amd.com> Signed-off-by: Venkatesh Yadav Abbarapu <venkatesh.abbarapu@amd.com> Acked-by: Senthil Nathan Thangaraj <senthilnathan.thangaraj@amd.com> Signed-off-by: Michal Simek <michal.simek@amd.com> Link: https://lore.kernel.org/r/5ae6b560741f3ca8b89059c4ebb87acf75b4718e.1756388537.git.michal.simek@amd.com
236 lines
6.4 KiB
C
236 lines
6.4 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* (C) Copyright 2018 Xilinx
|
|
*
|
|
* Cadence QSPI controller DMA operations
|
|
*/
|
|
|
|
#include <clk.h>
|
|
#include <memalign.h>
|
|
#include <wait_bit.h>
|
|
#include <asm/io.h>
|
|
#include <asm/gpio.h>
|
|
#include <asm/cache.h>
|
|
#include <cpu_func.h>
|
|
#include <zynqmp_firmware.h>
|
|
#include <asm/arch/hardware.h>
|
|
#include "cadence_qspi.h"
|
|
#include <dt-bindings/power/xlnx-versal-power.h>
|
|
|
|
int cadence_qspi_apb_dma_read(struct cadence_spi_priv *priv,
|
|
const struct spi_mem_op *op)
|
|
{
|
|
u32 reg, ret, rx_rem, n_rx, bytes_to_dma, data, status;
|
|
u8 opcode, addr_bytes, *rxbuf, dummy_cycles;
|
|
|
|
n_rx = op->data.nbytes;
|
|
|
|
if (op->addr.dtr && (op->addr.val % 2)) {
|
|
n_rx += 1;
|
|
writel(op->addr.val & ~0x1,
|
|
priv->regbase + CQSPI_REG_INDIRECTRDSTARTADDR);
|
|
}
|
|
|
|
rxbuf = op->data.buf.in;
|
|
rx_rem = n_rx % 4;
|
|
bytes_to_dma = n_rx - rx_rem;
|
|
|
|
if (bytes_to_dma) {
|
|
cadence_qspi_apb_enable_linear_mode(false);
|
|
reg = readl(priv->regbase + CQSPI_REG_CONFIG);
|
|
reg |= CQSPI_REG_CONFIG_ENBL_DMA;
|
|
writel(reg, priv->regbase + CQSPI_REG_CONFIG);
|
|
|
|
writel(bytes_to_dma, priv->regbase + CQSPI_REG_INDIRECTRDBYTES);
|
|
|
|
writel(CQSPI_DFLT_INDIR_TRIG_ADDR_RANGE,
|
|
priv->regbase + CQSPI_REG_INDIR_TRIG_ADDR_RANGE);
|
|
writel(CQSPI_DFLT_DMA_PERIPH_CFG,
|
|
priv->regbase + CQSPI_REG_DMA_PERIPH_CFG);
|
|
writel(lower_32_bits((unsigned long)rxbuf), priv->regbase +
|
|
CQSPI_DMA_DST_ADDR_REG);
|
|
writel(upper_32_bits((unsigned long)rxbuf), priv->regbase +
|
|
CQSPI_DMA_DST_ADDR_MSB_REG);
|
|
writel(priv->trigger_address, priv->regbase +
|
|
CQSPI_DMA_SRC_RD_ADDR_REG);
|
|
writel(bytes_to_dma, priv->regbase +
|
|
CQSPI_DMA_DST_SIZE_REG);
|
|
flush_dcache_range((unsigned long)rxbuf,
|
|
(unsigned long)rxbuf + bytes_to_dma);
|
|
writel(CQSPI_DFLT_DST_CTRL_REG_VAL,
|
|
priv->regbase + CQSPI_DMA_DST_CTRL_REG);
|
|
|
|
/* Start the indirect read transfer */
|
|
writel(CQSPI_REG_INDIRECTRD_START, priv->regbase +
|
|
CQSPI_REG_INDIRECTRD);
|
|
/* Wait for dma to complete transfer */
|
|
ret = cadence_qspi_apb_wait_for_dma_cmplt(priv);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* Clear indirect completion status */
|
|
writel(CQSPI_REG_INDIRECTRD_DONE, priv->regbase +
|
|
CQSPI_REG_INDIRECTRD);
|
|
rxbuf += bytes_to_dma;
|
|
}
|
|
|
|
if (rx_rem) {
|
|
reg = readl(priv->regbase + CQSPI_REG_CONFIG);
|
|
reg &= ~CQSPI_REG_CONFIG_ENBL_DMA;
|
|
writel(reg, priv->regbase + CQSPI_REG_CONFIG);
|
|
|
|
reg = readl(priv->regbase + CQSPI_REG_INDIRECTRDSTARTADDR);
|
|
reg += bytes_to_dma;
|
|
writel(reg, priv->regbase + CQSPI_REG_CMDADDRESS);
|
|
|
|
addr_bytes = readl(priv->regbase + CQSPI_REG_SIZE) &
|
|
CQSPI_REG_SIZE_ADDRESS_MASK;
|
|
|
|
opcode = CMD_4BYTE_FAST_READ;
|
|
|
|
/* Set up command opcode extension. */
|
|
status = readl(priv->regbase + CQSPI_REG_CONFIG);
|
|
if (status & CQSPI_REG_CONFIG_DTR_PROTO) {
|
|
ret = cadence_qspi_setup_opcode_ext(priv, op,
|
|
CQSPI_REG_OP_EXT_STIG_LSB);
|
|
if (ret)
|
|
return ret;
|
|
}
|
|
|
|
dummy_cycles = 8;
|
|
writel((dummy_cycles << CQSPI_REG_RD_INSTR_DUMMY_LSB) | opcode,
|
|
priv->regbase + CQSPI_REG_RD_INSTR);
|
|
|
|
reg = opcode << CQSPI_REG_CMDCTRL_OPCODE_LSB;
|
|
reg |= (0x1 << CQSPI_REG_CMDCTRL_RD_EN_LSB);
|
|
reg |= (addr_bytes & CQSPI_REG_CMDCTRL_ADD_BYTES_MASK) <<
|
|
CQSPI_REG_CMDCTRL_ADD_BYTES_LSB;
|
|
reg |= (0x1 << CQSPI_REG_CMDCTRL_ADDR_EN_LSB);
|
|
dummy_cycles = (readl(priv->regbase + CQSPI_REG_RD_INSTR) >>
|
|
CQSPI_REG_RD_INSTR_DUMMY_LSB) &
|
|
CQSPI_REG_RD_INSTR_DUMMY_MASK;
|
|
reg |= (dummy_cycles & CQSPI_REG_CMDCTRL_DUMMY_MASK) <<
|
|
CQSPI_REG_CMDCTRL_DUMMY_LSB;
|
|
reg |= (((rx_rem - 1) & CQSPI_REG_CMDCTRL_RD_BYTES_MASK) <<
|
|
CQSPI_REG_CMDCTRL_RD_BYTES_LSB);
|
|
ret = cadence_qspi_apb_exec_flash_cmd(priv->regbase, reg);
|
|
if (ret)
|
|
return ret;
|
|
|
|
data = readl(priv->regbase + CQSPI_REG_CMDREADDATALOWER);
|
|
memcpy(rxbuf, &data, rx_rem);
|
|
}
|
|
|
|
if (op->addr.dtr && (op->addr.val % 2)) {
|
|
rxbuf -= bytes_to_dma;
|
|
memcpy(rxbuf, rxbuf + 1, n_rx - 1);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
int cadence_qspi_apb_wait_for_dma_cmplt(struct cadence_spi_priv *priv)
|
|
{
|
|
u32 timeout = CQSPI_DMA_TIMEOUT;
|
|
|
|
while (!(readl(priv->regbase + CQSPI_DMA_DST_I_STS_REG) &
|
|
CQSPI_DMA_DST_I_STS_DONE) && timeout--)
|
|
udelay(1);
|
|
|
|
if (!timeout) {
|
|
printf("DMA timeout\n");
|
|
return -ETIMEDOUT;
|
|
}
|
|
|
|
writel(readl(priv->regbase + CQSPI_DMA_DST_I_STS_REG),
|
|
priv->regbase + CQSPI_DMA_DST_I_STS_REG);
|
|
return 0;
|
|
}
|
|
|
|
#if !CONFIG_IS_ENABLED(DM_GPIO)
|
|
int cadence_qspi_flash_reset(struct udevice *dev)
|
|
{
|
|
/* CRP WPROT */
|
|
writel(0, WPROT_CRP);
|
|
/* GPIO Reset */
|
|
writel(0, RST_GPIO);
|
|
|
|
/* disable IOU write protection */
|
|
writel(0, WPROT_LPD_MIO);
|
|
|
|
/* set direction as output */
|
|
writel((readl(BOOT_MODE_DIR) | BIT(FLASH_RESET_GPIO)),
|
|
BOOT_MODE_DIR);
|
|
|
|
/* Data output enable */
|
|
writel((readl(BOOT_MODE_OUT) | BIT(FLASH_RESET_GPIO)),
|
|
BOOT_MODE_OUT);
|
|
|
|
/* IOU SLCR write enable */
|
|
writel(0, WPROT_PMC_MIO);
|
|
|
|
/* set MIO as GPIO */
|
|
writel(0x60, MIO_PIN_12);
|
|
|
|
/* Set value 1 to pin */
|
|
writel((readl(BANK0_OUTPUT) | BIT(FLASH_RESET_GPIO)), BANK0_OUTPUT);
|
|
udelay(10);
|
|
|
|
/* Disable Tri-state */
|
|
writel((readl(BANK0_TRI) & ~BIT(FLASH_RESET_GPIO)), BANK0_TRI);
|
|
udelay(1);
|
|
|
|
/* Set value 0 to pin */
|
|
writel((readl(BANK0_OUTPUT) & ~BIT(FLASH_RESET_GPIO)), BANK0_OUTPUT);
|
|
udelay(10);
|
|
|
|
/* Set value 1 to pin */
|
|
writel((readl(BANK0_OUTPUT) | BIT(FLASH_RESET_GPIO)), BANK0_OUTPUT);
|
|
udelay(10);
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
void cadence_qspi_apb_enable_linear_mode(bool enable)
|
|
{
|
|
if (IS_ENABLED(CONFIG_ZYNQMP_FIRMWARE)) {
|
|
if (enable)
|
|
/* ahb read mode */
|
|
xilinx_pm_request(PM_IOCTL, PM_DEV_OSPI,
|
|
IOCTL_OSPI_MUX_SELECT,
|
|
PM_OSPI_MUX_SEL_LINEAR, 0, 0, 0, NULL);
|
|
else
|
|
/* DMA mode */
|
|
xilinx_pm_request(PM_IOCTL, PM_DEV_OSPI,
|
|
IOCTL_OSPI_MUX_SELECT,
|
|
PM_OSPI_MUX_SEL_DMA, 0, 0, 0, NULL);
|
|
} else {
|
|
if (enable)
|
|
writel(readl(VERSAL_AXI_MUX_SEL) |
|
|
VERSAL_OSPI_LINEAR_MODE, VERSAL_AXI_MUX_SEL);
|
|
else
|
|
writel(readl(VERSAL_AXI_MUX_SEL) &
|
|
~VERSAL_OSPI_LINEAR_MODE, VERSAL_AXI_MUX_SEL);
|
|
}
|
|
}
|
|
|
|
int cadence_device_reset(struct udevice *bus)
|
|
{
|
|
struct cadence_spi_priv *priv = dev_get_priv(bus);
|
|
u32 reg;
|
|
|
|
reg = readl(priv->regbase + CQSPI_REG_CONFIG);
|
|
reg |= CQSPI_REG_CONFIG_RESET_CFG_FLD_MASK;
|
|
writel(reg, priv->regbase + CQSPI_REG_CONFIG);
|
|
|
|
writel(reg & ~CQSPI_REG_CONFIG_RESET_PIN_FLD_MASK, priv->regbase + CQSPI_REG_CONFIG);
|
|
udelay(5);
|
|
writel(reg | CQSPI_REG_CONFIG_RESET_PIN_FLD_MASK, priv->regbase + CQSPI_REG_CONFIG);
|
|
udelay(150);
|
|
writel(reg & ~CQSPI_REG_CONFIG_RESET_PIN_FLD_MASK, priv->regbase + CQSPI_REG_CONFIG);
|
|
udelay(1200);
|
|
|
|
return 0;
|
|
}
|