// SPDX-License-Identifier: GPL-2.0+ /* * Copyright (C) 2012-2015 Panasonic Corporation * Copyright (C) 2015-2017 Socionext Inc. * Author: Masahiro Yamada */ #include #include #include #include #include #include #include #include #include "init.h" #include "sg-regs.h" #include "soc-info.h" DECLARE_GLOBAL_DATA_PTR; struct uniphier_dram_map { unsigned long base; unsigned long size; }; static int uniphier_memconf_decode(struct uniphier_dram_map *dram_map, unsigned long sparse_ch1_base, bool have_ch2) { unsigned long size; u32 val; val = readl(sg_base + SG_MEMCONF); /* set up ch0 */ dram_map[0].base = 0x80000000; switch (val & SG_MEMCONF_CH0_SZ_MASK) { case SG_MEMCONF_CH0_SZ_64M: size = SZ_64M; break; case SG_MEMCONF_CH0_SZ_128M: size = SZ_128M; break; case SG_MEMCONF_CH0_SZ_256M: size = SZ_256M; break; case SG_MEMCONF_CH0_SZ_512M: size = SZ_512M; break; case SG_MEMCONF_CH0_SZ_1G: size = SZ_1G; break; default: pr_err("error: invalid value is set to MEMCONF ch0 size\n"); return -EINVAL; } if ((val & SG_MEMCONF_CH0_NUM_MASK) == SG_MEMCONF_CH0_NUM_2) size *= 2; dram_map[0].size = size; /* set up ch1 */ dram_map[1].base = dram_map[0].base + size; if (val & SG_MEMCONF_SPARSEMEM) { if (dram_map[1].base > sparse_ch1_base) { pr_warn("Sparse mem is enabled, but ch0 and ch1 overlap\n"); pr_warn("Only ch0 is available\n"); dram_map[1].base = 0; return 0; } dram_map[1].base = sparse_ch1_base; } switch (val & SG_MEMCONF_CH1_SZ_MASK) { case SG_MEMCONF_CH1_SZ_64M: size = SZ_64M; break; case SG_MEMCONF_CH1_SZ_128M: size = SZ_128M; break; case SG_MEMCONF_CH1_SZ_256M: size = SZ_256M; break; case SG_MEMCONF_CH1_SZ_512M: size = SZ_512M; break; case SG_MEMCONF_CH1_SZ_1G: size = SZ_1G; break; default: pr_err("error: invalid value is set to MEMCONF ch1 size\n"); return -EINVAL; } if ((val & SG_MEMCONF_CH1_NUM_MASK) == SG_MEMCONF_CH1_NUM_2) size *= 2; dram_map[1].size = size; if (!have_ch2 || val & SG_MEMCONF_CH2_DISABLE) return 0; /* set up ch2 */ dram_map[2].base = dram_map[1].base + size; switch (val & SG_MEMCONF_CH2_SZ_MASK) { case SG_MEMCONF_CH2_SZ_64M: size = SZ_64M; break; case SG_MEMCONF_CH2_SZ_128M: size = SZ_128M; break; case SG_MEMCONF_CH2_SZ_256M: size = SZ_256M; break; case SG_MEMCONF_CH2_SZ_512M: size = SZ_512M; break; case SG_MEMCONF_CH2_SZ_1G: size = SZ_1G; break; default: pr_err("error: invalid value is set to MEMCONF ch2 size\n"); return -EINVAL; } if ((val & SG_MEMCONF_CH2_NUM_MASK) == SG_MEMCONF_CH2_NUM_2) size *= 2; dram_map[2].size = size; return 0; } static int uniphier_ld4_dram_map_get(struct uniphier_dram_map dram_map[]) { return uniphier_memconf_decode(dram_map, 0xc0000000, false); } static int uniphier_pro4_dram_map_get(struct uniphier_dram_map dram_map[]) { return uniphier_memconf_decode(dram_map, 0xa0000000, false); } static int uniphier_pxs2_dram_map_get(struct uniphier_dram_map dram_map[]) { return uniphier_memconf_decode(dram_map, 0xc0000000, true); } struct uniphier_dram_init_data { unsigned int soc_id; int (*dram_map_get)(struct uniphier_dram_map dram_map[]); }; static const struct uniphier_dram_init_data uniphier_dram_init_data[] = { { .soc_id = UNIPHIER_LD4_ID, .dram_map_get = uniphier_ld4_dram_map_get, }, { .soc_id = UNIPHIER_PRO4_ID, .dram_map_get = uniphier_pro4_dram_map_get, }, { .soc_id = UNIPHIER_SLD8_ID, .dram_map_get = uniphier_ld4_dram_map_get, }, { .soc_id = UNIPHIER_PRO5_ID, .dram_map_get = uniphier_ld4_dram_map_get, }, { .soc_id = UNIPHIER_PXS2_ID, .dram_map_get = uniphier_pxs2_dram_map_get, }, { .soc_id = UNIPHIER_LD6B_ID, .dram_map_get = uniphier_pxs2_dram_map_get, }, { .soc_id = UNIPHIER_LD11_ID, .dram_map_get = uniphier_ld4_dram_map_get, }, { .soc_id = UNIPHIER_LD20_ID, .dram_map_get = uniphier_pxs2_dram_map_get, }, { .soc_id = UNIPHIER_PXS3_ID, .dram_map_get = uniphier_pxs2_dram_map_get, }, }; UNIPHIER_DEFINE_SOCDATA_FUNC(uniphier_get_dram_init_data, uniphier_dram_init_data) static int uniphier_dram_map_get(struct uniphier_dram_map *dram_map) { const struct uniphier_dram_init_data *data; data = uniphier_get_dram_init_data(); if (!data) { pr_err("unsupported SoC\n"); return -ENOTSUPP; } return data->dram_map_get(dram_map); } int dram_init(void) { struct uniphier_dram_map dram_map[3] = {}; bool valid_bank_found = false; unsigned long prev_top; int ret, i; gd->ram_size = 0; ret = uniphier_dram_map_get(dram_map); if (ret) return ret; for (i = 0; i < ARRAY_SIZE(dram_map); i++) { unsigned long max_size; if (!dram_map[i].size) continue; /* * U-Boot relocates itself to the tail of the memory region, * but it does not expect sparse memory. We use the first * contiguous chunk here. */ if (valid_bank_found && prev_top < dram_map[i].base) break; /* * Do not use memory that exceeds 32bit address range. U-Boot * relocates itself to the end of the effectively available RAM. * This could be a problem for DMA engines that do not support * 64bit address (SDMA of SDHCI, UniPhier AV-ether, etc.) */ if (dram_map[i].base >= 1ULL << 32) break; max_size = (1ULL << 32) - dram_map[i].base; gd->ram_size = min(dram_map[i].size, max_size); if (!valid_bank_found) gd->ram_base = dram_map[i].base; prev_top = dram_map[i].base + dram_map[i].size; valid_bank_found = true; } /* * LD20 uses the last 64 byte for each channel for dynamic * DDR PHY training */ if (uniphier_get_soc_id() == UNIPHIER_LD20_ID) gd->ram_size -= 64; return 0; } int dram_init_banksize(void) { struct uniphier_dram_map dram_map[3] = {}; unsigned long base, top; bool valid_bank_found = false; int ret, i; ret = uniphier_dram_map_get(dram_map); if (ret) return ret; for (i = 0; i < ARRAY_SIZE(dram_map); i++) { if (i < ARRAY_SIZE(gd->bd->bi_dram)) { gd->bd->bi_dram[i].start = dram_map[i].base; gd->bd->bi_dram[i].size = dram_map[i].size; } if (!dram_map[i].size) continue; if (!valid_bank_found) base = dram_map[i].base; top = dram_map[i].base + dram_map[i].size; valid_bank_found = true; } if (!valid_bank_found) return -EINVAL; /* map all the DRAM regions */ uniphier_mem_map_init(base, top - base); return 0; }