• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2005-2017 Andes Technology Corporation
3 
4 #include <linux/cpu.h>
5 #include <linux/memblock.h>
6 #include <linux/seq_file.h>
7 #include <linux/console.h>
8 #include <linux/screen_info.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/of_fdt.h>
12 #include <linux/of_platform.h>
13 #include <asm/setup.h>
14 #include <asm/sections.h>
15 #include <asm/proc-fns.h>
16 #include <asm/cache_info.h>
17 #include <asm/elf.h>
18 #include <asm/fpu.h>
19 #include <nds32_intrinsic.h>
20 
21 #define HWCAP_MFUSR_PC		0x000001
22 #define HWCAP_EXT		0x000002
23 #define HWCAP_EXT2		0x000004
24 #define HWCAP_FPU		0x000008
25 #define HWCAP_AUDIO		0x000010
26 #define HWCAP_BASE16		0x000020
27 #define HWCAP_STRING		0x000040
28 #define HWCAP_REDUCED_REGS	0x000080
29 #define HWCAP_VIDEO		0x000100
30 #define HWCAP_ENCRYPT		0x000200
31 #define HWCAP_EDM		0x000400
32 #define HWCAP_LMDMA		0x000800
33 #define HWCAP_PFM		0x001000
34 #define HWCAP_HSMP		0x002000
35 #define HWCAP_TRACE		0x004000
36 #define HWCAP_DIV		0x008000
37 #define HWCAP_MAC		0x010000
38 #define HWCAP_L2C		0x020000
39 #define HWCAP_FPU_DP		0x040000
40 #define HWCAP_V2		0x080000
41 #define HWCAP_DX_REGS		0x100000
42 #define HWCAP_HWPRE		0x200000
43 
44 unsigned long cpu_id, cpu_rev, cpu_cfgid;
45 bool has_fpu = false;
46 char cpu_series;
47 char *endianness = NULL;
48 
49 unsigned int __atags_pointer __initdata;
50 unsigned int elf_hwcap;
51 EXPORT_SYMBOL(elf_hwcap);
52 
53 /*
54  * The following string table, must sync with HWCAP_xx bitmask,
55  * which is defined in <asm/procinfo.h>
56  */
57 static const char *hwcap_str[] = {
58 	"mfusr_pc",
59 	"perf1",
60 	"perf2",
61 	"fpu",
62 	"audio",
63 	"16b",
64 	"string",
65 	"reduced_regs",
66 	"video",
67 	"encrypt",
68 	"edm",
69 	"lmdma",
70 	"pfm",
71 	"hsmp",
72 	"trace",
73 	"div",
74 	"mac",
75 	"l2c",
76 	"fpu_dp",
77 	"v2",
78 	"dx_regs",
79 	"hw_pre",
80 	NULL,
81 };
82 
83 #ifdef CONFIG_CPU_DCACHE_WRITETHROUGH
84 #define WRITE_METHOD "write through"
85 #else
86 #define WRITE_METHOD "write back"
87 #endif
88 
89 struct cache_info L1_cache_info[2];
dump_cpu_info(int cpu)90 static void __init dump_cpu_info(int cpu)
91 {
92 	int i, p = 0;
93 	char str[sizeof(hwcap_str) + 16];
94 
95 	for (i = 0; hwcap_str[i]; i++) {
96 		if (elf_hwcap & (1 << i)) {
97 			sprintf(str + p, "%s ", hwcap_str[i]);
98 			p += strlen(hwcap_str[i]) + 1;
99 		}
100 	}
101 
102 	pr_info("CPU%d Features: %s\n", cpu, str);
103 
104 	L1_cache_info[ICACHE].ways = CACHE_WAY(ICACHE);
105 	L1_cache_info[ICACHE].line_size = CACHE_LINE_SIZE(ICACHE);
106 	L1_cache_info[ICACHE].sets = CACHE_SET(ICACHE);
107 	L1_cache_info[ICACHE].size =
108 	    L1_cache_info[ICACHE].ways * L1_cache_info[ICACHE].line_size *
109 	    L1_cache_info[ICACHE].sets / 1024;
110 	pr_info("L1I:%dKB/%dS/%dW/%dB\n", L1_cache_info[ICACHE].size,
111 		L1_cache_info[ICACHE].sets, L1_cache_info[ICACHE].ways,
112 		L1_cache_info[ICACHE].line_size);
113 	L1_cache_info[DCACHE].ways = CACHE_WAY(DCACHE);
114 	L1_cache_info[DCACHE].line_size = CACHE_LINE_SIZE(DCACHE);
115 	L1_cache_info[DCACHE].sets = CACHE_SET(DCACHE);
116 	L1_cache_info[DCACHE].size =
117 	    L1_cache_info[DCACHE].ways * L1_cache_info[DCACHE].line_size *
118 	    L1_cache_info[DCACHE].sets / 1024;
119 	pr_info("L1D:%dKB/%dS/%dW/%dB\n", L1_cache_info[DCACHE].size,
120 		L1_cache_info[DCACHE].sets, L1_cache_info[DCACHE].ways,
121 		L1_cache_info[DCACHE].line_size);
122 	pr_info("L1 D-Cache is %s\n", WRITE_METHOD);
123 	if (L1_cache_info[DCACHE].size != L1_CACHE_BYTES)
124 		pr_crit
125 		    ("The cache line size(%d) of this processor is not the same as L1_CACHE_BYTES(%d).\n",
126 		     L1_cache_info[DCACHE].size, L1_CACHE_BYTES);
127 #ifdef CONFIG_CPU_CACHE_ALIASING
128 	{
129 		int aliasing_num;
130 		aliasing_num =
131 		    L1_cache_info[ICACHE].size * 1024 / PAGE_SIZE /
132 		    L1_cache_info[ICACHE].ways;
133 		L1_cache_info[ICACHE].aliasing_num = aliasing_num;
134 		L1_cache_info[ICACHE].aliasing_mask =
135 		    (aliasing_num - 1) << PAGE_SHIFT;
136 		aliasing_num =
137 		    L1_cache_info[DCACHE].size * 1024 / PAGE_SIZE /
138 		    L1_cache_info[DCACHE].ways;
139 		L1_cache_info[DCACHE].aliasing_num = aliasing_num;
140 		L1_cache_info[DCACHE].aliasing_mask =
141 		    (aliasing_num - 1) << PAGE_SHIFT;
142 	}
143 #endif
144 #ifdef CONFIG_FPU
145 	/* Disable fpu and enable when it is used. */
146 	if (has_fpu)
147 		disable_fpu();
148 #endif
149 }
150 
setup_cpuinfo(void)151 static void __init setup_cpuinfo(void)
152 {
153 	unsigned long tmp = 0, cpu_name;
154 
155 	cpu_dcache_inval_all();
156 	cpu_icache_inval_all();
157 	__nds32__isb();
158 
159 	cpu_id = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskCPUID) >> CPU_VER_offCPUID;
160 	cpu_name = ((cpu_id) & 0xf0) >> 4;
161 	cpu_series = cpu_name ? cpu_name - 10 + 'A' : 'N';
162 	cpu_id = cpu_id & 0xf;
163 	cpu_rev = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskREV) >> CPU_VER_offREV;
164 	cpu_cfgid = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskCFGID) >> CPU_VER_offCFGID;
165 
166 	pr_info("CPU:%c%ld, CPU_VER 0x%08x(id %lu, rev %lu, cfg %lu)\n",
167 		cpu_series, cpu_id, __nds32__mfsr(NDS32_SR_CPU_VER), cpu_id, cpu_rev, cpu_cfgid);
168 
169 	elf_hwcap |= HWCAP_MFUSR_PC;
170 
171 	if (((__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskBASEV) >> MSC_CFG_offBASEV) == 0) {
172 		if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskDIV)
173 			elf_hwcap |= HWCAP_DIV;
174 
175 		if ((__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskMAC)
176 		    || (cpu_id == 12 && cpu_rev < 4))
177 			elf_hwcap |= HWCAP_MAC;
178 	} else {
179 		elf_hwcap |= HWCAP_V2;
180 		elf_hwcap |= HWCAP_DIV;
181 		elf_hwcap |= HWCAP_MAC;
182 	}
183 
184 	if (cpu_cfgid & 0x0001)
185 		elf_hwcap |= HWCAP_EXT;
186 
187 	if (cpu_cfgid & 0x0002)
188 		elf_hwcap |= HWCAP_BASE16;
189 
190 	if (cpu_cfgid & 0x0004)
191 		elf_hwcap |= HWCAP_EXT2;
192 
193 	if (cpu_cfgid & 0x0008) {
194 		elf_hwcap |= HWCAP_FPU;
195 		has_fpu = true;
196 	}
197 	if (cpu_cfgid & 0x0010)
198 		elf_hwcap |= HWCAP_STRING;
199 
200 	if (__nds32__mfsr(NDS32_SR_MMU_CFG) & MMU_CFG_mskDE)
201 		endianness = "MSB";
202 	else
203 		endianness = "LSB";
204 
205 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskEDM)
206 		elf_hwcap |= HWCAP_EDM;
207 
208 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskLMDMA)
209 		elf_hwcap |= HWCAP_LMDMA;
210 
211 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskPFM)
212 		elf_hwcap |= HWCAP_PFM;
213 
214 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskHSMP)
215 		elf_hwcap |= HWCAP_HSMP;
216 
217 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskTRACE)
218 		elf_hwcap |= HWCAP_TRACE;
219 
220 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskAUDIO)
221 		elf_hwcap |= HWCAP_AUDIO;
222 
223 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskL2C)
224 		elf_hwcap |= HWCAP_L2C;
225 
226 #ifdef CONFIG_HW_PRE
227 	if (__nds32__mfsr(NDS32_SR_MISC_CTL) & MISC_CTL_makHWPRE_EN)
228 		elf_hwcap |= HWCAP_HWPRE;
229 #endif
230 
231 	tmp = __nds32__mfsr(NDS32_SR_CACHE_CTL);
232 	if (!IS_ENABLED(CONFIG_CPU_DCACHE_DISABLE))
233 		tmp |= CACHE_CTL_mskDC_EN;
234 
235 	if (!IS_ENABLED(CONFIG_CPU_ICACHE_DISABLE))
236 		tmp |= CACHE_CTL_mskIC_EN;
237 	__nds32__mtsr_isb(tmp, NDS32_SR_CACHE_CTL);
238 
239 	dump_cpu_info(smp_processor_id());
240 }
241 
setup_memory(void)242 static void __init setup_memory(void)
243 {
244 	unsigned long ram_start_pfn;
245 	unsigned long free_ram_start_pfn;
246 	phys_addr_t memory_start, memory_end;
247 	struct memblock_region *region;
248 
249 	memory_end = memory_start = 0;
250 
251 	/* Find main memory where is the kernel */
252 	memory_start = memblock_start_of_DRAM();
253 	memory_end = memblock_end_of_DRAM();
254 
255 	if (!memory_end) {
256 		panic("No memory!");
257 	}
258 
259 	ram_start_pfn = PFN_UP(memblock_start_of_DRAM());
260 	/* free_ram_start_pfn is first page after kernel */
261 	free_ram_start_pfn = PFN_UP(__pa(&_end));
262 	max_pfn = PFN_DOWN(memblock_end_of_DRAM());
263 	/* it could update max_pfn */
264 	if (max_pfn - ram_start_pfn <= MAXMEM_PFN)
265 		max_low_pfn = max_pfn;
266 	else {
267 		max_low_pfn = MAXMEM_PFN + ram_start_pfn;
268 		if (!IS_ENABLED(CONFIG_HIGHMEM))
269 			max_pfn = MAXMEM_PFN + ram_start_pfn;
270 	}
271 	/* high_memory is related with VMALLOC */
272 	high_memory = (void *)__va(max_low_pfn * PAGE_SIZE);
273 	min_low_pfn = free_ram_start_pfn;
274 
275 	/*
276 	 * initialize the boot-time allocator (with low memory only).
277 	 *
278 	 * This makes the memory from the end of the kernel to the end of
279 	 * RAM usable.
280 	 */
281 	memblock_set_bottom_up(true);
282 	memblock_reserve(PFN_PHYS(ram_start_pfn), PFN_PHYS(free_ram_start_pfn - ram_start_pfn));
283 
284 	early_init_fdt_reserve_self();
285 	early_init_fdt_scan_reserved_mem();
286 
287 	memblock_dump_all();
288 }
289 
setup_arch(char ** cmdline_p)290 void __init setup_arch(char **cmdline_p)
291 {
292 	early_init_devtree(__atags_pointer ? \
293 		phys_to_virt(__atags_pointer) : __dtb_start);
294 
295 	setup_cpuinfo();
296 
297 	init_mm.start_code = (unsigned long)&_stext;
298 	init_mm.end_code = (unsigned long)&_etext;
299 	init_mm.end_data = (unsigned long)&_edata;
300 	init_mm.brk = (unsigned long)&_end;
301 
302 	/* setup bootmem allocator */
303 	setup_memory();
304 
305 	/* paging_init() sets up the MMU and marks all pages as reserved */
306 	paging_init();
307 
308 	/* invalidate all TLB entries because the new mapping is created */
309 	__nds32__tlbop_flua();
310 
311 	/* use generic way to parse */
312 	parse_early_param();
313 
314 	unflatten_and_copy_device_tree();
315 
316 	*cmdline_p = boot_command_line;
317 	early_trap_init();
318 }
319 
c_show(struct seq_file * m,void * v)320 static int c_show(struct seq_file *m, void *v)
321 {
322 	int i;
323 
324 	seq_printf(m, "Processor\t: %c%ld (id %lu, rev %lu, cfg %lu)\n",
325 		   cpu_series, cpu_id, cpu_id, cpu_rev, cpu_cfgid);
326 
327 	seq_printf(m, "L1I\t\t: %luKB/%luS/%luW/%luB\n",
328 		   CACHE_SET(ICACHE) * CACHE_WAY(ICACHE) *
329 		   CACHE_LINE_SIZE(ICACHE) / 1024, CACHE_SET(ICACHE),
330 		   CACHE_WAY(ICACHE), CACHE_LINE_SIZE(ICACHE));
331 
332 	seq_printf(m, "L1D\t\t: %luKB/%luS/%luW/%luB\n",
333 		   CACHE_SET(DCACHE) * CACHE_WAY(DCACHE) *
334 		   CACHE_LINE_SIZE(DCACHE) / 1024, CACHE_SET(DCACHE),
335 		   CACHE_WAY(DCACHE), CACHE_LINE_SIZE(DCACHE));
336 
337 	seq_printf(m, "BogoMIPS\t: %lu.%02lu\n",
338 		   loops_per_jiffy / (500000 / HZ),
339 		   (loops_per_jiffy / (5000 / HZ)) % 100);
340 
341 	/* dump out the processor features */
342 	seq_puts(m, "Features\t: ");
343 
344 	for (i = 0; hwcap_str[i]; i++)
345 		if (elf_hwcap & (1 << i))
346 			seq_printf(m, "%s ", hwcap_str[i]);
347 
348 	seq_puts(m, "\n\n");
349 
350 	return 0;
351 }
352 
c_start(struct seq_file * m,loff_t * pos)353 static void *c_start(struct seq_file *m, loff_t * pos)
354 {
355 	return *pos < 1 ? (void *)1 : NULL;
356 }
357 
c_next(struct seq_file * m,void * v,loff_t * pos)358 static void *c_next(struct seq_file *m, void *v, loff_t * pos)
359 {
360 	++*pos;
361 	return NULL;
362 }
363 
c_stop(struct seq_file * m,void * v)364 static void c_stop(struct seq_file *m, void *v)
365 {
366 }
367 
368 struct seq_operations cpuinfo_op = {
369 	.start = c_start,
370 	.next = c_next,
371 	.stop = c_stop,
372 	.show = c_show
373 };
374