• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 // SPDX-License-Identifier: GPL-2.0
2 // Copyright (C) 2005-2017 Andes Technology Corporation
3 
4 #include <linux/cpu.h>
5 #include <linux/memblock.h>
6 #include <linux/seq_file.h>
7 #include <linux/console.h>
8 #include <linux/screen_info.h>
9 #include <linux/delay.h>
10 #include <linux/dma-mapping.h>
11 #include <linux/of_fdt.h>
12 #include <linux/of_platform.h>
13 #include <asm/setup.h>
14 #include <asm/sections.h>
15 #include <asm/proc-fns.h>
16 #include <asm/cache_info.h>
17 #include <asm/elf.h>
18 #include <asm/fpu.h>
19 #include <nds32_intrinsic.h>
20 
21 #define HWCAP_MFUSR_PC		0x000001
22 #define HWCAP_EXT		0x000002
23 #define HWCAP_EXT2		0x000004
24 #define HWCAP_FPU		0x000008
25 #define HWCAP_AUDIO		0x000010
26 #define HWCAP_BASE16		0x000020
27 #define HWCAP_STRING		0x000040
28 #define HWCAP_REDUCED_REGS	0x000080
29 #define HWCAP_VIDEO		0x000100
30 #define HWCAP_ENCRYPT		0x000200
31 #define HWCAP_EDM		0x000400
32 #define HWCAP_LMDMA		0x000800
33 #define HWCAP_PFM		0x001000
34 #define HWCAP_HSMP		0x002000
35 #define HWCAP_TRACE		0x004000
36 #define HWCAP_DIV		0x008000
37 #define HWCAP_MAC		0x010000
38 #define HWCAP_L2C		0x020000
39 #define HWCAP_FPU_DP		0x040000
40 #define HWCAP_V2		0x080000
41 #define HWCAP_DX_REGS		0x100000
42 #define HWCAP_HWPRE		0x200000
43 
44 unsigned long cpu_id, cpu_rev, cpu_cfgid;
45 bool has_fpu = false;
46 char cpu_series;
47 char *endianness = NULL;
48 
49 unsigned int __atags_pointer __initdata;
50 unsigned int elf_hwcap;
51 EXPORT_SYMBOL(elf_hwcap);
52 
53 /*
54  * The following string table, must sync with HWCAP_xx bitmask,
55  * which is defined in <asm/procinfo.h>
56  */
57 static const char *hwcap_str[] = {
58 	"mfusr_pc",
59 	"perf1",
60 	"perf2",
61 	"fpu",
62 	"audio",
63 	"16b",
64 	"string",
65 	"reduced_regs",
66 	"video",
67 	"encrypt",
68 	"edm",
69 	"lmdma",
70 	"pfm",
71 	"hsmp",
72 	"trace",
73 	"div",
74 	"mac",
75 	"l2c",
76 	"fpu_dp",
77 	"v2",
78 	"dx_regs",
79 	"hw_pre",
80 	NULL,
81 };
82 
83 #ifdef CONFIG_CPU_DCACHE_WRITETHROUGH
84 #define WRITE_METHOD "write through"
85 #else
86 #define WRITE_METHOD "write back"
87 #endif
88 
89 struct cache_info L1_cache_info[2];
dump_cpu_info(int cpu)90 static void __init dump_cpu_info(int cpu)
91 {
92 	int i, p = 0;
93 	char str[sizeof(hwcap_str) + 16];
94 
95 	for (i = 0; hwcap_str[i]; i++) {
96 		if (elf_hwcap & (1 << i)) {
97 			sprintf(str + p, "%s ", hwcap_str[i]);
98 			p += strlen(hwcap_str[i]) + 1;
99 		}
100 	}
101 
102 	pr_info("CPU%d Features: %s\n", cpu, str);
103 
104 	L1_cache_info[ICACHE].ways = CACHE_WAY(ICACHE);
105 	L1_cache_info[ICACHE].line_size = CACHE_LINE_SIZE(ICACHE);
106 	L1_cache_info[ICACHE].sets = CACHE_SET(ICACHE);
107 	L1_cache_info[ICACHE].size =
108 	    L1_cache_info[ICACHE].ways * L1_cache_info[ICACHE].line_size *
109 	    L1_cache_info[ICACHE].sets / 1024;
110 	pr_info("L1I:%dKB/%dS/%dW/%dB\n", L1_cache_info[ICACHE].size,
111 		L1_cache_info[ICACHE].sets, L1_cache_info[ICACHE].ways,
112 		L1_cache_info[ICACHE].line_size);
113 	L1_cache_info[DCACHE].ways = CACHE_WAY(DCACHE);
114 	L1_cache_info[DCACHE].line_size = CACHE_LINE_SIZE(DCACHE);
115 	L1_cache_info[DCACHE].sets = CACHE_SET(DCACHE);
116 	L1_cache_info[DCACHE].size =
117 	    L1_cache_info[DCACHE].ways * L1_cache_info[DCACHE].line_size *
118 	    L1_cache_info[DCACHE].sets / 1024;
119 	pr_info("L1D:%dKB/%dS/%dW/%dB\n", L1_cache_info[DCACHE].size,
120 		L1_cache_info[DCACHE].sets, L1_cache_info[DCACHE].ways,
121 		L1_cache_info[DCACHE].line_size);
122 	pr_info("L1 D-Cache is %s\n", WRITE_METHOD);
123 	if (L1_cache_info[DCACHE].size != L1_CACHE_BYTES)
124 		pr_crit
125 		    ("The cache line size(%d) of this processor is not the same as L1_CACHE_BYTES(%d).\n",
126 		     L1_cache_info[DCACHE].size, L1_CACHE_BYTES);
127 #ifdef CONFIG_CPU_CACHE_ALIASING
128 	{
129 		int aliasing_num;
130 		aliasing_num =
131 		    L1_cache_info[ICACHE].size * 1024 / PAGE_SIZE /
132 		    L1_cache_info[ICACHE].ways;
133 		L1_cache_info[ICACHE].aliasing_num = aliasing_num;
134 		L1_cache_info[ICACHE].aliasing_mask =
135 		    (aliasing_num - 1) << PAGE_SHIFT;
136 		aliasing_num =
137 		    L1_cache_info[DCACHE].size * 1024 / PAGE_SIZE /
138 		    L1_cache_info[DCACHE].ways;
139 		L1_cache_info[DCACHE].aliasing_num = aliasing_num;
140 		L1_cache_info[DCACHE].aliasing_mask =
141 		    (aliasing_num - 1) << PAGE_SHIFT;
142 	}
143 #endif
144 #ifdef CONFIG_FPU
145 	/* Disable fpu and enable when it is used. */
146 	if (has_fpu)
147 		disable_fpu();
148 #endif
149 }
150 
setup_cpuinfo(void)151 static void __init setup_cpuinfo(void)
152 {
153 	unsigned long tmp = 0, cpu_name;
154 
155 	cpu_dcache_inval_all();
156 	cpu_icache_inval_all();
157 	__nds32__isb();
158 
159 	cpu_id = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskCPUID) >> CPU_VER_offCPUID;
160 	cpu_name = ((cpu_id) & 0xf0) >> 4;
161 	cpu_series = cpu_name ? cpu_name - 10 + 'A' : 'N';
162 	cpu_id = cpu_id & 0xf;
163 	cpu_rev = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskREV) >> CPU_VER_offREV;
164 	cpu_cfgid = (__nds32__mfsr(NDS32_SR_CPU_VER) & CPU_VER_mskCFGID) >> CPU_VER_offCFGID;
165 
166 	pr_info("CPU:%c%ld, CPU_VER 0x%08x(id %lu, rev %lu, cfg %lu)\n",
167 		cpu_series, cpu_id, __nds32__mfsr(NDS32_SR_CPU_VER), cpu_id, cpu_rev, cpu_cfgid);
168 
169 	elf_hwcap |= HWCAP_MFUSR_PC;
170 
171 	if (((__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskBASEV) >> MSC_CFG_offBASEV) == 0) {
172 		if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskDIV)
173 			elf_hwcap |= HWCAP_DIV;
174 
175 		if ((__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskMAC)
176 		    || (cpu_id == 12 && cpu_rev < 4))
177 			elf_hwcap |= HWCAP_MAC;
178 	} else {
179 		elf_hwcap |= HWCAP_V2;
180 		elf_hwcap |= HWCAP_DIV;
181 		elf_hwcap |= HWCAP_MAC;
182 	}
183 
184 	if (cpu_cfgid & 0x0001)
185 		elf_hwcap |= HWCAP_EXT;
186 
187 	if (cpu_cfgid & 0x0002)
188 		elf_hwcap |= HWCAP_BASE16;
189 
190 	if (cpu_cfgid & 0x0004)
191 		elf_hwcap |= HWCAP_EXT2;
192 
193 	if (cpu_cfgid & 0x0008) {
194 		elf_hwcap |= HWCAP_FPU;
195 		has_fpu = true;
196 	}
197 	if (cpu_cfgid & 0x0010)
198 		elf_hwcap |= HWCAP_STRING;
199 
200 	if (__nds32__mfsr(NDS32_SR_MMU_CFG) & MMU_CFG_mskDE)
201 		endianness = "MSB";
202 	else
203 		endianness = "LSB";
204 
205 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskEDM)
206 		elf_hwcap |= HWCAP_EDM;
207 
208 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskLMDMA)
209 		elf_hwcap |= HWCAP_LMDMA;
210 
211 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskPFM)
212 		elf_hwcap |= HWCAP_PFM;
213 
214 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskHSMP)
215 		elf_hwcap |= HWCAP_HSMP;
216 
217 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskTRACE)
218 		elf_hwcap |= HWCAP_TRACE;
219 
220 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskAUDIO)
221 		elf_hwcap |= HWCAP_AUDIO;
222 
223 	if (__nds32__mfsr(NDS32_SR_MSC_CFG) & MSC_CFG_mskL2C)
224 		elf_hwcap |= HWCAP_L2C;
225 
226 #ifdef CONFIG_HW_PRE
227 	if (__nds32__mfsr(NDS32_SR_MISC_CTL) & MISC_CTL_makHWPRE_EN)
228 		elf_hwcap |= HWCAP_HWPRE;
229 #endif
230 
231 	tmp = __nds32__mfsr(NDS32_SR_CACHE_CTL);
232 	if (!IS_ENABLED(CONFIG_CPU_DCACHE_DISABLE))
233 		tmp |= CACHE_CTL_mskDC_EN;
234 
235 	if (!IS_ENABLED(CONFIG_CPU_ICACHE_DISABLE))
236 		tmp |= CACHE_CTL_mskIC_EN;
237 	__nds32__mtsr_isb(tmp, NDS32_SR_CACHE_CTL);
238 
239 	dump_cpu_info(smp_processor_id());
240 }
241 
setup_memory(void)242 static void __init setup_memory(void)
243 {
244 	unsigned long ram_start_pfn;
245 	unsigned long free_ram_start_pfn;
246 	phys_addr_t memory_start, memory_end;
247 	struct memblock_region *region;
248 
249 	memory_end = memory_start = 0;
250 
251 	/* Find main memory where is the kernel */
252 	for_each_memblock(memory, region) {
253 		memory_start = region->base;
254 		memory_end = region->base + region->size;
255 		pr_info("%s: Memory: 0x%x-0x%x\n", __func__,
256 			memory_start, memory_end);
257 	}
258 
259 	if (!memory_end) {
260 		panic("No memory!");
261 	}
262 
263 	ram_start_pfn = PFN_UP(memblock_start_of_DRAM());
264 	/* free_ram_start_pfn is first page after kernel */
265 	free_ram_start_pfn = PFN_UP(__pa(&_end));
266 	max_pfn = PFN_DOWN(memblock_end_of_DRAM());
267 	/* it could update max_pfn */
268 	if (max_pfn - ram_start_pfn <= MAXMEM_PFN)
269 		max_low_pfn = max_pfn;
270 	else {
271 		max_low_pfn = MAXMEM_PFN + ram_start_pfn;
272 		if (!IS_ENABLED(CONFIG_HIGHMEM))
273 			max_pfn = MAXMEM_PFN + ram_start_pfn;
274 	}
275 	/* high_memory is related with VMALLOC */
276 	high_memory = (void *)__va(max_low_pfn * PAGE_SIZE);
277 	min_low_pfn = free_ram_start_pfn;
278 
279 	/*
280 	 * initialize the boot-time allocator (with low memory only).
281 	 *
282 	 * This makes the memory from the end of the kernel to the end of
283 	 * RAM usable.
284 	 */
285 	memblock_set_bottom_up(true);
286 	memblock_reserve(PFN_PHYS(ram_start_pfn), PFN_PHYS(free_ram_start_pfn - ram_start_pfn));
287 
288 	early_init_fdt_reserve_self();
289 	early_init_fdt_scan_reserved_mem();
290 
291 	memblock_dump_all();
292 }
293 
setup_arch(char ** cmdline_p)294 void __init setup_arch(char **cmdline_p)
295 {
296 	early_init_devtree(__atags_pointer ? \
297 		phys_to_virt(__atags_pointer) : __dtb_start);
298 
299 	setup_cpuinfo();
300 
301 	init_mm.start_code = (unsigned long)&_stext;
302 	init_mm.end_code = (unsigned long)&_etext;
303 	init_mm.end_data = (unsigned long)&_edata;
304 	init_mm.brk = (unsigned long)&_end;
305 
306 	/* setup bootmem allocator */
307 	setup_memory();
308 
309 	/* paging_init() sets up the MMU and marks all pages as reserved */
310 	paging_init();
311 
312 	/* invalidate all TLB entries because the new mapping is created */
313 	__nds32__tlbop_flua();
314 
315 	/* use generic way to parse */
316 	parse_early_param();
317 
318 	unflatten_and_copy_device_tree();
319 
320 	if(IS_ENABLED(CONFIG_VT)) {
321 		if(IS_ENABLED(CONFIG_DUMMY_CONSOLE))
322 			conswitchp = &dummy_con;
323 	}
324 
325 	*cmdline_p = boot_command_line;
326 	early_trap_init();
327 }
328 
c_show(struct seq_file * m,void * v)329 static int c_show(struct seq_file *m, void *v)
330 {
331 	int i;
332 
333 	seq_printf(m, "Processor\t: %c%ld (id %lu, rev %lu, cfg %lu)\n",
334 		   cpu_series, cpu_id, cpu_id, cpu_rev, cpu_cfgid);
335 
336 	seq_printf(m, "L1I\t\t: %luKB/%luS/%luW/%luB\n",
337 		   CACHE_SET(ICACHE) * CACHE_WAY(ICACHE) *
338 		   CACHE_LINE_SIZE(ICACHE) / 1024, CACHE_SET(ICACHE),
339 		   CACHE_WAY(ICACHE), CACHE_LINE_SIZE(ICACHE));
340 
341 	seq_printf(m, "L1D\t\t: %luKB/%luS/%luW/%luB\n",
342 		   CACHE_SET(DCACHE) * CACHE_WAY(DCACHE) *
343 		   CACHE_LINE_SIZE(DCACHE) / 1024, CACHE_SET(DCACHE),
344 		   CACHE_WAY(DCACHE), CACHE_LINE_SIZE(DCACHE));
345 
346 	seq_printf(m, "BogoMIPS\t: %lu.%02lu\n",
347 		   loops_per_jiffy / (500000 / HZ),
348 		   (loops_per_jiffy / (5000 / HZ)) % 100);
349 
350 	/* dump out the processor features */
351 	seq_puts(m, "Features\t: ");
352 
353 	for (i = 0; hwcap_str[i]; i++)
354 		if (elf_hwcap & (1 << i))
355 			seq_printf(m, "%s ", hwcap_str[i]);
356 
357 	seq_puts(m, "\n\n");
358 
359 	return 0;
360 }
361 
c_start(struct seq_file * m,loff_t * pos)362 static void *c_start(struct seq_file *m, loff_t * pos)
363 {
364 	return *pos < 1 ? (void *)1 : NULL;
365 }
366 
c_next(struct seq_file * m,void * v,loff_t * pos)367 static void *c_next(struct seq_file *m, void *v, loff_t * pos)
368 {
369 	++*pos;
370 	return NULL;
371 }
372 
c_stop(struct seq_file * m,void * v)373 static void c_stop(struct seq_file *m, void *v)
374 {
375 }
376 
377 struct seq_operations cpuinfo_op = {
378 	.start = c_start,
379 	.next = c_next,
380 	.stop = c_stop,
381 	.show = c_show
382 };
383