GNU Linux-libre 5.15.137-gnu
[releases.git] / arch / sh / kernel / setup.c
1 // SPDX-License-Identifier: GPL-2.0
2 /*
3  * arch/sh/kernel/setup.c
4  *
5  * This file handles the architecture-dependent parts of initialization
6  *
7  *  Copyright (C) 1999  Niibe Yutaka
8  *  Copyright (C) 2002 - 2010 Paul Mundt
9  */
10 #include <linux/screen_info.h>
11 #include <linux/ioport.h>
12 #include <linux/init.h>
13 #include <linux/initrd.h>
14 #include <linux/console.h>
15 #include <linux/root_dev.h>
16 #include <linux/utsname.h>
17 #include <linux/nodemask.h>
18 #include <linux/cpu.h>
19 #include <linux/pfn.h>
20 #include <linux/fs.h>
21 #include <linux/mm.h>
22 #include <linux/kexec.h>
23 #include <linux/module.h>
24 #include <linux/smp.h>
25 #include <linux/err.h>
26 #include <linux/crash_dump.h>
27 #include <linux/mmzone.h>
28 #include <linux/clk.h>
29 #include <linux/delay.h>
30 #include <linux/platform_device.h>
31 #include <linux/memblock.h>
32 #include <linux/of.h>
33 #include <linux/of_fdt.h>
34 #include <linux/uaccess.h>
35 #include <uapi/linux/mount.h>
36 #include <asm/io.h>
37 #include <asm/page.h>
38 #include <asm/elf.h>
39 #include <asm/sections.h>
40 #include <asm/irq.h>
41 #include <asm/setup.h>
42 #include <asm/clock.h>
43 #include <asm/smp.h>
44 #include <asm/mmu_context.h>
45 #include <asm/mmzone.h>
46 #include <asm/processor.h>
47 #include <asm/sparsemem.h>
48 #include <asm/platform_early.h>
49
50 /*
51  * Initialize loops_per_jiffy as 10000000 (1000MIPS).
52  * This value will be used at the very early stage of serial setup.
53  * The bigger value means no problem.
54  */
55 struct sh_cpuinfo cpu_data[NR_CPUS] __read_mostly = {
56         [0] = {
57                 .type                   = CPU_SH_NONE,
58                 .family                 = CPU_FAMILY_UNKNOWN,
59                 .loops_per_jiffy        = 10000000,
60                 .phys_bits              = MAX_PHYSMEM_BITS,
61         },
62 };
63 EXPORT_SYMBOL(cpu_data);
64
65 /*
66  * The machine vector. First entry in .machvec.init, or clobbered by
67  * sh_mv= on the command line, prior to .machvec.init teardown.
68  */
69 struct sh_machine_vector sh_mv = { .mv_name = "generic", };
70 EXPORT_SYMBOL(sh_mv);
71
72 #ifdef CONFIG_VT
73 struct screen_info screen_info;
74 #endif
75
76 extern int root_mountflags;
77
78 #define RAMDISK_IMAGE_START_MASK        0x07FF
79 #define RAMDISK_PROMPT_FLAG             0x8000
80 #define RAMDISK_LOAD_FLAG               0x4000
81
82 static char __initdata command_line[COMMAND_LINE_SIZE] = { 0, };
83
84 static struct resource code_resource = {
85         .name = "Kernel code",
86         .flags = IORESOURCE_BUSY | IORESOURCE_SYSTEM_RAM,
87 };
88
89 static struct resource data_resource = {
90         .name = "Kernel data",
91         .flags = IORESOURCE_BUSY | IORESOURCE_SYSTEM_RAM,
92 };
93
94 static struct resource bss_resource = {
95         .name   = "Kernel bss",
96         .flags  = IORESOURCE_BUSY | IORESOURCE_SYSTEM_RAM,
97 };
98
99 unsigned long memory_start;
100 EXPORT_SYMBOL(memory_start);
101 unsigned long memory_end = 0;
102 EXPORT_SYMBOL(memory_end);
103 unsigned long memory_limit = 0;
104
105 static struct resource mem_resources[MAX_NUMNODES];
106
107 int l1i_cache_shape, l1d_cache_shape, l2_cache_shape;
108
109 static int __init early_parse_mem(char *p)
110 {
111         if (!p)
112                 return 1;
113
114         memory_limit = PAGE_ALIGN(memparse(p, &p));
115
116         pr_notice("Memory limited to %ldMB\n", memory_limit >> 20);
117
118         return 0;
119 }
120 early_param("mem", early_parse_mem);
121
122 void __init check_for_initrd(void)
123 {
124 #ifdef CONFIG_BLK_DEV_INITRD
125         unsigned long start, end;
126
127         /*
128          * Check for the rare cases where boot loaders adhere to the boot
129          * ABI.
130          */
131         if (!LOADER_TYPE || !INITRD_START || !INITRD_SIZE)
132                 goto disable;
133
134         start = INITRD_START + __MEMORY_START;
135         end = start + INITRD_SIZE;
136
137         if (unlikely(end <= start))
138                 goto disable;
139         if (unlikely(start & ~PAGE_MASK)) {
140                 pr_err("initrd must be page aligned\n");
141                 goto disable;
142         }
143
144         if (unlikely(start < __MEMORY_START)) {
145                 pr_err("initrd start (%08lx) < __MEMORY_START(%x)\n",
146                         start, __MEMORY_START);
147                 goto disable;
148         }
149
150         if (unlikely(end > memblock_end_of_DRAM())) {
151                 pr_err("initrd extends beyond end of memory "
152                        "(0x%08lx > 0x%08lx)\ndisabling initrd\n",
153                        end, (unsigned long)memblock_end_of_DRAM());
154                 goto disable;
155         }
156
157         /*
158          * If we got this far in spite of the boot loader's best efforts
159          * to the contrary, assume we actually have a valid initrd and
160          * fix up the root dev.
161          */
162         ROOT_DEV = Root_RAM0;
163
164         /*
165          * Address sanitization
166          */
167         initrd_start = (unsigned long)__va(start);
168         initrd_end = initrd_start + INITRD_SIZE;
169
170         memblock_reserve(__pa(initrd_start), INITRD_SIZE);
171
172         return;
173
174 disable:
175         pr_info("initrd disabled\n");
176         initrd_start = initrd_end = 0;
177 #endif
178 }
179
180 #ifndef CONFIG_GENERIC_CALIBRATE_DELAY
181 void calibrate_delay(void)
182 {
183         struct clk *clk = clk_get(NULL, "cpu_clk");
184
185         if (IS_ERR(clk))
186                 panic("Need a sane CPU clock definition!");
187
188         loops_per_jiffy = (clk_get_rate(clk) >> 1) / HZ;
189
190         printk(KERN_INFO "Calibrating delay loop (skipped)... "
191                          "%lu.%02lu BogoMIPS PRESET (lpj=%lu)\n",
192                          loops_per_jiffy/(500000/HZ),
193                          (loops_per_jiffy/(5000/HZ)) % 100,
194                          loops_per_jiffy);
195 }
196 #endif
197
198 void __init __add_active_range(unsigned int nid, unsigned long start_pfn,
199                                                 unsigned long end_pfn)
200 {
201         struct resource *res = &mem_resources[nid];
202         unsigned long start, end;
203
204         WARN_ON(res->name); /* max one active range per node for now */
205
206         start = start_pfn << PAGE_SHIFT;
207         end = end_pfn << PAGE_SHIFT;
208
209         res->name = "System RAM";
210         res->start = start;
211         res->end = end - 1;
212         res->flags = IORESOURCE_SYSTEM_RAM | IORESOURCE_BUSY;
213
214         if (request_resource(&iomem_resource, res)) {
215                 pr_err("unable to request memory_resource 0x%lx 0x%lx\n",
216                        start_pfn, end_pfn);
217                 return;
218         }
219
220         /*
221          * We don't know which RAM region contains kernel data or
222          * the reserved crashkernel region, so try it repeatedly
223          * and let the resource manager test it.
224          */
225         request_resource(res, &code_resource);
226         request_resource(res, &data_resource);
227         request_resource(res, &bss_resource);
228 #ifdef CONFIG_KEXEC
229         request_resource(res, &crashk_res);
230 #endif
231
232         /*
233          * Also make sure that there is a PMB mapping that covers this
234          * range before we attempt to activate it, to avoid reset by MMU.
235          * We can hit this path with NUMA or memory hot-add.
236          */
237         pmb_bolt_mapping((unsigned long)__va(start), start, end - start,
238                          PAGE_KERNEL);
239
240         memblock_set_node(PFN_PHYS(start_pfn), PFN_PHYS(end_pfn - start_pfn),
241                           &memblock.memory, nid);
242 }
243
244 void __init __weak plat_early_device_setup(void)
245 {
246 }
247
248 #ifdef CONFIG_OF_EARLY_FLATTREE
249 void __ref sh_fdt_init(phys_addr_t dt_phys)
250 {
251         static int done = 0;
252         void *dt_virt;
253
254         /* Avoid calling an __init function on secondary cpus. */
255         if (done) return;
256
257 #ifdef CONFIG_USE_BUILTIN_DTB
258         dt_virt = __dtb_start;
259 #else
260         dt_virt = phys_to_virt(dt_phys);
261 #endif
262
263         if (!dt_virt || !early_init_dt_scan(dt_virt)) {
264                 pr_crit("Error: invalid device tree blob"
265                         " at physical address %p\n", (void *)dt_phys);
266
267                 while (true)
268                         cpu_relax();
269         }
270
271         done = 1;
272 }
273 #endif
274
275 void __init setup_arch(char **cmdline_p)
276 {
277         enable_mmu();
278
279         ROOT_DEV = old_decode_dev(ORIG_ROOT_DEV);
280
281         printk(KERN_NOTICE "Boot params:\n"
282                            "... MOUNT_ROOT_RDONLY - %08lx\n"
283                            "... RAMDISK_FLAGS     - %08lx\n"
284                            "... ORIG_ROOT_DEV     - %08lx\n"
285                            "... LOADER_TYPE       - %08lx\n"
286                            "... INITRD_START      - %08lx\n"
287                            "... INITRD_SIZE       - %08lx\n",
288                            MOUNT_ROOT_RDONLY, RAMDISK_FLAGS,
289                            ORIG_ROOT_DEV, LOADER_TYPE,
290                            INITRD_START, INITRD_SIZE);
291
292 #ifdef CONFIG_BLK_DEV_RAM
293         rd_image_start = RAMDISK_FLAGS & RAMDISK_IMAGE_START_MASK;
294 #endif
295
296         if (!MOUNT_ROOT_RDONLY)
297                 root_mountflags &= ~MS_RDONLY;
298         setup_initial_init_mm(_text, _etext, _edata, _end);
299
300         code_resource.start = virt_to_phys(_text);
301         code_resource.end = virt_to_phys(_etext)-1;
302         data_resource.start = virt_to_phys(_etext);
303         data_resource.end = virt_to_phys(_edata)-1;
304         bss_resource.start = virt_to_phys(__bss_start);
305         bss_resource.end = virt_to_phys(__bss_stop)-1;
306
307 #ifdef CONFIG_CMDLINE_OVERWRITE
308         strlcpy(command_line, CONFIG_CMDLINE, sizeof(command_line));
309 #else
310         strlcpy(command_line, COMMAND_LINE, sizeof(command_line));
311 #ifdef CONFIG_CMDLINE_EXTEND
312         strlcat(command_line, " ", sizeof(command_line));
313         strlcat(command_line, CONFIG_CMDLINE, sizeof(command_line));
314 #endif
315 #endif
316
317         /* Save unparsed command line copy for /proc/cmdline */
318         memcpy(boot_command_line, command_line, COMMAND_LINE_SIZE);
319         *cmdline_p = command_line;
320
321         parse_early_param();
322
323         plat_early_device_setup();
324
325         sh_mv_setup();
326
327         /* Let earlyprintk output early console messages */
328         sh_early_platform_driver_probe("earlyprintk", 1, 1);
329
330 #ifdef CONFIG_OF_EARLY_FLATTREE
331 #ifdef CONFIG_USE_BUILTIN_DTB
332         unflatten_and_copy_device_tree();
333 #else
334         unflatten_device_tree();
335 #endif
336 #endif
337
338         paging_init();
339
340         /* Perform the machine specific initialisation */
341         if (likely(sh_mv.mv_setup))
342                 sh_mv.mv_setup(cmdline_p);
343
344         plat_smp_setup();
345 }
346
347 /* processor boot mode configuration */
348 int generic_mode_pins(void)
349 {
350         pr_warn("generic_mode_pins(): missing mode pin configuration\n");
351         return 0;
352 }
353
354 int test_mode_pin(int pin)
355 {
356         return sh_mv.mv_mode_pins() & pin;
357 }
358
359 void __init arch_cpu_finalize_init(void)
360 {
361         char *p = &init_utsname()->machine[2]; /* "sh" */
362
363         select_idle_routine();
364
365         current_cpu_data.loops_per_jiffy = loops_per_jiffy;
366
367         switch (current_cpu_data.family) {
368         case CPU_FAMILY_SH2:
369                 *p++ = '2';
370                 break;
371         case CPU_FAMILY_SH2A:
372                 *p++ = '2';
373                 *p++ = 'a';
374                 break;
375         case CPU_FAMILY_SH3:
376                 *p++ = '3';
377                 break;
378         case CPU_FAMILY_SH4:
379                 *p++ = '4';
380                 break;
381         case CPU_FAMILY_SH4A:
382                 *p++ = '4';
383                 *p++ = 'a';
384                 break;
385         case CPU_FAMILY_SH4AL_DSP:
386                 *p++ = '4';
387                 *p++ = 'a';
388                 *p++ = 'l';
389                 *p++ = '-';
390                 *p++ = 'd';
391                 *p++ = 's';
392                 *p++ = 'p';
393                 break;
394         case CPU_FAMILY_UNKNOWN:
395                 /*
396                  * Specifically use CPU_FAMILY_UNKNOWN rather than
397                  * default:, so we're able to have the compiler whine
398                  * about unhandled enumerations.
399                  */
400                 break;
401         }
402
403         pr_info("CPU: %s\n", get_cpu_subtype(&current_cpu_data));
404
405 #ifndef __LITTLE_ENDIAN__
406         /* 'eb' means 'Endian Big' */
407         *p++ = 'e';
408         *p++ = 'b';
409 #endif
410         *p = '\0';
411 }