• Home
  • Line#
  • Scopes#
  • Navigate#
  • Raw
  • Download
1 /******************************************************************************
2  * Copyright (c) 2008 Isaku Yamahata <yamahata at valinux co jp>
3  *                    VA Linux Systems Japan K.K.
4  *
5  * This program is free software; you can redistribute it and/or modify
6  * it under the terms of the GNU General Public License as published by
7  * the Free Software Foundation; either version 2 of the License, or
8  * (at your option) any later version.
9  *
10  * This program is distributed in the hope that it will be useful,
11  * but WITHOUT ANY WARRANTY; without even the implied warranty of
12  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.  See the
13  * GNU General Public License for more details.
14  *
15  * You should have received a copy of the GNU General Public License
16  * along with this program; if not, write to the Free Software
17  * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA  02111-1307  USA
18  *
19  */
20 
21 #ifndef _ASM_IA64_PARAVIRT_PRIVOP_H
22 #define _ASM_IA64_PARAVIRT_PRIVOP_H
23 
24 #ifdef CONFIG_PARAVIRT
25 
26 #ifndef __ASSEMBLY__
27 
28 #include <linux/types.h>
29 #include <asm/kregs.h> /* for IA64_PSR_I */
30 
31 /******************************************************************************
32  * replacement of intrinsics operations.
33  */
34 
35 struct pv_cpu_ops {
36 	void (*fc)(void *addr);
37 	unsigned long (*thash)(unsigned long addr);
38 	unsigned long (*get_cpuid)(int index);
39 	unsigned long (*get_pmd)(int index);
40 	unsigned long (*getreg)(int reg);
41 	void (*setreg)(int reg, unsigned long val);
42 	void (*ptcga)(unsigned long addr, unsigned long size);
43 	unsigned long (*get_rr)(unsigned long index);
44 	void (*set_rr)(unsigned long index, unsigned long val);
45 	void (*set_rr0_to_rr4)(unsigned long val0, unsigned long val1,
46 			       unsigned long val2, unsigned long val3,
47 			       unsigned long val4);
48 	void (*ssm_i)(void);
49 	void (*rsm_i)(void);
50 	unsigned long (*get_psr_i)(void);
51 	void (*intrin_local_irq_restore)(unsigned long flags);
52 };
53 
54 extern struct pv_cpu_ops pv_cpu_ops;
55 
56 extern void ia64_native_setreg_func(int regnum, unsigned long val);
57 extern unsigned long ia64_native_getreg_func(int regnum);
58 
59 /************************************************/
60 /* Instructions paravirtualized for performance */
61 /************************************************/
62 
63 #ifndef ASM_SUPPORTED
64 #define paravirt_ssm_i()	pv_cpu_ops.ssm_i()
65 #define paravirt_rsm_i()	pv_cpu_ops.rsm_i()
66 #define __paravirt_getreg()	pv_cpu_ops.getreg()
67 #endif
68 
69 /* mask for ia64_native_ssm/rsm() must be constant.("i" constraing).
70  * static inline function doesn't satisfy it. */
71 #define paravirt_ssm(mask)			\
72 	do {					\
73 		if ((mask) == IA64_PSR_I)	\
74 			paravirt_ssm_i();	\
75 		else				\
76 			ia64_native_ssm(mask);	\
77 	} while (0)
78 
79 #define paravirt_rsm(mask)			\
80 	do {					\
81 		if ((mask) == IA64_PSR_I)	\
82 			paravirt_rsm_i();	\
83 		else				\
84 			ia64_native_rsm(mask);	\
85 	} while (0)
86 
87 /* returned ip value should be the one in the caller,
88  * not in __paravirt_getreg() */
89 #define paravirt_getreg(reg)					\
90 	({							\
91 		unsigned long res;				\
92 		if ((reg) == _IA64_REG_IP)			\
93 			res = ia64_native_getreg(_IA64_REG_IP); \
94 		else						\
95 			res = __paravirt_getreg(reg);		\
96 		res;						\
97 	})
98 
99 /******************************************************************************
100  * replacement of hand written assembly codes.
101  */
102 struct pv_cpu_asm_switch {
103 	unsigned long switch_to;
104 	unsigned long leave_syscall;
105 	unsigned long work_processed_syscall;
106 	unsigned long leave_kernel;
107 };
108 void paravirt_cpu_asm_init(const struct pv_cpu_asm_switch *cpu_asm_switch);
109 
110 #endif /* __ASSEMBLY__ */
111 
112 #define IA64_PARAVIRT_ASM_FUNC(name)	paravirt_ ## name
113 
114 #else
115 
116 /* fallback for native case */
117 #define IA64_PARAVIRT_ASM_FUNC(name)	ia64_native_ ## name
118 
119 #endif /* CONFIG_PARAVIRT */
120 
121 #if defined(CONFIG_PARAVIRT) && defined(ASM_SUPPORTED)
122 #define paravirt_dv_serialize_data()	ia64_dv_serialize_data()
123 #else
124 #define paravirt_dv_serialize_data()	/* nothing */
125 #endif
126 
127 /* these routines utilize privilege-sensitive or performance-sensitive
128  * privileged instructions so the code must be replaced with
129  * paravirtualized versions */
130 #define ia64_switch_to			IA64_PARAVIRT_ASM_FUNC(switch_to)
131 #define ia64_leave_syscall		IA64_PARAVIRT_ASM_FUNC(leave_syscall)
132 #define ia64_work_processed_syscall	\
133 	IA64_PARAVIRT_ASM_FUNC(work_processed_syscall)
134 #define ia64_leave_kernel		IA64_PARAVIRT_ASM_FUNC(leave_kernel)
135 
136 
137 #if defined(CONFIG_PARAVIRT)
138 /******************************************************************************
139  * binary patching infrastructure
140  */
141 #define PARAVIRT_PATCH_TYPE_FC				1
142 #define PARAVIRT_PATCH_TYPE_THASH			2
143 #define PARAVIRT_PATCH_TYPE_GET_CPUID			3
144 #define PARAVIRT_PATCH_TYPE_GET_PMD			4
145 #define PARAVIRT_PATCH_TYPE_PTCGA			5
146 #define PARAVIRT_PATCH_TYPE_GET_RR			6
147 #define PARAVIRT_PATCH_TYPE_SET_RR			7
148 #define PARAVIRT_PATCH_TYPE_SET_RR0_TO_RR4		8
149 #define PARAVIRT_PATCH_TYPE_SSM_I			9
150 #define PARAVIRT_PATCH_TYPE_RSM_I			10
151 #define PARAVIRT_PATCH_TYPE_GET_PSR_I			11
152 #define PARAVIRT_PATCH_TYPE_INTRIN_LOCAL_IRQ_RESTORE	12
153 
154 /* PARAVIRT_PATY_TYPE_[GS]ETREG + _IA64_REG_xxx */
155 #define PARAVIRT_PATCH_TYPE_GETREG			0x10000000
156 #define PARAVIRT_PATCH_TYPE_SETREG			0x20000000
157 
158 /*
159  * struct task_struct* (*ia64_switch_to)(void* next_task);
160  * void *ia64_leave_syscall;
161  * void *ia64_work_processed_syscall
162  * void *ia64_leave_kernel;
163  */
164 
165 #define PARAVIRT_PATCH_TYPE_BR_START			0x30000000
166 #define PARAVIRT_PATCH_TYPE_BR_SWITCH_TO		\
167 	(PARAVIRT_PATCH_TYPE_BR_START + 0)
168 #define PARAVIRT_PATCH_TYPE_BR_LEAVE_SYSCALL		\
169 	(PARAVIRT_PATCH_TYPE_BR_START + 1)
170 #define PARAVIRT_PATCH_TYPE_BR_WORK_PROCESSED_SYSCALL	\
171 	(PARAVIRT_PATCH_TYPE_BR_START + 2)
172 #define PARAVIRT_PATCH_TYPE_BR_LEAVE_KERNEL		\
173 	(PARAVIRT_PATCH_TYPE_BR_START + 3)
174 
175 #ifdef ASM_SUPPORTED
176 #include <asm/paravirt_patch.h>
177 
178 /*
179  * pv_cpu_ops calling stub.
180  * normal function call convension can't be written by gcc
181  * inline assembly.
182  *
183  * from the caller's point of view,
184  * the following registers will be clobbered.
185  * r2, r3
186  * r8-r15
187  * r16, r17
188  * b6, b7
189  * p6-p15
190  * ar.ccv
191  *
192  * from the callee's point of view ,
193  * the following registers can be used.
194  * r2, r3: scratch
195  * r8: scratch, input argument0 and return value
196  * r0-r15: scratch, input argument1-5
197  * b6: return pointer
198  * b7: scratch
199  * p6-p15: scratch
200  * ar.ccv: scratch
201  *
202  * other registers must not be changed. especially
203  * b0: rp: preserved. gcc ignores b0 in clobbered register.
204  * r16: saved gp
205  */
206 /* 5 bundles */
207 #define __PARAVIRT_BR							\
208 	";;\n"								\
209 	"{ .mlx\n"							\
210 	"nop 0\n"							\
211 	"movl r2 = %[op_addr]\n"/* get function pointer address */	\
212 	";;\n"								\
213 	"}\n"								\
214 	"1:\n"								\
215 	"{ .mii\n"							\
216 	"ld8 r2 = [r2]\n"	/* load function descriptor address */	\
217 	"mov r17 = ip\n"	/* get ip to calc return address */	\
218 	"mov r16 = gp\n"	/* save gp */				\
219 	";;\n"								\
220 	"}\n"								\
221 	"{ .mii\n"							\
222 	"ld8 r3 = [r2], 8\n"	/* load entry address */		\
223 	"adds r17 =  1f - 1b, r17\n"	/* calculate return address */	\
224 	";;\n"								\
225 	"mov b7 = r3\n"		/* set entry address */			\
226 	"}\n"								\
227 	"{ .mib\n"							\
228 	"ld8 gp = [r2]\n"	/* load gp value */			\
229 	"mov b6 = r17\n"	/* set return address */		\
230 	"br.cond.sptk.few b7\n"	/* intrinsics are very short isns */	\
231 	"}\n"								\
232 	"1:\n"								\
233 	"{ .mii\n"							\
234 	"mov gp = r16\n"	/* restore gp value */			\
235 	"nop 0\n"							\
236 	"nop 0\n"							\
237 	";;\n"								\
238 	"}\n"
239 
240 #define PARAVIRT_OP(op)				\
241 	[op_addr] "i"(&pv_cpu_ops.op)
242 
243 #define PARAVIRT_TYPE(type)			\
244 	PARAVIRT_PATCH_TYPE_ ## type
245 
246 #define PARAVIRT_REG_CLOBBERS0					\
247 	"r2", "r3", /*"r8",*/ "r9", "r10", "r11", "r14",	\
248 		"r15", "r16", "r17"
249 
250 #define PARAVIRT_REG_CLOBBERS1					\
251 	"r2","r3", /*"r8",*/ "r9", "r10", "r11", "r14",		\
252 		"r15", "r16", "r17"
253 
254 #define PARAVIRT_REG_CLOBBERS2					\
255 	"r2", "r3", /*"r8", "r9",*/ "r10", "r11", "r14",	\
256 		"r15", "r16", "r17"
257 
258 #define PARAVIRT_REG_CLOBBERS5					\
259 	"r2", "r3", /*"r8", "r9", "r10", "r11", "r14",*/	\
260 		"r15", "r16", "r17"
261 
262 #define PARAVIRT_BR_CLOBBERS			\
263 	"b6", "b7"
264 
265 #define PARAVIRT_PR_CLOBBERS						\
266 	"p6", "p7", "p8", "p9", "p10", "p11", "p12", "p13", "p14", "p15"
267 
268 #define PARAVIRT_AR_CLOBBERS			\
269 	"ar.ccv"
270 
271 #define PARAVIRT_CLOBBERS0			\
272 		PARAVIRT_REG_CLOBBERS0,		\
273 		PARAVIRT_BR_CLOBBERS,		\
274 		PARAVIRT_PR_CLOBBERS,		\
275 		PARAVIRT_AR_CLOBBERS,		\
276 		"memory"
277 
278 #define PARAVIRT_CLOBBERS1			\
279 		PARAVIRT_REG_CLOBBERS1,		\
280 		PARAVIRT_BR_CLOBBERS,		\
281 		PARAVIRT_PR_CLOBBERS,		\
282 		PARAVIRT_AR_CLOBBERS,		\
283 		"memory"
284 
285 #define PARAVIRT_CLOBBERS2			\
286 		PARAVIRT_REG_CLOBBERS2,		\
287 		PARAVIRT_BR_CLOBBERS,		\
288 		PARAVIRT_PR_CLOBBERS,		\
289 		PARAVIRT_AR_CLOBBERS,		\
290 		"memory"
291 
292 #define PARAVIRT_CLOBBERS5			\
293 		PARAVIRT_REG_CLOBBERS5,		\
294 		PARAVIRT_BR_CLOBBERS,		\
295 		PARAVIRT_PR_CLOBBERS,		\
296 		PARAVIRT_AR_CLOBBERS,		\
297 		"memory"
298 
299 #define PARAVIRT_BR0(op, type)					\
300 	register unsigned long ia64_clobber asm ("r8");		\
301 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
302 					  PARAVIRT_TYPE(type))	\
303 		      :	"=r"(ia64_clobber)			\
304 		      : PARAVIRT_OP(op)				\
305 		      : PARAVIRT_CLOBBERS0)
306 
307 #define PARAVIRT_BR0_RET(op, type)				\
308 	register unsigned long ia64_intri_res asm ("r8");	\
309 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
310 					  PARAVIRT_TYPE(type))	\
311 		      : "=r"(ia64_intri_res)			\
312 		      : PARAVIRT_OP(op)				\
313 		      : PARAVIRT_CLOBBERS0)
314 
315 #define PARAVIRT_BR1(op, type, arg1)				\
316 	register unsigned long __##arg1 asm ("r8") = arg1;	\
317 	register unsigned long ia64_clobber asm ("r8");		\
318 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
319 					  PARAVIRT_TYPE(type))	\
320 		      :	"=r"(ia64_clobber)			\
321 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
322 		      : PARAVIRT_CLOBBERS1)
323 
324 #define PARAVIRT_BR1_RET(op, type, arg1)			\
325 	register unsigned long ia64_intri_res asm ("r8");	\
326 	register unsigned long __##arg1 asm ("r8") = arg1;	\
327 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
328 					  PARAVIRT_TYPE(type))	\
329 		      : "=r"(ia64_intri_res)			\
330 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
331 		      : PARAVIRT_CLOBBERS1)
332 
333 #define PARAVIRT_BR1_VOID(op, type, arg1)			\
334 	register void *__##arg1 asm ("r8") = arg1;		\
335 	register unsigned long ia64_clobber asm ("r8");		\
336 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
337 					  PARAVIRT_TYPE(type))	\
338 		      :	"=r"(ia64_clobber)			\
339 		      : PARAVIRT_OP(op), "0"(__##arg1)		\
340 		      : PARAVIRT_CLOBBERS1)
341 
342 #define PARAVIRT_BR2(op, type, arg1, arg2)				\
343 	register unsigned long __##arg1 asm ("r8") = arg1;		\
344 	register unsigned long __##arg2 asm ("r9") = arg2;		\
345 	register unsigned long ia64_clobber1 asm ("r8");		\
346 	register unsigned long ia64_clobber2 asm ("r9");		\
347 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,		\
348 					  PARAVIRT_TYPE(type))		\
349 		      : "=r"(ia64_clobber1), "=r"(ia64_clobber2)	\
350 		      : PARAVIRT_OP(op), "0"(__##arg1), "1"(__##arg2)	\
351 		      : PARAVIRT_CLOBBERS2)
352 
353 
354 #define PARAVIRT_DEFINE_CPU_OP0(op, type)		\
355 	static inline void				\
356 	paravirt_ ## op (void)				\
357 	{						\
358 		PARAVIRT_BR0(op, type);			\
359 	}
360 
361 #define PARAVIRT_DEFINE_CPU_OP0_RET(op, type)		\
362 	static inline unsigned long			\
363 	paravirt_ ## op (void)				\
364 	{						\
365 		PARAVIRT_BR0_RET(op, type);		\
366 		return ia64_intri_res;			\
367 	}
368 
369 #define PARAVIRT_DEFINE_CPU_OP1_VOID(op, type)		\
370 	static inline void				\
371 	paravirt_ ## op (void *arg1)			\
372 	{						\
373 		PARAVIRT_BR1_VOID(op, type, arg1);	\
374 	}
375 
376 #define PARAVIRT_DEFINE_CPU_OP1(op, type)		\
377 	static inline void				\
378 	paravirt_ ## op (unsigned long arg1)		\
379 	{						\
380 		PARAVIRT_BR1(op, type, arg1);		\
381 	}
382 
383 #define PARAVIRT_DEFINE_CPU_OP1_RET(op, type)		\
384 	static inline unsigned long			\
385 	paravirt_ ## op (unsigned long arg1)		\
386 	{						\
387 		PARAVIRT_BR1_RET(op, type, arg1);	\
388 		return ia64_intri_res;			\
389 	}
390 
391 #define PARAVIRT_DEFINE_CPU_OP2(op, type)		\
392 	static inline void				\
393 	paravirt_ ## op (unsigned long arg1,		\
394 			 unsigned long arg2)		\
395 	{						\
396 		PARAVIRT_BR2(op, type, arg1, arg2);	\
397 	}
398 
399 
400 PARAVIRT_DEFINE_CPU_OP1_VOID(fc, FC);
PARAVIRT_DEFINE_CPU_OP1_RET(thash,THASH)401 PARAVIRT_DEFINE_CPU_OP1_RET(thash, THASH)
402 PARAVIRT_DEFINE_CPU_OP1_RET(get_cpuid, GET_CPUID)
403 PARAVIRT_DEFINE_CPU_OP1_RET(get_pmd, GET_PMD)
404 PARAVIRT_DEFINE_CPU_OP2(ptcga, PTCGA)
405 PARAVIRT_DEFINE_CPU_OP1_RET(get_rr, GET_RR)
406 PARAVIRT_DEFINE_CPU_OP2(set_rr, SET_RR)
407 PARAVIRT_DEFINE_CPU_OP0(ssm_i, SSM_I)
408 PARAVIRT_DEFINE_CPU_OP0(rsm_i, RSM_I)
409 PARAVIRT_DEFINE_CPU_OP0_RET(get_psr_i, GET_PSR_I)
410 PARAVIRT_DEFINE_CPU_OP1(intrin_local_irq_restore, INTRIN_LOCAL_IRQ_RESTORE)
411 
412 static inline void
413 paravirt_set_rr0_to_rr4(unsigned long val0, unsigned long val1,
414 			unsigned long val2, unsigned long val3,
415 			unsigned long val4)
416 {
417 	register unsigned long __val0 asm ("r8") = val0;
418 	register unsigned long __val1 asm ("r9") = val1;
419 	register unsigned long __val2 asm ("r10") = val2;
420 	register unsigned long __val3 asm ("r11") = val3;
421 	register unsigned long __val4 asm ("r14") = val4;
422 
423 	register unsigned long ia64_clobber0 asm ("r8");
424 	register unsigned long ia64_clobber1 asm ("r9");
425 	register unsigned long ia64_clobber2 asm ("r10");
426 	register unsigned long ia64_clobber3 asm ("r11");
427 	register unsigned long ia64_clobber4 asm ("r14");
428 
429 	asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,
430 					  PARAVIRT_TYPE(SET_RR0_TO_RR4))
431 		      : "=r"(ia64_clobber0),
432 			"=r"(ia64_clobber1),
433 			"=r"(ia64_clobber2),
434 			"=r"(ia64_clobber3),
435 			"=r"(ia64_clobber4)
436 		      : PARAVIRT_OP(set_rr0_to_rr4),
437 			"0"(__val0), "1"(__val1), "2"(__val2),
438 			"3"(__val3), "4"(__val4)
439 		      : PARAVIRT_CLOBBERS5);
440 }
441 
442 /* unsigned long paravirt_getreg(int reg) */
443 #define __paravirt_getreg(reg)						\
444 	({								\
445 		register unsigned long ia64_intri_res asm ("r8");	\
446 		register unsigned long __reg asm ("r8") = (reg);	\
447 									\
448 		asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
449 						  PARAVIRT_TYPE(GETREG) \
450 						  + (reg))		\
451 			      : "=r"(ia64_intri_res)			\
452 			      : PARAVIRT_OP(getreg), "0"(__reg)		\
453 			      : PARAVIRT_CLOBBERS1);			\
454 									\
455 		ia64_intri_res;						\
456 	})
457 
458 /* void paravirt_setreg(int reg, unsigned long val) */
459 #define paravirt_setreg(reg, val)					\
460 	do {								\
461 		register unsigned long __val asm ("r8") = val;		\
462 		register unsigned long __reg asm ("r9") = reg;		\
463 		register unsigned long ia64_clobber1 asm ("r8");	\
464 		register unsigned long ia64_clobber2 asm ("r9");	\
465 									\
466 		asm volatile (paravirt_alt_bundle(__PARAVIRT_BR,	\
467 						  PARAVIRT_TYPE(SETREG) \
468 						  + (reg))		\
469 			      : "=r"(ia64_clobber1),			\
470 				"=r"(ia64_clobber2)			\
471 			      : PARAVIRT_OP(setreg),			\
472 				"1"(__reg), "0"(__val)			\
473 			      : PARAVIRT_CLOBBERS2);			\
474 	} while (0)
475 
476 #endif /* ASM_SUPPORTED */
477 #endif /* CONFIG_PARAVIRT && ASM_SUPPOTED */
478 
479 #endif /* _ASM_IA64_PARAVIRT_PRIVOP_H */
480