Linux kernel mirror (for testing)
git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
kernel
os
linux
1/* SPDX-License-Identifier: GPL-2.0-or-later */
2#ifndef _ASM_POWERPC_CODE_PATCHING_H
3#define _ASM_POWERPC_CODE_PATCHING_H
4
5/*
6 * Copyright 2008, Michael Ellerman, IBM Corporation.
7 */
8
9#include <asm/types.h>
10#include <asm/ppc-opcode.h>
11#include <linux/string.h>
12#include <linux/kallsyms.h>
13#include <asm/asm-compat.h>
14#include <asm/inst.h>
15
16/* Flags for create_branch:
17 * "b" == create_branch(addr, target, 0);
18 * "ba" == create_branch(addr, target, BRANCH_ABSOLUTE);
19 * "bl" == create_branch(addr, target, BRANCH_SET_LINK);
20 * "bla" == create_branch(addr, target, BRANCH_ABSOLUTE | BRANCH_SET_LINK);
21 */
22#define BRANCH_SET_LINK 0x1
23#define BRANCH_ABSOLUTE 0x2
24
25/*
26 * Powerpc branch instruction is :
27 *
28 * 0 6 30 31
29 * +---------+----------------+---+---+
30 * | opcode | LI |AA |LK |
31 * +---------+----------------+---+---+
32 * Where AA = 0 and LK = 0
33 *
34 * LI is a signed 24 bits integer. The real branch offset is computed
35 * by: imm32 = SignExtend(LI:'0b00', 32);
36 *
37 * So the maximum forward branch should be:
38 * (0x007fffff << 2) = 0x01fffffc = 0x1fffffc
39 * The maximum backward branch should be:
40 * (0xff800000 << 2) = 0xfe000000 = -0x2000000
41 */
42static inline bool is_offset_in_branch_range(long offset)
43{
44 return (offset >= -0x2000000 && offset <= 0x1fffffc && !(offset & 0x3));
45}
46
47static inline bool is_offset_in_cond_branch_range(long offset)
48{
49 return offset >= -0x8000 && offset <= 0x7fff && !(offset & 0x3);
50}
51
52static inline int create_branch(ppc_inst_t *instr, const u32 *addr,
53 unsigned long target, int flags)
54{
55 long offset;
56
57 *instr = ppc_inst(0);
58 offset = target;
59 if (! (flags & BRANCH_ABSOLUTE))
60 offset = offset - (unsigned long)addr;
61
62 /* Check we can represent the target in the instruction format */
63 if (!is_offset_in_branch_range(offset))
64 return 1;
65
66 /* Mask out the flags and target, so they don't step on each other. */
67 *instr = ppc_inst(0x48000000 | (flags & 0x3) | (offset & 0x03FFFFFC));
68
69 return 0;
70}
71
72int create_cond_branch(ppc_inst_t *instr, const u32 *addr,
73 unsigned long target, int flags);
74int patch_branch(u32 *addr, unsigned long target, int flags);
75int patch_instruction(u32 *addr, ppc_inst_t instr);
76int raw_patch_instruction(u32 *addr, ppc_inst_t instr);
77int patch_instructions(u32 *addr, u32 *code, size_t len, bool repeat_instr);
78
79/*
80 * The data patching functions patch_uint() and patch_ulong(), etc., must be
81 * called on aligned addresses.
82 *
83 * The instruction patching functions patch_instruction() and similar must be
84 * called on addresses satisfying instruction alignment requirements.
85 */
86
87#ifdef CONFIG_PPC64
88
89int patch_uint(void *addr, unsigned int val);
90int patch_ulong(void *addr, unsigned long val);
91
92#define patch_u64 patch_ulong
93
94#else
95
96static inline int patch_uint(void *addr, unsigned int val)
97{
98 if (!IS_ALIGNED((unsigned long)addr, sizeof(unsigned int)))
99 return -EINVAL;
100
101 return patch_instruction(addr, ppc_inst(val));
102}
103
104static inline int patch_ulong(void *addr, unsigned long val)
105{
106 if (!IS_ALIGNED((unsigned long)addr, sizeof(unsigned long)))
107 return -EINVAL;
108
109 return patch_instruction(addr, ppc_inst(val));
110}
111
112#endif
113
114#define patch_u32 patch_uint
115
116static inline unsigned long patch_site_addr(s32 *site)
117{
118 return (unsigned long)site + *site;
119}
120
121static inline int patch_instruction_site(s32 *site, ppc_inst_t instr)
122{
123 return patch_instruction((u32 *)patch_site_addr(site), instr);
124}
125
126static inline int patch_branch_site(s32 *site, unsigned long target, int flags)
127{
128 return patch_branch((u32 *)patch_site_addr(site), target, flags);
129}
130
131static inline int modify_instruction(unsigned int *addr, unsigned int clr,
132 unsigned int set)
133{
134 return patch_instruction(addr, ppc_inst((*addr & ~clr) | set));
135}
136
137static inline int modify_instruction_site(s32 *site, unsigned int clr, unsigned int set)
138{
139 return modify_instruction((unsigned int *)patch_site_addr(site), clr, set);
140}
141
142static inline unsigned int branch_opcode(ppc_inst_t instr)
143{
144 return ppc_inst_primary_opcode(instr) & 0x3F;
145}
146
147static inline int instr_is_branch_iform(ppc_inst_t instr)
148{
149 return branch_opcode(instr) == 18;
150}
151
152static inline int instr_is_branch_bform(ppc_inst_t instr)
153{
154 return branch_opcode(instr) == 16;
155}
156
157int instr_is_relative_branch(ppc_inst_t instr);
158int instr_is_relative_link_branch(ppc_inst_t instr);
159unsigned long branch_target(const u32 *instr);
160int translate_branch(ppc_inst_t *instr, const u32 *dest, const u32 *src);
161bool is_conditional_branch(ppc_inst_t instr);
162
163#define OP_RT_RA_MASK 0xffff0000UL
164#define LIS_R2 (PPC_RAW_LIS(_R2, 0))
165#define ADDIS_R2_R12 (PPC_RAW_ADDIS(_R2, _R12, 0))
166#define ADDI_R2_R2 (PPC_RAW_ADDI(_R2, _R2, 0))
167
168
169static inline unsigned long ppc_function_entry(void *func)
170{
171#ifdef CONFIG_PPC64_ELF_ABI_V2
172 u32 *insn = func;
173
174 /*
175 * A PPC64 ABIv2 function may have a local and a global entry
176 * point. We need to use the local entry point when patching
177 * functions, so identify and step over the global entry point
178 * sequence.
179 *
180 * The global entry point sequence is always of the form:
181 *
182 * addis r2,r12,XXXX
183 * addi r2,r2,XXXX
184 *
185 * A linker optimisation may convert the addis to lis:
186 *
187 * lis r2,XXXX
188 * addi r2,r2,XXXX
189 */
190 if ((((*insn & OP_RT_RA_MASK) == ADDIS_R2_R12) ||
191 ((*insn & OP_RT_RA_MASK) == LIS_R2)) &&
192 ((*(insn+1) & OP_RT_RA_MASK) == ADDI_R2_R2))
193 return (unsigned long)(insn + 2);
194 else
195 return (unsigned long)func;
196#elif defined(CONFIG_PPC64_ELF_ABI_V1)
197 /*
198 * On PPC64 ABIv1 the function pointer actually points to the
199 * function's descriptor. The first entry in the descriptor is the
200 * address of the function text.
201 */
202 return ((struct func_desc *)func)->addr;
203#else
204 return (unsigned long)func;
205#endif
206}
207
208static inline unsigned long ppc_global_function_entry(void *func)
209{
210#ifdef CONFIG_PPC64_ELF_ABI_V2
211 /* PPC64 ABIv2 the global entry point is at the address */
212 return (unsigned long)func;
213#else
214 /* All other cases there is no change vs ppc_function_entry() */
215 return ppc_function_entry(func);
216#endif
217}
218
219/*
220 * Wrapper around kallsyms_lookup() to return function entry address:
221 * - For ABIv1, we lookup the dot variant.
222 * - For ABIv2, we return the local entry point.
223 */
224static inline unsigned long ppc_kallsyms_lookup_name(const char *name)
225{
226 unsigned long addr;
227#ifdef CONFIG_PPC64_ELF_ABI_V1
228 /* check for dot variant */
229 char dot_name[1 + KSYM_NAME_LEN];
230 bool dot_appended = false;
231
232 if (strnlen(name, KSYM_NAME_LEN) >= KSYM_NAME_LEN)
233 return 0;
234
235 if (name[0] != '.') {
236 dot_name[0] = '.';
237 dot_name[1] = '\0';
238 strlcat(dot_name, name, sizeof(dot_name));
239 dot_appended = true;
240 } else {
241 dot_name[0] = '\0';
242 strlcat(dot_name, name, sizeof(dot_name));
243 }
244 addr = kallsyms_lookup_name(dot_name);
245 if (!addr && dot_appended)
246 /* Let's try the original non-dot symbol lookup */
247 addr = kallsyms_lookup_name(name);
248#elif defined(CONFIG_PPC64_ELF_ABI_V2)
249 addr = kallsyms_lookup_name(name);
250 if (addr)
251 addr = ppc_function_entry((void *)addr);
252#else
253 addr = kallsyms_lookup_name(name);
254#endif
255 return addr;
256}
257
258/*
259 * Some instruction encodings commonly used in dynamic ftracing
260 * and function live patching.
261 */
262
263/* This must match the definition of STK_GOT in <asm/ppc_asm.h> */
264#ifdef CONFIG_PPC64_ELF_ABI_V2
265#define R2_STACK_OFFSET 24
266#else
267#define R2_STACK_OFFSET 40
268#endif
269
270#define PPC_INST_LD_TOC PPC_RAW_LD(_R2, _R1, R2_STACK_OFFSET)
271
272/* usually preceded by a mflr r0 */
273#define PPC_INST_STD_LR PPC_RAW_STD(_R0, _R1, PPC_LR_STKOFF)
274
275#endif /* _ASM_POWERPC_CODE_PATCHING_H */