|
12 | 12 | #include <asm/msr-index.h>
|
13 | 13 | #include <asm/unwind_hints.h>
|
14 | 14 |
|
15 |
| -/* |
16 |
| - * This should be used immediately before a retpoline alternative. It tells |
17 |
| - * objtool where the retpolines are so that it can make sense of the control |
18 |
| - * flow by just reading the original instruction(s) and ignoring the |
19 |
| - * alternatives. |
20 |
| - */ |
21 |
| -#define ANNOTATE_NOSPEC_ALTERNATIVE \ |
22 |
| - ANNOTATE_IGNORE_ALTERNATIVE |
23 |
| - |
24 | 15 | /*
|
25 | 16 | * Fill the CPU return stack buffer.
|
26 | 17 | *
|
|
82 | 73 | .popsection
|
83 | 74 | .endm
|
84 | 75 |
|
85 |
| -/* |
86 |
| - * These are the bare retpoline primitives for indirect jmp and call. |
87 |
| - * Do not use these directly; they only exist to make the ALTERNATIVE |
88 |
| - * invocation below less ugly. |
89 |
| - */ |
90 |
| -.macro RETPOLINE_JMP reg:req |
91 |
| - call .Ldo_rop_\@ |
92 |
| -.Lspec_trap_\@: |
93 |
| - pause |
94 |
| - lfence |
95 |
| - jmp .Lspec_trap_\@ |
96 |
| -.Ldo_rop_\@: |
97 |
| - mov \reg, (%_ASM_SP) |
98 |
| - ret |
99 |
| -.endm |
100 |
| - |
101 |
| -/* |
102 |
| - * This is a wrapper around RETPOLINE_JMP so the called function in reg |
103 |
| - * returns to the instruction after the macro. |
104 |
| - */ |
105 |
| -.macro RETPOLINE_CALL reg:req |
106 |
| - jmp .Ldo_call_\@ |
107 |
| -.Ldo_retpoline_jmp_\@: |
108 |
| - RETPOLINE_JMP \reg |
109 |
| -.Ldo_call_\@: |
110 |
| - call .Ldo_retpoline_jmp_\@ |
111 |
| -.endm |
112 |
| - |
113 | 76 | /*
|
114 | 77 | * JMP_NOSPEC and CALL_NOSPEC macros can be used instead of a simple
|
115 | 78 | * indirect jmp/call which may be susceptible to the Spectre variant 2
|
116 | 79 | * attack.
|
117 | 80 | */
|
118 | 81 | .macro JMP_NOSPEC reg:req
|
119 | 82 | #ifdef CONFIG_RETPOLINE
|
120 |
| - ANNOTATE_NOSPEC_ALTERNATIVE |
121 |
| - ALTERNATIVE_2 __stringify(ANNOTATE_RETPOLINE_SAFE; jmp *%\reg), \ |
122 |
| - __stringify(RETPOLINE_JMP %\reg), X86_FEATURE_RETPOLINE,\ |
123 |
| - __stringify(lfence; ANNOTATE_RETPOLINE_SAFE; jmp *%\reg), X86_FEATURE_RETPOLINE_AMD |
| 83 | + ALTERNATIVE_2 __stringify(ANNOTATE_RETPOLINE_SAFE; jmp *%\reg), \ |
| 84 | + __stringify(jmp __x86_retpoline_\reg), X86_FEATURE_RETPOLINE, \ |
| 85 | + __stringify(lfence; ANNOTATE_RETPOLINE_SAFE; jmp *%\reg), X86_FEATURE_RETPOLINE_AMD |
124 | 86 | #else
|
125 | 87 | jmp *%\reg
|
126 | 88 | #endif
|
127 | 89 | .endm
|
128 | 90 |
|
129 | 91 | .macro CALL_NOSPEC reg:req
|
130 | 92 | #ifdef CONFIG_RETPOLINE
|
131 |
| - ANNOTATE_NOSPEC_ALTERNATIVE |
132 |
| - ALTERNATIVE_2 __stringify(ANNOTATE_RETPOLINE_SAFE; call *%\reg),\ |
133 |
| - __stringify(RETPOLINE_CALL %\reg), X86_FEATURE_RETPOLINE,\ |
134 |
| - __stringify(lfence; ANNOTATE_RETPOLINE_SAFE; call *%\reg), X86_FEATURE_RETPOLINE_AMD |
| 93 | + ALTERNATIVE_2 __stringify(ANNOTATE_RETPOLINE_SAFE; call *%\reg), \ |
| 94 | + __stringify(call __x86_retpoline_\reg), X86_FEATURE_RETPOLINE, \ |
| 95 | + __stringify(lfence; ANNOTATE_RETPOLINE_SAFE; call *%\reg), X86_FEATURE_RETPOLINE_AMD |
135 | 96 | #else
|
136 | 97 | call *%\reg
|
137 | 98 | #endif
|
|
165 | 126 | * which is ensured when CONFIG_RETPOLINE is defined.
|
166 | 127 | */
|
167 | 128 | # define CALL_NOSPEC \
|
168 |
| - ANNOTATE_NOSPEC_ALTERNATIVE \ |
169 | 129 | ALTERNATIVE_2( \
|
170 | 130 | ANNOTATE_RETPOLINE_SAFE \
|
171 | 131 | "call *%[thunk_target]\n", \
|
172 |
| - "call __x86_indirect_thunk_%V[thunk_target]\n", \ |
| 132 | + "call __x86_retpoline_%V[thunk_target]\n", \ |
173 | 133 | X86_FEATURE_RETPOLINE, \
|
174 | 134 | "lfence;\n" \
|
175 | 135 | ANNOTATE_RETPOLINE_SAFE \
|
176 | 136 | "call *%[thunk_target]\n", \
|
177 | 137 | X86_FEATURE_RETPOLINE_AMD)
|
| 138 | + |
178 | 139 | # define THUNK_TARGET(addr) [thunk_target] "r" (addr)
|
179 | 140 |
|
180 | 141 | #else /* CONFIG_X86_32 */
|
|
184 | 145 | * here, anyway.
|
185 | 146 | */
|
186 | 147 | # define CALL_NOSPEC \
|
187 |
| - ANNOTATE_NOSPEC_ALTERNATIVE \ |
188 | 148 | ALTERNATIVE_2( \
|
189 | 149 | ANNOTATE_RETPOLINE_SAFE \
|
190 | 150 | "call *%[thunk_target]\n", \
|
|
0 commit comments