3030extern const char __hyp_panic_string [];
3131
3232/* Check whether the FP regs were dirtied while in the host-side run loop: */
33- static inline bool __hyp_text update_fp_enabled (struct kvm_vcpu * vcpu )
33+ static inline bool update_fp_enabled (struct kvm_vcpu * vcpu )
3434{
3535/*
3636 * When the system doesn't support FP/SIMD, we cannot rely on
@@ -48,15 +48,15 @@ static inline bool __hyp_text update_fp_enabled(struct kvm_vcpu *vcpu)
4848}
4949
5050/* Save the 32-bit only FPSIMD system register state */
51- static inline void __hyp_text __fpsimd_save_fpexc32 (struct kvm_vcpu * vcpu )
51+ static inline void __fpsimd_save_fpexc32 (struct kvm_vcpu * vcpu )
5252{
5353if (!vcpu_el1_is_32bit (vcpu ))
5454return ;
5555
5656vcpu -> arch .ctxt .sys_regs [FPEXC32_EL2 ]= read_sysreg (fpexc32_el2 );
5757}
5858
59- static inline void __hyp_text __activate_traps_fpsimd32 (struct kvm_vcpu * vcpu )
59+ static inline void __activate_traps_fpsimd32 (struct kvm_vcpu * vcpu )
6060{
6161/*
6262 * We are about to set CPTR_EL2.TFP to trap all floating point
@@ -73,7 +73,7 @@ static inline void __hyp_text __activate_traps_fpsimd32(struct kvm_vcpu *vcpu)
7373}
7474}
7575
76- static inline void __hyp_text __activate_traps_common (struct kvm_vcpu * vcpu )
76+ static inline void __activate_traps_common (struct kvm_vcpu * vcpu )
7777{
7878/* Trap on AArch32 cp15 c15 (impdef sysregs) accesses (EL1 or EL0) */
7979write_sysreg (1 <<15 ,hstr_el2 );
@@ -89,13 +89,13 @@ static inline void __hyp_text __activate_traps_common(struct kvm_vcpu *vcpu)
8989write_sysreg (vcpu -> arch .mdcr_el2 ,mdcr_el2 );
9090}
9191
92- static inline void __hyp_text __deactivate_traps_common (void )
92+ static inline void __deactivate_traps_common (void )
9393{
9494write_sysreg (0 ,hstr_el2 );
9595write_sysreg (0 ,pmuserenr_el0 );
9696}
9797
98- static inline void __hyp_text ___activate_traps (struct kvm_vcpu * vcpu )
98+ static inline void ___activate_traps (struct kvm_vcpu * vcpu )
9999{
100100u64 hcr = vcpu -> arch .hcr_el2 ;
101101
@@ -108,7 +108,7 @@ static inline void __hyp_text ___activate_traps(struct kvm_vcpu *vcpu)
108108write_sysreg_s (vcpu -> arch .vsesr_el2 ,SYS_VSESR_EL2 );
109109}
110110
111- static inline void __hyp_text ___deactivate_traps (struct kvm_vcpu * vcpu )
111+ static inline void ___deactivate_traps (struct kvm_vcpu * vcpu )
112112{
113113/*
114114 * If we pended a virtual abort, preserve it until it gets
@@ -122,12 +122,12 @@ static inline void __hyp_text ___deactivate_traps(struct kvm_vcpu *vcpu)
122122}
123123}
124124
125- static inline void __hyp_text __activate_vm (struct kvm * kvm )
125+ static inline void __activate_vm (struct kvm * kvm )
126126{
127127__load_guest_stage2 (kvm );
128128}
129129
130- static inline bool __hyp_text __translate_far_to_hpfar (u64 far ,u64 * hpfar )
130+ static inline bool __translate_far_to_hpfar (u64 far ,u64 * hpfar )
131131{
132132u64 par ,tmp ;
133133
@@ -156,7 +156,7 @@ static inline bool __hyp_text __translate_far_to_hpfar(u64 far, u64 *hpfar)
156156return true;
157157}
158158
159- static inline bool __hyp_text __populate_fault_info (struct kvm_vcpu * vcpu )
159+ static inline bool __populate_fault_info (struct kvm_vcpu * vcpu )
160160{
161161u8 ec ;
162162u64 esr ;
@@ -196,7 +196,7 @@ static inline bool __hyp_text __populate_fault_info(struct kvm_vcpu *vcpu)
196196}
197197
198198/* Check for an FPSIMD/SVE trap and handle as appropriate */
199- static inline bool __hyp_text __hyp_handle_fpsimd (struct kvm_vcpu * vcpu )
199+ static inline bool __hyp_handle_fpsimd (struct kvm_vcpu * vcpu )
200200{
201201bool vhe ,sve_guest ,sve_host ;
202202u8 hsr_ec ;
@@ -283,7 +283,7 @@ static inline bool __hyp_text __hyp_handle_fpsimd(struct kvm_vcpu *vcpu)
283283return true;
284284}
285285
286- static inline bool __hyp_text handle_tx2_tvm (struct kvm_vcpu * vcpu )
286+ static inline bool handle_tx2_tvm (struct kvm_vcpu * vcpu )
287287{
288288u32 sysreg = esr_sys64_to_sysreg (kvm_vcpu_get_hsr (vcpu ));
289289int rt = kvm_vcpu_sys_get_rt (vcpu );
@@ -338,7 +338,7 @@ static inline bool __hyp_text handle_tx2_tvm(struct kvm_vcpu *vcpu)
338338return true;
339339}
340340
341- static inline bool __hyp_text esr_is_ptrauth_trap (u32 esr )
341+ static inline bool esr_is_ptrauth_trap (u32 esr )
342342{
343343u32 ec = ESR_ELx_EC (esr );
344344
@@ -371,7 +371,7 @@ static inline bool __hyp_text esr_is_ptrauth_trap(u32 esr)
371371regs[key ## KEYHI_EL1] = read_sysreg_s(SYS_ ## key ## KEYHI_EL1);\
372372})
373373
374- static inline bool __hyp_text __hyp_handle_ptrauth (struct kvm_vcpu * vcpu )
374+ static inline bool __hyp_handle_ptrauth (struct kvm_vcpu * vcpu )
375375{
376376struct kvm_cpu_context * ctxt ;
377377u64 val ;
@@ -401,7 +401,7 @@ static inline bool __hyp_text __hyp_handle_ptrauth(struct kvm_vcpu *vcpu)
401401 * the guest, false when we should restore the host state and return to the
402402 * main run loop.
403403 */
404- static inline bool __hyp_text fixup_guest_exit (struct kvm_vcpu * vcpu ,u64 * exit_code )
404+ static inline bool fixup_guest_exit (struct kvm_vcpu * vcpu ,u64 * exit_code )
405405{
406406if (ARM_EXCEPTION_CODE (* exit_code )!= ARM_EXCEPTION_IRQ )
407407vcpu -> arch .fault .esr_el2 = read_sysreg_el2 (SYS_ESR );
@@ -473,15 +473,15 @@ static inline bool __hyp_text fixup_guest_exit(struct kvm_vcpu *vcpu, u64 *exit_
473473return false;
474474}
475475
476- static inline bool __hyp_text __needs_ssbd_off (struct kvm_vcpu * vcpu )
476+ static inline bool __needs_ssbd_off (struct kvm_vcpu * vcpu )
477477{
478478if (!cpus_have_final_cap (ARM64_SSBD ))
479479return false;
480480
481481return !(vcpu -> arch .workaround_flags & VCPU_WORKAROUND_2_FLAG );
482482}
483483
484- static inline void __hyp_text __set_guest_arch_workaround_state (struct kvm_vcpu * vcpu )
484+ static inline void __set_guest_arch_workaround_state (struct kvm_vcpu * vcpu )
485485{
486486#ifdef CONFIG_ARM64_SSBD
487487/*
@@ -494,7 +494,7 @@ static inline void __hyp_text __set_guest_arch_workaround_state(struct kvm_vcpu
494494#endif
495495}
496496
497- static inline void __hyp_text __set_host_arch_workaround_state (struct kvm_vcpu * vcpu )
497+ static inline void __set_host_arch_workaround_state (struct kvm_vcpu * vcpu )
498498{
499499#ifdef CONFIG_ARM64_SSBD
500500/*