[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
[PULL 02/72] plugins: optimize cpu_index code generation
From: |
Richard Henderson |
Subject: |
[PULL 02/72] plugins: optimize cpu_index code generation |
Date: |
Tue, 24 Dec 2024 12:04:11 -0800 |
From: Pierrick Bouvier <pierrick.bouvier@linaro.org>
When running with a single vcpu, we can return a constant instead of a
load when accessing cpu_index.
A side effect is that all tcg operations using it are optimized, most
notably scoreboard access.
When running a simple loop in user-mode, the speedup is around 20%.
Signed-off-by: Pierrick Bouvier <pierrick.bouvier@linaro.org>
Reviewed-by: Richard Henderson <richard.henderson@linaro.org>
Signed-off-by: Richard Henderson <richard.henderson@linaro.org>
Message-ID: <20241128213843.1023080-1-pierrick.bouvier@linaro.org>
---
accel/tcg/plugin-gen.c | 9 +++++++++
1 file changed, 9 insertions(+)
diff --git a/accel/tcg/plugin-gen.c b/accel/tcg/plugin-gen.c
index 1ef075552c..7e5f040bf7 100644
--- a/accel/tcg/plugin-gen.c
+++ b/accel/tcg/plugin-gen.c
@@ -102,6 +102,15 @@ static void gen_disable_mem_helper(void)
static TCGv_i32 gen_cpu_index(void)
{
+ /*
+ * Optimize when we run with a single vcpu. All values using cpu_index,
+ * including scoreboard index, will be optimized out.
+ * User-mode calls tb_flush when setting this flag. In system-mode, all
+ * vcpus are created before generating code.
+ */
+ if (!tcg_cflags_has(current_cpu, CF_PARALLEL)) {
+ return tcg_constant_i32(current_cpu->cpu_index);
+ }
TCGv_i32 cpu_index = tcg_temp_ebb_new_i32();
tcg_gen_ld_i32(cpu_index, tcg_env,
-offsetof(ArchCPU, env) + offsetof(CPUState, cpu_index));
--
2.43.0
- [PULL 18/72] tcg/optimize: Use finish_folding in fold_divide, (continued)
- [PULL 18/72] tcg/optimize: Use finish_folding in fold_divide, Richard Henderson, 2024/12/24
- [PULL 21/72] tcg/optimize: Use fold_masks_z in fold_extract, Richard Henderson, 2024/12/24
- [PULL 24/72] tcg/optimize: Use fold_masks_z in fold_extu, Richard Henderson, 2024/12/24
- [PULL 26/72] tcg/optimize: Use finish_folding in fold_mul*, Richard Henderson, 2024/12/24
- [PULL 29/72] tcg/optimize: Use fold_masks_s in fold_nor, Richard Henderson, 2024/12/24
- [PULL 30/72] tcg/optimize: Use fold_masks_s in fold_not, Richard Henderson, 2024/12/24
- [PULL 32/72] tcg/optimize: Use fold_masks_zs in fold_orc, Richard Henderson, 2024/12/24
- [PULL 31/72] tcg/optimize: Use fold_masks_zs in fold_or, Richard Henderson, 2024/12/24
- [PULL 34/72] tcg/optimize: Return true from fold_qemu_st, fold_tcg_st, Richard Henderson, 2024/12/24
- [PULL 35/72] tcg/optimize: Use finish_folding in fold_remainder, Richard Henderson, 2024/12/24
- [PULL 02/72] plugins: optimize cpu_index code generation,
Richard Henderson <=
- [PULL 23/72] tcg/optimize: Use fold_masks_zs in fold_exts, Richard Henderson, 2024/12/24
- [PULL 38/72] tcg/optimize: Use fold_masks_s in fold_negsetcond, Richard Henderson, 2024/12/24
- [PULL 39/72] tcg/optimize: Use fold_masks_z in fold_setcond2, Richard Henderson, 2024/12/24
- [PULL 11/72] tcg/optimize: Use fold_masks_zs in fold_and, Richard Henderson, 2024/12/24
- [PULL 33/72] tcg/optimize: Use fold_masks_zs in fold_qemu_ld, Richard Henderson, 2024/12/24
- [PULL 20/72] tcg/optimize: Use fold_masks_s in fold_eqv, Richard Henderson, 2024/12/24
- [PULL 22/72] tcg/optimize: Use finish_folding in fold_extract2, Richard Henderson, 2024/12/24
- [PULL 28/72] tcg/optimize: Use fold_masks_z in fold_neg_no_const, Richard Henderson, 2024/12/24
- [PULL 27/72] tcg/optimize: Use fold_masks_s in fold_nand, Richard Henderson, 2024/12/24
- [PULL 36/72] tcg/optimize: Distinguish simplification in fold_setcond_zmask, Richard Henderson, 2024/12/24