options
********************************************************************************
MAQAO 2025.1.2 - ad4b42c12cfbc289a7a711f3ded92abe2eb90c0a::20250917-142411 || 2025/09/17
/beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao oneview -R1 WP=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so -WS -c=/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/config.json --with-FLOPS object-coverage-threshold=0.1 lprof_params=btm=fp --replace xp=/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241 -of=html 
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/binaries/aocc_4/exec --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/build/llama.cpp/../aocc_4/bin/libggml-base.so --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/libs/libggml-base.so
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/build/llama.cpp/../aocc_4/bin/libggml-blas.so --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/libs/libggml-blas.so
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/build/llama.cpp/../aocc_4/bin/libggml-cpu.so --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/libs/libggml-cpu.so
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/build/llama.cpp/../aocc_4/bin/libggml.so --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/libs/libggml.so
CPY:  [true] /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/build/llama.cpp/../aocc_4/bin/libllama.so --> /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/libs/libllama.so
CMD:  OMP_NUM_THREADS=6  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_0" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=6  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 6 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=72  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_1" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=72  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 72 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=96  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_2" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=96  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 96 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=120  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_3" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=120  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 120 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=128  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_4" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=128  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 128 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=144  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_5" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=144  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 144 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=168  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_6" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=168  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 168 -n 512 -p \"what is a LLM?\" --seed 0
CMD:  OMP_NUM_THREADS=192  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   /beegfs/hackathon/users/eoseret/MAQAO_ad4b42/bin/maqao lprof _caller=oneview btm=fp --xp="/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/tools/lprof_npsu_run_7" --mpi-command="mpirun -n 1  " --collect-CPU-time-intervals -p=SSE_AVX_FLOP  --collect-topology tpp=192  -ldi=libggml-base.so,libggml-blas.so,libggml-cpu.so,libggml.so,libllama.so  -- /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 192 -n 512 -p \"what is a LLM?\" --seed 0
In run 1x6, 36 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.29940117313527% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
69 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.33137605665252% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x72, 28 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.055277734063565% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
33 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.027892434736714% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x96, 28 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.038121946243336% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
15 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.0064807308954186% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x120, 28 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.045512658427469% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
9 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.003525628359057% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x128, 31 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.047603967541363% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
16 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.0052210801513866% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x144, 30 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.030328955268487% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
13 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.0035983505076729% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x168, 29 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.043141451547854% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
10 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.0018831586348825% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
In run 1x192, 34 loops were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.047768252567039% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
9 functions were discarded from static analysis because their coverage
are lower than object_coverage_threshold value (0.1%).
That represents 0.0012988347007195% of the execution time. To include them, change the value
in the experiment directory configuration file, then rerun the command with the additionnal parameter
--force-static-analysis
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=6  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 6 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=72  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 72 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=96  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 96 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=120  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 120 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=128  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 128 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=144  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 144 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=168  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 168 -n 512 -p "what is a LLM?" --seed 0
CMD: cd "/beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_run_1759256241";  OMP_TOOL_LIBRARIES=/home/eoseret/MAQAO_src/src/maqao/libprompt/libprompt_clang.so TARGET_PARALLEL_SECTION=ALL  OMP_NUM_THREADS=192  I_MPI_PIN_ORDER=bunch  OMP_DISPLAY_AFFINITY=TRUE  OMP_PROC_BIND=spread  OMP_AFFINITY_FORMAT='OMP: pid %P tid %i thread %n bound to OS proc set {%A}'  OMP_DISPLAY_ENV=TRUE  I_MPI_PIN_DOMAIN=auto  I_MPI_DEBUG=4  OMP_PLACES=threads   mpirun -n 1   /beegfs/hackathon/users/eoseret/qaas_runs_test/175-924-9259/intel/llama.cpp/run/oneview_runs/multicore/aocc_4/oneview_results_1759256241/binaries/exec -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 192 -n 512 -p "what is a LLM?" --seed 0
×