options

Help is available by moving the cursor above any symbol or by checking MAQAO website.

Global Metrics

Metricr0r1r2r3r4r5
Total Time (s)29.1529.0729.0429.1428.9129.12
Max (Thread Active Time) (s)28.2727.9528.2328.2827.8128.06
Average Active Time (s)27.7527.4127.8227.7627.2727.72
Activity Ratio (%)97.796.998.197.796.998.0
Average number of active threads182.801181.050183.933182.895181.096182.767
Affinity Stability (%)99.599.199.699.599.199.5
GFLOPS94.22794.29394.69994.22994.69894.294
Time in analyzed loops (%)86.287.886.586.487.986.9
Time in analyzed innermost loops (%)86.187.486.086.087.786.4
Time in user code (%)86.888.486.886.788.687.2
Compilation Options Score (%)99.774.9100.099.799.9100.0
Array Access Efficiency (%)10078.6Not Available10077.8Not Available
Potential Speedups
Perfect Flow Complexity1.001.001.001.001.001.00
Perfect OpenMP + MPI + Pthread1.071.041.061.061.041.07
Perfect OpenMP + MPI + Pthread + Perfect Load Distribution1.171.151.161.171.151.16
No Scalar IntegerPotential Speedup1.001.001.001.001.00Not Available
Nb Loops to get 80%11111Not Available
FP VectorisedPotential Speedup1.001.001.001.001.00Not Available
Nb Loops to get 80%11111Not Available
Fully VectorisedPotential Speedup1.001.061.001.001.06Not Available
Nb Loops to get 80%11111Not Available
Only FP ArithmeticPotential Speedup1.004.091.001.004.16Not Available
Nb Loops to get 80%11111Not Available

Cumulated Speedup If No Scalar Integer

Cumulated Speedup If FP Vectorized

Cumulated Speedup If Fully Vectorized

Cumulated Speedup If Only FP Arithmetic

Loop Based Profiles

Innermost / Single Loops

Inbetween Loops

Outermost Loops

Cumulated Coverage With All Loops

Innermost Loop Based Profiles

Coverage

Count

Application Categorization

Time

Coverage

Compilation Options

Source ObjectIssue
libllama.so
-g is missing for some functions (possibly ones added by the compiler), it is needed to have more accurate reports. Other recommended flags are: -O2/-O3, -march=(target)
-O2, -O3 or -Ofast is missing.
-march=(target) is missing.
exec
-g is missing for some functions (possibly ones added by the compiler), it is needed to have more accurate reports. Other recommended flags are: -O2/-O3, -march=(target)
-O2, -O3 or -Ofast is missing.
-march=(target) is missing.
libggml-base.so
-g is missing for some functions (possibly ones added by the compiler), it is needed to have more accurate reports. Other recommended flags are: -O2/-O3, -march=(target)
-O2, -O3 or -Ofast is missing.
-march=(target) is missing.
libggml-cpu.so
mmq.cpp
quants.c

Path Count Profiles

Coverage

Count

Low Iteration Count Profiles

Coverage

Count

Average Number of Active Threads

Run 1 - orig_default

Run 2 - gcc_default

Run 3 - aocc_default

Run 4 - icx_10

Run 5 - gcc_4

Run 6 - aocc_6

Experiment Summaries

r0r1r2r3r4r5
Experiment Name
Application/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/oneview_runs/defaults/orig/exec/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/base_runs/defaults/gcc/exec/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/base_runs/defaults/aocc/exec/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/binaries/icx_10/exec/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/binaries/gcc_4/exec/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/binaries/aocc_6/exec
Timestamp2025-09-15 04:48:432025-09-15 04:52:062025-09-15 04:55:342025-09-15 06:05:552025-09-15 06:06:562025-09-15 06:07:57
Experiment TypeMPI; OpenMP; same as r0same as r0same as r0same as r0same as r0
Machinesdp772511same as r0same as r0same as r0same as r0same as r0
Architecturex86_64same as r0same as r0same as r0same as r0same as r0
Micro ArchitectureGRANITE_RAPIDSsame as r0same as r0same as r0same as r0same as r0
Model NameIntel(R) Xeon(R) 6972Psame as r0same as r0same as r0same as r0same as r0
Cache Size491520 KBsame as r0same as r0same as r0same as r0same as r0
Number of Cores96same as r0same as r0same as r0same as r0same as r0
Maximal Frequency3.9 GHzsame as r0same as r0same as r0same as r0same as r0
OS VersionLinux 6.8.0-78-generic #78-Ubuntu SMP PREEMPT_DYNAMIC Tue Aug 12 11:34:18 UTC 2025same as r0same as r0same as r0same as r0same as r0
Architecture used during static analysisx86_64same as r0same as r0same as r0same as r0same as r0
Micro Architecture used during static analysisGRANITE_RAPIDSsame as r0same as r0same as r0same as r0same as r0
Compilation Options exec: N/A
libggml-base.so: N/A
libggml-cpu.so: clang based Intel(R) oneAPI DPC++/C++ Compiler 2025.1.1 (2025.1.1.20250418) /opt/intel/oneapi/compiler/2025.1/bin/compiler/clang --driver-mode=g++ --intel -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O3 -g -fno-omit-frame-pointer -fcf-protection=none -no-pie -grecord-command-line -O3 -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -march=native -fno-associative-math -fiopenmp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp -fveclib=SVML --driver-mode=g++ --intel -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O3 -g -fno-omit-frame-pointer -fcf-protection=none -no-pie -grecord-command-line -O3 -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -march=native -fno-associative-math -fiopenmp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp -fveclib=SVML
libllama.so: N/A
exec: N/A
libggml-base.so: N/A
libggml-cpu.so: GNU C++17 13.3.0 -march=graniterapids -mmmx -mpopcnt -msse -msse2 -msse3 -mssse3 -msse4.1 -msse4.2 -mavx -mavx2 -mno-sse4a -mno-fma4 -mno-xop -mfma -mavx512f -mbmi -mbmi2 -maes -mpclmul -mavx512vl -mavx512bw -mavx512dq -mavx512cd -mno-avx512er -mno-avx512pf -mavx512vbmi -mavx512ifma -mno-avx5124vnniw -mno-avx5124fmaps -mavx512vpopcntdq -mavx512vbmi2 -mgfni -mvpclmulqdq -mavx512vnni -mavx512bitalg -mavx512bf16 -mno-avx512vp2intersect -mno-3dnow -madx -mabm -mcldemote -mclflushopt -mclwb -mno-clzero -mcx16 -menqcmd -mf16c -mfsgsbase -mfxsr -mno-hle -msahf -mno-lwp -mlzcnt -mmovbe -mmovdir64b -mmovdiri -mno-mwaitx -mpconfig -mpku -mno-prefetchwt1 -mprfchw -mptwrite -mrdpid -mrdrnd -mrdseed -mno-rtm -mserialize -msgx -msha -mshstk -mno-tbm -mtsxldtrk -mvaes -mwaitpkg -mwbnoinvd -mxsave -mxsavec -mxsaveopt -mxsaves -mamx-tile -mamx-int8 -mamx-bf16 -muintr -mhreset -mno-kl -mno-widekl -mavxvnni -mavx512fp16 -mno-avxifma -mno-avxvnniint8 -mno-avxneconvert -mno-cmpccxadd -mamx-fp16 -mprefetchi -mno-raoint -mno-amx-complex --param=l1-cache-size=48 --param=l1-cache-line-size=64 --param=l2-cache-size=491520 -mtune=graniterapids -g -O3 -O3 -std=gnu++17 -fno-omit-frame-pointer -fcf-protection=none -fPIC -fopenmp -fasynchronous-unwind-tables -fstack-protector-strong -fstack-clash-protection
libllama.so: N/A
+ [vdso]: N/A
exec: N/A
libggml-base.so: N/A
libggml-cpu.so: AMD clang version 17.0.6 (CLANG: AOCC_5.0.0-Build#1377 2024_09_24) /scratch/users/amazouz/Tools/x86_64/compilers/AOCC/aocc-compiler-5.0.0/bin/clang-17 --driver-mode=g++ -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O3 -g -fno-omit-frame-pointer -fcf-protection=none -nopie -grecord-command-line -stdlib=libc++ -O3 -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -march=native -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp
libllama.so: N/A
exec: N/A
libggml-base.so: N/A
libggml-cpu.so: clang based Intel(R) oneAPI DPC++/C++ Compiler 2025.1.1 (2025.1.1.20250418) /opt/intel/oneapi/compiler/2025.1/bin/compiler/clang --driver-mode=g++ --intel -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O2 -x GRANITERAPIDS -fno-vectorize -fno-slp-vectorize -fno-iopenmp-simd -g -fno-omit-frame-pointer -fcf-protection=none -no-pie -grecord-command-line -fno-finite-math-only -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -fno-associative-math -fiopenmp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp -fveclib=SVML --driver-mode=g++ --intel -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O2 -x GRANITERAPIDS -fno-vectorize -fno-slp-vectorize -fno-iopenmp-simd -g -fno-omit-frame-pointer -fcf-protection=none -no-pie -grecord-command-line -fno-finite-math-only -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -fno-associative-math -fiopenmp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp -fveclib=SVML
libllama.so: N/A
exec: N/A
libggml-base.so: N/A
libggml-blas.so: N/A
libggml-cpu.so: GNU C++17 13.3.0 -march=graniterapids -mprefer-vector-width=128 -g -O3 -O3 -O3 -std=gnu++17 -funroll-loops -ffast-math -fno-omit-frame-pointer -fcf-protection=none -fno-finite-math-only -fPIC -fopenmp -fasynchronous-unwind-tables -fstack-protector-strong -fstack-clash-protection
libllama.so: N/A
+ [vdso]: N/A
exec: N/A
libggml-base.so: N/A
libggml-cpu.so: AMD clang version 17.0.6 (CLANG: AOCC_5.0.0-Build#1377 2024_09_24) /scratch/users/amazouz/Tools/x86_64/compilers/AOCC/aocc-compiler-5.0.0/bin/clang-17 --driver-mode=g++ -D GGML_BACKEND_BUILD -D GGML_BACKEND_SHARED -D GGML_SCHED_MAX_COPIES=4 -D GGML_SHARED -D GGML_USE_CPU_REPACK -D GGML_USE_LLAMAFILE -D GGML_USE_OPENMP -D _GNU_SOURCE -D _XOPEN_SOURCE=600 -D ggml_cpu_EXPORTS -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/.. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/. -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu -I /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/../include -O2 -march=graniterapids -ffast-math -g -fno-omit-frame-pointer -fcf-protection=none -nopie -grecord-command-line -fno-finite-math-only -stdlib=libc++ -D NDEBUG -std=gnu++17 -fPIC -Wmissing-declarations -Wmissing-noreturn -Wall -Wextra -Wpedantic -Wcast-qual -Wno-unused-function -Wunreachable-code-break -Wunreachable-code-return -Wmissing-prototypes -Wextra-semi -fopenmp=libomp -MD -MT ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -MF ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o.d -o ggml/src/CMakeFiles/ggml-cpu.dir/ggml-cpu/amx/mmq.cpp.o -c /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/ggml/src/ggml-cpu/amx/mmq.cpp
libllama.so: N/A
Number of processes observed6same as r0same as r0same as r0same as r0same as r0
Number of threads observed192same as r0same as r0same as r0same as r0same as r0
Frequency Driverintel_pstatesame as r0same as r0same as r0same as r0same as r0
Frequency Governorperformancesame as r0same as r0same as r0same as r0same as r0
Huge Pagesalwayssame as r0same as r0same as r0same as r0same as r0
Hyperthreadingonsame as r0same as r0same as r0same as r0same as r0
Number of sockets2same as r0same as r0same as r0same as r0same as r0
Number of cores per socket96same as r0same as r0same as r0same as r0same as r0
MAQAO version2025.1.1same as r0same as r0same as r0same as r0same as r0
MAQAO build3f8f629befdd61fbefd06c681d308ad28e2e93f8::20250630-093426same as r0same as r0same as r0same as r0same as r0
Commentssame as r0same as r0same as r0same as r0same as r0
×