options

Stylizer

orig_defaultgcc_defaultarmclang_3gcc_3

[ 0 / 3 ] Security settings from the host restrict profiling. Some metrics will be missing or incomplete.

Current value for kernel.perf_event_paranoid is 4. If possible, set it to 1 or check with your system administrator which flag can be used to achieve this.

[ 0 / 3 ] Security settings from the host restrict profiling. Some metrics will be missing or incomplete.

Current value for kernel.perf_event_paranoid is 4. If possible, set it to 1 or check with your system administrator which flag can be used to achieve this.

[ 0 / 3 ] Security settings from the host restrict profiling. Some metrics will be missing or incomplete.

Current value for kernel.perf_event_paranoid is 4. If possible, set it to 1 or check with your system administrator which flag can be used to achieve this.

[ 0 / 3 ] Security settings from the host restrict profiling. Some metrics will be missing or incomplete.

Current value for kernel.perf_event_paranoid is 4. If possible, set it to 1 or check with your system administrator which flag can be used to achieve this.

Not available for this run

[ 0 / 0 ] Fastmath not used

Consider to add ffast-math to compilation flags (or replace -O3 with -Ofast) to unlock potential extra speedup by relaxing floating-point computation consistency. Warning: floating-point accuracy may be reduced and the compliance to IEEE/ISO rules/specifications for math functions will be relaxed, typically 'errno' will no longer be set after calling some math functions.

Not available for this run

Not available for this run

Not available for this run

[ 0 / 3 ] Compilation of some functions is not optimized for the target processor

Architecture specific options are needed to produce efficient code for a specific processor ( -mcpu=native ).

Not available for this run

Not available for this run

Not available for this run

[ 2.32 / 3 ] Most of time spent in analyzed modules comes from functions without compilation information

Functions without compilation information (typically not compiled with -g) cumulate 22.81% of the time spent in analyzed modules. Check that -g is present. Remark: if -g is indeed used, this can also be due to some compiler built-in functions (typically math) or statically linked libraries. This warning can be ignored in that case.

Not available for this run

Not available for this run

[ 4 / 4 ] Application profile is long enough (19.33 s)

To have good quality measurements, it is advised that the application profiling time is greater than 10 seconds.

[ 4 / 4 ] Application profile is long enough (25.63 s)

To have good quality measurements, it is advised that the application profiling time is greater than 10 seconds.

[ 4 / 4 ] Application profile is long enough (16.67 s)

To have good quality measurements, it is advised that the application profiling time is greater than 10 seconds.

[ 4 / 4 ] Application profile is long enough (23.95 s)

To have good quality measurements, it is advised that the application profiling time is greater than 10 seconds.

[ 2 / 2 ] Application is correctly profiled ("Others" category represents 0.52 % of the execution time)

To have a representative profiling, it is advised that the category "Others" represents less than 20% of the execution time in order to analyze as much as possible of the user code

[ 2 / 2 ] Application is correctly profiled ("Others" category represents 0.01 % of the execution time)

To have a representative profiling, it is advised that the category "Others" represents less than 20% of the execution time in order to analyze as much as possible of the user code

[ 2 / 2 ] Application is correctly profiled ("Others" category represents 0.51 % of the execution time)

To have a representative profiling, it is advised that the category "Others" represents less than 20% of the execution time in order to analyze as much as possible of the user code

[ 2 / 2 ] Application is correctly profiled ("Others" category represents 0.00 % of the execution time)

To have a representative profiling, it is advised that the category "Others" represents less than 20% of the execution time in order to analyze as much as possible of the user code

[ 0 / 9 ] Compilation options are not available

Compilation options are an important optimization leverage but ONE-View is not able to analyze them.

[ 3 / 3 ] Optimization level option is correctly used

[ 0 / 9 ] Compilation options are not available

Compilation options are an important optimization leverage but ONE-View is not able to analyze them.

[ 0 / 9 ] Compilation options are not available

Compilation options are an important optimization leverage but ONE-View is not able to analyze them.

[ 0 / 1 ] Lstopo was not found on the host (either not installed or not loaded). The Topology lstopo report will not be generated.

[ 0 / 1 ] Lstopo was not found on the host (either not installed or not loaded). The Topology lstopo report will not be generated.

[ 0 / 1 ] Lstopo was not found on the host (either not installed or not loaded). The Topology lstopo report will not be generated.

[ 0 / 1 ] Lstopo was not found on the host (either not installed or not loaded). The Topology lstopo report will not be generated.

Strategizer

orig_defaultgcc_defaultarmclang_3gcc_3

[ 3 / 4 ] CPU activity is below 90% (89.27%)

CPU cores are idle more than 10% of time. Threads supposed to run on these cores are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 0 / 4 ] CPU activity is below 90% (26.58%)

CPU cores are idle more than 10% of time. Threads supposed to run on these cores are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 3 / 4 ] CPU activity is below 90% (89.26%)

CPU cores are idle more than 10% of time. Threads supposed to run on these cores are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 0 / 4 ] CPU activity is below 90% (25.28%)

CPU cores are idle more than 10% of time. Threads supposed to run on these cores are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 0 / 4 ] Affinity stability is lower than 90% (0.00%)

Threads are often migrating to other CPU cores/threads. For OpenMP, typically set (OMP_PLACES=cores OMP_PROC_BIND=close) or (OMP_PLACES=threads OMP_PROC_BIND=spread). With OpenMPI + OpenMP, use --bind-to core --map-by node:PE=$OMP_NUM_THREADS --report-bindings. With IntelMPI + OpenMP, set I_MPI_PIN_DOMAIN=omp:compact or I_MPI_PIN_DOMAIN=omp:scatter and use -print-rank-map.

[ 0 / 4 ] Affinity stability is lower than 90% (0.00%)

Threads are often migrating to other CPU cores/threads. For OpenMP, typically set (OMP_PLACES=cores OMP_PROC_BIND=close) or (OMP_PLACES=threads OMP_PROC_BIND=spread). With OpenMPI + OpenMP, use --bind-to core --map-by node:PE=$OMP_NUM_THREADS --report-bindings. With IntelMPI + OpenMP, set I_MPI_PIN_DOMAIN=omp:compact or I_MPI_PIN_DOMAIN=omp:scatter and use -print-rank-map.

[ 0 / 4 ] Affinity stability is lower than 90% (0.00%)

Threads are often migrating to other CPU cores/threads. For OpenMP, typically set (OMP_PLACES=cores OMP_PROC_BIND=close) or (OMP_PLACES=threads OMP_PROC_BIND=spread). With OpenMPI + OpenMP, use --bind-to core --map-by node:PE=$OMP_NUM_THREADS --report-bindings. With IntelMPI + OpenMP, set I_MPI_PIN_DOMAIN=omp:compact or I_MPI_PIN_DOMAIN=omp:scatter and use -print-rank-map.

[ 0 / 4 ] Affinity stability is lower than 90% (0.00%)

Threads are often migrating to other CPU cores/threads. For OpenMP, typically set (OMP_PLACES=cores OMP_PROC_BIND=close) or (OMP_PLACES=threads OMP_PROC_BIND=spread). With OpenMPI + OpenMP, use --bind-to core --map-by node:PE=$OMP_NUM_THREADS --report-bindings. With IntelMPI + OpenMP, set I_MPI_PIN_DOMAIN=omp:compact or I_MPI_PIN_DOMAIN=omp:scatter and use -print-rank-map.

[ 3 / 3 ] Functions mostly use all threads

Functions running on a reduced number of threads (typically sequential code) cover less than 10% of application walltime (3.17%)

[ 3 / 3 ] Functions mostly use all threads

Functions running on a reduced number of threads (typically sequential code) cover less than 10% of application walltime (6.02%)

[ 3 / 3 ] Functions mostly use all threads

Functions running on a reduced number of threads (typically sequential code) cover less than 10% of application walltime (5.04%)

[ 3 / 3 ] Functions mostly use all threads

Functions running on a reduced number of threads (typically sequential code) cover less than 10% of application walltime (3.19%)

[ 3 / 3 ] Cumulative Outermost/In between loops coverage (0.00%) lower than cumulative innermost loop coverage (1.24%)

Having cumulative Outermost/In between loops coverage greater than cumulative innermost loop coverage will make loop optimization more complex

[ 3 / 3 ] Cumulative Outermost/In between loops coverage (0.08%) lower than cumulative innermost loop coverage (10.84%)

Having cumulative Outermost/In between loops coverage greater than cumulative innermost loop coverage will make loop optimization more complex

[ 3 / 3 ] Cumulative Outermost/In between loops coverage (0.01%) lower than cumulative innermost loop coverage (1.10%)

Having cumulative Outermost/In between loops coverage greater than cumulative innermost loop coverage will make loop optimization more complex

[ 3 / 3 ] Cumulative Outermost/In between loops coverage (0.07%) lower than cumulative innermost loop coverage (11.44%)

Having cumulative Outermost/In between loops coverage greater than cumulative innermost loop coverage will make loop optimization more complex

[ 3 / 4 ] A significant amount of threads are idle (12.54%)

On average, more than 10% of observed threads are idle. Such threads are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 0 / 4 ] A significant amount of threads are idle (73.69%)

On average, more than 10% of observed threads are idle. Such threads are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 3 / 4 ] A significant amount of threads are idle (12.94%)

On average, more than 10% of observed threads are idle. Such threads are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 0 / 4 ] A significant amount of threads are idle (74.98%)

On average, more than 10% of observed threads are idle. Such threads are probably IO/sync waiting. Some hints: use faster filesystems to read/write data, improve parallel load balancing and/or scheduling.

[ 2 / 2 ] Less than 10% (0.00%) is spend in BLAS2 operations

BLAS2 calls usually could make a poor cache usage and could benefit from inlining.

[ 2 / 2 ] Less than 10% (0.00%) is spend in BLAS2 operations

BLAS2 calls usually could make a poor cache usage and could benefit from inlining.

[ 2 / 2 ] Less than 10% (0.00%) is spend in BLAS2 operations

BLAS2 calls usually could make a poor cache usage and could benefit from inlining.

[ 2 / 2 ] Less than 10% (0.00%) is spend in BLAS2 operations

BLAS2 calls usually could make a poor cache usage and could benefit from inlining.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed innermost loops (1.24%)

If the time spent in analyzed innermost loops is less than 15%, standard innermost loop optimizations such as vectorisation will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed innermost loops (10.84%)

If the time spent in analyzed innermost loops is less than 15%, standard innermost loop optimizations such as vectorisation will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed innermost loops (1.10%)

If the time spent in analyzed innermost loops is less than 15%, standard innermost loop optimizations such as vectorisation will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed innermost loops (11.44%)

If the time spent in analyzed innermost loops is less than 15%, standard innermost loop optimizations such as vectorisation will have a limited impact on application performances.

[ 3 / 3 ] Less than 10% (0.00%) is spend in BLAS1 operations

It could be more efficient to inline by hand BLAS1 operations

[ 3 / 3 ] Less than 10% (0.00%) is spend in BLAS1 operations

It could be more efficient to inline by hand BLAS1 operations

[ 3 / 3 ] Less than 10% (0.00%) is spend in BLAS1 operations

It could be more efficient to inline by hand BLAS1 operations

[ 3 / 3 ] Less than 10% (0.00%) is spend in BLAS1 operations

It could be more efficient to inline by hand BLAS1 operations

[ 2 / 2 ] Less than 10% (0.00%) is spend in Libm/SVML (special functions)

[ 2 / 2 ] Less than 10% (0.02%) is spend in Libm/SVML (special functions)

[ 2 / 2 ] Less than 10% (0.00%) is spend in Libm/SVML (special functions)

[ 2 / 2 ] Less than 10% (0.01%) is spend in Libm/SVML (special functions)

[ 0 / 4 ] Loop profile is flat

No hotspot found in the application (greatest loop coverage is 1.22%), and the twenty hottest loops cumulated coverage is lower than 20% of the application profiled time (1.24%)

[ 4 / 4 ] Loop profile is not flat

At least one loop coverage is greater than 4% (10.60%), representing an hotspot for the application

[ 0 / 4 ] Loop profile is flat

No hotspot found in the application (greatest loop coverage is 1.07%), and the twenty hottest loops cumulated coverage is lower than 20% of the application profiled time (1.10%)

[ 4 / 4 ] Loop profile is not flat

At least one loop coverage is greater than 4% (11.32%), representing an hotspot for the application

[ 0 / 4 ] Too little time of the experiment time spent in analyzed loops (1.24%)

If the time spent in analyzed loops is less than 30%, standard loop optimizations will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed loops (10.92%)

If the time spent in analyzed loops is less than 30%, standard loop optimizations will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed loops (1.11%)

If the time spent in analyzed loops is less than 30%, standard loop optimizations will have a limited impact on application performances.

[ 0 / 4 ] Too little time of the experiment time spent in analyzed loops (11.51%)

If the time spent in analyzed loops is less than 30%, standard loop optimizations will have a limited impact on application performances.

Optimizer

Analysisr0r1r2r3
Loop Computation IssuesPresence of expensive FP instructions0100
Less than 10% of the FP ADD/SUB/MUL arithmetic operations are performed using FMA0001
Presence of a large number of scalar integer instructions0111
Control Flow IssuesPresence of 2 to 4 paths0100
Data Access IssuesPresence of constant non-unit stride data access0201
Vectorization RoadblocksPresence of 2 to 4 paths0100
Presence of constant non-unit stride data access0201
×