options

ONE-View Configuration File

batch_command
basebinaryexec
is_all_external_libraries_in_ccfalse
ranges_count20
excluded_areas{ },
decan_multi_varianttrue
dataset_handlerlink
cqa_params{ },
localbinary/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/oneview_runs/compilers/aocc_6/oneview_results_1757941675/binaries/exec
lprof_post_process_params{ },
lprof_params
filter_decan{ type = all ; },
batch_script
maximal_path_number4
base_run_index0
_scalability_bins{ },
multiruns_params{ },
delay0
_is_loaded/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/oneview_runs/compilers/aocc_6/oneview_run_1757941675/config.json
repetitions31
outliers_count0
thread_filter_threshold1%
object_coverage_threshold0.01
job_submission_threshold0s
environment_variables{ },
script_variables{ },
scalability_referencemain
decan_threshold500
is_sudo_availablefalse
number_nodes1
keep_executable_locationfalse
number_processes6
base_run_nameaocc_6
vprof_params
optimizer_loop_count10
throughput_max_core0
__filtertrue
binary/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/binaries/aocc_6/exec
throughput_core-1
bucket_threshold1
external_libraries{ 1 = /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/../aocc_6/bin/libggml-base.so ; 2 = /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/../aocc_6/bin/libggml-blas.so ; 3 = /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/../aocc_6/bin/libggml-cpu.so ; 4 = /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/../aocc_6/bin/libggml.so ; 5 = /scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/build/llama.cpp/../aocc_6/bin/libllama.so ; },
mpi_commandmpirun -n <number_processes>
run_directory/scratch/users/amazouz/QAAS/service/Llama.cpp/sdp772511/175-793-6543/llama.cpp/run/oneview_runs/compilers/aocc_6/oneview_run_1757941675
included_areas{ },
custom_categories{ { type = library ; value = libggml-base.so ; },
{ type = library ; value = libggml-blas.so ; },
{ type = library ; value = libggml-cpu.so ; },
{ type = library ; value = libggml.so ; },
{ type = library ; value = libllama.so ; },
},
_is_custom_categoriesfalse
dataset
filter{ type = number ; value = 1 ; },
additional_hwc{ },
decan_all_variantstrue
decan_params
run_command<executable> -m meta-llama-3.1-8b-instruct-Q8_0.gguf -no-cnv -t 32 -n 512 -p "what is a LLM?" --seed 0
profile_start{ unit = none ; value = 0 ; },
pinning_command
frequencies{ 1 = 0 ; },
×