* installing *source* package ‘llamaR’ ... ** this is package ‘llamaR’ version ‘0.2.2’ ** package ‘llamaR’ successfully unpacked and MD5 sums checked ** using staged installation Checking for ggmlR... Found ggmlR at: /Users/ripley/R/Library/ggmlR/lib Vulkan: not found (building without GPU support) Configuration complete. ** libs specified C++17 using C++ compiler: ‘Apple clang version 21.0.0 (clang-2100.0.121.2)’ using C++17 using SDK: ‘MacOSX26.4.sdk’ clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama.cpp -o llama.o In file included from llama.cpp:6: In file included from ./llama-context.h:5: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama.cpp:9: ./llama-model-loader.h:35:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 35 | const int tensor_idx = gguf_find_tensor(gguf_ctx, ggml_get_name(tensor)); | ~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama.cpp:420:55: warning: implicit conversion from 'const uint32_t' (aka 'const unsigned int') to 'float' may lose precision [-Wimplicit-int-float-conversion] 420 | tensor_split[id] = ngl_per_device[id].n_layer; | ~ ~~~~~~~~~~~~~~~~~~~^~~~~~~ llama.cpp:537:22: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 537 | for (int id = nd - 1; id >= 0; id--) { | ~~ ~~~^~~ llama.cpp:556:83: warning: implicit conversion loses integer precision: 'long long' to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 556 | uint32_t step_size = int64_t(delta) * (targets[id] - mem[id]) / (mem_high[id] - mem[id]); | ~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ llama.cpp:603:22: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 603 | for (int id = nd - 1; id >= 0; id--) { | ~~ ~~~^~~ llama.cpp:628:79: warning: implicit conversion loses integer precision: 'long long' to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 628 | uint32_t step_size = int64_t(delta) * (targets[id] - mem[id]) / (mem_high[id] - mem[id]); | ~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~ llama.cpp:758:94: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 758 | LLAMA_LOG_INFO("%s: fitting params to free memory took %.2f seconds\n", __func__, (t1_us - t0_us) * 1e-6); | ~~~~~~^~~~~~~ ~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ 9 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-adapter.cpp -o llama-adapter.o In file included from llama-adapter.cpp:1: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ llama-adapter.cpp:67:61: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 67 | ggml_backend_buffer_type_t buft = model.select_buft(il); | ~~~~~~~~~~~ ^~ llama-adapter.cpp:193:22: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 193 | int id = gguf_find_key(gguf_ctx, key.c_str()); | ~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-adapter.cpp:197:22: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 197 | int id = gguf_find_key(gguf_ctx, key.c_str()); | ~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-adapter.cpp:389:118: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 389 | LLAMA_LOG_INFO("%s: %10s LoRA buffer size = %8.2f MiB\n", __func__, ggml_backend_buffer_name(buf.get()), ggml_backend_buffer_get_size(buf.get())/1024.0/1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-adapter.cpp:222:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 222 | const int kid = gguf_find_key(ctx_gguf.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-adapter.cpp:241:21: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 241 | int n_tensors = gguf_get_n_tensors(ctx_gguf.get()); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 8 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-arch.cpp -o llama-arch.o clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-batch.cpp -o llama-batch.o llama-batch.cpp:76:36: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'value_type' (aka 'int') [-Wshorten-64-to-32] 76 | n_seq_id[i] = seq_id_0.size(); | ~ ~~~~~~~~~^~~~~~ llama-batch.cpp:203:41: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'value_type' (aka 'int') [-Wshorten-64-to-32] 203 | seq_idx[s] = seq_id_unq.size(); | ~ ~~~~~~~~~~~^~~~~~ llama-batch.cpp:503:34: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 503 | return ubatch_add(idxs, idxs.size(), false); | ~~~~~~~~~~ ~~~~~^~~~~~ llama-batch.cpp:549:41: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 549 | const uint32_t n_seqs = cur_seq_set.size(); | ~~~~~~ ~~~~~~~~~~~~^~~~~~ llama-batch.cpp:680:36: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 680 | const uint32_t n_tokens = idxs.size(); | ~~~~~~~~ ~~~~~^~~~~~ llama-batch.cpp:743:51: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'value_type' (aka 'int') [-Wshorten-64-to-32] 743 | udata->seq_idx[s] = udata->seq_id_unq.size(); | ~ ~~~~~~~~~~~~~~~~~~^~~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-chat.cpp -o llama-chat.o llama-chat.cpp:416:27: warning: implicit conversion loses integer precision: 'int' to 'value_type' (aka 'char') [-Wimplicit-int-conversion] 416 | role[0] = toupper(role[0]); | ~ ^~~~~~~~~~~~~~~~ llama-chat.cpp:884:17: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 884 | return dest.size(); | ~~~~~~ ~~~~~^~~~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-context.cpp -o llama-context.o In file included from llama-context.cpp:1: In file included from ./llama-context.h:5: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ llama-context.cpp:262:21: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 262 | ggml_backend_buffer_get_size(buf_output.get()) / 1024.0 / 1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-context.cpp:369:89: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 369 | __func__, ggml_backend_buft_name(buft), size_act / (1024.0*1024.0), size_exp / (1024.0*1024.0)); | ^~~~~~~~ ~ llama-context.cpp:369:61: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 369 | __func__, ggml_backend_buft_name(buft), size_act / (1024.0*1024.0), size_exp / (1024.0*1024.0)); | ^~~~~~~~ ~ llama-context.cpp:372:89: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 372 | __func__, ggml_backend_buft_name(buft), size_act / (1024.0*1024.0), size_exp / (1024.0*1024.0)); | ^~~~~~~~ ~ llama-context.cpp:372:61: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 372 | __func__, ggml_backend_buft_name(buft), size_act / (1024.0*1024.0), size_exp / (1024.0*1024.0)); | ^~~~~~~~ ~ llama-context.cpp:518:21: warning: implicit conversion from 'value_type' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 518 | backend_buf_exp_size[i] / 1024.0 / 1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~ ~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-context.cpp:537:33: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 537 | __func__, (t_end_us - t_start_us)/1000.0, ggml_backend_sched_get_n_copies(sched.get())); | ~~~~~~~~~^~~~~~~~~~~~ ~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-context.cpp:402:95: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 402 | sched.reset(ggml_backend_sched_new(backend_ptrs.data(), backend_buft.data(), backend_ptrs.size(), max_nodes, cparams.pipeline_parallel, cparams.op_offload)); | ~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~^~~~~~ llama-context.cpp:474:107: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 474 | sched.reset(ggml_backend_sched_new(backend_ptrs.data(), backend_buft.data(), backend_ptrs.size(), max_nodes, false, cparams.op_offload)); | ~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~^~~~~~ llama-context.cpp:1198:56: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 1198 | if (!balloc->init(batch_inp, model.vocab, nullptr, n_embd, cparams.kv_unified ? LLAMA_MAX_SEQ : cparams.n_seq_max, true)) { | ~~~~ ^~~~~~ llama-context.cpp:1420:23: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'value_type' (aka 'unsigned int') [-Wshorten-64-to-32] 1420 | counts[row] = ggml_nelements(tensor); | ~ ^~~~~~~~~~~~~~~~~~~~~~ llama-context.cpp:1451:23: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'value_type' (aka 'unsigned int') [-Wshorten-64-to-32] 1451 | counts[row] = ggml_nelements(tensor); | ~ ^~~~~~~~~~~~~~~~~~~~~~ llama-context.cpp:1521:55: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 1521 | if (!balloc->init(batch_inp, vocab, memory.get(), n_embd, n_seq_max, output_all)) { | ~~~~ ^~~~~~ llama-context.cpp:1749:76: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 1749 | const auto seq_to_output_row = build_seq_to_output_row(ubatch, n_outputs_prev); | ~~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~~~~ llama-context.cpp:1776:34: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'value_type' (aka 'int') [-Wshorten-64-to-32] 1776 | output_ids[out_id] = i; | ~ ^ llama-context.cpp:1897:98: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1897 | LLAMA_LOG_ERROR("%s: failed to allocate output buffer of size %.2f MiB\n", __func__, new_size / (1024.0 * 1024.0)); | ^~~~~~~~ ~ llama-context.cpp:1953:12: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 1953 | return n_outputs_max; | ~~~~~~ ^~~~~~~~~~~~~ llama-context.cpp:2020:54: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'const unsigned int' [-Wshorten-64-to-32] 2020 | return std::max(n_tokens * 40, 32u * model.n_tensors()); | ~~~ ~~~~^~~~~~~~~~~~~~~~~~~ llama-context.cpp:2022:48: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'const unsigned int' [-Wshorten-64-to-32] 2022 | uint32_t res = std::max(1024u, 8u*model.n_tensors()); | ~~~ ~~^~~~~~~~~~~~~~~~~~ llama-context.cpp:2523:37: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'value_type' (aka 'int') [-Wshorten-64-to-32] 2523 | w_output_pos[pos] = i; | ~ ^ llama-context.cpp:2689:31: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2689 | data.t_start_ms = 1e-3 * t_start_us; | ~ ^~~~~~~~~~ llama-context.cpp:2690:31: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2690 | data.t_load_ms = 1e-3 * t_load_us; | ~ ^~~~~~~~~ llama-context.cpp:2691:31: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2691 | data.t_p_eval_ms = 1e-3 * t_p_eval_us; | ~ ^~~~~~~~~~~ llama-context.cpp:2692:31: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2692 | data.t_eval_ms = 1e-3 * t_eval_us; | ~ ^~~~~~~~~ llama-context.cpp:3017:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 3017 | const uint32_t blck_size = ggml_blck_size(params.type_k); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-context.cpp:3026:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 3026 | const uint32_t blck_size = ggml_blck_size(params.type_v); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-context.cpp:3523:36: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 3523 | const double t_end_ms = 1e-3 * ggml_time_us(); | ~ ^~~~~~~~~~~~~~ llama-context.cpp:3576:29: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 3576 | i_dev = i; | ~ ^ 30 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-cparams.cpp -o llama-cparams.o clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -include r_llama_compat.h -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-grammar.cpp -o llama-grammar.o llama-grammar.cpp:23:44: warning: implicit conversion loses integer precision: 'int' to 'uint8_t' (aka 'unsigned char') [-Wimplicit-int-conversion] 23 | uint8_t mask = (1 << (8 - len)) - 1; | ~~~~ ~~~~~~~~~~~~~~~~~^~~ llama-grammar.cpp:75:47: warning: implicit conversion loses integer precision: 'int' to 'uint8_t' (aka 'unsigned char') [-Wimplicit-int-conversion] 75 | uint8_t mask = (1 << (7 - n_remain)) - 1; | ~~~~ ~~~~~~~~~~~~~~~~~~~~~~^~~ llama-grammar.cpp:195:29: warning: implicit conversion loses integer precision: 'unsigned long' to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 195 | uint32_t token_id = std::stoul(std::string(pos, int_end - pos)); | ~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 3 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-graph.cpp -o llama-graph.o In file included from llama-graph.cpp:1: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-graph.cpp:7: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ llama-graph.cpp:82:31: warning: implicit conversion from 'llama_pos' (aka 'int') to 'const float' may lose precision [-Wimplicit-int-float-conversion] 82 | const float pos = ubatch->pos[i]; | ~~~ ^~~~~~~~~~~~~~ llama-graph.cpp:85:35: warning: implicit conversion loses floating-point precision: 'double' to 'value_type' (aka 'float') [-Wimplicit-float-conversion] 83 | attn_scale_data[i] = std::log( | ~ ~~~~~~~~~ 84 | std::floor((pos + f_attn_temp_offset) / n_attn_temp_floor_scale) + 1.0 | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 85 | ) * f_attn_temp_scale + 1.0; | ~~~~~~~~~~~~~~~~~~~~~~^~~~~ llama-graph.cpp:84:57: warning: implicit conversion from 'const uint32_t' (aka 'const unsigned int') to 'float' may lose precision [-Wimplicit-int-float-conversion] 84 | std::floor((pos + f_attn_temp_offset) / n_attn_temp_floor_scale) + 1.0 | ~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-graph.cpp:351:55: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 351 | data[idst + i0] = hparams.use_alibi ? -std::abs(p0 - p1) : 0.0f; | ~ ^~~~~~~~~~~~~~~~~~ llama-graph.cpp:864:29: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 864 | const float scale = alpha ? lora.second * alpha / rank : lora.second; | ^~~~~ ~ llama-graph.cpp:1195:80: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 1195 | ggml_tensor * selected_experts = ggml_argsort_top_k(ctx0, selection_probs, n_expert_used); // [n_expert_used, n_tokens] | ~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~~~ llama-graph.cpp:1409:60: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 1409 | cur = ggml_pad(ctx0, cur, hparams.n_embd_inp() - n_embd, 0, 0, 0); | ~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~^~~~~~~~ llama-graph.cpp:2048:59: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'const int32_t' (aka 'const int') [-Wshorten-64-to-32] 2048 | const int32_t n_enc = !cross->v_embd.empty() ? cross->n_enc : hparams.n_ctx_train; | ~~~~~ ~~~~~~~^~~~~ llama-graph.cpp:2235:33: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 2233 | ggml_tensor * token_shift = build_rs( | ~~~~~~~~ 2234 | inp, token_shift_all, 2235 | hparams.n_embd_r(), n_seqs); | ^~~~~~ llama-graph.cpp:2520:42: warning: implicit conversion turns floating-point number into integer: 'float' to 'int32_t' (aka 'int') [-Wfloat-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-graph.cpp:2520:59: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~~~~~~ ~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-graph.cpp:2520:156: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~ ^~~~~~~~~ llama-graph.cpp:2520:116: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~ ~~~~~~~~~~^~~~~~~~~~~ llama-graph.cpp:2520:90: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~ llama-graph.cpp:2520:92: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ~ ^~~~~~~~~ llama-graph.cpp:2520:49: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2520 | int32_t relative_position_if_large = floorf(max_exact + logf(1.0 * relative_position / max_exact) * (n_buckets - max_exact) / log(1.0 * max_distance / max_exact)); | ^~~~~~~~~ ~ llama-graph.cpp:2521:90: warning: implicit conversion loses integer precision: 'uint64_t' (aka 'unsigned long long') to 'const int' [-Wshorten-64-to-32] 2521 | relative_position_if_large = std::min(relative_position_if_large, n_buckets - 1); | ~~~ ~~~~~~~~~~^~~ 23 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-hparams.cpp -o llama-hparams.o clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -include r_llama_compat.h -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-impl.cpp -o llama-impl.o llama-impl.cpp:145:29: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 145 | int arr_n = gguf_get_arr_n(ctx_gguf, i); | ~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~ 1 warning generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-io.cpp -o llama-io.o llama-io.cpp:4:29: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 4 | uint32_t str_size = str.size(); | ~~~~~~~~ ~~~~^~~~~~ 1 warning generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-kv-cache.cpp -o llama-kv-cache.o In file included from llama-kv-cache.cpp:1: In file included from ./llama-kv-cache.h:4: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-kv-cache.cpp:1: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ llama-kv-cache.cpp:196:106: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 196 | LLAMA_LOG_INFO("%s: %10s KV buffer size = %8.2f MiB\n", __func__, ggml_backend_buffer_name(buf), ggml_backend_buffer_get_size(buf)/1024.0/1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-kv-cache.cpp:152:36: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'mapped_type' (aka 'int') [-Wshorten-64-to-32] 152 | map_layer_ids[il] = layers.size(); | ~ ~~~~~~~^~~~~~ llama-kv-cache.cpp:747:39: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const int' [-Wshorten-64-to-32] 747 | const int n = cur.size(); | ~ ~~~~^~~~~~ llama-kv-cache.cpp:915:47: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 915 | const uint32_t i = s*sinfo.size() + ii; | ~ ~~~~~~~~~~~~~~~^~~~ llama-kv-cache.cpp:1148:47: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 1148 | v_cur = ggml_pad(ctx, v_cur, v->ne[0] - n_embd_gqa, 0, 0, 0); | ~~~~~~~~ ~~~~~~~~~^~~~~~~~~~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1489:26: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int32_t' (aka 'const int') [-Wshorten-64-to-32] 1489 | const int32_t n_kv = dst->ne[0]; | ~~~~ ^~~~~~~~~~ llama-kv-cache.cpp:1751:47: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1751 | const uint32_t n_seq_id = seq_ids.size(); | ~~~~~~~~ ~~~~~~~~^~~~~~ llama-kv-cache.cpp:1770:37: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1770 | const uint32_t n_layer = layers.size(); | ~~~~~~~ ~~~~~~~^~~~~~ llama-kv-cache.cpp:1845:40: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1845 | const uint32_t v_size_el = ggml_type_size(v->type); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1468:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1468 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1433:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1433 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1423:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1423 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1413:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1413 | set_input_kq_mask_impl (args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ llama-kv-cache.cpp:1396:38: warning: implicit conversion from 'int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 1396 | data[idst + j] = -std::abs(p0 - p1); | ~ ^~~~~~~~~~~~~~~~~~ llama-kv-cache.cpp:1415:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1415 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1425:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1425 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1435:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1435 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1470:9: note: in instantiation of function template specialization 'set_input_kq_mask_impl' requested here 1470 | set_input_kq_mask_impl(args, data); | ^ llama-kv-cache.cpp:1294:40: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 1294 | const uint32_t i = s*n_tps + ii; | ~ ~~~~~~~~^~~~ 49 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-kv-cache-iswa.cpp -o llama-kv-cache-iswa.o In file included from llama-kv-cache-iswa.cpp:1: In file included from ./llama-kv-cache-iswa.h:3: In file included from ./llama-kv-cache.h:4: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-kv-cache-iswa.cpp:1: In file included from ./llama-kv-cache-iswa.h:3: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-memory.cpp -o llama-memory.o clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-memory-hybrid.cpp -o llama-memory-hybrid.o In file included from llama-memory-hybrid.cpp:1: In file included from ./llama-memory-hybrid.h:4: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-memory-hybrid.cpp:1: In file included from ./llama-memory-hybrid.h:5: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-memory-hybrid-iswa.cpp -o llama-memory-hybrid-iswa.o In file included from llama-memory-hybrid-iswa.cpp:1: In file included from ./llama-memory-hybrid-iswa.h:4: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-memory-hybrid-iswa.cpp:1: In file included from ./llama-memory-hybrid-iswa.h:5: In file included from ./llama-kv-cache-iswa.h:3: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-memory-recurrent.cpp -o llama-memory-recurrent.o In file included from llama-memory-recurrent.cpp:1: In file included from ./llama-memory-recurrent.h:4: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ llama-memory-recurrent.cpp:109:106: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 109 | LLAMA_LOG_INFO("%s: %10s RS buffer size = %8.2f MiB\n", __func__, ggml_backend_buffer_name(buf), ggml_backend_buffer_get_size(buf)/1024.0/1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-memory-recurrent.cpp:656:12: warning: implicit conversion loses integer precision: 'typename iterator_traits>::difference_type' (aka 'long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 656 | used = std::count_if(cells.begin(), cells.end(), | ~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ 657 | [](const mem_cell & cell){ return !cell.is_empty(); }); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-memory-recurrent.cpp:767:67: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const uint32_t' (aka 'const unsigned int') [-Wshorten-64-to-32] 767 | const uint32_t n_seq_id = seq_id == -1 ? cell.seq_id.size() : 0; | ~~~~~~~~ ~~~~~~~~~~~~^~~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-mmap.cpp -o llama-mmap.o llama-mmap.cpp:464:25: warning: implicit conversion loses integer precision: 'long' to 'int' [-Wshorten-64-to-32] 464 | int page_size = sysconf(_SC_PAGESIZE); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~ 1 warning generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-model.cpp -o llama-model.o In file included from llama-model.cpp:1: In file included from ./llama-model.h:5: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-model.cpp:6: ./llama-model-loader.h:35:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 35 | const int tensor_idx = gguf_find_tensor(gguf_ctx, ggml_get_name(tensor)); | ~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from llama-model.cpp:8: In file included from ./llama-kv-cache.h:5: ./llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ ./llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ ./llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ llama-model.cpp:407:61: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 407 | auto * buft = ggml_backend_split_buffer_type_fn(dev_index, tensor_split); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~ llama-model.cpp:2108:38: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 2108 | hparams.f_norm_eps = 1e-5; // eps for qk-norm, torch default | ~ ^~~~ llama-model.cpp:632:47: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 632 | hparams.n_cls_out = classifier_labels.size(); | ~ ~~~~~~~~~~~~~~~~~~^~~~~~ llama-model.cpp:2517:25: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'value_type' (aka 'float') may lose precision [-Wimplicit-int-float-conversion] 2517 | splits[i] = free; | ~ ^~~~ llama-model.cpp:2541:118: warning: implicit conversion from 'const int' to 'float' may lose precision [-Wimplicit-int-float-conversion] 2541 | const int layer_gpu = std::upper_bound(splits.begin(), splits.begin() + n_devices(), float(il - i_gpu_start)/act_gpu_layers) - splits.begin(); | ~^~~~~~~~~~~~~~ llama-model.cpp:2541:134: warning: implicit conversion loses integer precision: 'decltype(__x.base() - __y.base())' (aka 'long') to 'const int' [-Wshorten-64-to-32] 2541 | const int layer_gpu = std::upper_bound(splits.begin(), splits.begin() + n_devices(), float(il - i_gpu_start)/act_gpu_layers) - splits.begin(); | ~~~~~~~~~ ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~ llama-model.cpp:7100:64: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7100 | __func__, ggml_backend_buffer_name(buf.get()), ggml_backend_buffer_get_size(buf.get()) / 1024.0 / 1024.0); | ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-model.cpp:5650:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'const int' [-Wshorten-64-to-32] 5650 | const int attn_hidden_size = n_embd; | ~~~~~~~~~~~~~~~~ ^~~~~~ llama-model.cpp:5707:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'const int' [-Wshorten-64-to-32] 5707 | const int attn_hidden_size = n_embd; | ~~~~~~~~~~~~~~~~ ^~~~~~ llama-model.cpp:5766:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'const int' [-Wshorten-64-to-32] 5766 | const int attn_hidden_size = n_embd; | ~~~~~~~~~~~~~~~~ ^~~~~~ llama-model.cpp:5833:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'const int' [-Wshorten-64-to-32] 5833 | const int attn_hidden_size = n_embd; | ~~~~~~~~~~~~~~~~ ^~~~~~ llama-model.cpp:7294:16: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7294 | if (pimpl->n_elements >= 1e12) { | ~~~~~~~^~~~~~~~~~ ~~ llama-model.cpp:7295:81: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7295 | LLAMA_LOG_INFO("%s: model params = %.2f T\n", __func__, pimpl->n_elements*1e-12); | ~~~~~~~^~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-model.cpp:7296:23: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7296 | } else if (pimpl->n_elements >= 1e9) { | ~~~~~~~^~~~~~~~~~ ~~ llama-model.cpp:7297:81: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7297 | LLAMA_LOG_INFO("%s: model params = %.2f B\n", __func__, pimpl->n_elements*1e-9); | ~~~~~~~^~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-model.cpp:7298:23: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7298 | } else if (pimpl->n_elements >= 1e6) { | ~~~~~~~^~~~~~~~~~ ~~ llama-model.cpp:7299:81: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7299 | LLAMA_LOG_INFO("%s: model params = %.2f M\n", __func__, pimpl->n_elements*1e-6); | ~~~~~~~^~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-model.cpp:7301:81: warning: implicit conversion from 'uint64_t' (aka 'unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 7301 | LLAMA_LOG_INFO("%s: model params = %.2f K\n", __func__, pimpl->n_elements*1e-3); | ~~~~~~~^~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ 25 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-model-loader.cpp -o llama-model-loader.o In file included from llama-model-loader.cpp:1: ./llama-model-loader.h:35:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 35 | const int tensor_idx = gguf_find_tensor(gguf_ctx, ggml_get_name(tensor)); | ~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:266:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 266 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:294:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 294 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:335:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 335 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:430:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 430 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:476:24: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 476 | const int id = gguf_find_key(meta.get(), key.c_str()); | ~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:683:25: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 683 | ggml_nbytes(tensor)/1024.0f/1024.0f); | ^~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-model-loader.cpp:611:33: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 611 | const int kid = gguf_find_key(ctx_gguf.get(), kv_split_no.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:650:17: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 650 | n_kv = gguf_get_n_kv(meta.get()); | ~ ^~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:651:29: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 651 | n_tensors = weights_map.size(); | ~ ~~~~~~~~~~~~^~~~~~ llama-model-loader.cpp:1088:56: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 1088 | if (!progress_callback((float) size_done / size_data, progress_callback_user_data)) { | ~ ^~~~~~~~~ llama-model-loader.cpp:1257:120: warning: implicit conversion from 'const uint64_t' (aka 'const unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1257 | LLAMA_LOG_INFO("%s: file size = %.2f MiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0, n_bytes*8.0/n_elements); | ~^~~~~~~~~~ llama-model-loader.cpp:1257:108: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1257 | LLAMA_LOG_INFO("%s: file size = %.2f MiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0, n_bytes*8.0/n_elements); | ^~~~~~~~ llama-model-loader.cpp:1257:78: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1257 | LLAMA_LOG_INFO("%s: file size = %.2f MiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0, n_bytes*8.0/n_elements); | ^~~~~~~~ llama-model-loader.cpp:1259:120: warning: implicit conversion from 'const uint64_t' (aka 'const unsigned long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1259 | LLAMA_LOG_INFO("%s: file size = %.2f GiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0/1024.0, n_bytes*8.0/n_elements); | ~^~~~~~~~~~ llama-model-loader.cpp:1259:108: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1259 | LLAMA_LOG_INFO("%s: file size = %.2f GiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0/1024.0, n_bytes*8.0/n_elements); | ^~~~~~~~ llama-model-loader.cpp:1259:78: warning: implicit conversion from 'const size_t' (aka 'const unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1259 | LLAMA_LOG_INFO("%s: file size = %.2f GiB (%.2f BPW) \n", __func__, n_bytes/1024.0/1024.0/1024.0, n_bytes*8.0/n_elements); | ^~~~~~~~ llama-model-loader.cpp:266:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 266 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:286:16: note: in instantiation of function template specialization 'llama_model_loader::get_arr_n' requested here 286 | return get_arr_n(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:279:27: warning: implicit conversion loses integer precision: 'const size_t' (aka 'const unsigned long') to 'unsigned int' [-Wshorten-64-to-32] 279 | result = arr_info.length; | ~ ~~~~~~~~~^~~~~~ llama-model-loader.cpp:294:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 294 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:384:16: note: in instantiation of function template specialization 'llama_model_loader::get_arr' requested here 384 | return get_arr(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:228:32: warning: implicit conversion loses floating-point precision: 'const double' to 'float' [-Wimplicit-float-conversion] 228 | target = ovrd->val_f64; | ~ ~~~~~~^~~~~~~ llama-model-loader.cpp:245:17: note: in instantiation of function template specialization 'GGUFMeta::GKV::try_override' requested here 245 | if (try_override(target, ovrd)) { | ^ llama-model-loader.cpp:254:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ^ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:218:32: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'unsigned int' [-Wshorten-64-to-32] 218 | target = ovrd->val_i64; | ~ ~~~~~~^~~~~~~ llama-model-loader.cpp:245:17: note: in instantiation of function template specialization 'GGUFMeta::GKV::try_override' requested here 245 | if (try_override(target, ovrd)) { | ^ llama-model-loader.cpp:254:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ^ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:407:16: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 407 | return get_key(llm_kv(kid), result, required); | ^ llama-model-loader.cpp:430:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 430 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:335:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 335 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:451:20: note: in instantiation of function template specialization 'llama_model_loader::get_arr' requested here 451 | return get_arr(key, result, required); | ^ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:430:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 430 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:335:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 335 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:451:20: note: in instantiation of function template specialization 'llama_model_loader::get_arr' requested here 451 | return get_arr(key, result, required); | ^ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:430:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 430 | const int kid = gguf_find_key(meta.get(), key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:335:25: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 335 | const int kid = gguf_find_key(ctx, key.c_str()); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:451:20: note: in instantiation of function template specialization 'llama_model_loader::get_arr' requested here 451 | return get_arr(key, result, required); | ^ llama-model-loader.cpp:470:16: note: in instantiation of function template specialization 'llama_model_loader::get_key_or_arr' requested here 470 | return get_key_or_arr(llm_kv(kid), result, n, required); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:570:5: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 570 | get_key(llm_kv(LLM_KV_SPLIT_COUNT), n_split, false); | ^ llama-model-loader.cpp:218:32: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'unsigned short' [-Wimplicit-int-conversion] 218 | target = ovrd->val_i64; | ~ ~~~~~~^~~~~~~ llama-model-loader.cpp:245:17: note: in instantiation of function template specialization 'GGUFMeta::GKV::try_override' requested here 245 | if (try_override(target, ovrd)) { | ^ llama-model-loader.cpp:254:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ^ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:570:5: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 570 | get_key(llm_kv(LLM_KV_SPLIT_COUNT), n_split, false); | ^ llama-model-loader.cpp:254:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'int' [-Wshorten-64-to-32] 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ~~~ ^~~~~~~~~~~~~~~~~~~~~~~ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:637:9: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 637 | get_key(llm_kv(LLM_KV_SPLIT_TENSORS_COUNT), n_tensors); | ^ llama-model-loader.cpp:218:32: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 218 | target = ovrd->val_i64; | ~ ~~~~~~^~~~~~~ llama-model-loader.cpp:245:17: note: in instantiation of function template specialization 'GGUFMeta::GKV::try_override' requested here 245 | if (try_override(target, ovrd)) { | ^ llama-model-loader.cpp:254:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 254 | return set(ctx, gguf_find_key(ctx, key), target, ovrd); | ^ llama-model-loader.cpp:258:20: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 258 | return set(ctx, key.c_str(), target, ovrd); | ^ llama-model-loader.cpp:396:46: note: in instantiation of member function 'GGUFMeta::GKV::set' requested here 396 | const bool found = GGUFMeta::GKV::set(meta.get(), key, result, override); | ^ llama-model-loader.cpp:637:9: note: in instantiation of function template specialization 'llama_model_loader::get_key' requested here 637 | get_key(llm_kv(LLM_KV_SPLIT_TENSORS_COUNT), n_tensors); | ^ 36 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-model-saver.cpp -o llama-model-saver.o In file included from llama-model-saver.cpp:7: In file included from ./llama-model.h:5: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -include r_llama_compat.h -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-quant.cpp -o llama-quant.o In file included from llama-quant.cpp:3: In file included from ./llama-model.h:5: In file included from ./llama-graph.h:6: ./llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ ./llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from llama-quant.cpp:4: ./llama-model-loader.h:35:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 35 | const int tensor_idx = gguf_find_tensor(gguf_ctx, ggml_get_name(tensor)); | ~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ In file included from llama-quant.cpp:10: /Library/Developer/CommandLineTools/SDKs/MacOSX.sdk/usr/include/c++/v1/fstream:1030:14: error: expected unqualified-id 1030 | if (std::fflush(__file_)) | ^ ./r_llama_compat.h:38:5: note: expanded from macro 'fflush' 38 | ((stream == r_llama_dummy_stream_) ? 0 : fflush(stream)) | ^ llama-quant.cpp:602:58: warning: implicit conversion loses floating-point precision: 'const double' to 'float' [-Wimplicit-float-conversion] 602 | gguf_set_val_f32(ctx_out.get(), o.key, o.val_f64); | ~~~~~~~~~~~~~~~~ ~~^~~~~~~ llama-quant.cpp:702:85: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'uint16_t' (aka 'unsigned short') [-Wimplicit-int-conversion] 702 | gguf_set_val_u16(ctx_outs[i].get(), ml.llm_kv(LLM_KV_SPLIT_NO).c_str(), i); | ~~~~~~~~~~~~~~~~ ^ llama-quant.cpp:907:50: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 907 | LLAMA_LOG_INFO("size = %8.3f MiB\n", ggml_nbytes(tensor)/1024.0/1024.0); | ^~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-quant.cpp:1008:98: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1008 | LLAMA_LOG_INFO("size = %8.2f MiB -> %8.2f MiB\n", ggml_nbytes(tensor)/1024.0/1024.0, new_size/1024.0/1024.0); | ^~~~~~~~~ llama-quant.cpp:1008:63: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1008 | LLAMA_LOG_INFO("size = %8.2f MiB -> %8.2f MiB\n", ggml_nbytes(tensor)/1024.0/1024.0, new_size/1024.0/1024.0); | ^~~~~~~~~~~~~~~~~~~~ ./llama-impl.h:27:71: note: expanded from macro 'LLAMA_LOG_INFO' 27 | #define LLAMA_LOG_INFO(...) llama_log_internal(GGML_LOG_LEVEL_INFO , __VA_ARGS__) | ^~~~~~~~~~~ llama-quant.cpp:1024:63: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1024 | LLAMA_LOG_INFO("%s: model size = %8.2f MiB\n", __func__, total_size_org/1024.0/1024.0); | ^~~~~~~~~~~~~~~ llama-quant.cpp:1025:63: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 1025 | LLAMA_LOG_INFO("%s: quant size = %8.2f MiB\n", __func__, total_size_new/1024.0/1024.0); | ^~~~~~~~~~~~~~~ llama-quant.cpp:983:143: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 983 | new_size += llama_tensor_quantize_impl(new_type, f32_data_03, new_data_03, chunk_size, nrows, n_per_row, imatrix_03, workers, nthread_use); | ~~~~~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~ 11 warnings and 1 error generated. make[1]: *** [llama-quant.o] Error 1 clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-sampling.cpp -o llama-sampling.o llama-sampling.cpp:290:67: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 290 | llama_token_data_array_partial_sort_inplace(cur_p, cur_p->size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~^~~~ llama-sampling.cpp:918:24: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 918 | return p->samplers.size(); | ~~~~~~ ~~~~~~~~~~~~^~~~~~ llama-sampling.cpp:1090:29: warning: implicit conversion when assigning computation result loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 1090 | cur_p->data[i].p /= sum_cum; | ~~ ^~~~~~~ llama-sampling.cpp:1207:81: warning: implicit conversion from 'int64_t' (aka 'long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 1207 | struct ggml_tensor * idx = ggml_cast(ctx, ggml_scale_bias(ctx, idxf, -1.0f, mask->ne[0]), GGML_TYPE_I32); | ~~~~~~~~~~~~~~~ ^~~~~~~~~~~ llama-sampling.cpp:1233:23: warning: implicit conversion loses floating-point precision: 'result_type' (aka 'double') to 'const float' [-Wimplicit-float-conversion] 1233 | const float rnd = dist(sctx->rng); | ~~~ ^~~~~~~~~~~~~~~ llama-sampling.cpp:1391:53: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 1391 | llama_token_data_array_partial_sort(*cur_p, k, buf_sort); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ llama-sampling.cpp:1395:60: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 1395 | llama_token_data_array_partial_sort_inplace(cur_p, k); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ llama-sampling.cpp:1415:57: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 1415 | llama_token_data_array_partial_sort(*cur_p, k, buf_sort); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^ llama-sampling.cpp:1500:52: warning: implicit conversion from 'int64_t' (aka 'long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 1500 | idxf = ggml_clamp(ctx, idxf, 0.0f, mask->ne[0] - 1); | ~~~~~~~~~~ ~~~~~~~~~~~~^~~ llama-sampling.cpp:1603:71: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 1603 | llama_token_data_array_partial_sort_inplace(cur_p, cur_p->size); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~^~~~ llama-sampling.cpp:1950:49: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 1950 | float max_entropy = -logf(1.0f / cur_p->size); | ~ ~~~~~~~^~~~ llama-sampling.cpp:1987:32: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 1987 | cur_p->data[i].p = p; // Store the scaled probability | ~ ^ llama-sampling.cpp:1992:33: warning: implicit conversion when assigning computation result loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 1992 | cur_p->data[i].p /= cum_sum_double; // Re-normalize the probabilities | ~~ ^~~~~~~~~~~~~~ llama-sampling.cpp:2044:36: warning: implicit conversion from 'int64_t' (aka 'long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 2044 | const float max_entropy = logf(data->logits->ne[0]); | ~~~~ ^~~~~~~~~~~~~~~~~~~ llama-sampling.cpp:2162:24: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 2162 | pos_last = i; | ~ ^ llama-sampling.cpp:2274:70: warning: implicit conversion from 'const int32_t' (aka 'const int') to 'float' may lose precision [-Wimplicit-int-float-conversion] 2274 | float k = powf((epsilon_hat * powf(2, ctx->mu)) / (1 - powf(ctx->n_vocab, -epsilon_hat)), 1 / s_hat); | ~~~~ ~~~~~^~~~~~~ llama-sampling.cpp:2821:47: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 2821 | float mean = valid_count > 0 ? logits_sum/valid_count : 0; | ~^~~~~~~~~~~ llama-sampling.cpp:2828:20: warning: implicit conversion when assigning computation result loses floating-point precision: '__promote_t' (aka 'double') to 'float' [-Wimplicit-float-conversion] 2828 | acc += pow(cur_p->data[i].logit - mean, 2); | ~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-sampling.cpp:2831:44: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 2831 | float std = valid_count > 0 ? sqrt(acc/valid_count) : 0; | ~^~~~~~~~~~~ llama-sampling.cpp:3129:38: warning: implicit conversion turns floating-point number into integer: 'float' to 'int' [-Wfloat-conversion] 3129 | max_exponent = FLOAT_MAX_LOG / std::log(ctx->dry_base); | ~ ~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~ llama-sampling.cpp:3152:53: warning: implicit conversion loses floating-point precision: '__promote_t' (aka 'double') to 'float' [-Wimplicit-float-conversion] 3152 | float penalty = ctx->dry_multiplier * std::pow(ctx->dry_base, repeat_exp); | ~~~~~~~ ~~~~~~~~~~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-sampling.cpp:3505:72: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3505 | return llama_sampler_init_logit_bias(ctx->n_vocab, ctx->logit_bias.size(), ctx->logit_bias.data()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~~~~~~~~~^~~~~~ llama-sampling.cpp:3678:28: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 3678 | if (3*p_eog_sum*cur_p->size > p_txt_sum) { | ~~~~~~~~^~~~ llama-sampling.cpp:3803:16: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 3803 | thold = 1.0/(n_non_eog + 1); | ~ ~~~^~~~~~~~~~~~~~~~ llama-sampling.cpp:3803:28: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 3803 | thold = 1.0/(n_non_eog + 1); | ~ ~~~~~~~~~~^~~ llama-sampling.cpp:3717:99: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3717 | int len0 = ctx->vocab->token_to_piece(cur_p->data[i0].id, ctx->buf0.data(), ctx->buf0.size(), 0, false); | ~~~~~~~~~~~~~~ ~~~~~~~~~~^~~~~~ llama-sampling.cpp:3720:99: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3720 | len0 = ctx->vocab->token_to_piece(cur_p->data[i0].id, ctx->buf0.data(), ctx->buf0.size(), 0, false); | ~~~~~~~~~~~~~~ ~~~~~~~~~~^~~~~~ llama-sampling.cpp:3724:99: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3724 | int len1 = ctx->vocab->token_to_piece(cur_p->data[i1].id, ctx->buf1.data(), ctx->buf1.size(), 0, false); | ~~~~~~~~~~~~~~ ~~~~~~~~~~^~~~~~ llama-sampling.cpp:3727:99: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3727 | len1 = ctx->vocab->token_to_piece(cur_p->data[i1].id, ctx->buf1.data(), ctx->buf1.size(), 0, false); | ~~~~~~~~~~~~~~ ~~~~~~~~~~^~~~~~ llama-sampling.cpp:3733:27: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 3733 | int dst = i0; | ~~~ ^~ llama-sampling.cpp:3734:27: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 3734 | int src = i1; | ~~~ ^~ llama-sampling.cpp:3903:36: warning: implicit conversion from 'int64_t' (aka 'long long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 3903 | data.t_sample_ms = 1e-3 * ctx->t_sample_us; | ~ ~~~~~^~~~~~~~~~~ 32 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c llama-vocab.cpp -o llama-vocab.o In file included from llama-vocab.cpp:6: ./llama-model-loader.h:35:36: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 35 | const int tensor_idx = gguf_find_tensor(gguf_ctx, ggml_get_name(tensor)); | ~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:596:78: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'index' (aka 'int') [-Wshorten-64-to-32] 596 | symbols_final[final_prev_index].next = symbols_final.size(); | ~ ~~~~~~~~~~~~~~^~~~~~ llama-vocab.cpp:599:61: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 599 | final_prev_index = symbols_final.size() - 1; | ~ ~~~~~~~~~~~~~~~~~~~~~^~~ llama-vocab.cpp:691:33: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'const int' [-Wshorten-64-to-32] 691 | const int n = word1.size(); | ~ ~~~~~~^~~~~~ llama-vocab.cpp:1150:38: warning: implicit conversion loses integer precision: 'int' to 'uint8_t' (aka 'unsigned char') [-Wimplicit-int-conversion] 1150 | hex_acc = (hex_acc << 4) + value; | ~ ~~~~~~~~~~~~~~~^~~~~~~ llama-vocab.cpp:1259:55: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'llama_token' (aka 'int') [-Wshorten-64-to-32] 1259 | const auto & entry = vocab.get_token_data(token_id); | ~~~~~~~~~~~~~~ ^~~~~~~~ llama-vocab.cpp:1264:31: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'llama_token' (aka 'int') [-Wshorten-64-to-32] 1264 | if (vocab.is_byte(token_id)) { | ~~~~~~~ ^~~~~~~~ llama-vocab.cpp:1461:29: warning: implicit conversion loses integer precision: 'uint32_t' (aka 'unsigned int') to 'uint8_t' (aka 'unsigned char') [-Wimplicit-int-conversion] 1461 | b = c; | ~ ^ llama-vocab.cpp:2543:79: warning: implicit conversion from 'size_t' (aka 'unsigned long') to 'double' may lose precision [-Wimplicit-int-float-conversion] 2543 | LLAMA_LOG_INFO("%s: token to piece cache size = %.4f MB\n", __func__, size_cache / 1024.0 / 1024.0); | ^~~~~~~~~~ ~ llama-vocab.cpp:1754:39: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 1754 | const int merges_keyidx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_MERGES).c_str()); | ~~~~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:1759:34: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'const int' [-Wshorten-64-to-32] 1759 | const int n_merges = gguf_get_arr_n(ctx, merges_keyidx); | ~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:1795:53: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 1795 | const int precompiled_charsmap_keyidx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_PRECOMPILED_CHARSMAP).c_str()); | ~~~~~~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2078:27: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 2078 | const int token_idx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_LIST).c_str()); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2084:27: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 2084 | const int score_idx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_SCORES).c_str()); | ~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2090:29: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 2090 | const int toktype_idx = gguf_find_key(ctx, kv(LLM_KV_TOKENIZER_TOKEN_TYPE).c_str()); | ~~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2095:25: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 2095 | uint32_t n_tokens = gguf_get_arr_n(ctx, token_idx); | ~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2675:20: warning: implicit conversion loses integer precision: 'long' to 'uint8_t' (aka 'unsigned char') [-Wimplicit-int-conversion] 2675 | return strtol(buf.c_str(), NULL, 16); | ~~~~~~ ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~ llama-vocab.cpp:2848:70: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 2848 | const int n_chars = vocab.token_to_piece(token, &piece[0], piece.size(), 0, special); | ~~~~~~~~~~~~~~ ~~~~~~^~~~~~ llama-vocab.cpp:2851:66: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 2851 | int check = vocab.token_to_piece(token, &piece[0], piece.size(), 0, special); | ~~~~~~~~~~~~~~ ~~~~~~^~~~~~ llama-vocab.cpp:3624:16: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 3624 | return res.size(); | ~~~~~~ ~~~~^~~~~~ 20 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -include r_llama_compat.h -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c unicode.cpp -o unicode.o unicode.cpp:154:13: warning: implicit conversion loses integer precision: 'int' to 'key_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 154 | map[ch] = unicode_cpt_to_utf8(ch); | ~~~ ^~ unicode.cpp:158:13: warning: implicit conversion loses integer precision: 'int' to 'key_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 158 | map[ch] = unicode_cpt_to_utf8(ch); | ~~~ ^~ unicode.cpp:162:13: warning: implicit conversion loses integer precision: 'int' to 'key_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 162 | map[ch] = unicode_cpt_to_utf8(ch); | ~~~ ^~ unicode.cpp:166:22: warning: implicit conversion loses integer precision: 'int' to 'const key_type' (aka 'const unsigned char') [-Wimplicit-int-conversion] 166 | if (map.find(ch) == map.end()) { | ~~~~ ^~ unicode.cpp:167:17: warning: implicit conversion loses integer precision: 'int' to 'key_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 167 | map[ch] = unicode_cpt_to_utf8(256 + n); | ~~~ ^~ unicode.cpp:178:40: warning: implicit conversion loses integer precision: 'int' to 'mapped_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 178 | map[unicode_cpt_to_utf8(ch)] = ch; | ~ ^~ unicode.cpp:182:40: warning: implicit conversion loses integer precision: 'int' to 'mapped_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 182 | map[unicode_cpt_to_utf8(ch)] = ch; | ~ ^~ unicode.cpp:186:40: warning: implicit conversion loses integer precision: 'int' to 'mapped_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 186 | map[unicode_cpt_to_utf8(ch)] = ch; | ~ ^~ unicode.cpp:191:49: warning: implicit conversion loses integer precision: 'int' to 'mapped_type' (aka 'unsigned char') [-Wimplicit-int-conversion] 191 | map[unicode_cpt_to_utf8(256 + n)] = ch; | ~ ^~ unicode.cpp:843:26: warning: implicit conversion loses integer precision: 'uint32_t' (aka 'unsigned int') to 'value_type' (aka 'char') [-Wimplicit-int-conversion] 843 | result.push_back(cpt); | ~~~~~~~~~ ^~~ unicode.cpp:1022:37: warning: implicit conversion loses integer precision: 'const value_type' (aka 'const unsigned int') to 'value_type' (aka 'char') [-Wimplicit-int-conversion] 1022 | text_collapsed[i] = cpts[i]; | ~ ^~~~~~~ unicode.cpp:1033:48: warning: implicit conversion loses integer precision: 'const mapped_type' (aka 'const int') to 'value_type' (aka 'char') [-Wimplicit-int-conversion] 1033 | text_collapsed[i] = k_ucat_cpt.at(flags.category_flag()); | ~ ~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~ unicode.cpp:1102:68: warning: implicit conversion loses integer precision: 'const mapped_type' (aka 'const int') to 'value_type' (aka 'char') [-Wimplicit-int-conversion] 1102 | regex_expr_collapsed += k_ucat_cpt.at(k_ucat_enum.at(pat)); | ~~ ~~~~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~ 13 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c unicode-data.cpp -o unicode-data.o clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c r_llama_interface.cpp -o r_llama_interface.o r_llama_interface.cpp:754:99: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 754 | int actual = llama_chat_apply_template(tmpl, messages.data(), n_msg, add_ass, buf.data(), buf.size()); | ~~~~~~~~~~~~~~~~~~~~~~~~~ ~~~~^~~~~~ 1 warning generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/afmoe.cpp -o models/afmoe.o In file included from models/afmoe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/afmoe.cpp:130:29: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 121 | ggml_tensor * moe_out = build_moe_ffn(cur, | ~~~~~~~~~~~~~ 122 | model.layers[il].ffn_gate_inp, 123 | model.layers[il].ffn_up_exps, 124 | model.layers[il].ffn_gate_exps, 125 | model.layers[il].ffn_down_exps, 126 | model.layers[il].ffn_exp_probs_b, 127 | n_expert, n_expert_used, 128 | LLM_FFN_SILU, 129 | hparams.expert_weights_norm, // norm_w (route_norm=True) 130 | hparams.expert_weights_scale, // scale_w | ~~~~~~~~^~~~~~~~~~~~~~~~~~~~ models/afmoe.cpp:69:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 67 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 68 | ctx0, Qcur, inp_pos, nullptr, 69 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/afmoe.cpp:75:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 73 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 74 | ctx0, Kcur, inp_pos, nullptr, 75 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/apertus.cpp -o models/apertus.o In file included from models/apertus.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/apertus.cpp:54:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 54 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/apertus.cpp:57:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/arcee.cpp -o models/arcee.o In file included from models/arcee.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/arcee.cpp:66:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 65 | ctx0, Qcur, inp_pos, rope_factors, 66 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/arcee.cpp:72:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 70 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 71 | ctx0, Kcur, inp_pos, rope_factors, 72 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/arctic.cpp -o models/arctic.o In file included from models/arctic.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/arctic.cpp:49:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 48 | ctx0, Qcur, inp_pos, nullptr, 49 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/arctic.cpp:55:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 53 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 54 | ctx0, Kcur, inp_pos, nullptr, 55 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/arwkv7.cpp -o models/arwkv7.o In file included from models/arwkv7.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/baichuan.cpp -o models/baichuan.o In file included from models/baichuan.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/baichuan.cpp:49:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 48 | ctx0, Qcur, inp_pos, nullptr, 49 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/baichuan.cpp:54:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 53 | ctx0, Kcur, inp_pos, nullptr, 54 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/bailingmoe.cpp -o models/bailingmoe.o In file included from models/bailingmoe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/bailingmoe.cpp:59:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 58 | ctx0, Qcur, inp_pos, rope_factors, 59 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/bailingmoe.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Kcur, inp_pos, rope_factors, 65 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/bailingmoe2.cpp -o models/bailingmoe2.o In file included from models/bailingmoe2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/bailingmoe2.cpp:24:46: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 24 | const int n_transformer_layers = n_layer - hparams.nextn_predict_layers; | ~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ models/bailingmoe2.cpp:47:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/bailingmoe2.cpp:53:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 53 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/bert.cpp -o models/bert.o In file included from models/bert.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/bert.cpp:93:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 93 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/bert.cpp:96:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 96 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/bitnet.cpp -o models/bitnet.o In file included from models/bitnet.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/bitnet.cpp:70:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 68 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 69 | ctx0, Qcur, inp_pos, nullptr, 70 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/bitnet.cpp:76:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 75 | ctx0, Kcur, inp_pos, nullptr, 76 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/bloom.cpp -o models/bloom.o In file included from models/bloom.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/chameleon.cpp -o models/chameleon.o In file included from models/chameleon.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/chameleon.cpp:82:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 80 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 81 | ctx0, Qcur, inp_pos, nullptr, 82 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/chameleon.cpp:88:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 86 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 87 | ctx0, Kcur, inp_pos, nullptr, 88 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/chatglm.cpp -o models/chatglm.o In file included from models/chatglm.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/chatglm.cpp:68:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 67 | ctx0, Qcur, inp_pos, nullptr, 68 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/chatglm.cpp:74:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 72 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 73 | ctx0, Kcur, inp_pos, nullptr, 74 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/codeshell.cpp -o models/codeshell.o In file included from models/codeshell.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/codeshell.cpp:43:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 41 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 42 | ctx0, Qcur, inp_pos, nullptr, 43 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/codeshell.cpp:49:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 48 | ctx0, Kcur, inp_pos, nullptr, 49 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/cogvlm.cpp -o models/cogvlm.o In file included from models/cogvlm.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/cogvlm.cpp:63:51: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Qcur = ggml_rope(ctx0, Qcur, inp_pos, n_embd_head, rope_type); | ~~~~~~~~~ ^~~~~~~~~~~ models/cogvlm.cpp:64:51: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope(ctx0, Kcur, inp_pos, n_embd_head, rope_type); | ~~~~~~~~~ ^~~~~~~~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/cohere2-iswa.cpp -o models/cohere2-iswa.o In file included from models/cohere2-iswa.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/cohere2-iswa.cpp:126:9: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 126 | if (f_logit_scale) { | ~~ ^~~~~~~~~~~~~ models/cohere2-iswa.cpp:67:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 65 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 66 | ctx0, Qcur, inp_pos, rope_factors, 67 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/cohere2-iswa.cpp:73:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 71 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 72 | ctx0, Kcur, inp_pos, rope_factors, 73 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/command-r.cpp -o models/command-r.o In file included from models/command-r.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/command-r.cpp:115:9: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 115 | if (f_logit_scale) { | ~~ ^~~~~~~~~~~~~ models/command-r.cpp:61:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/command-r.cpp:68:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 68 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/dbrx.cpp -o models/dbrx.o In file included from models/dbrx.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/dbrx.cpp:50:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 48 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 49 | ctx0, Qcur, inp_pos, nullptr, 50 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/dbrx.cpp:56:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 54 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 55 | ctx0, Kcur, inp_pos, nullptr, 56 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/deci.cpp -o models/deci.o In file included from models/deci.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/deci.cpp:72:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 72 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/deci.cpp:75:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 75 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/deepseek.cpp -o models/deepseek.o In file included from models/deepseek.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/deepseek.cpp:62:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/deepseek.cpp:65:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 65 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/deepseek2.cpp -o models/deepseek2.o In file included from models/deepseek2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/deepseek2.cpp:218:25: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 210 | ggml_tensor * moe_out = build_moe_ffn(cur, | ~~~~~~~~~~~~~ 211 | model.layers[il].ffn_gate_inp, 212 | model.layers[il].ffn_up_exps, 213 | model.layers[il].ffn_gate_exps, 214 | model.layers[il].ffn_down_exps, 215 | model.layers[il].ffn_exp_probs_b, 216 | n_expert, n_expert_used, 217 | LLM_FFN_SILU, hparams.expert_weights_norm, 218 | hparams.expert_weights_scale, hparams.expert_weights_scale, | ~~~~~~~~^~~~~~~~~~~~~~~~~~~~ models/deepseek2.cpp:102:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 102 | q_pe = ggml_rope_ext(ctx0, q_pe, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/deepseek2.cpp:106:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 106 | k_pe = ggml_rope_ext(ctx0, k_pe, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/dots1.cpp -o models/dots1.o In file included from models/dots1.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/dots1.cpp:50:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/dots1.cpp:56:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/dream.cpp -o models/dream.o In file included from models/dream.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/dream.cpp:51:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 51 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/dream.cpp:54:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 54 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/ernie4-5-moe.cpp -o models/ernie4-5-moe.o In file included from models/ernie4-5-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/ernie4-5-moe.cpp:57:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/ernie4-5-moe.cpp:60:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 60 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/ernie4-5.cpp -o models/ernie4-5.o In file included from models/ernie4-5.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/ernie4-5.cpp:54:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 54 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/ernie4-5.cpp:57:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/exaone-moe.cpp -o models/exaone-moe.o In file included from models/exaone-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/exaone-moe.cpp:23:46: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 23 | const int n_transformer_layers = n_layer - hparams.nextn_predict_layers; | ~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ models/exaone-moe.cpp:58:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ models/exaone-moe.cpp:61:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/exaone.cpp -o models/exaone.o In file included from models/exaone.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/exaone.cpp:59:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 59 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/exaone.cpp:62:69: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/exaone4.cpp -o models/exaone4.o In file included from models/exaone4.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/exaone4.cpp:61:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ models/exaone4.cpp:122:17: note: in instantiation of member function 'llm_build_exaone4::llm_build_exaone4' requested here 122 | template struct llm_build_exaone4; | ^ models/exaone4.cpp:64:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ models/exaone4.cpp:61:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ models/exaone4.cpp:123:17: note: in instantiation of member function 'llm_build_exaone4::llm_build_exaone4' requested here 123 | template struct llm_build_exaone4; | ^ models/exaone4.cpp:64:73: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, | ~~~~~~~~~~~~~ ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/falcon-h1.cpp -o models/falcon-h1.o In file included from models/falcon-h1.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/falcon-h1.cpp:46:60: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 46 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, hparams.rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/falcon-h1.cpp:49:60: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 49 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, hparams.rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/falcon.cpp -o models/falcon.o In file included from models/falcon.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/falcon.cpp:55:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 53 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 54 | ctx0, Qcur, inp_pos, nullptr, 55 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/falcon.cpp:61:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 59 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 60 | ctx0, Kcur, inp_pos, nullptr, 61 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gemma-embedding.cpp -o models/gemma-embedding.o In file included from models/gemma-embedding.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gemma-embedding.cpp:13:56: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 13 | inpL = ggml_scale(ctx0, inpL, ubatch.token ? sqrtf(n_embd) : 1.0f); | ~~~~~ ^~~~~~ models/gemma-embedding.cpp:50:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/gemma-embedding.cpp:56:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gemma.cpp -o models/gemma.o In file included from models/gemma.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gemma.cpp:12:41: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 12 | inpL = ggml_scale(ctx0, inpL, sqrtf(n_embd)); | ~~~~~ ^~~~~~ models/gemma.cpp:47:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 45 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 46 | ctx0, Qcur, inp_pos, nullptr, 47 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/gemma.cpp:52:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 51 | ctx0, Kcur, inp_pos, nullptr, 52 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gemma2-iswa.cpp -o models/gemma2-iswa.o In file included from models/gemma2-iswa.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gemma2-iswa.cpp:11:41: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 11 | inpL = ggml_scale(ctx0, inpL, sqrtf(n_embd)); | ~~~~~ ^~~~~~ models/gemma2-iswa.cpp:49:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 48 | ctx0, Qcur, inp_pos, nullptr, 49 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/gemma2-iswa.cpp:54:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 53 | ctx0, Kcur, inp_pos, nullptr, 54 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gemma3.cpp -o models/gemma3.o In file included from models/gemma3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gemma3.cpp:13:56: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 13 | inpL = ggml_scale(ctx0, inpL, ubatch.token ? sqrtf(n_embd) : 1.0f); | ~~~~~ ^~~~~~ models/gemma3.cpp:154:17: note: in instantiation of member function 'llm_build_gemma3::llm_build_gemma3' requested here 154 | template struct llm_build_gemma3; | ^ models/gemma3.cpp:68:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 67 | ctx0, Qcur, inp_pos, nullptr, 68 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/gemma3.cpp:76:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 75 | ctx0, Kcur, inp_pos, nullptr, 76 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/gemma3.cpp:142:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 142 | if (hparams.f_final_logit_softcapping) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~ models/gemma3.cpp:13:56: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 13 | inpL = ggml_scale(ctx0, inpL, ubatch.token ? sqrtf(n_embd) : 1.0f); | ~~~~~ ^~~~~~ models/gemma3.cpp:155:17: note: in instantiation of member function 'llm_build_gemma3::llm_build_gemma3' requested here 155 | template struct llm_build_gemma3; | ^ models/gemma3.cpp:68:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 67 | ctx0, Qcur, inp_pos, nullptr, 68 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/gemma3.cpp:76:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 75 | ctx0, Kcur, inp_pos, nullptr, 76 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/gemma3.cpp:142:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 142 | if (hparams.f_final_logit_softcapping) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~ 10 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gemma3n-iswa.cpp -o models/gemma3n-iswa.o In file included from models/gemma3n-iswa.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gemma3n-iswa.cpp:16:56: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 16 | inpL = ggml_scale(ctx0, inpL, ubatch.token ? sqrtf(n_embd) : 1.0f); | ~~~~~ ^~~~~~ models/gemma3n-iswa.cpp:87:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 87 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/gemma3n-iswa.cpp:90:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 90 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/gemma3n-iswa.cpp:108:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 108 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/glm4-moe.cpp -o models/glm4-moe.o In file included from models/glm4-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/glm4-moe.cpp:31:46: warning: implicit conversion loses integer precision: 'int64_t' (aka 'long long') to 'const int' [-Wshorten-64-to-32] 31 | const int n_transformer_layers = n_layer - hparams.nextn_predict_layers; | ~~~~~~~~~~~~~~~~~~~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ models/glm4-moe.cpp:75:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Qcur = ggml_rope_multi(ctx0, Qcur, inp_pos, nullptr, | ~~~~~~~~~~~~~~~ 75 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/glm4-moe.cpp:79:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 78 | Kcur = ggml_rope_multi(ctx0, Kcur, inp_pos, nullptr, | ~~~~~~~~~~~~~~~ 79 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/glm4-moe.cpp:83:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 83 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, | ~~~~~~~~~~~~~ ^~~~~ models/glm4-moe.cpp:87:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 87 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, | ~~~~~~~~~~~~~ ^~~~~ 7 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/glm4.cpp -o models/glm4.o In file included from models/glm4.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/glm4.cpp:78:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 77 | Qcur = ggml_rope_multi(ctx0, Qcur, inp_pos, nullptr, | ~~~~~~~~~~~~~~~ 78 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/glm4.cpp:82:29: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 81 | Kcur = ggml_rope_multi(ctx0, Kcur, inp_pos, nullptr, | ~~~~~~~~~~~~~~~ 82 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/glm4.cpp:86:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 86 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, | ~~~~~~~~~~~~~ ^~~~~ models/glm4.cpp:90:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 90 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, | ~~~~~~~~~~~~~ ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gpt2.cpp -o models/gpt2.o In file included from models/gpt2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/gptneox.cpp -o models/gptneox.o In file included from models/gptneox.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/gptneox.cpp:43:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 41 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 42 | ctx0, Qcur, inp_pos, nullptr, 43 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/gptneox.cpp:49:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 47 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 48 | ctx0, Kcur, inp_pos, nullptr, 49 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/granite-hybrid.cpp -o models/granite-hybrid.o In file included from models/granite-hybrid.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/granite-hybrid.cpp:62:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 62 | if (hparams.f_logit_scale) { | ~~ ~~~~~~~~^~~~~~~~~~~~~ models/granite-hybrid.cpp:106:65: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 106 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/granite-hybrid.cpp:109:65: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 109 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, rope_factors, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/granite-hybrid.cpp:131:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 131 | if (hparams.f_residual_scale) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~ models/granite-hybrid.cpp:186:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 186 | if (hparams.f_residual_scale) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~ 7 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/granite.cpp -o models/granite.o In file included from models/granite.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/granite.cpp:111:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 109 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 110 | ctx0, Qcur, inp_pos, rope_factors, 111 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/granite.cpp:117:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 115 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 116 | ctx0, Kcur, inp_pos, rope_factors, 117 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/granite.cpp:141:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 141 | if (hparams.f_residual_scale) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~ models/granite.cpp:201:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 201 | if (hparams.f_residual_scale) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/graph-context-mamba.cpp -o models/graph-context-mamba.o In file included from models/graph-context-mamba.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/graph-context-mamba.cpp:35:77: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 35 | ggml_tensor * conv = build_rs(inp, conv_states_all, hparams.n_embd_r(), n_seqs); | ~~~~~~~~ ^~~~~~ models/graph-context-mamba.cpp:172:77: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 172 | ggml_tensor * conv = build_rs(inp, conv_states_all, hparams.n_embd_r(), n_seqs); | ~~~~~~~~ ^~~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/grok.cpp -o models/grok.o In file included from models/grok.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/grok.cpp:116:83: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 116 | cur = ggml_scale(ctx0, ggml_add(ctx0, ffn_out, moe_out), std::sqrt(2) / 2); | ~~~~~~~~~~ ~~~~~~~~~~~~~^~~ models/grok.cpp:150:17: warning: implicit conversion turns floating-point number into integer: 'const float' to 'bool' [-Wfloat-conversion] 150 | if (hparams.f_final_logit_softcapping) { | ~~ ~~~~~~~~^~~~~~~~~~~~~~~~~~~~~~~~~ models/grok.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/grok.cpp:63:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 62 | ctx0, Kcur, inp_pos, nullptr, 63 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/grovemoe.cpp -o models/grovemoe.o In file included from models/grovemoe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/grovemoe.cpp:51:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 51 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/grovemoe.cpp:57:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/hunyuan-dense.cpp -o models/hunyuan-dense.o In file included from models/hunyuan-dense.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/hunyuan-dense.cpp:61:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 59 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 60 | ctx0, Qcur, inp_pos, rope_factors, 61 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/hunyuan-dense.cpp:71:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 70 | ctx0, Kcur, inp_pos, rope_factors, 71 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/hunyuan-moe.cpp -o models/hunyuan-moe.o In file included from models/hunyuan-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/hunyuan-moe.cpp:62:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 60 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 61 | ctx0, Qcur, inp_pos, rope_factors, 62 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/hunyuan-moe.cpp:72:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 70 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 71 | ctx0, Kcur, inp_pos, rope_factors, 72 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/internlm2.cpp -o models/internlm2.o In file included from models/internlm2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/internlm2.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/internlm2.cpp:63:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 62 | ctx0, Kcur, inp_pos, nullptr, 63 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/jais.cpp -o models/jais.o In file included from models/jais.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/jamba.cpp -o models/jamba.o In file included from models/jamba.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/lfm2.cpp -o models/lfm2.o In file included from models/lfm2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ In file included from models/lfm2.cpp:3: In file included from models/../llama-memory-hybrid.h:5: In file included from models/../llama-kv-cache.h:5: models/../llama-kv-cells.h:60:20: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 60 | return pos.size(); | ~~~~~~ ~~~~^~~~~~ models/../llama-kv-cells.h:80:21: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 80 | return used.size(); | ~~~~~~ ~~~~~^~~~~~ models/../llama-kv-cells.h:144:25: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 144 | res.resize(idxs.size()); | ~~~~~~ ~~~~~^~~~~~ models/../llama-kv-cells.h:297:23: warning: implicit conversion loses integer precision: 'size_t' (aka 'unsigned long') to 'int' [-Wshorten-64-to-32] 297 | return seq[i].count(); | ~~~~~~ ~~~~~~~^~~~~~~ models/lfm2.cpp:102:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 102 | q = ggml_rope_ext(ctx0, q, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, ext_factor, | ~~~~~~~~~~~~~ ^~~~~ models/lfm2.cpp:104:50: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 104 | k = ggml_rope_ext(ctx0, k, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, ext_factor, | ~~~~~~~~~~~~~ ^~~~~ models/lfm2.cpp:148:76: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 148 | auto * conv_rs = build_rs(inp_recr, conv_state, hparams.n_embd_r(), n_seqs); | ~~~~~~~~ ^~~~~~ 9 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/llada-moe.cpp -o models/llada-moe.o In file included from models/llada-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/llada-moe.cpp:54:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 53 | ctx0, Qcur, inp_pos, nullptr, 54 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/llada-moe.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 59 | ctx0, Kcur, inp_pos, nullptr, 60 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/llada.cpp -o models/llada.o In file included from models/llada.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/llada.cpp:45:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 45 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/llada.cpp:48:64: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 48 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/llama-iswa.cpp -o models/llama-iswa.o In file included from models/llama-iswa.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/llama-iswa.cpp:74:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 72 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 73 | ctx0, Qcur, inp_pos, rope_factors, 74 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/llama-iswa.cpp:80:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 78 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 79 | ctx0, Kcur, inp_pos, rope_factors, 80 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/llama.cpp -o models/llama.o In file included from models/llama.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/llama.cpp:70:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 68 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 69 | ctx0, Qcur, inp_pos, rope_factors, 70 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/llama.cpp:167:17: note: in instantiation of member function 'llm_build_llama::llm_build_llama' requested here 167 | template struct llm_build_llama; | ^ models/llama.cpp:76:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 75 | ctx0, Kcur, inp_pos, rope_factors, 76 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/llama.cpp:70:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 68 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 69 | ctx0, Qcur, inp_pos, rope_factors, 70 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/llama.cpp:168:17: note: in instantiation of member function 'llm_build_llama::llm_build_llama' requested here 168 | template struct llm_build_llama; | ^ models/llama.cpp:76:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 74 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 75 | ctx0, Kcur, inp_pos, rope_factors, 76 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/maincoder.cpp -o models/maincoder.o In file included from models/maincoder.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/maincoder.cpp:48:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 46 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 47 | ctx0, Qcur, inp_pos, nullptr, 48 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/maincoder.cpp:54:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 53 | ctx0, Kcur, inp_pos, nullptr, 54 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/mamba.cpp -o models/mamba.o In file included from models/mamba.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/mimo2-iswa.cpp -o models/mimo2-iswa.o In file included from models/mimo2-iswa.cpp:2: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/mimo2-iswa.cpp:45:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 43 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 44 | ctx0, Qcur, inp_pos, nullptr, 45 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/mimo2-iswa.cpp:51:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 49 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 50 | ctx0, Kcur, inp_pos, nullptr, 51 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/minicpm3.cpp -o models/minicpm3.o In file included from models/minicpm3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/minicpm3.cpp:117:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 115 | q_pe = ggml_rope_ext( | ~~~~~~~~~~~~~ 116 | ctx0, q_pe, inp_pos, rope_factors, 117 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/minicpm3.cpp:125:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 123 | k_pe = ggml_rope_ext( | ~~~~~~~~~~~~~ 124 | ctx0, k_pe, inp_pos, rope_factors, 125 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/minimax-m2.cpp -o models/minimax-m2.o In file included from models/minimax-m2.cpp:2: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/minimax-m2.cpp:53:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 51 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 52 | ctx0, Qcur, inp_pos, nullptr, 53 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/minimax-m2.cpp:59:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 58 | ctx0, Kcur, inp_pos, nullptr, 59 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/mistral3.cpp -o models/mistral3.o In file included from models/mistral3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/mistral3.cpp:68:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 67 | ctx0, Qcur, inp_pos, rope_factors, 68 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/mistral3.cpp:74:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 72 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 73 | ctx0, Kcur, inp_pos, rope_factors, 74 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/modern-bert.cpp -o models/modern-bert.o In file included from models/modern-bert.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/modern-bert.cpp:52:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 51 | ctx0, Qcur, inp_pos, nullptr, 52 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/modern-bert.cpp:58:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 57 | ctx0, Kcur, inp_pos, nullptr, 58 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/mpt.cpp -o models/mpt.o In file included from models/mpt.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/nemotron-h.cpp -o models/nemotron-h.o In file included from models/nemotron-h.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/nemotron.cpp -o models/nemotron.o In file included from models/nemotron.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/nemotron.cpp:58:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 57 | ctx0, Qcur, inp_pos, nullptr, 58 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/nemotron.cpp:64:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 63 | ctx0, Kcur, inp_pos, nullptr, 64 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/neo-bert.cpp -o models/neo-bert.o In file included from models/neo-bert.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/neo-bert.cpp:45:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 43 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 44 | ctx0, Qcur, inp_pos, nullptr, 45 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/neo-bert.cpp:51:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 49 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 50 | ctx0, Kcur, inp_pos, nullptr, 51 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/olmo.cpp -o models/olmo.o In file included from models/olmo.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/olmo.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/olmo.cpp:63:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 62 | ctx0, Kcur, inp_pos, nullptr, 63 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/olmo2.cpp -o models/olmo2.o In file included from models/olmo2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/olmo2.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Qcur, inp_pos, nullptr, 65 | n_rot, rope_type, n_ctx_orig, freq_base, 1.0, | ^~~~~ models/olmo2.cpp:149:17: note: in instantiation of member function 'llm_build_olmo2::llm_build_olmo2' requested here 149 | template struct llm_build_olmo2; | ^ models/olmo2.cpp:71:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 70 | ctx0, Kcur, inp_pos, nullptr, 71 | n_rot, rope_type, n_ctx_orig, freq_base, 1.0, | ^~~~~ models/olmo2.cpp:77:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 75 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 76 | ctx0, Qcur, inp_pos, nullptr, 77 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/olmo2.cpp:83:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 81 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 82 | ctx0, Kcur, inp_pos, nullptr, 83 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/olmo2.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Qcur, inp_pos, nullptr, 65 | n_rot, rope_type, n_ctx_orig, freq_base, 1.0, | ^~~~~ models/olmo2.cpp:150:17: note: in instantiation of member function 'llm_build_olmo2::llm_build_olmo2' requested here 150 | template struct llm_build_olmo2; | ^ models/olmo2.cpp:71:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 70 | ctx0, Kcur, inp_pos, nullptr, 71 | n_rot, rope_type, n_ctx_orig, freq_base, 1.0, | ^~~~~ models/olmo2.cpp:77:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 75 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 76 | ctx0, Qcur, inp_pos, nullptr, 77 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/olmo2.cpp:83:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 81 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 82 | ctx0, Kcur, inp_pos, nullptr, 83 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 10 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/olmoe.cpp -o models/olmoe.o In file included from models/olmoe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/olmoe.cpp:56:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 54 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 55 | ctx0, Qcur, inp_pos, nullptr, 56 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/olmoe.cpp:62:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 60 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 61 | ctx0, Kcur, inp_pos, nullptr, 62 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/openai-moe-iswa.cpp -o models/openai-moe-iswa.o In file included from models/openai-moe-iswa.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/openai-moe-iswa.cpp:55:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 53 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 54 | ctx0, Qcur, inp_pos, nullptr, 55 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/openai-moe-iswa.cpp:61:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 59 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 60 | ctx0, Kcur, inp_pos, nullptr, 61 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/openelm.cpp -o models/openelm.o In file included from models/openelm.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/openelm.cpp:61:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 59 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 60 | ctx0, Qcur, inp_pos, NULL, 61 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/openelm.cpp:67:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 65 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 66 | ctx0, Kcur, inp_pos, NULL, 67 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/orion.cpp -o models/orion.o In file included from models/orion.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/orion.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 59 | ctx0, Qcur, inp_pos, nullptr, 60 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/orion.cpp:66:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 65 | ctx0, Kcur, inp_pos, nullptr, 66 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/pangu-embedded.cpp -o models/pangu-embedded.o In file included from models/pangu-embedded.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/pangu-embedded.cpp:52:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 51 | ctx0, Qcur, inp_pos, nullptr, 52 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/pangu-embedded.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, | ~~~~~~~~~~~~~ 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/phi2.cpp -o models/phi2.o In file included from models/phi2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/phi2.cpp:58:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 56 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 57 | ctx0, Qcur, inp_pos, nullptr, 58 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/phi2.cpp:64:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 63 | ctx0, Kcur, inp_pos, nullptr, 64 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/phi3.cpp -o models/phi3.o In file included from models/phi3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/phi3.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Qcur, inp_pos, rope_factors, 65 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/phi3.cpp:151:17: note: in instantiation of member function 'llm_build_phi3::llm_build_phi3' requested here 151 | template struct llm_build_phi3; | ^ models/phi3.cpp:71:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 70 | ctx0, Kcur, inp_pos, rope_factors, 71 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/phi3.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Qcur, inp_pos, rope_factors, 65 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/phi3.cpp:152:17: note: in instantiation of member function 'llm_build_phi3::llm_build_phi3' requested here 152 | template struct llm_build_phi3; | ^ models/phi3.cpp:71:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 70 | ctx0, Kcur, inp_pos, rope_factors, 71 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/plamo.cpp -o models/plamo.o In file included from models/plamo.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/plamo.cpp:48:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 46 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 47 | ctx0, Qcur, inp_pos, nullptr, 48 | n_embd_head, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~~~~~~~ models/plamo.cpp:54:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 53 | ctx0, Kcur, inp_pos, nullptr, 54 | n_embd_head, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~~~~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/plamo2.cpp -o models/plamo2.o In file included from models/plamo2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/plamo2.cpp:131:60: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 131 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/plamo2.cpp:137:60: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 137 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ~~~~~~~~~~~~~ ^~~~~ models/plamo2.cpp:176:77: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 176 | ggml_tensor * conv = build_rs(inp, conv_states_all, hparams.n_embd_r(), n_seqs); | ~~~~~~~~ ^~~~~~ 5 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/plamo3.cpp -o models/plamo3.o In file included from models/plamo3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/plamo3.cpp:67:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, | ~~~~~~~~~~~~~ 67 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/plamo3.cpp:127:17: note: in instantiation of member function 'llm_build_plamo3::llm_build_plamo3' requested here 127 | template struct llm_build_plamo3; | ^ models/plamo3.cpp:70:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, | ~~~~~~~~~~~~~ 70 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/plamo3.cpp:67:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, | ~~~~~~~~~~~~~ 67 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ models/plamo3.cpp:128:17: note: in instantiation of member function 'llm_build_plamo3::llm_build_plamo3' requested here 128 | template struct llm_build_plamo3; | ^ models/plamo3.cpp:70:17: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 69 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, | ~~~~~~~~~~~~~ 70 | n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/plm.cpp -o models/plm.o In file included from models/plm.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/plm.cpp:103:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 101 | q_pe = ggml_rope_ext( | ~~~~~~~~~~~~~ 102 | ctx0, q_pe, inp_pos, nullptr, 103 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/plm.cpp:111:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 109 | k_pe = ggml_rope_ext( | ~~~~~~~~~~~~~ 110 | ctx0, k_pe, inp_pos, nullptr, 111 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen.cpp -o models/qwen.o In file included from models/qwen.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen.cpp:44:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 42 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 43 | ctx0, Qcur, inp_pos, nullptr, 44 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen.cpp:50:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 48 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 49 | ctx0, Kcur, inp_pos, nullptr, 50 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen2.cpp -o models/qwen2.o In file included from models/qwen2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen2.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 59 | ctx0, Qcur, inp_pos, nullptr, 60 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen2.cpp:66:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 65 | ctx0, Kcur, inp_pos, nullptr, 66 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen2moe.cpp -o models/qwen2moe.o In file included from models/qwen2moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen2moe.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen2moe.cpp:63:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 62 | ctx0, Kcur, inp_pos, nullptr, 63 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen2vl.cpp -o models/qwen2vl.o In file included from models/qwen2vl.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen2vl.cpp:54:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 52 | Qcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 53 | ctx0, Qcur, inp_pos, nullptr, 54 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen2vl.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Kcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 59 | ctx0, Kcur, inp_pos, nullptr, 60 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen3.cpp -o models/qwen3.o In file included from models/qwen3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen3.cpp:51:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 49 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 50 | ctx0, Qcur, inp_pos, nullptr, 51 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen3.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 59 | ctx0, Kcur, inp_pos, nullptr, 60 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen3moe.cpp -o models/qwen3moe.o In file included from models/qwen3moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen3moe.cpp:51:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 49 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 50 | ctx0, Qcur, inp_pos, nullptr, 51 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen3moe.cpp:60:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 58 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 59 | ctx0, Kcur, inp_pos, nullptr, 60 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen3next.cpp -o models/qwen3next.o In file included from models/qwen3next.cpp:2: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen3next.cpp:132:38: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 132 | const float scale = 1.0f / sqrtf(S_v); | ~~~~~ ^~~ models/qwen3next.cpp:170:30: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 170 | q = ggml_pad(ctx0, q, 0, pad, 0, 0); | ~~~~~~~~ ^~~ models/qwen3next.cpp:171:30: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 171 | k = ggml_pad(ctx0, k, 0, pad, 0, 0); | ~~~~~~~~ ^~~ models/qwen3next.cpp:172:30: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 172 | v = ggml_pad(ctx0, v, 0, pad, 0, 0); | ~~~~~~~~ ^~~ models/qwen3next.cpp:173:27: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 173 | g = ggml_pad(ctx0, g, pad, 0, 0, 0); | ~~~~~~~~ ^~~ models/qwen3next.cpp:174:36: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 174 | beta = ggml_pad(ctx0, beta, 0, pad, 0, 0); | ~~~~~~~~ ^~~ models/qwen3next.cpp:385:38: warning: implicit conversion from 'const int64_t' (aka 'const long long') to 'float' may lose precision [-Wimplicit-int-float-conversion] 385 | const float scale = 1.0f / sqrtf(S_v); | ~~~~~ ^~~ models/qwen3next.cpp:504:13: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 502 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 503 | ctx0, Qcur, inp_pos, nullptr, 504 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen3next.cpp:509:13: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 507 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 508 | ctx0, Kcur, inp_pos, nullptr, 509 | n_rot, rope_type, n_ctx_orig, freq_base, | ^~~~~ models/qwen3next.cpp:686:84: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 686 | ggml_tensor * conv_states = build_rs(inp, conv_states_all, hparams.n_embd_r(), n_seqs); | ~~~~~~~~ ^~~~~~ models/qwen3next.cpp:748:77: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int32_t' (aka 'int') [-Wshorten-64-to-32] 748 | ggml_tensor * state = build_rs(inp, ssm_states_all, hparams.n_embd_s(), n_seqs); | ~~~~~~~~ ^~~~~~ 13 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen3vl-moe.cpp -o models/qwen3vl-moe.o In file included from models/qwen3vl-moe.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen3vl-moe.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen3vl-moe.cpp:66:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 65 | ctx0, Kcur, inp_pos, nullptr, 66 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/qwen3vl.cpp -o models/qwen3vl.o In file included from models/qwen3vl.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/qwen3vl.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/qwen3vl.cpp:66:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 64 | Kcur = ggml_rope_multi( | ~~~~~~~~~~~~~~~ 65 | ctx0, Kcur, inp_pos, nullptr, 66 | n_rot, sections, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/refact.cpp -o models/refact.o In file included from models/refact.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rnd1.cpp -o models/rnd1.o In file included from models/rnd1.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/rnd1.cpp:53:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 51 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 52 | ctx0, Qcur, inp_pos, nullptr, 53 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/rnd1.cpp:62:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 60 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 61 | ctx0, Kcur, inp_pos, nullptr, 62 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rwkv6-base.cpp -o models/rwkv6-base.o In file included from models/rwkv6-base.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/rwkv6-base.cpp:135:74: warning: implicit conversion loses floating-point precision: '__promote_t' (aka 'double') to 'float' [-Wimplicit-float-conversion] 135 | wkv_output = ggml_gated_linear_attn(ctx0, k, v, r, w, wkv_state, pow(head_size, -0.5f)); | ~~~~~~~~~~~~~~~~~~~~~~ ^~~~~~~~~~~~~~~~~~~~~ 3 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rwkv6.cpp -o models/rwkv6.o In file included from models/rwkv6.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rwkv6qwen2.cpp -o models/rwkv6qwen2.o In file included from models/rwkv6qwen2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rwkv7-base.cpp -o models/rwkv7-base.o In file included from models/rwkv7-base.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/rwkv7-base.cpp:68:64: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 68 | w = ggml_exp(ctx0, ggml_scale(ctx0, ggml_sigmoid(ctx0, w), -0.606531)); | ~~~~~~~~~~ ^~~~~~~~~ models/rwkv7-base.cpp:92:47: warning: implicit conversion loses floating-point precision: 'double' to 'float' [-Wimplicit-float-conversion] 92 | kk = ggml_l2_norm(ctx0, kk, 1e-12); | ~~~~~~~~~~~~ ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/rwkv7.cpp -o models/rwkv7.o In file included from models/rwkv7.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/seed-oss.cpp -o models/seed-oss.o In file included from models/seed-oss.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/seed-oss.cpp:59:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 57 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 58 | ctx0, Qcur, inp_pos, nullptr, 59 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/seed-oss.cpp:65:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 63 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 64 | ctx0, Kcur, inp_pos, nullptr, 65 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/smallthinker.cpp -o models/smallthinker.o In file included from models/smallthinker.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/smallthinker.cpp:62:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/smallthinker.cpp:125:17: note: in instantiation of member function 'llm_build_smallthinker::llm_build_smallthinker' requested here 125 | template struct llm_build_smallthinker; | ^ models/smallthinker.cpp:65:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 65 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/smallthinker.cpp:62:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 62 | Qcur = ggml_rope_ext(ctx0, Qcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ models/smallthinker.cpp:126:17: note: in instantiation of member function 'llm_build_smallthinker::llm_build_smallthinker' requested here 126 | template struct llm_build_smallthinker; | ^ models/smallthinker.cpp:65:68: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 65 | Kcur = ggml_rope_ext(ctx0, Kcur, inp_pos, nullptr, n_rot, rope_type, n_ctx_orig, freq_base_l, freq_scale_l, | ~~~~~~~~~~~~~ ^~~~~ 6 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/smollm3.cpp -o models/smollm3.o In file included from models/smollm3.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/smollm3.cpp:62:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 60 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 61 | ctx0, Qcur, inp_pos, nullptr, 62 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/smollm3.cpp:68:25: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 66 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 67 | ctx0, Kcur, inp_pos, nullptr, 68 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/stablelm.cpp -o models/stablelm.o In file included from models/stablelm.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/stablelm.cpp:75:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 73 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 74 | ctx0, Qcur, inp_pos, nullptr, 75 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/stablelm.cpp:81:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 79 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 80 | ctx0, Kcur, inp_pos, nullptr, 81 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/starcoder.cpp -o models/starcoder.o In file included from models/starcoder.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/starcoder2.cpp -o models/starcoder2.o In file included from models/starcoder2.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/starcoder2.cpp:57:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 55 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 56 | ctx0, Qcur, inp_pos, nullptr, 57 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/starcoder2.cpp:63:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 61 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 62 | ctx0, Kcur, inp_pos, nullptr, 63 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/t5-dec.cpp -o models/t5-dec.o In file included from models/t5-dec.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/t5-enc.cpp -o models/t5-enc.o In file included from models/t5-enc.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/wavtokenizer-dec.cpp -o models/wavtokenizer-dec.o In file included from models/wavtokenizer-dec.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ 2 warnings generated. clang++ -mmacos-version-min=26 -std=gnu++17 -I"/Users/ripley/R/R-devel/include" -DNDEBUG -I. -I/Users/ripley/R/Library/ggmlR/include -DGGML_USE_CPU -I'/Users/ripley/R/Library/ggmlR/include' -isystem /opt/R/arm64/include -fPIC -g -O2 -Wall -pedantic -Wconversion -Wno-sign-conversion -c models/xverse.cpp -o models/xverse.o In file included from models/xverse.cpp:1: In file included from models/models.h:3: In file included from models/../llama-model.h:5: In file included from models/../llama-graph.h:6: models/../llama-adapter.h:53:29: warning: implicit conversion turns floating-point number into integer: 'float' to 'bool' [-Wfloat-conversion] 53 | const float scale = alpha ? adapter_scale * alpha / rank : adapter_scale; | ^~~~~ ~ models/../llama-adapter.h:82:30: warning: implicit conversion loses integer precision: 'size_type' (aka 'unsigned long') to 'uint32_t' (aka 'unsigned int') [-Wshorten-64-to-32] 82 | return ab_map.size() * 6u; // a, b, scale, add, 2 x mul_mat | ~~~~~~ ~~~~~~~~~~~~~~^~~~ models/xverse.cpp:46:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 44 | Qcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 45 | ctx0, Qcur, inp_pos, nullptr, 46 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ models/xverse.cpp:52:21: warning: implicit conversion loses integer precision: 'const int64_t' (aka 'const long long') to 'int' [-Wshorten-64-to-32] 50 | Kcur = ggml_rope_ext( | ~~~~~~~~~~~~~ 51 | ctx0, Kcur, inp_pos, nullptr, 52 | n_rot, rope_type, n_ctx_orig, freq_base, freq_scale, | ^~~~~ 4 warnings generated. make[1]: Target `all' not remade because of errors. ERROR: compilation failed for package ‘llamaR’ * removing ‘/Users/ripley/R/packages/tests-devel/llamaR.Rcheck/llamaR’