Skip to content

Commit 21551c5

Browse files
committed
embedding : handle --parallel argument
ggml-ci
1 parent fdede39 commit 21551c5

File tree

2 files changed

+5
-1
lines changed

2 files changed

+5
-1
lines changed

examples/embedding/embedding.cpp

Lines changed: 4 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -81,6 +81,10 @@ int main(int argc, char ** argv) {
8181

8282
params.embedding = true;
8383

84+
if (params.n_parallel == 1) {
85+
params.kv_unified = true;
86+
}
87+
8488
// utilize the full context
8589
if (params.n_batch < params.n_ctx) {
8690
LOG_WRN("%s: setting batch size to %d\n", __func__, params.n_ctx);

src/llama-batch.cpp

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -59,7 +59,7 @@ bool llama_batch_allocr::init(
5959
for (int32_t i = 0; i < batch.n_tokens; ++i) {
6060
for (int32_t s = 0; s < batch.n_seq_id[i]; ++s) {
6161
if (batch.seq_id && (batch.seq_id[i][s] < 0 || batch.seq_id[i][s] >= (llama_seq_id) n_seq_max)) {
62-
LLAMA_LOG_ERROR("%s: invalid seq_id[%d][%d] = %d > %d\n", __func__, i, s, batch.seq_id[i][s], (llama_seq_id) n_seq_max);
62+
LLAMA_LOG_ERROR("%s: invalid seq_id[%d][%d] = %d >= %d\n", __func__, i, s, batch.seq_id[i][s], (llama_seq_id) n_seq_max);
6363
return false;
6464
}
6565
}

0 commit comments

Comments
 (0)