From a175d9563f60e515eb3ed6d4f1238e2693587cfe Mon Sep 17 00:00:00 2001 From: Alex Stan Date: Wed, 28 Feb 2024 23:21:01 +0200 Subject: [PATCH] e Signed-off-by: Alex Stan --- llama.log | 53 ----------------------------------------------------- 1 file changed, 53 deletions(-) delete mode 100644 llama.log diff --git a/llama.log b/llama.log deleted file mode 100644 index c3f6094..0000000 --- a/llama.log +++ /dev/null @@ -1,53 +0,0 @@ -[1706516367] -llama server listening at http://127.0.0.1:62322 - -[1706516367] warming up the model with an empty run -[1706516367] Available slots: -[1706516367] -> Slot 0 - max context: 2048 -[1706516367] all slots are idle and system prompt is empty, clear the KV cache -[1706516368] slot 0 is processing [task id: 0] -[1706516368] slot 0 : kv cache rm - [0, end) -[1706516384] sampled token: 3492: 'You' -[1706516384] sampled token: 526: ' are' -[1706516384] sampled token: 263: ' a' -[1706516384] sampled token: 8444: ' helpful' -[1706516384] sampled token: 319: ' A' -[1706516384] sampled token: 29902: 'I' -[1706516384] sampled token: 20255: ' assistant' -[1706516384] sampled token: 29991: '!' -[1706516384] sampled token: 306: ' I' -[1706516384] sampled token: 626: ' am' -[1706516384] sampled token: 2307: ' already' -[1706516385] sampled token: 2734: ' running' -[1706516385] sampled token: 373: ' on' -[1706516385] sampled token: 596: ' your' -[1706516385] sampled token: 3271: ' home' -[1706516385] sampled token: 9609: ' PC' -[1706516385] sampled token: 29889: '.' -[1706516385] sampled token: 887: ' You' -[1706516385] sampled token: 508: ' can' -[1706516385] sampled token: 1423: ' check' -[1706516385] sampled token: 590: ' my' -[1706516385] sampled token: 4660: ' status' -[1706516385] sampled token: 491: ' by' -[1706516385] sampled token: 19229: ' typing' -[1706516385] sampled token: 376: ' "' -[1706516385] sampled token: 1004: 'me' -[1706516385] sampled token: 29908: '"' -[1706516385] sampled token: 472: ' at' -[1706516385] sampled token: 278: ' the' -[1706516385] sampled token: 9508: ' prompt' -[1706516385] sampled token: 470: ' or' -[1706516385] sampled token: 3763: ' simply' -[1706516385] sampled token: 6721: ' asking' -[1706516385] sampled token: 363: ' for' -[1706516385] sampled token: 372: ' it' -[1706516385] sampled token: 297: ' in' -[1706516385] sampled token: 13563: ' chat' -[1706516385] sampled token: 29889: '.' -[1706516385] sampled token: 2: '' -[1706516385] -[1706516385] print_timings: prompt eval time = 16766.02 ms / 598 tokens ( 28.04 ms per token, 35.67 tokens per second) -[1706516385] print_timings: eval time = 718.30 ms / 39 runs ( 18.42 ms per token, 54.29 tokens per second) -[1706516385] print_timings: total time = 17484.32 ms -[1706516385] slot 0 released (637 tokens in cache)