Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Misc. bug: llama-cli '--log-disable' parameter omits response #11983

Open
nmandic78 opened this issue Feb 20, 2025 · 0 comments
Open

Misc. bug: llama-cli '--log-disable' parameter omits response #11983

nmandic78 opened this issue Feb 20, 2025 · 0 comments

Comments

@nmandic78
Copy link

Name and Version

version: 4526 (a94f3b2)
built with cc (Ubuntu 12.3.0-1ubuntu1~22.04) 12.3.0 for x86_64-linux-gnu

Not sure from when this started, but before, when using llama-cli with --log-disable, I would get the response printed without other verbose info.
Now when used with --log-disabled, there is no response printed in terminal.

Example:

llama-cli -m '/mnt/disk2/LLM_MODELS/models/Phi-3.5-mini-instruct-Q5_K_M.gguf' -p "Write short joke." -ngl 99 -no-cnv --log-disable
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes

Additionally, interactive mode is always on, without passing -i parameter. I need to put -no-cnv to dissable it. For example, this:
llama-cli -m '/mnt/disk2/LLM_MODELS/models/Phi-3.5-mini-instruct-Q5_K_M.gguf' -p "Write short joke." -ngl 99
puts it in interactive mode ('main: interactive mode on.' is visible in output).

Operating systems

Linux

Which llama.cpp modules do you know to be affected?

llama-cli

Command line

Problem description & steps to reproduce

llama-cli -m '/mnt/disk2/LLM_MODELS/models/Phi-3.5-mini-instruct-Q5_K_M.gguf' -p "Write short joke." -ngl 99 -no-cnv --log-disable
ggml_cuda_init: GGML_CUDA_FORCE_MMQ:    no
ggml_cuda_init: GGML_CUDA_FORCE_CUBLAS: no
ggml_cuda_init: found 1 CUDA devices:
  Device 0: NVIDIA GeForce RTX 3090, compute capability 8.6, VMM: yes

Omits LLM response in terminal output.

First Bad Commit

No response

Relevant log output

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

1 participant