Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix logging issues #1494

Merged
merged 1 commit into from
Oct 30, 2023
Merged

Fix logging issues #1494

merged 1 commit into from
Oct 30, 2023

Conversation

Tostino
Copy link
Contributor

@Tostino Tostino commented Oct 28, 2023

Currently, info logs are not actually being logged.

For example the args log in the OpenAI api_server:

logger.info(f"args: {args}")

doesn't actually print anything when you start the server.

This pull request fixes that.

Copy link
Member

@zhuohan123 zhuohan123 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for fixing this!! I was confused about the behavior of this logger for a while and met the issue that I could not print logs from the openai API server.

@zhuohan123 zhuohan123 merged commit 2f3d36a into vllm-project:main Oct 30, 2023
2 checks passed
@WoosukKwon
Copy link
Collaborator

@Tostino @zhuohan123 It seems this PR makes the logs printed twice:

$ python examples/llm_engine_example.py 
INFO 10-31 15:40:46 llm_engine.py:72] Initializing an LLM engine with config: model='facebook/opt-125m', tokenizer='facebook/opt-125m', tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=auto, tensor_parallel_size=1, quantization=None, seed=0)
INFO 10-31 15:40:46 llm_engine.py:72] Initializing an LLM engine with config: model='facebook/opt-125m', tokenizer='facebook/opt-125m', tokenizer_mode=auto, revision=None, tokenizer_revision=None, trust_remote_code=False, dtype=torch.float16, max_seq_len=2048, download_dir=None, load_format=auto, tensor_parallel_size=1, quantization=None, seed=0)
INFO 10-31 15:40:52 llm_engine.py:207] # GPU blocks: 127960, # CPU blocks: 7281
INFO 10-31 15:40:52 llm_engine.py:207] # GPU blocks: 127960, # CPU blocks: 7281

@Tostino
Copy link
Contributor Author

Tostino commented Oct 31, 2023

Hmm, is that with multiple GPUs that it is happening? Does it happen with a single GPU if so?

I'd check myself, but I'm out of town right now and don't have a machine I can test on locally.

sjchoi1 pushed a commit to casys-kaist-internal/vllm that referenced this pull request May 7, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants