Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Running locally on linux always says address already in use #477

Open
2 of 4 tasks
Datta0 opened this issue Jan 22, 2025 · 1 comment
Open
2 of 4 tasks

Running locally on linux always says address already in use #477

Datta0 opened this issue Jan 22, 2025 · 1 comment

Comments

@Datta0
Copy link

Datta0 commented Jan 22, 2025

System Info

  • Python 3.10.12
  • text-embeddings-router 1.6.0, installed using cargo install --path router -F candle-cuda -F http --no-default-features
  • platform: Ubuntu 2204 with Nvidia A100 GPU.
  • Model: Happens with both nomic-ai/nomic-embed-text-v1.5 and BAAI/bge-reranker-v2-m3. Haven't checked others
❯ lsof -i :12345 # shows that nothing is running on the said port.
❯ text-embeddings-router --model-id nomic-ai/nomic-embed-text-v1.5 --port 12345
2025-01-22T10:39:14.154099Z  INFO text_embeddings_router: router/src/main.rs:175: Args { model_id: "nom**-**/*****-*****-****-v1.5", revision: None, tokenization_workers: None, dtype: None, pooling: None, max_concurrent_requests: 512, max_batch_tokens: 16384, max_batch_requests: None, max_client_batch_size: 32, auto_truncate: false, default_prompt_name: None, default_prompt: None, hf_api_token: None, hostname: "0.0.0.0", port: 12345, uds_path: "/tmp/text-embeddings-inference-server", huggingface_hub_cache: None, payload_limit: 2000000, api_key: None, json_output: false, otlp_endpoint: None, otlp_service_name: "text-embeddings-inference.server", cors_allow_origin: None }
2025-01-22T10:39:14.236075Z  INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:20: Starting download
2025-01-22T10:39:14.236096Z  INFO download_artifacts:download_pool_config: text_embeddings_core::download: core/src/download.rs:53: Downloading `1_Pooling/config.json`
2025-01-22T10:39:14.236141Z  INFO download_artifacts:download_new_st_config: text_embeddings_core::download: core/src/download.rs:77: Downloading `config_sentence_transformers.json`
2025-01-22T10:39:14.236155Z  INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:40: Downloading `config.json`
2025-01-22T10:39:14.236167Z  INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:43: Downloading `tokenizer.json`
2025-01-22T10:39:14.236179Z  INFO download_artifacts: text_embeddings_core::download: core/src/download.rs:47: Model artifacts downloaded in 106.335µs
2025-01-22T10:39:14.247515Z  INFO text_embeddings_router: router/src/lib.rs:188: Maximum number of tokens per request: 8192
2025-01-22T10:39:14.247554Z  INFO text_embeddings_core::tokenization: core/src/tokenization.rs:28: Starting 24 tokenization workers
2025-01-22T10:39:14.343914Z  INFO text_embeddings_router: router/src/lib.rs:230: Starting model backend
2025-01-22T10:39:14.344999Z  INFO text_embeddings_backend: backends/src/lib.rs:486: Downloading `model.safetensors`
2025-01-22T10:39:14.345044Z  INFO text_embeddings_backend: backends/src/lib.rs:370: Model weights downloaded in 46.73µs
2025-01-22T10:39:15.568001Z  INFO text_embeddings_backend_candle: backends/candle/src/lib.rs:332: Starting FlashNomicBert model on Cuda(CudaDevice(DeviceId(1)))
2025-01-22T10:39:16.131682Z  INFO text_embeddings_router: router/src/lib.rs:248: Warming up model
Error: failed to build prometheus recorder

Caused by:
    failed to create HTTP listener: Address already in use (os error 98)
❯ lsof -i :12345

Information

  • Docker
  • The CLI directly

Tasks

  • An officially supported command
  • My own modifications

Reproduction

  1. Install text-embeddings-router on linux
  2. Run text-embeddings-router --model-id nomic-ai/nomic-embed-text-v1.5 --port 8087

Expected behavior

TEI loads up fine for inference without any errors.

@Datta0
Copy link
Author

Datta0 commented Jan 24, 2025

Seems like switching to python 3.12 solved my issue. I noticed in the Dockerfile that python was set to 3.11 and that hinted me to try python 3.12

Should this be documented somewhere?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant