Skip to content

Commit

Permalink
Merge pull request #3 from sjmonson/develop
Browse files Browse the repository at this point in the history
Rename to llm-eval-test
  • Loading branch information
sjmonson authored Oct 15, 2024
2 parents 735947c + bfc29fe commit 2cef539
Show file tree
Hide file tree
Showing 44 changed files with 11 additions and 11 deletions.
10 changes: 5 additions & 5 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,4 +1,4 @@
# perf-llm-eval
# llm-eval-test

A wrapper around [lm-eval-harness](https://github.com/EleutherAI/lm-evaluation-harness) and [Unitxt](https://github.com/IBM/unitxt) designed for evaluation of a local inference endpoint.

Expand All @@ -25,16 +25,16 @@ python -m venv venv
source venv/bin/activate

# Install the package
pip install git+https://github.com/sjmonson/perf-llm-eval.git
pip install git+https://github.com/sjmonson/llm-eval-test.git

# View run options
perf-llm-eval run --help
llm-eval-test run --help
```

## Usage

```
usage: perf-llm-eval run [-h] [--catalog_path PATH] [--tasks_path PATH] [-v | -q] -H ENDPOINT -m MODEL -t TASKS -d PATH [-b INT] [-o OUTPUT]
usage: llm-eval-test run [-h] [--catalog_path PATH] [--tasks_path PATH] [-v | -q] -H ENDPOINT -m MODEL -t TASKS -d PATH [-b INT] [-o OUTPUT]
Run tasks
Expand Down Expand Up @@ -74,5 +74,5 @@ huggingface-cli download $DATASET --repo-type dataset --local-dir $DATASETS_DIR/
# Run the benchmark
ENDPOINT=http://127.0.0.1:8000/v1/completions # An OpenAI API-compatable completions endpoint
MODEL_NAME=meta-llama/Llama-3.1-8B # Name of the model hosted on the inference server
perf-llm-eval run --endpoint $ENDPOINT --model $MODEL_NAME --datasets $DATASETS_DIR --tasks mmlu_pro
llm-eval-test run --endpoint $ENDPOINT --model $MODEL_NAME --datasets $DATASETS_DIR --tasks mmlu_pro
```
6 changes: 3 additions & 3 deletions perf_llm_eval/__main__.py → llm_eval_test/__main__.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,9 +4,9 @@
import logging
import tempfile

from perf_llm_eval.parser import setup_parser
from llm_eval_test.parser import setup_parser

logger = logging.getLogger("perf-llm-eval")
logger = logging.getLogger("llm-eval-test")


def eval_cli():
Expand Down Expand Up @@ -35,7 +35,7 @@ def eval_cli():
os.environ["UNITXT_ARTIFACTORIES"] = args.catalog_path

# Late import to avoid slow cli
from perf_llm_eval.lm_eval_wrapper import LMEvalWrapper
from llm_eval_test.lm_eval_wrapper import LMEvalWrapper

if args.command == 'list':
LMEvalWrapper.list_tasks(args.tasks_path)
Expand Down
File renamed without changes.
File renamed without changes.
Original file line number Diff line number Diff line change
Expand Up @@ -8,7 +8,7 @@
from lm_eval.tasks import TaskManager # type: ignore
from lm_eval.utils import handle_non_serializable, make_table

logger = logging.getLogger("perf-llm-eval")
logger = logging.getLogger("llm-eval-test")

class LMEvalWrapper(object):
@staticmethod
Expand Down
File renamed without changes.
File renamed without changes.
4 changes: 2 additions & 2 deletions pyproject.toml
Original file line number Diff line number Diff line change
@@ -1,5 +1,5 @@
[tool.poetry]
name = "perf-llm-eval"
name = "llm-eval-test"
version = "0.1.0"
description = ""
authors = ["Samuel Monson <[email protected]>"]
Expand All @@ -12,7 +12,7 @@ unitxt = "^1.13.1"


[tool.poetry.scripts]
perf-llm-eval = "perf_llm_eval.__main__:eval_cli"
llm-eval-test = "llm_eval_test.__main__:eval_cli"

[build-system]
requires = ["poetry-core"]
Expand Down

0 comments on commit 2cef539

Please sign in to comment.