diff --git a/README.md b/README.md index f092b4e..cdc43ac 100644 --- a/README.md +++ b/README.md @@ -10,7 +10,7 @@ We offer a suite of prompt-decoders, albeit with focus on open source LLMs compatible with [FastChat](https://github.com/lm-sys/FastChat?tab=readme-ov-file#supported-models) (e.g., Vicuna, Zephyr, etc.). Some of the code in this repository is borrowed from [RankGPT](https://github.com/sunnweiwei/RankGPT)! # Releases -current_version = 0.2.8 +current_version = 0.12.8 ## 📟 Instructions @@ -26,6 +26,11 @@ conda activate rankllm pip3 install torch torchvision torchaudio --index-url https://download.pytorch.org/whl/cu118 ``` +### Install openjdk with maven if you want to use the retriever +```bash +conda install -c conda-forge openjdk=21 maven -y +``` + ### Install Dependencies ```bash pip install -r requirements.txt diff --git a/pyproject.toml b/pyproject.toml index 49b2e40..4d1c434 100644 --- a/pyproject.toml +++ b/pyproject.toml @@ -6,7 +6,7 @@ build-backend = "setuptools.build_meta" [project] name = "rank-llm" -version = "0.2.8" +version = "0.12.8" description = "A Package for running prompt decoders like RankVicuna" readme = "README.md" authors = [ @@ -30,7 +30,7 @@ dependencies = {file = ["requirements.txt"]} Homepage = "https://github.com/castorini/rank_llm" [tool.bumpver] -current_version = "0.2.8" +current_version = "0.12.8" version_pattern = "MAJOR.MINOR.PATCH" commit_message = "Bump version {old_version} -> {new_version}" commit = true diff --git a/src/rank_llm/demo/rerank_stored_retrieved_results.py b/src/rank_llm/demo/rerank_stored_retrieved_results.py index 909e93a..669b6ae 100644 --- a/src/rank_llm/demo/rerank_stored_retrieved_results.py +++ b/src/rank_llm/demo/rerank_stored_retrieved_results.py @@ -11,7 +11,7 @@ from rank_llm.rerank.zephyr_reranker import ZephyrReranker file_name = ( - "retrieve_results/BM25/retrieve_results_msmarco-v2.1-doc-segmented-dl21_top20.jsonl" + "retrieve_results/BM25/retrieve_results_dl23_top20.json" ) requests = read_requests_from_file(file_name)