--- license: apache-2.0 language: - en library_name: transformers tags: - bert - information retrieval - learned sparse model --- Paper: [DeeperImpact: Optimizing Sparse Learned Index Structures](https://arxiv.org/abs/2405.17093) This repository contains the DeeperImpact model trained on the MS-MARCO passage dataset expanded using a [fine-tuned Llama 2 model](https://huggingface.co/soyuj/llama2-doc2query) with hard negatives, distillation, and pre-trained CoCondenser model initialization. The code to train and run inferences using DeeperImpact can be found in the [DeeperImpact Repo](https://github.com/basnetsoyuj/improving-learned-index). Please refer to the following notebook to understand how to use the model: [inference_deeper_impact.ipynb](https://github.com/basnetsoyuj/improving-learned-index/blob/master/inference_deeper_impact.ipynb) For running inference on a larger collection of documents, use the following command: ```bash python -m src.deep_impact.index \ --collection_path \ --output_file_path \ --model_checkpoint_path soyuj/deeper-impact \ --num_processes \ --process_batch_size \ --model_batch_size ``` It distributes the inference across multiple GPUs in the machine. To manually set the GPUs, use `CUDA_VISIBLE_DEVICES` environment variable.