Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

MS MARCO v2.1 and v2.1 segmented for TREC 2024 RAG #267

Open
5 of 8 tasks
mam10eks opened this issue Jun 21, 2024 · 3 comments
Open
5 of 8 tasks

MS MARCO v2.1 and v2.1 segmented for TREC 2024 RAG #267

mam10eks opened this issue Jun 21, 2024 · 3 comments

Comments

@mam10eks
Copy link
Contributor

mam10eks commented Jun 21, 2024

Dataset Information:

It would be awesome to have the document corpus (and the segmented counterpart) used in TREC RAG 2024 as integration to ir_datasets. From the description on the web page, it should be no problem to add this, random access to documents should also be very efficient as the file and byte offset are already encoded in the document identifiers, so I think there should be no problem.

The only question that I would have is: As the document identifiers contain the offsets in the file where a document starts (but not the end), is there maybe already a functionality that seeks to the start and readys the json entry until the closing bracket? If not, I could add this as well with unit tests, should be no problem.

Links to Resources:

Dataset ID(s) & supported entities:

  • msmarco-document-v2.1: for the original documents
  • msmarco-document-v2.1/segmented: for the segmented documents

Checklist

Mark each task once completed. All should be checked prior to merging a new dataset.

  • Dataset definition (in ir_datasets/datasets/[topid].py)
  • Tests (in tests/integration/[topid].py)
  • Metadata generated (using ir_datasets generate_metadata command, should appear in ir_datasets/etc/metadata.json)
  • Documentation (in ir_datasets/etc/[topid].yaml)
  • Downloadable content (in ir_datasets/etc/downloads.json)
    • Download verification action (in .github/workflows/verify_downloads.yml). Only one needed per topid.
    • Any small public files from NIST (or other potentially troublesome files) mirrored in https://github.com/seanmacavaney/irds-mirror/. Mirrored status properly reflected in downloads.json.

Additional comments/concerns/ideas/etc.

@mam10eks
Copy link
Contributor Author

Dear all, I would be open to make a first proposal for an implementation here.

@mam10eks
Copy link
Contributor Author

mam10eks commented Aug 5, 2024

Dear all, I started a draft pull request (only to indicate that there is some progress): #269

Mainly documentation todos are pending, but as the deadline is close, this might be already useful for others even when the documentation is not yet finalized.

I.e., the main thing for iterating over documents could be already done via (e.g., as covered in the unit tests):

for doc in ir_datasets.load('msmarco-document-v2.1/segmented').docs_iter():
    print(doc)
    break

@seanmacavaney
Copy link
Collaborator

Awesome, thanks! I'll take a look at it tomorrow and see if I can tick some of the other tasks :)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Projects
None yet
Development

No branches or pull requests

2 participants