Show / Hide Table of Contents

    run-trec-eval

    Name

    benchmark-run-trec-eval - Runs a TREC evaluation.

    Synopsis

    lucene benchmark run-trec-eval <INPUT_TOPICS_FILE> <INPUT_QUERY_RELEVANCE_FILE> <OUTPUT_SUBMISSION_FILE> <INDEX_DIRECTORY> [-t|--query-on-title] [-d|--query-on-description] [-n|--query-on-narrative] [?|-h|--help]

    Arguments

    INPUT_TOPICS_FILE

    Input file containing queries.

    INPUT_QUERY_RELEVANCE_FILE

    Input file conataining relevance judgements.

    OUTPUT_SUBMISSION_FILE

    Output submission file for TREC evaluation.

    INDEX_DIRECTORY

    The index directory.

    Options

    ?|-h|--help

    Prints out a short help for the command.

    -t|--query-on-title

    Use title field in query. This flag will automatically be on if no other field is specified.

    -d|--query-on-description

    Use description field in query.

    -n|--query-on-narrative

    Use narrative field in query.

    Example

    Runs a TREC evaluation on the c:\topics queries file and the c:\queries relevance judgements on the c:\lucene-index index using the description and narrative fields and places the output in c:\output.txt.

    lucene benchmark run-trec-eval c:\topics.txt c:\queries.txt c:\submissions.txt c:\output.txt c:\lucene-index -d -n

    • Improve this Doc
    Back to top Copyright © 2019 Licensed to the Apache Software Foundation (ASF)