Skip to content

Runs - Legal 2006

humL06dvo

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06dvo
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: Same as humL06tvo except that the terms were from the production request instead of from the final boolean. (Stemming still was not applied.)

humL06t

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06t
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: manual
  • Run description: The final boolean query was used, respecting the boolean operators such as AND, phrase, proximity, NOT, etc. Full wildcard matching was supported. A relevance-ranking algorithm was applied to the matching rows. Some hand-editing was needed to convert the queries to our syntax, but the run is automatic in spirit because it just implements the final boolean query intended by the negotiators.

humL06t0

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06t0
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: manual
  • Run description: Same as humL06t except that the defendant boolean was used instead of the final boolean.

humL06tv

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06tv
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: Vectorized use of the final boolean query. Operators such as AND, phrase, proximity were dropped. Punctuation was dropped. Full wildcarding was still respected.

humL06tvc

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06tvc
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Same as humL06tv except that a duplicate filtering heuristic was applied.

humL06tve

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06tve
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: Blind feedback using top-2 rows of humL06tv.

humL06tvo

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06tvo
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: Same as humL06tv except metadata not indexed (also, some other minor indexing differences, such as a stopword list applied, and records without a docid-tag skipped).

humL06tvz

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: humL06tvz
  • Participant: hummingbird.tomlinson
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: One percent subset of first 9000 rows of humL06tv (rows 1, 101, 201, 301, ..., 8901) plus next 1000 rows of humL06tv (rows 9001-10000).

NUSCHUA1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: NUSCHUA1
  • Participant: nus.kor
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: The IITCDIP corpus was indexed with Lucene open source indexer. No attempts were made at correcting OCR errors. We used version 1.2 of the track topics as provided by Jianqiang Wang in order to avoid the Boolean Query errors in the original version. Other than this change, the system is fully automatic. We use a 2 phase approach to find the best documents for each topic. Details of the full system is provided in the NUSCHUA2 run description. For this run, NUSCHUA1, we only ran phase 1 of our system to obtain a high recall document set but do not run phase 2. Thus for us, this run is a baseline for comparing our phase 2 high precision reranking algorithm.

NUSCHUA2

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: NUSCHUA2
  • Participant: nus.kor
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: The IITCDIP corpus was indexed with Lucene open source indexer. No attempts were made at correcting OCR errors. We used version 1.2 of the track topics as provided by Jianqiang Wang in order to avoid the Boolean Query errors in the original version. Other than this change, the system is fully automatic. We use a 2 phase approach to find the best documents for each topic. In phase 1, we aim to produce a high recall document set by merging terms in the Boolean Query and Production Request text to form a high recall query. First, wildcard terms in the final Boolean Query are expanded. All ORed Boolean Query terms are then grouped together, forming groups of terms. Next, the production request sentences are parsed with Minipar in order to extract verb and noun phrases. In order to filter away junk words introduced by the wildcard expansion, we use WordNet to remove any wildcard term that are not synonymous with at least 1 production request phrase word. We next identify intersections between the filtered boolean query group terms and words in the production request phrases. These intersecting terms form the central query that all documents must match. All other terms are optionally matched. Now that we have a set of retrieved documents, we focus on improving precision of the retrieved results in phase 2. This is achieved by reranking the documents such that relevant documents are given a higher score than irrelevant documents. We use an information theoretic method to cluster retrieved documents based on their content words (bigrams), how often these words overlap and cooccurance information. We then boost the score for documents that are in relevant clusters and downweight the score for documents in irrelevant clusters.

SabLeg06aa1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06aa1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Production Request plus words of Negotiatedboolean and Complaint (heavily downweighted), no OCR expansion, no explicit use of metadata, no boolean operatives

SabLeg06ab1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06ab1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Production Request plus words of Negotiatedboolean. no OCR expansion, no explicit use of metadata, no boolean operatives. blind feedback based on top 30 docs.

sableg06ao1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: sableg06ao1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Production Request plus words of Negotiatedboolean. OCR expansion of request words only (share common prefix), no explicit use of metadata, no boolean operatives.

SabLeg06ao2

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06ao2
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/2/2006
  • Type: automatic
  • Run description: Production Request plus words of Negotiatedboolean. OCR expansion of request words only (share common prefix,suffix, massive expansion), no explicit use of metadata, no boolean operatives.

SabLeg06ar1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06ar1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic

SabLeg06arb1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06arb1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Production Request plus words of Finalboolean, no OCR expansion, no explicit use of metadata, no boolean operatives

SabLeg06arn1

Results | Participants | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: SabLeg06arn1
  • Participant: sabir.buckley
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: Production Request plus words of negotiatedboolean, no OCR expansion, no explicit use of metadata, no boolean operatives

UmdBase

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UmdBase
  • Participant: umaryland.oard
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: we used all words in the filed.

UmdBool

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UmdBool
  • Participant: umaryland.oard
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: manual
  • Run description: we automatically converted the Boolean queries into Indri Boolean queries and then manually fixed some syntax errors.

UmdBoolAuto

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UmdBoolAuto
  • Participant: umaryland.oard
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: we used all words contained in the Boolean and simply ignored the Boolean syntax

UmdComb

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UmdComb
  • Participant: umaryland.oard
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: we used all words in the filed plus all words contained in the Boolean and simply ignored the Boolean syntax

UMKCB

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCB
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: Basic Ranked Boolean Query from FinalQuery

UMKCB2

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCB2
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: ocr boolean auto w2and

UMKCBQE10

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCBQE10
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: ocr boolean auto query expansion for individual term to 10

UMKCBQE5

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCBQE5
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/31/2006
  • Type: automatic
  • Run description: ocr boolean auto query expansion for individual term to 5

UMKCQE100

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCQE100
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: Automatic Query Expansion from FinalQuery

UMKCQE25

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCQE25
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: Automatic Query Expansion from FinalQuery

UMKCSN

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCSN
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: Basic Ranked Surround Boolean Query Without Order from FinalQuery

UMKCSW

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: UMKCSW
  • Participant: umkc.zhao
  • Track: Legal
  • Year: 2006
  • Submission: 7/29/2006
  • Type: automatic
  • Run description: Basic Ranked Surround Boolean Query With Order from FinalQuery

york06la01

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: york06la01
  • Participant: yorku.huang
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: 1. Use Okapi BM25 for weighting and retrieval. 2. No relevance feedback is used.

york06la02

Results | Participants | Proceedings | Input | Summary (trec-eval) | Summary (prec-at-B) | Appendix

  • Run ID: york06la02
  • Participant: yorku.huang
  • Track: Legal
  • Year: 2006
  • Submission: 8/1/2006
  • Type: automatic
  • Run description: 1. Use Okapi BM25 for weighting and retrieval. 2. Use the beta-approximation term weighting method for new term selection to do automatic relevance feedback.