Skip to content

Runs - Crowdsourcing 2012

BUPTPRISZHS

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: BUPTPRISZHS
  • Participant: pris411
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/29/2012
  • Task: trat
  • MD5: 006e998ac6bbf33cc0e08ef50c8c6b13
  • Run description: This is our first run.Results are obtained on the CrowdFlower platform.

INFLB2012

Participants | Input | Summary | Appendix

  • Run ID: INFLB2012
  • Participant: INFLBSTF
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 330d193a67ef520c2c9417258087370e
  • Run description: Our approach includes hybrid strategy using computer calculated results and crowdsourced results, and some interface issues on TRAT task.

NEUElo2

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUElo2
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: b3cff6197ff1376211d1653b1b8b6f7b
  • Run description: We have used crowd sourcing to generate relevance judgements for pairs of documents in form of preference pair judgements. Elo rating algorithm is used to combine the preference judgements from crowds. The Elo rating system is a method for calculating the relative skill levels of players in two-player games such as chess.

NEUElo3

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUElo3
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: f393c26eed407caf592ac03d0bf1a90b
  • Run description: We have used crowd sourcing to generate relevance judgements for pairs of documents in form of preference pair judgements. Elo rating algorithm is used to combine the preference judgements from crowds. The Elo rating system is a method for calculating the relative skill levels of players in two-player games such as chess.

NEUElo4

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUElo4
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: bc5fc1aeda9e9da915d280b0cc2b873d
  • Run description: We have used crowd sourcing to generate relevance judgements for pairs of documents in form of preference pair judgements. Elo rating algorithm is used to combine the preference judgements from crowds. The Elo rating system is a method for calculating the relative skill levels of players in two-player games such as chess.

NEUElo5

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUElo5
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: d815ef0ddd29c2cc91bbb4b09784bda7
  • Run description: We have used crowd sourcing to generate relevance judgements for pairs of documents in form of preference pair judgements. Elo rating algorithm is used to combine the preference judgements from crowds. The Elo rating system is a method for calculating the relative skill levels of players in two-player games such as chess.

NEUEM1

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUEM1
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: c1d06c302304323af38de35709a95775
  • Run description: This is the EM algorithm for acquiring relevant judgements through crowdsourcing.

NEUNugget12

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: NEUNugget12
  • Participant: NEU
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 3f15c04bbb021b169ecbbb877a8db395
  • Run description: 1 assessor takes about 45 min/query to judge documents and nuggets (extracted automatically from relevant documents) in an iterative process. When done, all documents are ranked given their match with the assessed nuggets.

Orc2G

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: Orc2G
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: c790edf41c0822d8e227243df383773f
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. A simplified version of Independent Bayesian Classifier Combination was applied, learning from Topic features extracted from the text. Reliability of workers is also learnt from test examples and used to weight crowdsourced labels.

Orc2GUL

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: Orc2GUL
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 10fcb21dd1e707b287ca921155f2016b
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. A simplified version of Independent Bayesian Classifier Combination was applied, learning from Topic features extracted from the text. Reliability of workers is also learnt from test examples and used to weight crowdsourced labels. Crowdsourced labels are taken into account when creating the final classification, in contrast to Orc2G, which only uses them to train a classifier.

Orc2GULConf

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: Orc2GULConf
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 4ffc76fee627b82ed3023e582004fd85
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. A simplified version of Independent Bayesian Classifier Combination was applied, learning from Topic features extracted from the text. Reliability of workers is also learnt from test examples and used to weight crowdsourced labels. Crowdsourced labels are taken into account when creating the final classification, in contrast to Orc2G, which only uses them to train a classifier. Confidence labels for the individual responses from the crowd are used to weight more confident responses more strongly when using the labels in classification.

Orc2Stage

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: Orc2Stage
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 44f3211a11fb75191f84863a023c6302
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. A simplified version of Independent Bayesian Classifier Combination was applied, learning from Topic features extracted from the text.

OrcVB1

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVB1
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 386975a8ed10e38d6df48df873a7a630
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied, treating all crowd members as equal and learning from Topic features extracted from the text.

OrcVB1Conf

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVB1Conf
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 768ff24c9d31a19e114ab09ada2965da
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied, treating all crowd members as equal and learning from Topic features extracted from the text. The crowd give estimates of their confidence when providing a label, which is taken into account by the classifier.

OrcVBW16Conf

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVBW16Conf
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: ddeb22e6987c081b7f9bf930be02ecdc
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied to crowdsourced labels, learning from Topic features extracted from the text. Reliability of workers is also learnt from the data and from test examples and used to weight crowdsourced labels. Confidence labels for the individual responses from the crowd are used to weight more confident responses more strongly. This version of the classifier uses middling priors that balance the prior belief that crowd members are accurate with the ability to spot the few that are not from patterns in the data. This allows us to disregard or treat as expert some of the responses from the crowd.

OrcVBW80

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVBW80
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 7147eae96835b94814faedb9cf4217d4
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied, learning the reliability of individual crowd members and learning from Topic features extracted from the text. The crowd give estimates of their confidence when providing a label, which is taken into account by the classifier as it learns how to use the confidence label from test data.

OrcVBW80Conf

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVBW80Conf
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: b57ba0f289868314e59098e2a78a9e34
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied, learning the reliability of individual crowd members and learning from Topic features extracted from the text. The crowd give estimates of their confidence when providing a label, which is taken into account by the classifier.

OrcVBW9Conf

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: OrcVBW9Conf
  • Participant: HAC
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 1e69d3c636c9f88ff66acf3c868d29c7
  • Run description: Using topic analysis to select files to crowdsource, we obtained 2600 labels from Amazon Mechanical Turk workers. Independent Bayesian Classifier Combination was applied, learning the reliability of individual crowd members and learning from Topic features extracted from the text. The crowd give estimates of their confidence when providing a label, which is taken into account by the classifier as it learns how to use the confidence label from test data. This uses weaker priors so that worker reliability is inferred in a semi-supervised manner from the data.

SetuServtest

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SetuServtest
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/31/2012
  • Task: trat
  • MD5: 589d715b34113b188e691f90b6c05327
  • Run description: To test submission

SSEC3excl

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSEC3excl
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: a1942c728d00fb41970e57e7f9674b7d
  • Run description: Post Error Correction (likely errors are identified using M/L), counting "not sures" as irrelevant documents

SSEC3incl

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSEC3incl
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: b9e4dc6823dbdd021e33cfcde266bf53
  • Run description: Post Error Correction (likely errors are identified using M/L), counting "not sures" as relevant documents

SSEC3inclML

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSEC3inclML
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 43140dd7564b760e145b49df4baa1499
  • Run description: Post Error Correction (likely errors are identified using M/L), counting "not sures" with ML score > 50% as relevant docs & remaining "not sures" as irrelevant

SSECML2to99

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSECML2to99
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: 57163b7d0abc98876a0f61d38ab94207
  • Run description: Post Error Correction (likely errors are identified using M/L for document with confidence scores between 2% and 99%), counting "not sures" as relevant documents

SSECML50pct

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSECML50pct
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: f2073d1363b6ddd35f2b4031ac5400c1
  • Run description: Post Error Correction (likely errors are identified using M/L for document with confidence score of >50%), counting "not sures" as relevant documents

SSECML75pct

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSECML75pct
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/4/2012
  • Task: trat
  • MD5: afd2624c3ec54f02d1e4b73af5b63fbe
  • Run description: Post Error Correction (likely errors are identified using M/L for document with confidence score of >75%), counting "not sures" as relevant documents

SSML2pct

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSML2pct
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/3/2012
  • Task: trat
  • MD5: e87e2d05b7f58c02f2d3bf8dbf6de3cd
  • Run description: Crowdsourced using Skierarchy approach for pairs with machine learning score of greater than 2%. Remaining documents are marked as not relevant

SSNoEC

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSNoEC
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/3/2012
  • Task: trat
  • MD5: b9f8cb3cb05ef991c35c7b78d1ae4f83
  • Run description: Crowdsourced using Skierarchy approach for all pairs (no error correction & no plurality)

SSPostEC

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSPostEC
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/15/2012
  • Task: irat
  • MD5: 1e66d0d472b86052e04fa395d1e1edb8
  • Run description: Post Error Correction

SSPostECv2

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSPostECv2
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/16/2012
  • Task: irat
  • MD5: 8022401c7c17dd08fd34a3676824e372
  • Run description: Final - derived using 2 annotators and 1 person for error correction

SSPreEC

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: SSPreEC
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/15/2012
  • Task: irat
  • MD5: ad781a89b1611187f7479f498291e6c3
  • Run description: Prior to Error Correction

testrun

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: testrun
  • Participant: SetuServ
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/13/2012
  • Task: irat
  • MD5: c6ed210b6460e8b89497904c276b26a8
  • Run description: Test run to see if the submission goes through

UIowaS01r

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: UIowaS01r
  • Participant: UIowaS
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/3/2012
  • Task: trat
  • MD5: f90b2b642a2239c81c0016dcf437d68a
  • Run description: Ranked by a weighted score. Partitioned into fixed sizes and sent to the crowd in ranked order.

UIowaS02r

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: UIowaS02r
  • Participant: UIowaS
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/3/2012
  • Task: trat
  • MD5: 182d4b56504a06b3c9abc93784d75d4d
  • Run description: Used k-means text clustering. Each cluster ranked by an average of each document's weighted score. Sent to the crowd in ranked order first by cluster, then by order within the cluster.

UIowaS03r

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: UIowaS03r
  • Participant: UIowaS
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/3/2012
  • Task: trat
  • MD5: 89199dfeb48c12deda26e2020b7f96b7
  • Run description: Used k-means text clustering. Each cluster ranked by an average of each document's weighted score. Sent to the crowd in ranked order by cluster, then by a random sample of documents within the cluster.

UTAustinM

Participants | Input | Summary | Appendix

  • Run ID: UTAustinM
  • Participant: UTAustin
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 9/15/2012
  • Task: irat
  • MD5: b5d94b813eb76df004d3b4bb9814e2b6
  • Run description: Incremental Crowd-sourcing by improving the level of agreement per each HIT and increasing a pool of trustful workers.

yorku12cs01

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: yorku12cs01
  • Participant: york
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/28/2012
  • Task: trat
  • MD5: 35d4a682ea23892690ceb2b57097f14e
  • Run description: Judgements partially based on fusion with TF and partially from CrowdFlower.

yorku12cs02

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: yorku12cs02
  • Participant: york
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/28/2012
  • Task: trat
  • MD5: cad8f20805028ce5be02cade3276a845
  • Run description: Judgements based on fusion with extracted expansion words from crowdFlower.

yorku12cs03

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: yorku12cs03
  • Participant: york
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/28/2012
  • Task: trat
  • MD5: 63542a908ff906d6805ff32ce1fc5480
  • Run description: Judgements based on fusion with expansion words from term proximity.

yorku12cs04

Participants | Proceedings | Input | Summary | Appendix

  • Run ID: yorku12cs04
  • Participant: york
  • Track: Crowdsourcing
  • Year: 2012
  • Submission: 8/28/2012
  • Task: trat
  • MD5: 6a971cb0e2f35fe6db8d958c7f280e99
  • Run description: Judgements based on fusion with expansion words from pseudo feedback and partially from crowdFlower.