JADE/Open questions

Open questions
 * We need to define what type of feedback we accept from people.  Free-form text, but also some quantitative number that measures level of disagreement?  Prespecified, structured data and a comment from human evaluator.


 * Which entities will we eventually support as thread target?


 * What makes a Meta-ORES thread any different from a talk page topic?


 * We need terminology for each element of our system.  "target entity"?  "refutation", "review", "agreement"?  "comment"?  "reply"?


 * Since comments will likely reference ORES scores as part of a refutation, we'll want to store that relationship explicitly.


 * Should this affect the automated score, automatically?  After some arbitration process?


 * Meta ORES human judgments should have some kind of influence on the error function we use during ML training.  How is that feedback structured?  Is it a strong signal?  Do we respect all judgment equally? Do we need another level of human meta-meta-curation of feedback??


 * What do we expect clients will do with this information?