Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Very high false coding rate #52

Open
ahalterman opened this issue Oct 29, 2018 · 7 comments
Open

Very high false coding rate #52

ahalterman opened this issue Oct 29, 2018 · 7 comments
Labels
critical must address before program functions

Comments

@ahalterman
Copy link
Member

(Just making an issue for what @philip-schrodt has reported elsewhere so we can consolidate discussion):

UniversalPetrarch is enormously overproducing events. Not sure what the source of the false positives is. E.g.:

Records evaluated:    1018
Correct events:        565    44.21%
Uncoded events:        713    55.79%
Extra events:         3212   315.52%
@PTB-OEDA
Copy link
Member

PTB-OEDA commented Oct 30, 2018 via email

@ahalterman
Copy link
Member Author

Any update on this? This is one of the biggest obstacles to using UniversalPetrarch in production.

@PTB-OEDA
Copy link
Member

PTB-OEDA commented Nov 5, 2018 via email

@ahalterman
Copy link
Member Author

From talking with @JingL1014, it sounds like she has a couple theories for what could be causing the high false positives:

  1. other true events in the sentences (we only annotated protest and assault events, so events beyond those will be seen as "extra events")
  2. the actor spans could be too large, leading to too many actors being coded, and therefore too many events.

There's also the possibility there's some larger unknown problem in how it's handling the coding.

I think a few tests could help figure out which are going on:

  • what's the overproduction of protest or assault events only? (eliminate cases from theory 1)
  • how many of the extra events share an event code and just have different actors? (That would test theory 2)
  • look at the spans UP is returning (see Use GSR phrase info to improve UP #44) to make sure nothing wild is happening in the phrase extraction part.

@ahalterman
Copy link
Member Author

Is there an update on this? Have you looked into those tests? It sounds from @philip-schrodt like some sentences are still producing many events so something's still going wrong.

@PTB-OEDA
Copy link
Member

PTB-OEDA commented Dec 7, 2018 via email

@philip-schrodt
Copy link
Contributor

I added this in validation.py to

return_dict = petrarch_ud.do_coding(dict)
PETRwriter.write_events(return_dict, "evts.validation.txt")`

where write_events() is the routine that petrarch_ud uses to write the final output file, and for AFP_SPA_19940921.0205_7.0_0 is still produces 51 events in the regular output: I really need to be able to look at the code that Arizona is running which is producing fewer events, since all I'm doing is making calls to the petrarch_ud code I've downloaded from the -master branch: downloaded a new copy this morning and confirmed it is still generating this behavior.

@ahalterman ahalterman added the critical must address before program functions label Jan 25, 2019
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
critical must address before program functions
Projects
None yet
Development

No branches or pull requests

3 participants