Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Why is toxicity threshold so low? It is set to 0.00056. #9

Open
eitanturok opened this issue Dec 29, 2023 · 0 comments
Open

Why is toxicity threshold so low? It is set to 0.00056. #9

eitanturok opened this issue Dec 29, 2023 · 0 comments

Comments

@eitanturok
Copy link

eitanturok commented Dec 29, 2023

In configs/toxicity/conditional.yml, we have the line

dataset:
  conditional_training_config:
    threshold: 0.00056
    aligned_prefix: "<|aligned|>"
    misaligned_prefix: "<|misaligned|>"
    drop_token_fraction: 0.01

Why here is the toxicity threshold 0.00056? This is incredibly low. Only sentences with toxicity scores lower than 0.00056 would be marked as non-toxic. Everything greater (or equal to) that would be marked as toxic.

Don't we only want documents to be marked as toxic when their toxicity is, let's say, 0.9 or greater? (I chose 0.9 arbitrarily as an example). Generally speaking, 0.00056 seems to be quite a low threshold and I'm worried that this might hurt performance.

Can you explain the thought process that went into making the toxicity threshold 0.00056? Is this simply what got the best results?

Thanks!

@eitanturok eitanturok changed the title Why is toxicity threshold set to 0.00056? Why is toxicity threshold so low (it is set to 0.00056)? Dec 29, 2023
@eitanturok eitanturok changed the title Why is toxicity threshold so low (it is set to 0.00056)? Why is toxicity threshold so low? It is set to 0.00056. Dec 29, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant