Large language models are biased. Can logic help save them?
MIT CSAIL researchers trained logic-aware language models to reduce harmful stereotypes like gender and racial biases using textual-entailment models.
MIT CSAIL researchers trained logic-aware language models to reduce harmful stereotypes like gender and racial biases using textual-entailment models.