
Why we need your help: To avoid subjectively biased datasets and obtain a larger set of samples, we’re reaching out to philosophers and those with related skills to obtain more examples of good and bad cases of informal reasoning, with the bad cases ideally exemplifying one or another of the most common informal fallacies typically taught in undergraduate reasoning courses. We kindly request that these be original examples not to be found anywhere online (in case they’ve already been part of the training data for the relevant models!). It shouldn’t take more than 5-10 minutes of your time, and you’re welcome to submit anywhere from 1 to 5 examples of fallacious and reasonable cases. Further details are available at the indicated link. All submitters who wish to make their names public will be acknowledged in materials submitted to BIG-bench.
Our specific hope at the moment is to assess the ability of these language models to detect informal fallacies—equivocations, ad hominem attacks, no-true-Scotsman arguments, faulty generalisations, and the like. While classifying fallacies sensibly and appropriately is far from trivial—many philosophy undergraduates struggle with it at the best of times—it has the advantage of being a relatively easy-to-operationalise aspect of informal reasoning that could help drive progress in AI natural language reasoning assessment in general. This latter project has the potential for significant social impact. If large scale language models could be developed that surpass human performance in categorising informal reasoning as good or bad, there may be many socially-impactful applications, ranging from more nuanced fact-checking to assistive tools to help students, policymakers, and journalists evaluate the cogency of their own arguments.
[Kjetil Golid, “Mint” (detail)]
