Opinion | A.I. Will Change Education. Don’t Let It Worsen Inequality. - The New York Times
▻https://www.nytimes.com/2022/12/15/opinion/chatgpt-education-ai-technology.html
It is in high schools and even college where some of ChatGPT’s most interesting and troubling aspects will become clear.
Essay writing is most often assigned not because the result has much value — proud parents putting good grades on the fridge aside — but because the process teaches crucial skills: researching a topic, judging claims, synthesizing knowledge and expressing it in a clear, coherent and persuasive manner. Those skills will be even more important because of advances in A.I.
When I asked ChatGPT a range of questions — about the ethical challenges faced by journalists who work with hacked materials, the necessity of cryptocurrency regulation, the possibility of democratic backsliding in the United States — the answers were cogent, well reasoned and clear. It’s also interactive: I could ask for more details or request changes.
But then, on trickier topics or more complicated concepts, ChatGPT sometimes gave highly plausible answers that were flat-out wrong — something its creators warn about in their disclaimers.
Unless you already knew the answer or were an expert in the field, you could be subjected to a high-quality intellectual snow job.
In flipped classrooms, students wouldn’t use ChatGPT to conjure up a whole essay. Instead, they’d use it as a tool to generate critically examined building blocks of essays. It would be similar to how students in advanced math classes are allowed to use calculators to solve complex equations without replicating tedious, previously mastered steps.
Teachers could assign a complicated topic and allow students to use such tools as part of their research. Assessing the veracity and reliability of these A.I.-generated notes and using them to create an essay would be done in the classroom, with guidance and instruction from teachers. The goal would be to increase the quality and the complexity of the argument.
This would require more teachers to provide detailed feedback. Unless sufficient resources are provided equitably, adapting to conversational A.I. in flipped classrooms could exacerbate inequalities.
Some school officials may treat this as a problem of merely plagiarism detection and expand the use of draconian surveillance systems. During the pandemic, many students were forced to take tests or write essays under the gaze of an automated eye-tracking system or on a locked-down computer to prevent cheating.
In a fruitless arms race against conversational A.I., automated plagiarism software may become supercharged, making school more punitive for monitored students. Worse, such systems will inevitably produce some false accusations, which damage trust and may even stymie the prospects of promising students.
Educational approaches that treat students like enemies may teach students to hate or subvert the controls. That’s not a recipe for human betterment.
As societies responded to previous technological advances, like mechanization, by eventually enacting a public safety net, a shorter workweek and a minimum wage, we will also need policies that allow more people to live with dignity as a basic right, even if their skills have been superseded. With so much more wealth generated now, we could unleash our imagination even more, expanding free time and better working conditions for more people.
The way forward is not to just lament supplanted skills, as Plato did, but also to recognize that as more complex skills become essential, our society must equitably educate people to develop them. And then it always goes back to the basics. Value people as people, not just as bundles of skills.