A new study by the Educational Testing Service (ETS) has raised concerns about potential racial bias in AI essay grading systems.
Researchers compared human grader scores with those of a ChatGPT-based AI model, GPT-4o. GPT-4o consistently scored essays lower than human graders, averaging 2.8 compared to 3.7.
This discrepancy was most pronounced for Asian American students, who received an average of 3.2 from GPT-4o versus 4.3 from human graders, a difference of 1.1 points. The gap for white, Black and Hispanic students was smaller, averaging 0.9 points.
🤳Click our linkinbio to sign up to our daily newsletter.
📰Follow us for the latest Asian News NextShark
#news #study #racialbias #aiessaygrading #chatgpt #aimodel