ParEval Leaderboard: Evaluating the Ability of Large Language Models to Generate Parallel Code
We introduced the ParEval benchmark in “Can Large Language Models Write Parallel Code?” to evaluate the capability of LLMs at parallel code generation. We fo...