Model,Accuracy Qwen2-7B-Instruct,0.9353258206761391 Meta-Llama-3.1-8B-Instruct,0.9039686428221461 Qwen2_5_32B_Instruct,0.9559039686428221 Qwen2_5_7B_Instruct,0.9348358647721705 Qwen2_5_1_5B_Instruct,0.8314551690347869 Qwen2-72B-Instruct,0.9612934835864773 cross_openhermes_llama3_8b_4096_inst,0.8613424791768741 Meta-Llama-3-8B-Instruct,0.8946594806467418 Meta-Llama-3.1-70B-Instruct,0.9559039686428221 Qwen2_5_3B_Instruct,0.9029887310142087 SeaLLMs-v3-7B-Chat,0.9265066144047036 Qwen2_5_72B_Instruct,0.9627633512983832 gemma-2-9b-it,0.9416952474277315 Meta-Llama-3-70B-Instruct,0.9480646741793238 Qwen2_5_14B_Instruct,0.9461048505634493 sg_llama3_8192_8b,0.9103380695737384 sg_llama3_70b_inst,0.9524742773150416 gemma-2-2b-it,0.8510534051935326 llama3-8b-cpt-sea-lionv2-instruct,0.8858402743753062 Qwen2_5_0_5B_Instruct,0.6526212640862322 GPT4o_0513,0.9583537481626654 cross_openhermes_llama3_70b_4096_inst,0.9514943655071043 cross_openhermes_llama3_8b_4096_2_inst,0.876531112199902