PEEM paper published in IEEE Access! ๐
Published:
I am thrilled to announce that our paper, โPEEM: Prompt Engineering Evaluation Metrics for Interpretable Joint Evaluation of Prompts and Responses,โ has been officially published in IEEE Access! ๐
This research introduces a novel evaluation framework, PEEM, designed to jointly assess the quality of prompts and the corresponding model responses. Unlike traditional metrics that often treat the model as a black box, PEEM provides interpretable insights into the interaction between prompt engineering and output performance.
๐ก Key Contributions
1๏ธโฃ Joint Evaluation Framework: We move beyond simple response scoring by simultaneously evaluating the instructional quality of prompts and the resulting response accuracy.
2๏ธโฃ Interpretable Metrics: PEEM offers granular metrics that help researchers and practitioners understand why a model succeeds or fails based on specific prompt characteristics.
3๏ธโฃ Extensive Validation: We demonstrated PEEMโs effectiveness through rigorous testing across multiple large language models (LLMs) and diverse prompt engineering strategies.
You can access the full paper through the following links:
- IEEE Xplore: https://ieeexplore.ieee.org/document/11460151
- arXiv (Preprint): https://arxiv.org/abs/2603.10477
Iโm incredibly grateful to my co-authors and the reviewers for their valuable feedback and support throughout this process! ๐
