PrediQT – Predicting question complexity

The complexity scores generated using this model are based on a Large Language Model trained on thousands of human responses. The model is strongly correlated with human ratings of question complexity.

PrediQT’s scoring is grounded in the hierarchical Bloom taxonomy framework, which classifies cognitive tasks from basic recall through creative synthesis. By aligning the LLM’s continuous outputs to Bloom’s six levels—Remember, Understand, Apply, Analyze, Evaluate, and Create—the app ensures that each complexity score reflects well-established educational standards.

Either upload an Excel file or paste questions below.

  • Upload → returns a table with a new Complexity Score column.
  • Paste → returns one score per line.

Disclaimer: All questions asked by users are collected anonymously and used only for scientific purposes. By uploading and scoring questions, you consent to your data being used for research purposes.

Questions & Complexity Scores

Questions & Complexity Scores
1 6

Admin only: download full question history

Cite

Raz, T., Luchini, S., Beaty, R., & Kenett, Y. N. (2024). Automated Scoring of Open-Ended Question Complexity: A Large Language Model Approach. Research Square https://doi.org/10.21203/rs.3.rs-3890828/v1