This study addresses the challenges of large-scale reading comprehension assessment using the Cloze test, a technique where students fill in gaps in a text. Traditionally, only exact words that match with the original text are considered correct, which limits the assessment of students' actual text comprehension. Our research proposes a practical solution: using Natural Language Processing (NLP) techniques, specifically word embedding models, to evaluate semantic similarity between the expected answer and the student's response. Word embeddings are numerical representations of words in a multi-dimensional space, where words with similar meanings are positioned close to each other. This approach allows for considering semantically similar answers as correct, even when they differ from the original word. We administered a Cloze test to elementary school students in Brazil and compared three different word embedding models (GloVe, Wang2Vec, and spaCy) to evaluate the semantic similarity of responses. To validate our approach, we engaged twelve human judges who ranked the students' answers.