Cast your vote
You can rate an item by clicking the amount of stars they wish to award to this item.
When enough users have cast their vote on this item, the average rating will also be shown.
Your vote was cast
Thank you for your feedback
Thank you for your feedback
AdvisorsOrasan, Constantin Dr.
Mitkov, Ruslan Prof.
Navigli, Roberto Prof.
MetadataShow full item record
AbstractOpen-domain question answering (QA) is an established NLP task which enables users to search for speciVc pieces of information in large collections of texts. Instead of using keyword-based queries and a standard information retrieval engine, QA systems allow the use of natural language questions and return the exact answer (or a list of plausible answers) with supporting snippets of text. In the past decade, open-domain QA research has been dominated by evaluation fora such as TREC and CLEF, where shallow techniques relying on information redundancy have achieved very good performance. However, this performance is generally limited to simple factoid and deVnition questions because the answer is usually explicitly present in the document collection. Current approaches are much less successful in Vnding implicit answers and are diXcult to adapt to more complex question types which are likely to be posed by users. In order to advance the Veld of QA, this thesis proposes a shift in focus from simple factoid questions to encyclopaedic questions: list questions composed of several constraints. These questions have more than one correct answer which usually cannot be extracted from one small snippet of text. To correctly interpret the question, systems need to combine classic knowledge-based approaches with advanced NLP techniques. To Vnd and extract answers, systems need to aggregate atomic facts from heterogeneous sources as opposed to simply relying on keyword-based similarity. Encyclopaedic questions promote QA systems which use basic reasoning, making them more robust and easier to extend with new types of constraints and new types of questions. A novel semantic architecture is proposed which represents a paradigm shift in open-domain QA system design, using semantic concepts and knowledge representation instead of words and information retrieval. The architecture consists of two phases, analysis – responsible for interpreting questions and Vnding answers, and feedback – responsible for interacting with the user. This architecture provides the basis for EQUAL, a semantic QA system developed as part of the thesis, which uses Wikipedia as a source of world knowledge and iii employs simple forms of open-domain inference to answer encyclopaedic questions. EQUAL combines the output of a syntactic parser with semantic information from Wikipedia to analyse questions. To address natural language ambiguity, the system builds several formal interpretations containing the constraints speciVed by the user and addresses each interpretation in parallel. To Vnd answers, the system then tests these constraints individually for each candidate answer, considering information from diUerent documents and/or sources. The correctness of an answer is not proved using a logical formalism, instead a conVdence-based measure is employed. This measure reWects the validation of constraints from raw natural language, automatically extracted entities, relations and available structured and semi-structured knowledge from Wikipedia and the Semantic Web. When searching for and validating answers, EQUAL uses the Wikipedia link graph to Vnd relevant information. This method achieves good precision and allows only pages of a certain type to be considered, but is aUected by the incompleteness of the existing markup targeted towards human readers. In order to address this, a semantic analysis module which disambiguates entities is developed to enrich Wikipedia articles with additional links to other pages. The module increases recall, enabling the system to rely more on the link structure of Wikipedia than on word-based similarity between pages. It also allows authoritative information from diUerent sources to be linked to the encyclopaedia, further enhancing the coverage of the system. The viability of the proposed approach was evaluated in an independent setting by participating in two competitions at CLEF 2008 and 2009. In both competitions, EQUAL outperformed standard textual QA systems as well as semi-automatic approaches. Having established a feasible way forward for the design of open-domain QA systems, future work will attempt to further improve performance to take advantage of recent advances in information extraction and knowledge representation, as well as by experimenting with formal reasoning and inferencing capabilities.
PublisherUniversity of Wolverhampton
TypeThesis or dissertation
Showing items related by title, author, creator and subject.
Automatic Generation of Factual Questions from Video DocumentariesMitkov, Ruslan; Specia, Lucia; Ha, Le An; Skalban, Yvonne (University of Wolverhampton, 2013-10)Questioning sessions are an essential part of teachers’ daily instructional activities. Questions are used to assess students’ knowledge and comprehension and to promote learning. The manual creation of such learning material is a laborious and time-consuming task. Research in Natural Language Processing (NLP) has shown that Question Generation (QG) systems can be used to efficiently create high-quality learning materials to support teachers in their work and students in their learning process. A number of successful QG applications for education and training have been developed, but these focus mainly on supporting reading materials. However, digital technology is always evolving; there is an ever-growing amount of multimedia content available, and more and more delivery methods for audio-visual content are emerging and easily accessible. At the same time, research provides empirical evidence that multimedia use in the classroom has beneficial effects on student learning. Thus, there is a need to investigate whether QG systems can be used to assist teachers in creating assessment materials from these different types of media that are being employed in classrooms. This thesis serves to explore how NLP tools and techniques can be harnessed to generate questions from non-traditional learning materials, in particular videos. A QG framework which allows the generation of factual questions from video documentaries has been developed and a number of evaluations to analyse the quality of the produced questions have been performed. The developed framework uses several readily available NLP tools to generate questions from the subtitles accompanying a video documentary. The reason for choosing video vii documentaries is two-fold: firstly, they are frequently used by teachers and secondly, their factual nature lends itself well to question generation, as will be explained within the thesis. The questions generated by the framework can be used as a quick way of testing students’ comprehension of what they have learned from the documentary. As part of this research project, the characteristics of documentary videos and their subtitles were analysed and the methodology has been adapted to be able to exploit these characteristics. An evaluation of the system output by domain experts showed promising results but also revealed that generating even shallow questions is a task which is far from trivial. To this end, the evaluation and subsequent error analysis contribute to the literature by highlighting the challenges QG from documentary videos can face. In a user study, it was investigated whether questions generated automatically by the system developed as part of this thesis and a state-of-the-art system can successfully be used to assist multimedia-based learning. Using a novel evaluation methodology, the feasibility of using a QG system’s output as ‘pre-questions’ with different types of prequestions (text-based and with images) used was examined. The psychometric parameters of the automatically generated questions by the two systems and of those generated manually were compared. The results indicate that the presence of pre-questions (preferably with images) improves the performance of test-takers and they highlight that the psychometric parameters of the questions generated by the system are comparable if not better than those of the state-of-the-art system. In another experiment, the productivity of questions in terms of time taken to generate questions manually vs. time taken to post-edit system-generated questions was analysed. A viii post-editing tool which allows for the tracking of several statistics such as edit distance measures, editing time, etc, was used. The quality of questions before and after postediting was also analysed. Not only did the experiments provide quantitative data about automatically and manually generated questions, but qualitative data in the form of user feedback, which provides an insight into how users perceived the quality of questions, was also gathered.
Automatically marked summative assessment using internet toolsPenfold, Brian (University of Wolverhampton, 2001)With very large groups, individual assessment is becoming increasingly difficult. We are constantly aware of the cost of the time taken in traditional forms of assessment and the effect of marking fatigue on quality. The system described here is a ‘home-grown’ system to present summative multiple-choice question (MCQ) papers in an efficient, cost effective and simple way. The system directly replaces manually marked MCQ tests and because of its nature opens up new more sophisticated multimedia assessment formats.
Reviewing the challenge for able students’: a participatory enquiry exploring the nature of pedagogy that can enhance cognitive engagement with homework.Devlin, Linda.; Badyal, Caroline (University of Wolverhampton, 2013-07)This thesis investigates and analyses the level of challenge for able students in an 11 - 18 Academy. It is addressed from my position as the Principal of the case study Academy and a novice researcher. Eight teachers who formed the Teaching and Learning group within the Academy participated in the study, as part of a community of practice with an interest in the issue addressed and the research process. The study focused on concerns arising from Learning Walks and Ofsted feedback about the perceived lack of challenge for able students. Using a three layer action research methodology, the views and practices of staff and students about challenge in ILTs (Independent Learning Tasks) were explored. An initial brainstorming activity was followed by questionnaires, lesson observations and focus group sessions with a sample of 100 students (Years 7, 9, 10 and 11). At the close of the first layer of research, data analysis revealed a range of levels of challenge in different subject areas, and from these a Year 10 Geography group was selected, with the support of the teacher. The second action research layer involved the Geography teacher and 15 Geography students who had identified a lack of challenge in their ILTs. This shifted the focus of the research to consider the cognitive challenge incorporated into tasks, focusing on thinking skills and questioning techniques. The third and final action research layer resulted in a newly developed, collaboratively-constructed ‘student friendly’ thinking skills analysis which provided powerful and formative insights to ‘label’ challenge. The teacher responded reflexively to the outcomes by trying out a redeveloped approach to ILTs (homework) and questioning techniques within the Academy. The findings from this investigation suggest that, cognitively challenging, problem-solving tasks, co-constructed with students to include opportunities for Socratic questioning provide for greater challenge in the classroom. Finally, the benefits to be gained from establishing a research community where the Principal is the lead researcher, include an increased emphasis on staff as change agents and the critical contribution of student voice in pursuit of challenging teaching and learning.