作者: Katunobu Itou , Atsushi Fujii , Tetsuya Ishikawa , Tomoyosi Akiba
DOI:
关键词:
摘要: ABSTRACTSpoken queries submitted to question answering systemsusually consist of query contents (e.g. about newspaper ar-ticles) and frozen patterns WH-words), which can bemodeled with N-gram models grammar-based models,respectively. We propose a method integrate those differ-ent types into single model. rep-resent the two language in wordnetwork. However, common smoothing methods, whichare effective for models, decrease grammatical con-straints patterns. For this problem, we proposea selective back-off method, controls adegree is applied depending net-work fragment. Additionally, resulting are compati-ble conventional thusexisting decoders easily be used. show theeffectiveness our by way experiments.1. INTRODUCTIONThe model has been used successfully as languagemodel large vocabulary continuous speech recognition(LVCSR) systems. The simple but ro-bust enough all word sequences vocabulary.However, it needs training corpus such corpuscannot constructed unless there already exists alarge text based on, example, articles.On other hand, alanguage tasks involving relatively small vocab-ulary. This does not need because ittakes advantage linguistic knowledge. It corre-lations more distant than possible model,which only local relations between words.Thus, some spoken sentences modeled suit-ably suitably bythe also true from an intra-sentence perspective – parts sentence best mod-eled bya model.For systems receive queriesthat often part that conveys various con-tents about, articles, thatrepresents pattern sentences. firstpart seems dealt using modeltrained secondpart