att_abstract={{Real-time speech-to-speech (S2S) translation of lectures and speeches require simultaneous translation with low latency to continually engage the listeners. However, simultaneous speech-to-speech translation systems have been predominantly repurposing translation models that are typically trained for 
consecutive translation without a motivated attempt to model incrementality. Furthermore, the notion of translation is simplified to translation plus simultaneity. In contrast, human interpreters are able to perform simultaneous interpretation by generating target speech incrementally with very low ear-voice span by 
using a variety of strategies such as compression (paraphrasing), incremental comprehension, and anticipation through discourse inference and expectation of discourse redundancies. Exploiting and modeling such phenomena can potentially improve automatic real-time translation of speech. As a first step, in this work 
we identify and present a systematic analysis of phenomena used by human interpreters to perform simultaneous interpretation and elucidate how it can be exploited in a conventional simultaneous translation framework. We 
perform our study on a corpus of simultaneous interpretation of Parliamentary speeches in English and Spanish. Specifically, we present an empirical analysis of factors such as time constraint, redundancy and inference as 
evidenced in the simultaneous interpretation corpus.}},
	att_authors={vk947h, sb7658, jc582p},
	att_copyright_notice={{The definitive version was published in 2013. {{, 2013-08-31}}
	author={Vivek kumar Rangarajan sridhar and Srinivas Bangalore and John Chen},
	institution={{Interspeech 2013}},
	title={{Corpus Analysis of Simultaneous Interpretation Data for Improving Real Time Speech Translation}},