MT providers claim that (customized) MT "helps you translate more words and grow your business". It "boosts productivity". And it can even "increase revenues" or "optimize customer service and support". What is the reality? Is MT improving or did we reach a plateau? Speakers at last year's QE Summit agreed: one of the main problems in the translation industry today is the lack of benchmarking. The output of MT engines cannot be compared to industry averages or standards because these are not yet available. Automated scores are meaningless outside the “laboratory”. At the same time, buyers of translation services are increasingly interested in translated content of different quality levels. They also want to know how the different engines are performing on different content types and in different language pairs. How do we know? Can we predict the output quality? Shouldn’t MT providers become more transparent to help buyers of these technologies make informed decisions? Session leader: Dag Schmidtke (Microsoft) Panelists: John Tinsley (Iconic), Olga Beregovaya (Welocalize), Olga Pospelova (eBay)