SDL recently announced some new initiatives in their machine translation product SDL Enterprise Translation Server (ETS). As with most press releases, there is very little specific detail in the release itself, and my initial impression was that there really was not much news here other than the mention that they are also doing Neural MT. Thus, I approached my SDL contacts and asked if they would be willing to share some more information with me for this blog, for the reader base who were curious about what this announcement really means. I initially set out with questions that were very focused on NMT, but the more I learned about SDL ETS, the more I felt that it was worth more attention. The SDL team were very forthcoming, and shared interesting material with me, thus allowing me to provide a clearer picture in this post of the substance behind this release, (which I think Enterprise Buyers, in particular, should take note of), and which I have summarized below. The press release basically focuses on two things:
The SDL ETS platform is an evolution of the on-premises MT product offering that has been in use in Government environments for over 15 years now and has been widely used in national security and intelligence agencies in particular, across the NATO block of countries. Given the nature of national security work, the product has had to be rock solid, and as support-free as possible as anti-terrorist analysts are not inclined to, or permitted to make calls for technical support to the MT vendor. Those of you who have struggled through clunky, almost-working-MT-onsite-software from other vendors, who are less prepared for this use-case, may appreciate the value of this long-term experience.
As we have seen of late, determined hackers can break into both government and corporate networks and do so on a regular basis. This graphic and interactive visualization is quite telling in how frequently hackers succeed, and how frequently large data sets of allegedly private data are accidentally exposed. So it is understandable that when the new SDL management surveyed large global enterprises priorities, they found that “Data Security and Data Privacy” were a key concern for many executives across several industries.
In a world where MT is a global resource, and 500 Billion words a day are being translated across the various public MT portals, data security and privacy have to be a concern for a responsible executive, and, any serious corporate governance initiative. While the public MT platforms have indeed made MT ubiquitous, they also generally reserve the right to run machine learning algorithms on our data, to try and pick up useful patterns from our usage and continue to improve their services. Today, MT is increasingly used to translate large volumes of customer and corporate communications, and it is likely that most responsible executives would rather not share the intimate details of their customer and intra-corporate global communications with the public MT services where privacy could be compromised.
If you think that your use of MT is not monitored or watched, at least at a machine learning level, you should perhaps take a look at the following graphic. This link provides a summary of what they collect.
The notion that these MT services are “free” is naive, and we cannot really be surprised that the public MT services try to capitalize on what they can learn from the widespread use of their MT services. The understanding gained from watching user behavior not only helps improve the MT technology, it also provides a basis to boost advertising revenues, since an MT service provider has detailed metrics on what people translate, and search on, in different parts of the world.
To adapt the original ETS platform to the different needs of the global enterprise market, SDL had to add several features and capabilities that were not required for national security information triage applications, where MT was invariably an embedded component service, interacting with other embedded components like classification and text analytics in a larger information analysis scenario. The key enhancements added are for the broader enterprise market, where MT can be an added corporate IT service for many different kinds of applications, and the MT service needs direct as well as embedded access. The new features include the following:
Based on my experience with, and observation of other on-premise MT systems, I think it is fair to say that the SDL ETS features are a significant step forward for the translation industry in bringing the capabilities of industrial strength MT up to modern enterprise IT standards and needs. In #americanbuffoon-speak we might even say it is tremendous and bigly good.
Based on what I was able to gather from my conversations, here is a list of distinctive features that come to mind. Note that most of these current updates relate to the improved UX, the elegant simplicity of SDL ETS, and, the ease of ongoing management of changing needs and requirements from an enterprise customer viewpoint.
Except for the NMT efforts at Facebook, Google, and Microsoft, most of the NMT initiatives we hear about in the translation industry today, are based on open source solutions built around the Torch and Theano frameworks. While using open source allows an MT practitioner to get started quickly, it also means that they have to submit to the black box nature of the framework. Very few are going to be able to go into the source code, to fundamentally alter the logic and mechanics of the framework, without potentially damaging or destabilizing the basic system. The NMT systems that practitioners are able to develop are only as good as the data they use, or their ability to modify the open source codebase. In contrast to this, the SDL NMT core engine is owned and developed 100% in-house by SDL, which allows much greater flexibility and deep underlying control of the basic logic and data processing flows of the NMT framework. This deeper control also allows a developer more alternatives in dealing with NMT challenges like limited vocabulary, performance/speed and changing the machine learning strategies and techniques as the market and the deep learning technology evolve e.g. switching from recurrent neural networks (RNN) to convolutional neural networks (CNN) deep learning strategies as Facebook just did. My sense, based on my limited understanding, is that owning your NMT code base very likely affords more powerful control options than open source alternatives allow, because problem areas can be approached at a more fundamental level, in the well-understood source code, rather than using workarounds to handle problematic output of open source black-box components. It is also possible that owning and understanding the code base also results in longer term integrity and stability in the code base. This is also probably why the big 3 choose to develop their own code base over using open source components and foundations. The SDL system architecture reflects 15+ years of experience with data-driven MT and is designed to allow rapid response to emerging changes in machine learning technology, like a possible change to CNN from the current RNN + Attention approach that everybody is using. In my conversations with SDL technical team members, it became apparent that they have a much greater ability to address several different NMT problem areas:
The NMT customization and adaptation options are currently being explored and benchmarked against well understood PB-SMT systems. Initial results are providing great insight into specific data combination and pruning strategies that result in the best custom NMT system output. SDL’s long-term experience building thousands of custom systems should be invaluable in driving the development of superior custom NMT solutions. The research methodology used to investigate this follows best practices (i.e. they are careful and conservative, unlike the over-the-top claims by Google) and we should expect that all production NMT systems will be significantly superior to most other alternatives. While I am not at liberty to share details of the benchmark comparisons, I can say that the improvements are significant and especially promising in language combinations that are especially important to global enterprises. The SDL team is also especially careful in making public claims about improved productivity and quality (unlike some MT vendors in the market), and are gathering multiple points of verification from both internal and customer tests to validate initial results which are very promising.
I expect that they will also start (or possibly have already started) exploring linking their very competitive Adaptive MT capabilities with high-quality NMT engines. I look forward to learning more about their NMT experience in production customer environments.
Happy 4th of July to my US friends, and here is an unusual presentation of the national anthem that is possibly more fitting for the current leadership.
Originally published at kv-emptypages.blogspot.com on July 19, 2017.