Learning Library

← Back to Library

LLMs Transforming Global Machine Translation

Key Points

  • The speaker stresses that understanding a message often depends on knowing the speaker’s language, highlighting the critical role of translation.
  • Only about 25 % of internet users have English as their primary language, while more than 65 % prefer content and support in their native languages, making machine translation essential for business.
  • Traditional machine‑translation approaches include rule‑based systems using linguistic rules and dictionaries, statistical methods that learn patterns from human‑translated data, and neural models that consider whole‑sentence structure.
  • Hybrid systems can combine these methods, but all rely on supervised learning and explicit linguistic resources.
  • Large language models (LLMs) represent a new, superior era of translation by leveraging massive pre‑trained knowledge to deliver more accurate, context‑aware multilingual output.

Full Transcript

# LLMs Transforming Global Machine Translation **Source:** [https://www.youtube.com/watch?v=TabyC8otFY8](https://www.youtube.com/watch?v=TabyC8otFY8) **Duration:** 00:06:18 ## Summary - The speaker stresses that understanding a message often depends on knowing the speaker’s language, highlighting the critical role of translation. - Only about 25 % of internet users have English as their primary language, while more than 65 % prefer content and support in their native languages, making machine translation essential for business. - Traditional machine‑translation approaches include rule‑based systems using linguistic rules and dictionaries, statistical methods that learn patterns from human‑translated data, and neural models that consider whole‑sentence structure. - Hybrid systems can combine these methods, but all rely on supervised learning and explicit linguistic resources. - Large language models (LLMs) represent a new, superior era of translation by leveraging massive pre‑trained knowledge to deliver more accurate, context‑aware multilingual output. ## Sections - [00:00:00](https://www.youtube.com/watch?v=TabyC8otFY8&t=0s) **The Business Case for LLM Translation** - The speaker highlights the widespread need for native‑language content and support online, cites user preference statistics, and outlines how large language models can enhance machine translation over traditional approaches. - [00:05:52](https://www.youtube.com/watch?v=TabyC8otFY8&t=352s) **Leveraging LLMs for Multilingual Outreach** - The speaker highlights using large language models to communicate with customers in their native languages and concludes with a request to subscribe and like. ## Full Transcript
0:00Hello. 0:01[foreign language...] 0:07Unless you know Hindi, you wouldn't understand what I just said. What I 0:12said was I wanted to tell you something very  important-- unless you know my language, you 0:17can't understand it. All of you must know and must  have experienced LLMs (large language models) in 0:23the recent times. Large language models are  very popularly known for generating text, but 0:30it is also important to know that LLMs can also  do a very good job of translating languages. Why 0:38is this important? It seems only about 25% of the  Internet users--their primary language is English. 0:47And more than 65% of the users on the Internet  prefer to be provided information in their 0:55primary languages--respective primary languages.  Also, more than 70% of Internet users would like 1:03to receive support, issue resolution, etc. in  their preferred languages. Now, because they do 1:12not receive the help in their primary languages,  more than 65% of these Internet users are using 1:20machine translations to get the help that they  need. So it seems that machine translations are 1:28essential for us to do business. So I'm going to  explain machine translations in two parts. First, 1:36I will talk about how we have been doing machine  translation so far, and then I will jump to the 1:42advantage of the large language models and how we  are going to do translations from them. So let's 1:48see how machine translations are done. Machine  translations use artificial intelligence to 1:54translate between languages automatically without  any human help. So how do they do that? Let's take 2:01an example here: English, Spanish, and Japanese.  To translate between any of these languages, 2:10you need linguistic rules and you need  dictionaries for each of these languages. 2:16And the machine translation are done in multiple  approaches. The rule based approach--the rule 2:22based approach is the one that predominantly  uses the linguistic rules and the dictionaries 2:28and also the parallel dictionaries that have the  meanings of two different languages, the source 2:34language as well as the target language. And then  we have the second approach called the statistical 2:41approach. It takes a totally different approach  of leveraging the human translations and learning 2:47the patterns from them and making very smart  guesses of those translations and delivering 2:53those translations. Both approaches work very,  very well by the way. We take it one notch up 3:00with the neural approach where, as in rule  based and statistical, it actually is looking 3:06at each word to get to the translations. Neural  takes it one notch up because it is actually 3:13looking at the sentence constructions to do the  translations. Now, as in any other approach, 3:19you can take a combination of these approaches and  make it a hybrid approach. So as we discussed, the 3:27traditional way makes use of the linguistic rules  as well as the dictionaries. And it goes through 3:34the supervised learning into one. Large language  models do the translations differently. They make 3:42use of the content that is already available in  different languages. We call it the large corpus 3:49of parallel text. That is the examples of the same  text in different languages like English, Spanish, 3:56Japanese and so on. And we feed it to the models.  So the large language models, as you all know, 4:02use the transformer models and they have both the  encoder and decoder capabilities. On top of that, 4:11the large languages models typically make use of  two approaches in doing the translations. Number 4:17one is the sequence-to-sequence approach.  And the sequence-to-sequence approach, 4:24you can take an input text and feed it to the  encoder. "Hello, how are you?" And the encoder 4:31goes through the text and creates the semantic  representation of the text and also captures 4:38the meaning of the text and passes it on to the  decoder. Now the decoder is capturing the semantic 4:47representation and the meaning and translating  it to the representative target language. You 4:54say "Hello, how are you?" in English, and if  your target language is going to be Spanish, 5:00you will get "Hola, cómo estás". The second  approach-- also interesting --is the attention 5:07model. The attention model is a little bit of a  lazy model compared to the sequence-to-sequence 5:13one. The attention model is focusing on the  main relevant vocabulary of the sentence. It 5:20is not going through the entire sentence. So  for example, it can pick up "hello" and "how 5:25are you" and focus on the "hola" and "cómo  estás". But it is still going to capture the 5:33meaning and the semantic representation essence  through the encoded and decoder. As you can see, 5:40the larger language models, instead of using  the linguistic rules and the dictionaries are 5:45really focusing on capturing the patterns and the  relationships between the data and translating it. 5:52It's quite obvious now that everybody wants to be  communicated in their own language, including our 5:58customers. Let's go leverage large language models  to meet them at the table and communicate in their 6:06own language. Thank you for watching. Before  you leave, please click subscribe and like.