Tuesday, March 11, 2025

New Research Explores How to Boost Large Language Models’ Multilingual Performance

In a February 20, 2025 paper, researchers Danni Liu and Jan Niehues from the Karlsruhe Institute of Technology proposed a way to improve how large language models (LLMs) perform across different languages.

New Research Explores How to Boost Large Language Models’ Multilingual Performance

They explained that LLMs like Llama 3 and Qwen 2.5, show strong performance in tasks like machine translation (MT) but often struggle with low-resource languages due to limited available data. Current fine-tuning processes do not effectively bridge the performance gaps across diverse languages, making it difficult for models to generalize effectively beyond high-resource settings.

The researchers focus on leveraging the middle layers of LLMs to enable better cross-lingual transfer across multiple tasks, including MT.

LLMs consist of multiple layers. The early (or bottom) layers handle basic patterns like individual words, while the final (or top) layers focus on producing a response. The middle layers play a key role in capturing the deeper meaning of sentences and how different words relate to each other.

Liu and Niehues found that these middle layers “exhibit the strongest potential for cross-lingual alignment,” meaning they help ensure that words and phrases with similar meanings are represented in a comparable way across languages. Strengthening this alignment helps the model transfer knowledge between languages more effectively.

By extracting embeddings (i.e., representations of text in vector form) from the model’s middle layers and adjusting them so that equivalent concepts are closer together across languages, the researchers aim to improve the model’s ability to understand and generate text in multiple languages.

Alternating Training Strategy

Rather than relying solely on task-specific fine-tuning, they introduce an “alternating training strategy” that switches between task-specific fine-tuning (e.g., for translation) and alignment training. Specifically, an additional step — middle-layer alignment — is integrated into the fine-tuning process to ensure that the representations learned in one language are more transferable to others.

Tests showed that this method improved both translation accuracy and performance across both high-resource and low-resource languages. Liu and Niehues noted that the models were also able to generalize their performance to languages not included in the initial alignment training.

One significant advantage of this method is its modular nature: “task-specific and alignment modules trained separately can be combined post-hoc to improve transfer performance” without requiring full model retraining. This makes it possible to improve existing models with enhanced multilingual capabilities while avoiding the high computational costs of retraining from scratch.

Additionally, this approach is faster and more cost-effective since “a few hundreds of parallel sentences as alignment data are sufficient.”

The researchers have made the code available on GitHub, allowing others to implement and test their approach.

Wednesday, March 5, 2025

CEOs React as Trump Declares English the Sole Official Language of the US

In response to President Trump’s executive order designating English as the official language of the US, SlatorPod gathered Dipak Patel, CEO of GLOBO, and Peter Argondizzo, CEO of Argo Translation, to discuss its implications for the US language industry.

The discussion highlighted that language access has long been a key part of US policy, particularly in healthcare, education, and legal services. Dipak pointed out that eliminating language services would create inefficiencies, making it harder for medical professionals to provide accurate care.

CEOs React as Trump Declares English the Sole Official Language of the US

Peter emphasized the broader uncertainty the order creates as many organizations rely on federal funding for language services, and a lack of clear guidance could lead to reduced support in schools, courts, and public services.

Both CEOs acknowledged that while this order presents challenges, the language services industry has historically adapted to change. Dipak suggested that financial pressures may push the industry to innovate, potentially accelerating AI adoption in interpreting.

CEOs React as Trump Declares English the Sole Official Language of the US

While the long-term impact remains unclear, the consensus is that language access will persist — driven by business needs and market demand.

Monday, March 3, 2025

Trump Makes English the Only Official Language of the US, Revokes Clinton Language Access Order

A new Executive Order published on March 1, 2025, by the US White House designates English as the only official language of the United States and revokes Executive Order 13166, “Improving Access to Services for Persons with Limited English Proficiency” (LEP), signed in 2000 during the Clinton administration.

Trump Makes English the Only Official Language of the US, Revokes Clinton Language Access Order

The new order states that it is in the country’s best interest for the federal government to designate only one official language. It also argues that “Establishing English as the official language will not only streamline communication but also reinforce shared national values, and create a more cohesive and efficient society.”

The order also specifies that agency heads, defined as “the highest-ranking official of an agency,” are “not required to amend, remove, or otherwise stop production of documents, products, or other services prepared or offered in languages other than English.”

While the text states that this Executive Order does not create new legal rights or benefits, and that it should be implemented subject to existing laws, the Trump administration has instructed the Attorney General to update policy guidance in line with the new official language designation.

A Shift in Language Access Policy?

Revoked order 13166 required federal agencies to examine their services and identify language assistance needs for LEP populations, mandating the development and implementation of systems to meet those needs.

Language access guidelines under 13166 aligned with Title VI of the Civil Rights Act of 1964, a federal law that prohibits discrimination on the basis of multiple criteria, including national origin (and by association, language). It expressly mandated recipients of federal financial assistance, such as healthcare providers and educational institutions, to offer language assistance to LEP applicants and beneficiaries.

In 2003, during George W. Bush’s administration, the LEP.gov website was launched to help federal agencies, advocates, and individuals access related information and services. And for a couple of decades, including during the previous Trump term (2017–2021), the US Department of Justice (DOJ) provided policy guidance to assist federal agencies for compliance.

Still standing, and not mentioned in the latest Executive Order, is the one formulated during the Biden administration: Executive Order 14091 (2023), “Further Advancing Racial Equity and Support for Underserved Communities Through the Federal Government.”

Executive Order 14091 instructs agencies to consider opportunities to “improve accessibility for people with disabilities and improve language access services to ensure that all communities can engage with agencies’ respective civil rights offices.”

At the time of writing, both the LEP.gov site and the US Commission on Civil Rights’ Limited English Proficiency Plan remain online.

Wednesday, February 12, 2025

Researchers Present DOLFIN, a New Test Set for AI Translation for Financial Content

On February 5, 2025, a team of researchers from Grenoble Alpes University and Lingua Custodia, a France-based company specializing in AI and natural language processing (NLP) for the finance sector, introduced DOLFIN, a new test set designed to evaluate document-level machine translation (MT) in the financial domain.


The researchers say that the financial domain presents unique challenges for MT due to its reliance on precise terminology and strict formatting rules. They describe it as “an interesting use-case for MT” since key terms often shift meaning depending on context.

For example, the French word couverture means blanket in a general setting but hedge in financial texts. Such nuances are difficult to capture without larger translation units.

Despite strong research interest in document-level MT, specialized test sets remain scarce, the researchers note. Most datasets focus on general topics rather than domains such as legal and financial translation.

Given that many financial documents “contain an explicit definition of terms used for the mentioned entities that must be respected throughout the document,” they argue that document-level evaluation is essential. 

DOLFIN allows researchers to assess how well MT models translate longer texts while maintaining context. 

Unlike traditional test sets that rely on sentence-level alignment, DOLFIN structures data into aligned sections, enabling the evaluation of broader linguistic challenges, such as information reorganization, terminology consistency, and formatting accuracy.

Context-Sensitive

To build the dataset, they sourced parallel documents from Fundinfo, a provider of investment fund data, and extracted and aligned financial sections rather than individual sentences. The dataset covers English-French, English-German, English-Spanish, English-Italian, and French-Spanish, with an average of 1,950 segments per language pair. 

The goal, according to the researchers, was to develop “a test set rich in context-sensitive phenomena to challenge MT models.”

To assess the usefulness of DOLFIN, the researchers evaluated large language models (LLMs) including GPT-4o, Llama-3-70b, and their smaller counterparts. They tested these models in two settings: translating sentence by sentence versus translating full document sections. 

They found that DOLFIN effectively distinguishes between context-aware and context-agnostic models, while also exposing model weaknesses in financial translation.

Larger models benefited from more context, producing more accurate and consistent translations, while smaller models often struggled. “For some segments, the generation enters a downhill, and with every token, the model’s predictions get worse,” the researchers observed, describing how smaller LLMs failed to maintain coherence over longer passages.

DOLFIN also reveals persistent weaknesses in financial MT, particularly in formatting and terminology consistency. Many models failed to properly localize currency formats, defaulting to English-style notation instead of adapting to European conventions.

The dataset is publicly available on Hugging Face.

Authors: Mariam Nakhlé, Marco Dinarelli, Raheel Qader, Emmanuelle Esperança-Rodier, and Hervé Blanchon

Monday, February 10, 2025

Off-Screen Drama Pits AI Dubbing Against French Voice Actors

How do US actor Sylvester Stallone, France’s minister for gender equality Aurore Bergé, and a multilingual/multibillion voice AI company collide in a tense drama? Since early January 2024, multiple online media sources have highlighted a clash that began with news that “Armor,” a film starring Stallone, would feature AI dubbing.

For 50 years, Alain Dorval was the familiar voice of Stallone in French-dubbed films, but he passed away in February 2024. Minister Bergé happens to be Dorval’s daughter. Enter ElevenLabs, which in January 2024 reached a USD 3bn valuation, and found itself at the center of a weeks-long controversy over the cloning of Dorval’s voice.

Bergé publicly opposed (article in French) the use of her father’s digitally recreated voice, despite acknowledging a prior agreement to a test. “It was just a trial run, with an agreement strictly guaranteeing that my mother and I would have final approval before any use or publication. And that nothing could be done without our consent.”

According to Variety, which has followed the story since the partnership around “Armor” between Lumiere Ventures and ElevenLabs came to light, Bergé’s move galvanized the French actors’ guild (FIA, in French). 

FIA’s representative, Jimmy Shuman, called the voice cloning attempt a “provocation” in the Variety article. That is because the union is in the midst of “negotiating agreements on limits for artificial intelligence and dubbing.”

The controversy over Stallone’s French voice underscores the potential for AI to displace voice actors, often celebrities in their own right across Europe.

ElevenLabs CEO, Mati Staniszewski, told Variety that “Recreating Alain Dorval’s voice is a chance to show how technology can honor tradition while creating new possibilities in film production.” 

Like their US counterparts after a few notable actions, voice-over artists in several European countries are taking a proactive stance through their unions, including AI clauses in their contracts to restrict AI voice use to specific projects or outright banning work for studios that do not offer adequate protections.

Per the latest Variety article on the subject, voice actor Michel Vigné will be the voice of Stallone for the French release. According to IMDB, Vigné has already voiced Stallone in French in the past.

The larger issue remains: the film industry acknowledges that AI voice cloning technology is rapidly advancing and the drama around Armor’s French dubs serves as a symbol of things to come in Europe and beyond.

One decision that perhaps many voice actors will need to grapple with is whether they want their voice to be immortalized with AI or simply be replaced by it or by another actor.

Thursday, January 30, 2025

Why Interpreting Remains a Growth Market with Boostlingo CEO Bryan Forrester

Bryan Forrester, Co-founder and CEO of Boostlingo, returns to SlatorPod for round 2 to talk about the company’s growth, the US interpreting market, and the evolving role of AI.

Bryan shares how the company has tripled in size since he last appeared on the pod, driven by strategic acquisitions, including VoiceBoxer and Interpreter Intelligence, and a rebranding effort to unify its product portfolio.

Bryan explains how Boostlingo balances innovation with practicality, ensuring that new features align with customer needs. He highlights the company’s three-pronged strategy: retaining existing customers, enabling growth, and making long-term bets on emerging trends.

While tools like real-time captions and transcription enhance efficiency, Bryan stresses that AI alone cannot replace human interpreters in complex industries like healthcare. He highlights privacy, compliance, and the nuanced expertise of human interpreters as critical factors, positioning AI as a supportive tool rather than a replacement.

https://youtu.be/fMNcJ5EV2zk

Bryan discusses market dynamics and regulatory changes, including how those under the new US administration could influence language access demand, particularly in areas like healthcare and public services. 

He describes Boostlingo’s strategy of leveraging third-party AI models, optimizing them with proprietary data, and rigorously testing to ensure quality and reliability. Looking ahead, Boostlingo plans to expand internationally and integrate AI ethically and effectively into its offerings, guided by its newly formed AI Advisory Board. 

Sunday, January 19, 2025

AI in Interpreting: Slator Pro Guide

Slator's Pro Guide: AI in Interpreting is an absolute must-read for all providers of interpreting services and solutions. Here, the authors give a quick snapshot of what the newest applications of AI and large language models (LLMs) look like in interpreting.

This Slator Pro Guide will bring you up to speed on what value AI can bring to your company and the new interpreting workflows, service models, and speech AI capabilities now available to you.

The guide covers 10 one-page, actionable case studies-thematically designed and presented as vibrant infographics drawn from research and interviews with some of the leading interpreting service providers in the industry.

The ten use cases highlight new areas of growth, innovative models for service delivery, and novel workflows in interpretation made possible by the recent developments in LLMs, speech-to-text, and speech synthesis.

We will illustrate how AI speech translation solutions are being leveraged to open up language access across corporate, government, and healthcare settings, cutting across a wide variety of settings and service delivery models.

The guide also discusses AI as an interpreter tool and co-pilot, as well as its capability to optimize operations and extract insights from interpreted interactions.

Each use case describes the underlying concept and practical implications. An adoption and value-add score is also provided to reflect the industry's current level of uptake for the application as well as the additional value that it delivers to end clients.

We explain how the technology works and offer a brief list of leading AI solution providers currently on the market.

We expand on the new opportunities and benefits that each use case presents for interpreting stakeholders and carry out an impact analysis for the interpreting sector.

We also identify key risks and limitations to watch out for, which need to be considered in the adoption process.

The guide provides a higher-level overview of the key and impactful applications that can serve as a launching pad for stakeholders to make strategic decisions about adopting AI in interpreting technology and service models.

This Pro Guide is a must-read and time-saving briefing on how AI is revolutionizing the interpretation landscape.


Language Discordance Raises Risk of Hospital Readmissions, U.S. Study Finds

  A June 2024 meta-analysis published in   BMJ Quality & Safety   was recently brought back into the spotlight by Dr. Lucy Shi, who disc...