Tuesday, 12 December 2017

The problem of localisation

Broadcast

The scale and complexity of producing different versions of a show often flies under the radar, yet the process can be critical for maximising international sales.
Content localisation – the process of creating compliant and culturally appropriate versions for inter national markets – is growing in complexity and volume.
The Media & Entertainment Services Alliance (MESA), which organises a Localisation Council of content owners, distributors and vendors, calculates that the market for services in Europe and the Middle East generated $2bn (£1.5bn) last year. This is expected to grow by 10% annually to hit $2.5bn (£1.9bn) by 2020.
“Localisation is the biggest single cost in our global distribution process and is, in many ways, the most complex,” says Brij Sharma, head of supply chain strategy and localisation at BBC Worldwide.
“Volumes have increased significantly over the past three years and continue to do so in emerging markets such as China and India. For example, a new language for a single channel fully localised can represent an increase in volume of around 850 new hours per year.”
In part a function of volume, complexity is also increasing as content owners look to deliver their programmes faster, in more languages, to different platforms. Localisation doesn’t merely create a new text or audio file. In order to comply with local rules, there’s a need to create new editorial versions, which are often delivered at a later date.
According to Sharma, this adds further complexity in terms of management across technical, commercial, legal, financial and creative teams, and in terms of keeping track of content versions.
“Knowing how multiple content elements for a given programme are related to each other is an increasing area of focus, pushing us to challenge the way we do things and to look at creative ways to use existing resources and new technologies.”
There is an effort among specialist suppliers, supported by MESA, to raise the profile of what has traditionally been bracketed as ‘sub and dub’ even among broadcaster and studio personnel.
A sales exec might offer a new series at Mip to a distributor and sign off on delivery in 17 languages by the end of the next week without understanding the pressure on turnaround times this creates.
“It’s not just taking a programme and slapping a language on it,” says TVT chief executive Ian Brotherston. “There are different technical standards ranging from pre- and post-watershed versions to removal of product placement.
”If you manage, say, The Walking Dead [which TVT handles for AMC], you need a downstream understanding of what is appropriate in each market to maximise worldwide sales for the content owner.”
Julian Day, business development director at Zoo Digital, says online platforms have raised the bar in terms of quality, even for catalogue sales.
“Previously, broadcasters were not paying much for subtitling and it was often the last thing they’d think about. Now, when it comes to licensing titles abroad, they find the original subtitles aren’t good enough. They have to recreate or conform them and get them up to spec.”
Dubbing accounts for the largest portion of revenues – MESA puts the figure at 70% – in this sector. “Experienced linguists and subtitlers will know how to create a lip-sync dubbing script that is as true as possible to the original,” says Day.
Acting as well as language skills are required to translate dialogue with the same nuances as the original. There’s a “significant shortage” of such skills in some markets, according to MESA, exacerbated by voice talent being tied into long-term contracts with suppliers or shows.
“It’s a major concern,” says executive director Jim Bottoms. “Given the growth in demand, content producers and their supply partners are keen to expand the talent pool to avoid serious capacity shortages.”
Zoo Digital has tried to address this with an online cloud-based service that aims to give casting directors access to more voice artists globally. Live video direction of dubbing sessions and real-time collaborative script review is also carried out online.
The method of translation is key in versioning. Whether to meet local language needs and expectations via voice over, lip-sync dubbing or subtitling is based on local cultural norms and regulations.
Operator brand identity may necessitate re-scripting for voiceovers. Some channels rescript and re-voice the narration to match the channel personality and house style – even in the same language.
“For example, two of our clients broadcast Ice Road Truckers, a show about drivers who take on some of the most dangerous frozen routes in the world,” says Andy Steele, vice-president of operations at TVT.
“One wants just the original US version – Americanisms and all – while the other wants a distinct UK version, with the narration done in a British accent, using terms the local audience would use.”
Another aspect is ensuring content is fit for the target platform. This includes viewing it then marking changes based on factors such as duration, what’s going on in each scene, age rating and local regulations, which can cover sex, violence, bad language, product placement and religion.
“The process must also ensure regulatory compliance for audio loudness and flash patterns or certain instances of varying luminance that can trigger epilepsy,” says Brotherston.
“All these guidelines need to be codified for each piece of content, depending on country and platform of transmission, along with the broadcasters’ own brand values. Time of day is also important.”
A+E Networks’ docudrama required compliance and versioning for European markets including the UK, the Netherlands, Poland, Russia and Croatia.
A key part of the brief for TVT was to create softer versions for daytime slots while maintaining the harder-hitting versions for a post-watershed audience.
This meant dealing with the extreme violence and gore depicted in many parts of the series – which is intended to reflect historical accuracy in the dramatisation sections – while remaining sensitive to the tension, drama and flow of the storytelling.
Special care was taken to deal with violence against children, which is depicted in several scenes in Barbarians Rising. In preparing the daytime versions, TVT editors made cuts just before or in the midst of very graphic scenes – for instance, retaining a sword combat visual right up to the moment a blade has struck a body, but before the spray of blood.
While photosensitive epilepsy (PSE) triggers usually mean strobe lighting or flash photography, in this case, variations between light and shade, and the glint of steel weapons, had to be monitored and dealt with.
Failure to pick up on even seemingly small taboos and rules can have a high price. TVT gives examples such as allowing use of the word ‘prick’ pre-watershed in Ireland, where research shows it is deemed more offensive than ‘fuck’; showing a tattoo on screen in South Korea; using ‘God’ as an exclamation in Poland; allowing US-style product placement in the more restrictive European market; or depicting conspicuous drug use for a broadcast in Japan.
Once the issues are identified, craft editors adjust the content while retaining production values. “The content needs to remain in keeping with the programme aesthetic and work for all the audio and subtitle language translations,” says Brotherston. “This requires continuity of skills and understanding of the content to make sure that edits don’t damage the storyline.”
Masks, cover shots (programme b-roll) and colour correction can be used to retain the integrity of content and voiceover.
Additional edits of on-screen graphics and captions to localise them for non-English-speaking markets may be required, as well as a credit for the localisation performers.
Increasingly, platforms are requesting ‘enhanced versions’, with extra scenes or graphics. “Linear viewers might be attracted to the enhanced OTT versions of their favourite show, giving content owners a second bite at the cherry,” says Brotherston.
“This generates additional complexity in version IDs and rights management, and can lead to content having a much longer tail [requiring it to go into longer-term archives with multiple versions that require additional storage systems].”
While compliance editing will generally involve removal of content, some firms take picture localisation to extraordinary lengths. Pixar Animation created 7,482 new shots for international versions of Cars 3, making artistic changes to frames or whole scenes.
For example, substituting language-appropriate text in newspaper headlines to ensure a joke or plot line is not missed.
For a scene in Inside Out featuring a child refusing to eat broccoli, Pixar changed the vegetable to peppers for distribution in Japan. “Everyone eats broccoli in Japan. But they hate green peppers,” says Eric Pearson, home entertainment supervisor at Pixar.
Underpinning the whole chain is a minutely detailed tracking and reporting process. Once manually administered on spreadsheets, it’s now increasingly part of a media asset management (MAM) system. For increased flexibility and ease of access, vendors like TVT and Zoo Digital have also centralised their MAM in the cloud.
With such complexity, it would be a logical step to introduce automation to streamline services. Indeed, some proponents of machine learning suggest that incumbent service providers are reluctant to introduce new technology for fear of damaging their revenue.
“For the same amount of money you spend on a transcription house, you can use machine learning today to deliver speech to text, localisation and an almost infinite other variety of data tasks,” says Josh Wiggins, chief commercial officer at metadata solutions provider GreyMeta.
“At the same time, you end up with richer content because of the comprehensive and detailed metadata that results from applying artificial intelligence.”
However, automatic transcription is still a long way from converting a soundtrack into accurate text in real-time. According to subtitling specialist Screen Systems, the challenges include sound effects and background music, different voices and accents – and switching between them – unexpected vocabulary and punctuation.
War For The Planet Of The Apes
Digital Media Services (DMS) was responsible for creating 500 promotional assets, including localised TV and social media spots, digital billboards and a variety of online content to support worldwide distribution of Fox’s theatrical release War For The Planet Of The Apes.
To work on the assets, DMS received master files from the Fox Media Cloud system and uploaded them to its own platform, Panther Online.
“Requirements vary considerably, with some countries requiring one or two traditional spots to support their marketing, while other countries order 20 and even up to 40 bespoke variations,” says Michael Petch, director of customer services.
Many of the briefs only specified the text needed for the call to action details per asset, leaving DMS responsible for incorporating local specifics such as voiceover, age ratings, URL and social media tags.
Voice artists recorded the localised voiceovers at ADR facilities. In most instances, content was also dubbed and subtitled, while graphics cards were created to match the quality of the studio’s masters. Locallanguage-speaking customer service co-ordinators checked every asset before sending to Fox for sign-off.
A glaring illustration of this occurred when Newcastle United was referred to as ‘the black and white scum’ in an automated mistranslation of a presenter’s words during Match Of The Day 2 in September.
“AI translation technologies have improved greatly in the past few years… [but] we do not believe that the toolsets on offer today can replace human input when it comes to adapting long-form content such as drama,” says Sharma. “In our experience to date, it’s often quicker to start a translation from scratch than to start from whatever base a machine might create.”
Zoo Digital’s Day suggests that a good translator can accept or edit suggestions from a machine, “but it’s a supporting tool at this stage. Ultimately, we still rely on a human check.”
Pinewood, which creates formats for theatrical content, is investigating the use of language-specific speech recognition as an extra step in its QC workflows, but says reliability is an issue – especially when it needs the tools to recognise 44 languages.
TVT has speech and image recognition software in beta. “We are starting to use AI from the EDLs and metadata we’ve previously created,” says Brotherston.
“Image recognition might be useful in a future editorial process – for example, if there’s a real-life plane crash we could search programming for incidents of plane crashes and alert broadcasters, or reedit and redistribute the video as needs be.”
Benefits from automation are most likely to be derived in workflow management and orchestration, to reduce duplication.
“Each piece of content is in effect a project, with its own sets of rules around which version to create, what voices to cast, approvals, merchandising, rights and so on,” says Sharma. “This is highly labour intensive, so a good candidate for automation.”


No comments:

Post a Comment