Friday, 13 July 2018

Artificial intelligence gets creative


IBC
From newsrooms to sports and TV production, AI is already being used to speed up and boost the creative process. But could it replace humans altogether?

We like to think of TV as a creative industry but it is already highly automated. Once an idea is conceived, there are many strict, even formulaic, tasks from robotic camera moves to compliance checks that are better performed by machine. The clue is in the term ‘industry’.
Few corners of any industry will be left untouched by machine learning (ML) and artificial intelligence (AI) but in media the highly programmatic activities surrounding the recording, processing and transmitting of information make the appliance of cognitive science particularly suitable.
Inevitably this throws up practical questions about the extent to which machines can or should supplant human jobs and broader philosophical ones about the nature of creativity itself.
One argument is that as these technologies are introduced, creatives will be free to spend more of their time actually being creative. The other follows the trend to its logical conclusion: that eventually the autonomous creation of content will be indistinguishable from the real thing.
Newsrooms embrace AI
No part of the industry has adopted AI more than the newsroom. A recent report from Reuters revealed that almost three-quarters of news publishers surveyed were already using some kind of AI.
Algorithms are being used to optimise marketing, to automate fact-checking, and to speed up tagging and metadata.
News organisations know they have to do more with less; they have to find ways of making journalists more productive without leading to burn out. Intelligent automation is one way to square this circle.
As one example, the Press Association (PA) in the UK has been working with Urbs Media to deliver hundreds of semi-automated stories for local newspaper clients. A journalist would find a story using one or more publicly available datasets (e.g NHS and population data), then write a generic story that is versioned automatically by a computer to create bespoke versions for myriad local publications.
Another potential use of AI in news is for reporting of live events. Dreamwriter, an automated newswriting programme developed by Chinese tech giant Tencent, uses speech to text software to turn conference speeches into stories. It apparently churns out 2500 stories daily.
The BBC says it does not currently publish stories generated by ‘robo-journalists’. But the corporation’s News Labs research team has worked on tools to automate “the transcription of interviews and identification of unusual trends in public data,” according to the division’s editor Robert McKenzie.
Automated workflows are similarly being applied to assemble video-led news packages. RTVE, the national Spanish TV channel, is testing an AI programme from media asset management (MAM) vendor VSN which can assemble video clips from an archive search into a timeline within seconds.
Broadcast news production is a particular focus of MAM software companies who assert that artificial intelligence will augment rather than replace human editorial oversight.
“Most of the value lies in combining multiple AI engines with human intelligence to achieve things that couldn’t be done before,” says Dalet director of product strategy Kevin Savina.
In the newsroom, for example, cognitive services are being used to auto tag thousands of pieces of archive content. In order to create a richer data set for search, other AIs for facial and object recognition might also be applied. Yet another AI might then applied to this material to recommend results in a form – a timeline for example – that can be of most value to the story the journalist has at hand.
 “Even then you need the filter of the journalist’s editorial judgement,” says Savina. “Only when you have a combination of multiple AI tools plus interaction with staff selecting and deselecting content of interest, plus a continuous feedback loop, will the AI be trained.”
“Cataloguing and search are very systematic tasks,” agrees VSN product manager Toni Vilalta. “Using AI means journalists can be more focussed on the creative editing decisions.”
However, AI is often expensive to use and few media organisations are set up to use it. What’s more, AI outcomes can still not be fully trusted. But for companies thinking of adding an AI to their workforce, there is also the issue of change management to be considered.
“An analogy for putting in place an AI is that of hiring a new employee,” says Savina. “You need to teach it how to work in the organisation and colleagues need to learn to adapt to it accordingly.”
Sports production automation
AI is also likely to have big future in sports production. Unmanned multi-camera systems covering an entire field of play have been on the market for years. Advanced auto-production algorithms, such as those used by Pixellot, track the flow of play, identify highlights, create replays and insert ads without human intervention.
However, Pixellot can now create highlight reels automatically, moving the system “one step closer to end-to-end television-like production” according to CEO Alon Werber.
Elsewhere in the industry, Tedial and TVU Networks among others are seeking to link the metadata captured live at a game with relevant content archives, using AI to automate the packaging of content into different lengths and for different platforms. Ultimately, such tools could be combined with information about the preferences and devices of individual viewers to auto-create and distribute personalised content – something which is not humanly possible today.
However, development is in its early days and a long way from being applied to flagship shows like Match of the Day.
“The quality drops off dramatically if there’s any kind of judgement [required of an AI],” says video search specialist Axle AI’s CEO Sam Bogoch. “If you’re prepared to bang out a thousand different versions out of the same event, then AI is probably more cost effective than a human. If you’re trying to do that at broadcast quality it would be foolhardy to think a machine will do any way near the job of a professional.”
That’s for now, but inevitably as AIs learn from more and more data, it follows that more and more of the production will be automated.
“It could well be that the editors of the future are really managers of AI bots,” suggests Bogoch. “If broadcasting is fundamentally changing from producing one definitive edit for millions of viewers to narrowcasting where we’re slicing and dicing an infinite number of permutations, then AI has a huge role to play. In that scenario the role of the editor would be to shape narrative and let the bots do the work.”
BBC explores AI production
This is a scenario which BBC R&D is toying with. “We need to use AI to shape world-class public service, and we need to use world-class public service to shape AI,” stressed BBC Chief Technology and Product Officer Matthew Postgate recently.
In production, the BBC is exploring how good machines can be at choosing the right shots within an edit in direct comparison with a human.
“The goal is to automate more of the process to cover more content rather than reduce the need for human producers,” BBC R&D emphasises in a blog post.
It has developed a system that automatically frames and cuts coverage from high-resolution crops of UHD camera feeds to deliver an edited video package. The system can be manually tweaked to change the way it puts the content together – for instance by adjusting the frequency with which it cuts between different crops.
So positive have the tests been that the BBC thinks the prototype may already generate packages “close enough in quality to that produced by a human editor to be usable - at least for simple events”.
Beyond this, the BBC is exploring how AI can perform a draft assembly job for editing more complex recorded programming. It acknowledges craft editing as a “deeply creative role” but suggests an editor’s first task usually involves finding good shots from rushes.
“Sorting through assets to find good shots isn’t the best use of the editor’s time - or the fun, creative part of their job,” says the BBC. “AI could help automate this for them.”
A commercial application using speech to text in order to index rushes is already offered by Lumberjack. Its system has been used on Danish semi-scripted series Klassen where the shoot ratios are typically high.
“I see Lumberjack as more of a producer’s and assistant editor’s tool to take the grunt work out of production,” says inventor Phil Hodgetts. “When production time is so limited we need all the help we can get for creative work.”
The AI commissioner
Internet-first players like Netflix have embedded data mining into their business to use information about audience preferences to commission original programming.
UK-based Epagogix helps Hollywood studios to identify “likely successes and probable turkeys” by running its algorithms over scripts. It says it helps risk assess potential investment “by delivering accurate predictive analysis of the Box Office value of individual film scripts.”
A related BBC experiment in this area explored how AI/ML could be used to mine the extensive BBC archive and to help schedule content targeted at consumers of high-brow channel BBC4.
“The AI liberates content we may not otherwise have come across in the BBC archive,” explained BBC4 editor Cassian Harrison, speaking at an RTS event in May. He also stressed [in a follow-up BBC blog post] that: “Humans have nuance, taste and judgement. These are qualities no algorithm or machine can replace. This is all about looking at whether technology can help them even further.”
Meanwhile, AI firm Novamente is creating a “knowledge engine” with the University of Southern California to analyse audience sentiment (a viewer’s emotional connection) across TV scripts.
“The aim is to link together scene-level attributes of narrative with character attributes and learn how they resonate or not with audiences,” explains Novamente CEO Yves Bergquist.
In the long term, he forecasts, “You’ll see lots of mass-produced content that’s extremely automated; higher level content will be less automated.”
AI with imagination
Netflix has also used AI to restore Orson Welles’ unfinished feature The Other Side of the Wind. Made in the 1970s but uncompleted and unreleased, the film was pieced together from rushes and scanned at 4K at Technicolor before being passed through an AI to improve the resolution of every pixel ahead of release this year.
“Everybody is scared of AI taking over the planet and destroying the industry but I look at this from a different perspective,” says Alex Zhukov, CTO of Video Gorillas, the company whose AI tools were used on the project.
“At the time Orson Welles was making that film you had to literally cut the negative to edit it. Now that job (neg cutter) doesn’t exist. [Likewise] all the intermediate jobs in content creation will go away but other jobs will arise to enable us to create content that we’ve not been able to before.
“AI can create art,” he insists. “The real question is when AI will create art that is indistinguishable from that of a human.”
Microsoft is part way there. It has developed an AI bot that can draw near pixel-perfect renditions of objects purely from a text description. Each image contains details that are absent from the text descriptions, indicating that this AI contains an artificial ‘imagination’.
“The model learns what humans call common sense from the training data, and it pulls on this learned notion to fill in details of images that are left to the imagination,” explains Microsoft researcher Xiadong He in a blogpost.
Microsoft envisions that, boosted by enough computer power, an AI like this will be able digitally animate a feature film from nothing other than a script.
Facebook is also teaching neural networks to automatically create images of objects, claiming that 40% of the time the results fool humans into believing they’re looking at a photograph of the real thing rather than a photoreal computer graphic.
Google is doing something similar, by teaching machines to look for familiar patterns in a photo, enhancing those patterns, and then repeating the process. The result is a kind of “machine-generated abstract art”, critiqued Wired.
The possibility of AI-powered music creation is also causing heads to turn. Facebook has devised an AI that is claimed to generate entire songs from little more than a hummed tune.
The Facebook Artificial Intelligence Research (FAIR) is part of a larger exploration by the AI community into unsupervised translation. Like Microsoft’s drawing bot, it’s a way of building in the sort of randomness or improvisation that inspires a lot of human art.
Facebook says its research “is a strong indicator of how AI could soon power human creativity”.
FAIR’s method still requires training to create different kinds of musical output - such as piano in the style of Beethoven. But the team intentionally distort the musical input by shifting it slightly out of tune to achieve a less supervised, more accidental, outcome.
“We are trying to build tools for helping anyone to create music,” says Ed Newton-Rex, founder and CEO at Soho-based outfit Jukedeck which offers royalty free machine written production music. “Our goal is not to create art.”
He says the Jukedeck platform, which can deliver a unique composition in as little as 20-seconds, is being increasingly used by video creators, brands and indie producers for social media, podcasts and games.
It was also used to compose the score for short film Zone Out, an experimental production written, performed, directed and edited by artificial intelligence in just 48 hours. The AI was fed on science fiction films to produce the script with footage from public domain films pieced together with green-screen footage of professional actors.
The line between science and art has already blurred.


No comments:

Post a Comment