FEED
When games engines meet live broadcast the real and
photoreal are interchangeable
Design and production tools that enable broadcasters to
create virtual objects that appear as if they’re actually in the studio have
been available for a few years, but improvements in fidelity, camera tracking
and, notably, the fusion of photoreal games engine renders with live footage
has seen Augmented Reality go mainstream.
Miguel Churruca, the marketing and communications director
at 3D graphics systems developer Brainstorm, explains, “AR is a very useful way
of providing in-context information and enhancing the live images while
improving and simplifying the storytelling. Examples of this can be found in
elections nights, plus entertainment and sports events, where a huge amount of
data must be shown in-context and in a format that is understandable and
appealing to the audience.”
Virtual studios typically broadcast from a green screen set.
AR comes into play where there is a physically-built set in the foreground, and
augmented graphics and props are placed behind the camera. Some scenarios could
have no physical props with everything behind and in front of the presenter being
graphics.
“Apart from the quality of the graphics and backgrounds, the
most important challenge is the integration and continuity of the whole scene,”
says Churruca. “Having tracked cameras, remote locations and graphics moving
accordingly, perfect integration, perspective matching and full broadcast
continuity are paramount to provide the audience with a perfect viewing
experience of AR graphics.”
The introduction of games engines, such as Epic’s Unreal
Engine or Unity has brought photorealism into the mix. Originally designed to
quickly render polygons, textures and lighting in video games, these engines
can seriously improve the graphics, animation, physics of conventional
broadcast character generators and graphics packages, but it’s complicated because
of the constraints of real-time rendering and operation.
That, though, has been cracked.
Virtual/real live music show
Last year a dragon made a virtual appearance as singer Jay
Chou performed at the opening ceremony for the League of Legends final at the
famous Birdsnest Stadium. This year, the esports’ developer Riot Games wanted
to go one better and unveil a virtual pop group singing live with their real
world counterparts.
It’s a bit like what Gorillaz and Jamie Hewlett have been up
to for years, only this isn’t as tongue in cheek.
K/DA, is a virtual girl group consisting of four of the most
popular characters in League of Legends. In reality, their vocals are provided
by a cross-continental line-up of accomplished music stars: US-based Madison
Beer and Jaira Burns, and Miyeon and Soyeon from actual K-pop girl group
(G)I-DLE.
Riot tapped Oslo-based The Future Group (TFG) to bring them
to life at November’s World Championship Finals opening ceremony from South
Korea’s Munhak stadium.
Riot Games provided art direction and a base CG imagery
model for K/DA’s lead member Ahri, and TFG transformed Ahri into her popstar
counterpart and originated models for her three group mates, based on concept
art designs from Riot.
LA postproduction house Digital Domain team supplied the
motion capture data for the characters, TFG completed their facial expressions,
hair, clothing, and realistic texturing and lighting.
Lawrence Jones, Executive Creative Director, TFG. “We didn’t
want to make the characters too photorealistic. They needed to be stylised yet
believable. That means getting them to track to camera and having the
reflections and shadows change realistically with the environment. It also
meant their interaction with the real pop stars had to look convincing.”
All the animation and the directing cuts were pre-planned,
pre-visualised and entirely driven by timecode to sync with the music.
“Frontier is our version of Unreal which we have made for
broadcast and realtime compositing. It enables us to synchronise the graphics
with the live signal frame accurately. It drove monitors in the stadium (for
fans to view the virtual event live) and it drove real world lighting and
pyrotechnics.”
Three cameras were used all with tracking data supplied by
Stype including a Steadicam, a PTZ cam and a camera on a 40ft jib.
“This methodology is fantastic for narrative driven AR
experiences and especially for elevating live music events,” he says. “The most
challenging aspect of AR is executing it for broadcast. Broadcast has such a
high-quality visual threshold that the technology has to be perfect. Any glitch
in the video not correlating to the CG may be fine for Pokemon on a phone but
will be a showstopper for broadcast.”
Over 200 million viewers watched the event on Twitch and
YouTube.
“The energy that these visuals created among the crowd live
in the stadium was amazing,” he adds. “Being able to see these characters in
the real world is awesome.”
WWE WrestleMania
The World Wrestling Entertainment (WWE) enhanced live stream
production of its annual WrestleMania pro wrestling event last April with
Augmented Reality content produced by WWE using Brainstorm’s InfinitySet
technology.
The overall graphic design was intended to be
indistinguishable from the live event staging at the Mercedes-Benz Superdome in
New Orleans
The graphics package included player avatars, logos,
refractions and virtual lighting and substantial amounts of glass and other
semi-transparent as well as reflective materials.
Using InfinitySet 3, WWE created a wide range of different
content, from on-camera wrap arounds to be inserted into long format shows, to
short self-contained pieces. Especially useful was a depth of
field/focus feature, and the ability to adjust the virtual contact shadows and
reflections to achieve realistic results.
Crucial to the Madrid-based firm’s technology is the
integration of Unreal Engine with the Brainstorm eStudio render engine. This
allows InfinitySet 3 (the brand name for Brainstorm’s top-end AR package) to
combine the high-quality scene rendering of Unreal with the graphics,
typography and external data management of eStudio and allows full control of
parameters such as 3D motion graphics, lower-thirds, tickers, and CG
The Virtual Studio in use by the WWE includes three cameras
with an InfinitySet Player renderer per camera with Unreal Engine plugins, all
controlled via a touchscreen. Chroma keying is by Blackmagic Ultimatte 12.
For receiving the live video signal, InfinitySet is
integrated with three Ross Furio robotics on curved rails, two of them on the
same track with collision detection.
WWE also use Brainstorm’s AR Studio, a compact version which
relies on a single camera on a jib with Mo-Sys StarTracker. There’s
a portable AR system too designed to be a plug and play option for on the road
events.
The tech played a role in creating the “hyper-realistic” 4K
AR elements that were broadcast as part of the opening ceremony of the 2018
Winter Olympic Games in PyeongChang.
The AR components included a dome made of stars and virtual
fireworks that were synchronised and matched with the real event footage and
inserted into the live signal for broadcast.
As with the WWE, Brainstorm combined the render engine
graphics of its eStudio virtual studio product with content from Unreal Engine
within InfinitySet. The setup also included two Ncam-tracked cameras and a
SpyderCam for tracked shots around and above the stadium.
InfinitySet 3 also comes with
a VirtualGate feature which allows for the integration of the
presenter not only in the virtual set but also inside additional
content within it, so the talent in the virtual world can be
‘teletransported’ to any video with full broadcast continuity.
ESPN
Last month, ESPN introduced AR to refresh presentation of
its long running sports discussion show, Around the Horn (ATH).
The format is in the style of a panel game and involves
sports pundits located all over the U.S talking with show host Tony Reali via
video conference link.
The new virtual studio environment, created by the DCTI
Technology Group using Vizrt graphics and Mo-sys camera tracking, gives the
illusion that the panellists are in the studio with Reali. Viz Virtual Studio
software can manage the tracking data coming in for any tracking system, works
in tandem with Viz Engine for rendering,
“Augmented reality is something we’ve wanted to try for
years,” Reali told Forbes. “The technology of this studio will take the
video-game element of Around the Horn to the next level while also enhancing
the debate and interplay of our panel.”
Sky Sports
Since the beginning of this season’s EPL Sky Sports has been
using a mobile AR studio for match presentation on its Super Sunday live
double-header and Saturday lunchtime live matches.
Sky Sports has worked with AR at its studio base in Osterley
for some time but moving into grounds is aimed to improve the output
aesthetically, editorially and analytically. A green screen is rigged and
de-rigged at each ground inside a standard matchday 5m x 5m presentation box
with a real window open to the pitch. Camera tracking for the AR studio is done
using Stype’s RedSpy with keying on Blackmagic Design Ultimatte 12.
Environment rendering is in Unreal 4 while editorial graphics are
produced using Vizrt and an NCam plugin.
Sky is exploring displaying AR team formations using player
avatars and displaying formations on the floor of the studio, having them
appear in front of the pundits.
Sky Sports head of football Gary Hughes says the set
initially looked “very CGI” and “not very real” but it’s improved a lot.
“With the amount of CGI and video games out there, people
can easily tell what is real and what is not,” he says. “If there is any
mystique to it, and people are asking if it is real or not, then I think you’ve
done the right thing with AR.”
Spanish sports
Spanish sports shows have taken to AR like a duck to water.
Specifically, multiple shows have been using systems and designs from Lisbon’s
wTVision, which is part of Mediapro the Spanish media group.
In a collaboration with Vàlencia Imagina
Televisió and the TV channel À Punt, wTVision manages all
virtual graphics for the live shows Tot Futbol and Tot Esport.
The project combines wTVision’s Studio CG and R³ Space
Engine (real-time 3D graphics engine). Augmented Reality graphics are
generated with camera tracking via Stype.
For Movistar+ shows like Noche de Champions
wTVision has created an AR ceiling with virtual video walls. Its Studio CG
product controls all the graphics. For this project, wTVision uses three
cameras tracked by Redspy with Viz Studio Manager and three Vizrt engines with
the AR output covering the ceiling of the real set and the virtual fourth wall.
The same solution is being used for the show Viva La
Liga, in a collaboration with La Liga TV International.
AR is also being used for analytical overlay during a live
soccer match. Launched in August, wTVision’s, AR³ Football is able to generate
AR graphics for analysis of offside lines and freekick distances from multiple
camera angles. The technology allows a director to switch cameras, the system
auto-recalibrates the AR and it takes a couple of seconds to have it on air.