Friday, 29 July 2016
Wednesday, 20 July 2016
Does the screen matter for live sports?
TVB Europe
As live sports pours online TVB Europe asks whether OTT streams are as good – or better – than the cable and satellite delivered TV experience.
p27
https://issuu.com/newbayeurope/docs/tvbe_july_2016_lowres/1
http://www.tvbeurope.com/does-the-screen-matter-for-live-sports/
http://www.tvbeurope.com/does-the-screen-matter-for-live-sports/
Sports rights are now a target of social media players just as much as extending the reach of sports properties onto OTT platforms is a strategy of sport franchises and broadcasters.
Broadly, traditional players need to expand into social and online to reach millennials who are deserting studio-bound linear presentations for interactive, informal, and mostly free TV anywhere experiences.
Social giants, on the other hand, are increasingly moving into premium sports because the audience and profile of these businesses rely on scale, suggests Richard Broughton, research director, Ampere Analysis. “Sports are mass market form of entertainment and capable of attracting large audiences with a high value to sponsors and advertisers.”
Examples of the former include BT's live coverage of the UEFA Champions League on YouTube; Discovery's deal with Snapchat to create a dedicated mobile channel around the Olympics (with content supplied by BuzzFeed); Sky's investment in online sports network Whistle Sports; and the use by broadcasters, including Sky and the BBC, of Facebook Live to augment linear sports programming. In the latter camp is Twitter's live stream of 10 NFL matches (which also meets the NFL's need to look beyond traditional networks).
All sides are dovetailing on internet delivery, which begs the question as to whether the quality of service between linear (satellite, cable) TV and broadband is indistinguishable.
“It can be equivalent – in all honesty I think OTT can be better in some cases – for example, OTT can do 4K now,” says Shawn Carnahan, CTO, Telestream. “The big question is what device is the viewer using and is their prime concern quality or convenience?
He adds, though, that OTT may not ever be able to achieve the very low latency of broadcast from a technical perspective. “In broadcast, each viewer has the exact same bandwidth and very low latency. I'm not sure OTT could ever achieve that same level.”
Telestream has introduced Lightspeed Live Stream to bring broadcasts and OTT together in the live space. The solution is designed to provide high quality encoding and control the amount of bandwidth available between production and distribution plus the amount available between distribution and the end user.
Neulion's EVP & Co-Founder Chris Wagner is in no doubt. “Online is better. Satellite and cable deliver at 30fps. We deliver at 4K at 60fps [Neulion streamed El Clásico’ Barcelona Real Madrid match live in 4K 60fps over the NeuLion Digital Platform to Sony 4K TVs in April].”
Neulion was signed by Eleven Sports Network to stream live (and VOD) coverage of La Liga, UEFA Euros, Formula 1 and the FA Cup Final to subscribers in Belgium, Luxembourg, Poland, Singapore and Taiwan. “Digital delivery of live video is better and looks better than cable,” says Wagner. “Satellite and cable platforms are being replaced by digitally delivered video.”
Ian Munford, Director of Product Marketing, Media Solutions, Akamai, says the industry has reached the point where good HD quality live streams can be delivered with reliability. “We have the luxury in the UK of a great HD TV service. We can easily surpass that.”
For Superbowl 50 Akamai saw a “dramatic increase in every single viewing metric” says Munford. In 2015 it counted 2.5 million viewers concurrently streaming the event live, this year it peaked at 4 million. “The average viewing time increased from 84 to 101 minutes and we saw a big jump in bit rate from a 3.5Mb average to 4.5Mb average. This tells us that there's a shift toward watching major sports using an OTT service and away from snacking to long form viewing.”
There were reports, though, that online viewers of SB50 did not receive a buffer-free experience (not necessarily to do with any of Akamai's involvement). The transition to HTTP-based streaming may have enabled OTT delivery but it inherently introduces latency.
A study by network performance analytics firm IneoQuest (conducted before SB50) found that that sports buffering inflicted rage in viewers with 2 out of 5 consumers likely to will wait only 10 seconds or less for the video to resume – or they leave the stream.
“When you're using HTTP streaming technology there can be a challenge from the camera through to the playing device,” says Munford. “Some things are not in the rights owner's control.”
One issue is the shift in bit rate, where the live stream pixelates or blurs on account of ABR. “Dramatic shifts do impact the viewer experience,” says Munford. “We're seeing quite a high abandonment rates as a result.”
The content delivery network (CDN) has a number of Media Services Live delivery technologies designed to reduce latency. Its accelerated ingest capabilities minimise the amount of time live video streams take to reach the CDN from their origination point. It uses HTTP/UDP to prevent packet loss and speed the transit of content, and make it easier to handle unpredictable peaks. It will also use multicasting, and peer-assisted delivery using WebRTC.
It also opened a Broadcast Operations Control Center at its Cambridge, Mass., headquarters, to monitor the reliability of OTT streams around major events like the Rio Olympics. “During London 2012 online traffic peaked at about a Terabits per second Tbps,” reveals Munford. “We expect peaks globally of between 15-18 Tbps during Rio to set new global records in terms of online streaming traffic.”
Akamai forecasts that 500 million viewers will soon be watching prime-time live sports online. “With 500 million online viewers, we need 1500 Tbps. Today we do 32 Tbps [at peak], so you can see the huge gap we have to bridge,” says Munford.
“Any organization looking to deliver high quality scaled events needs to plan how to deal with very large peaks of audience,” he says. “It's a bit like a power surge. Peaks can be unpredictable.”
Carnahan points out that no one would never know about latency unless there is a 'back channel' such as Twitter also providing information about a live sporting event. “Tweets from my friends may be talking about something I haven't seen yet, due to latency on my OTT feed to my device. However, it can be down as low as under 30 seconds - possibly under 15 seconds.”
Wagner counters, “You can't you compare video streaming to texting. It's like watching a match live bu listening to the radio commentary. Video quality and latency go hand in glove. If you want no latency then you'll get video quality at 800kbps.”
With 4K in particular the caveat is the last mile to the home. “When we deliver 4K we are reliant on end user bandwidth,” says Munford. “We are seeing an average 13Mbps [for 4K] and for that we need good fibre in the home. The technology is there, we can surpass [this speed] and we're confident we will continue to push those boundaries.”
Verizon claims to have reduced latency on live 4K online delivery to four milliseconds, for delivery of content including San Francisco 49ers pre-game coverage over the UltraFlix network. The industry norm, though, for streaming HD content is still only 720p.
Beyond the TV experience
Viewers will accept buffering and some pixelation due to ABR for the trade off of interactivity and anywhere viewing. “It's a bit like the transition of music from CDs to streams – the quality is down but the fact hat I can listen to it wherever I am is a bonus,” says Carlo De Marchis, Chief Product and Marketing Officer at online video sports specialist Deltatre.
Beyond a simple simulcast of the live video is the opportunity with OTT to create what De Marchis calls the “beyond TV experience”.
“In DIVA [deltatre's online platform] you can multi-angle synchronised feeds of up 12 cameras available for review after a few seconds. We have timeline markers for pausing the live stream and playback of key incidents. There is social media interaction. There will greater levels of audio choice and, in future, we will take the clean feed with no graphics and send it to a device where the user will define what graphics makes sense to them.”
For live events one of the immediate opportunities is to stream additional content (such as alternate camera angles, secondary audio, etc) in addition to the broadcast feed. “Eventually, there will be an opportunity to stream additional content that is intended for a VR environment,” says Carnahan, adding that Telestream is investigating this. “An interesting case is to imagine a crowd-sourced production where a central location could be getting feeds from mobile users. Multi-camera production sourced from the crowd - perfect for sports.”
Armed with its 10 match (non-exclusive) Thursday night NFL deal, Twitter is trying to turn a second screen experience into a first screen experience, suggests Carnahan. “Instead of watching TV and tweeting about it, it will all be on Twitter. It remains to be seen how many people will turn to Twitter to “watch TV.” It’s an experiment. The issue is the trade off between image quality for an enhanced user experience. Twitter is betting that the enhanced social experience of watching NFL football on their platform will, for some, outweigh the benefits of a traditional TV viewing experience. Twitter is not aiming to be just a second screen; they are changing the viewing experience. For some this may be worth it. Time will tell.”
Thursday, 14 July 2016
Live VR/360° Video Gets Social
Streaming Media Global
NextVR, LiveLikeVR, Greenfish Labs, and others are pushing to make live sports VR/360°viewing more social with avatars, spatial audio, and more.
Virtual reality and 360° video may provide the "best seat in the house" but it will be a fail for live sports if the experience of almost being there can't be shared. While multiple live streaming VR experiments are taking place, from the UEFA Euros to the Rio Olympics, developers are frantically trying to solve the issue of connecting the experience socially among fans. NextVR, arguably the leading producer of live VR, is on the verge of announcing a partnership with a social media platform.
"There is a prevailing view that VR is isolating but sports viewing is often shared," says Dave Cole, co-founder, NextVR. "We have a partnership with a social communications platform which will be our first foray into bringing communication with peers and friends into the virtual space."
The product is in beta test and primed to be announced ahead of a "marquee event late summer" by early September. It is likely that the initial form of social interaction will be via avatar.
"We plan to integrate APIs from gaming platforms like the PS4 into the the NextVR platform," says Cole. "It makes more sense for users to create one avatar and have the ability to port that to our platform, or other virtual spaces, than to have start from scratch each time. One idea is for users to invite friends inside a virtual lobby where they can assemble and share commentary on the experience."
Connecting friends and live experiences through VR is the long term goal of Facebook founder Mark Zukerberg. Facebook's Social VR team has demonstrated how "animated mannequins" of users/friends might look when augmented with a live view using Oculus Rift.
Other live streaming sports VR developers are working hard to offer at least the first stage of real-time social interaction.
LiveLikeVR, which has been paid to test its VR production pipe by ATP Media and Sky among others, is to launch live sharing capability later this year. It will also use avatars to represent a fellow LiveLikeVR user sitting adjacent to another in LiveLike's virtual lounge and watching a game.
"When you put on a headset and launch our app, then immediately around you is the VIP suite is CG," says Andre Lorenceau, founder and CEO of LiveLikeVR. "On the right or left of the couch is an avatar of a person—which in the early stages of this development will be a representation of a person. It will be just head and shoulders, and eventually hands, but it's not intended to be hyper-real. What really makes it feel as if they are there with you is localised audio. If you're watching a game and your friends are talking to you [via head mounted device MD mic] it will sound as if they are a couple feet away on your right. If you turn you will see them and they will sound like they are in front of you. The sound of someone talking next to you changes according to where you are looking and this makes an extremely powerful feeling of being present with someone."
One issue is to synchronise the live video stream of the game with the audio component of "friends" speaking. "We are fixing these problems," he says.
Greenish Labs, which live streamed the PIAA track and field event, college football, and Pennsylvania's ice hockey team the Hershey Bears in 360° video, is working on ways to incorporate picture-in-picture within the virtual view, plus live chat and realistic audio.
"We have a sound team working on developing our own plugin and software to make the audio sound realistic in the VR environment," says Greenfish CEO and founder Ben Duffey. "We have a 8-mic audio set up, which records spatial audio at a live event and software which translates that to different areas of the video, so that as your head moves around the video the audio will relocate accordingly. You can't just have stereo sound—when you turn your head you need to be able to hear those sounds accurately."
Gaming platform VREAL is in beta with an attempt to enable any number of viewers to experience esports in virtual reality. "The future of gaming is VR, and the future of VR is social," says CEO Todd Hooper. "The core of our technology is to re-render games in realtime on each specific machine. That enables a viewer to feel that they are inside the game and allows streamers to interact with their viewers. Streamers will be able to 'pass the mic,' interact with physical gestures, or even hand off digital items from in game to the viewers."
Hooper adds that this level of interaction is technically not possible with live action video.
NextVR Plans Expansion
Separately, NextVR has revealed its "domestic and international expansion" plans. This includes "talking with the largest live production companies on the planet" about adopting NextVR technology. These include outside broadcast production suppliers including NEP and Game Creek Video.
"The aim is to provide expertise our VR production vehicle as a reference platform for producers rather than partnering them on building trucks," says Cole. "We can showcase how live VR streams can be achieved."
The NextVR mobile facility captures 4K video from 8 cameras per rig with capacity for ten rigs. Per rig, that's 24,000 pixels horizontally and 6000 pixels vertically at 60hz. This totals 6 terabytes a second of raw data.
"A typical Netflix [HD] stream is 8Mbps, and we can deliver full 360 broadcast quality stereo video at less than that," claims Cole.
NextVR is busy with a five-year pact to test out VR with Fox Sports and also signed a deal to live stream VR music events for Live Nation. Cole says monetization is already here.
"Sponsor lift is already happening," he says. Lexus sponsored Fox Sports VR streams from the U.S. Open, for example. "We will test subscription and pay-per-view models this year. You will see a pay-per-view product [from a broadcaster] launching soon."
While NextVR's technical edge is in the compression technology first devised for broadcasting live stereoscopic 3D, its business model is in content, specifically in attracting as wide a user base as possible to its NextVR portal.
"We don't syndicate to other networks for both business and technical reasons," Cole explains. "We are building a platform with partners for consumers to come to and watch VR. We are highly incentivised to maintain content on our platform since that is how the company will be valued."
Sky Confirms August UHD Launch
Streaming Media Global
UHD service will go live on Sky Q Silver box with movies, drama, and Premier League, and could find traction where BT Sport has lagged.
As expected, pan-European pay TV broadcaster Sky has confirmed the launch of a new Ultra HD service for UK subscribers beginning August 13.
The service will go live on its Sky Q Silver receiver which was unveiled in February.
According to analyst Paolo Pescatore, director, multiplay and media at CCS Insight, the move will kickstart consumer appetite for 4K in the UK. "Though BT was the first provider to launch 4K in the UK, it has failed to see any meaningful uptake to its BT TV service," he says. "Despite a huge investment in 4K, BT has failed to build upon its early mover advantage."
Mobile operator EE, owned by BT, announced this week that BT Sport, would be available free to its mobile customers for six months in a bid to increase take-up.
The Sky UHD programme line up includes 70 movies including Spectre, drama and documentaries including the series of David Attenborough-fronted natural histories like Galapagos, which was produced by Atlantic Productions for Sky's 3D channel but shot in 4K or higher.
More significantly, a series of live sports including 124 English Premier League football matches will be shot and aired in the format. From 2017, Formula 1 motor racing will also be available for UHD badging.
It is not yet clear what Sky classifies as UHD, i.e. whether it will demand a high dynamic range finish on drama or in the live sports coverage.
Sky is among broadcasters known to be interested in introducing Dolby Atmos,, an object-based audio format, to the home. This technology was tested by Telegenic, a Sky outside broadcast supplier, for UEFA at the UEFA Euros last month.
Sky is also testing virtual reality, particularly around sports, and it would be logical to see it stream live VR from F1 circuits next year in tandem with the sport's new sponsor Heineken, which has stated its intent to shake up the sport's broadcast with VR.
BT Sport has a year's lead on its main rival and plans to up its quota of live sports for the coming season to include all EPL matches plus games from the UEFA Champions League. Virgin Media is also believed to be planning to introduce Ultra HD channels this year.
Sky's new product range is corralled into a TV Everywhere ecosystem. Sky Q, Sky Q Silver, Sky Q Mini, a Sky Q touch remote, Sky Q Hub and a Sky Q app all connect together under the term Fluid Viewing.
Set-up costs for Sky Q start at £99, and for those new to Sky the monthly cost starts at £42. The UHD programming will not cost users any further premium.
"Though the concept of fluid viewing has failed to resonate with consumers, premium content in 4K such as Premier League will drive awareness [and] appetite, and enhance their viewing experience," says Pescatore. "Sky's close relationship with content and rights owners still puts it in a far stronger position than its competitors. It comes when all providers will start to heavily promote their respective multi-play bundles ahead of the new Premier League football season. It's going to be a very busy second half of the year with Sky's entry into mobile, Vodafone's debut in TV and Virgin Media's new set-top box. Sky's latest move throws down the gauntlet to others, so let battle commence."
Hack the Future of Media
IBC
http://www.ibc.org/hot-news/hack-the-future-of-media
The world’s media is in a state of rapid flux and the rules are constantly being rewritten. When internet players are coming to TV and TV is moving to the web, defending the status quo is no longer an option. While traditional broadcast and pay TV models convulse with unprecedented disruption the best tactic may be to break convention and innovate a way to prepare for the future.
IBC meets this industry challenge head on with the return of the IBC Hackfest. Following the success of its debut last year, the event will double in size for IBC2016 with exciting new partners and prizes taking to the next level. Sponsors can set challenges for the best in international development talent at this high octane event, essentially outsourcing their R&D and encouraging outsiders to look at their brands in an un-blinkered light.
This year, the IBC Hackfest invites 100 designers, developers and entrepreneurs to engage with the theme 'Re-imaginging the Future of Cities Through Education, Entertainment and Sports'.
By tapping into proprietary API's and SDK's made available by technology companies uniquely for this event, as well as scores of publicly available APIs, developers will brainstorm ways in which we might live and work together in the smart urban net-connected environments of tomorrow.
They will draw on Open Data, the idea that information should be freely available without restrictions from copyright patents or other mechanisms of control, and work with established and emerging social media networks, artificial intelligence concepts and Internet of Things (IoT) technologies to unlock commercially oriented solutions that others have found hard to crack.
The IBC2016 event is augmented by AngelHack – the world's largest hackathon organiser – who will be promoting the Hackfest to their community of more than 98,000 developers worldwide. Participating sponsors will have the unrivalled opportunity to present a brief to teams of coders, user experience designers, hardware hackers and data scientists to achieve a specific software development goal.
The inaugural IBC Hackfest in 2015 produced some incredible results including Emoment, a means of capturing moments through emotion; Old News, a second screen engagement platform for news programmes; and Tapball, a gamification for sports in which fans would play along with friends by tapping their smart screen during a live streamed match to 'bet' on goals or other events. Partners included Twitter, Amazon Web Services, Streamzilla and Monterosa.
Held in Amsterdam over 36 hours on 10-11 September in the Diamond Lounge, the IBC Hackfest is a fun and creative space where hackers join with over 55,000 attendees at the heart of IBC2016.
For innovators looking to win high value prizes and to showcase their ideas to a hugely influential jury of broadcast industry professionals, participation at IBC Hackfest is empowering. Use your bright ideas to inspire and build something new and sponsors could invest in you and take your IP to market.
For industry companies trying to build a developer community and access uncharted expertise, the IBC Hackfest is a necessity. Don't get stuck in a corner. Think your way out of the problem at IBC, the nucleus for creative technology invention.
http://www.ibc.org/hot-news/hack-the-future-of-media
The world’s media is in a state of rapid flux and the rules are constantly being rewritten. When internet players are coming to TV and TV is moving to the web, defending the status quo is no longer an option. While traditional broadcast and pay TV models convulse with unprecedented disruption the best tactic may be to break convention and innovate a way to prepare for the future.
IBC meets this industry challenge head on with the return of the IBC Hackfest. Following the success of its debut last year, the event will double in size for IBC2016 with exciting new partners and prizes taking to the next level. Sponsors can set challenges for the best in international development talent at this high octane event, essentially outsourcing their R&D and encouraging outsiders to look at their brands in an un-blinkered light.
This year, the IBC Hackfest invites 100 designers, developers and entrepreneurs to engage with the theme 'Re-imaginging the Future of Cities Through Education, Entertainment and Sports'.
By tapping into proprietary API's and SDK's made available by technology companies uniquely for this event, as well as scores of publicly available APIs, developers will brainstorm ways in which we might live and work together in the smart urban net-connected environments of tomorrow.
They will draw on Open Data, the idea that information should be freely available without restrictions from copyright patents or other mechanisms of control, and work with established and emerging social media networks, artificial intelligence concepts and Internet of Things (IoT) technologies to unlock commercially oriented solutions that others have found hard to crack.
The IBC2016 event is augmented by AngelHack – the world's largest hackathon organiser – who will be promoting the Hackfest to their community of more than 98,000 developers worldwide. Participating sponsors will have the unrivalled opportunity to present a brief to teams of coders, user experience designers, hardware hackers and data scientists to achieve a specific software development goal.
The inaugural IBC Hackfest in 2015 produced some incredible results including Emoment, a means of capturing moments through emotion; Old News, a second screen engagement platform for news programmes; and Tapball, a gamification for sports in which fans would play along with friends by tapping their smart screen during a live streamed match to 'bet' on goals or other events. Partners included Twitter, Amazon Web Services, Streamzilla and Monterosa.
Held in Amsterdam over 36 hours on 10-11 September in the Diamond Lounge, the IBC Hackfest is a fun and creative space where hackers join with over 55,000 attendees at the heart of IBC2016.
For innovators looking to win high value prizes and to showcase their ideas to a hugely influential jury of broadcast industry professionals, participation at IBC Hackfest is empowering. Use your bright ideas to inspire and build something new and sponsors could invest in you and take your IP to market.
For industry companies trying to build a developer community and access uncharted expertise, the IBC Hackfest is a necessity. Don't get stuck in a corner. Think your way out of the problem at IBC, the nucleus for creative technology invention.
Wednesday, 13 July 2016
TV Journeys to Virtual Reality
IBC
It's still very early days for Virtual Reality (VR) and there have been predictions for its disruptive impact on everything from filmed entertainment to journalism. Although it is difficult to predict the impact, it is reasonable to expect that VR will not repeat the failure of stereo 3D.
According to Ampere analyst Andrew White, VR does not compete with standard video in the same way that 3D did since there is no way to convert 360-video to 2D while retaining the original context. VR, he suggests, should be seen as an entirely new medium, running in parallel or as a companion to TV and movies, rather than as an evolution of them.
JPMorgan Securities forecasts VR to be a $13.5bn industry by 2020 mainly comprised of hardware sales topping 89.3m units. Growing sales of consumer gear are predicated on content to watch but here too production seems to be moving at an astonishing pace.
While video games remain the big initial content draw for consumer VR, likely to be given a boost when Sony debuts Playstation VR in October, movie studios and filmmakers are extending their ambition from short marketing promos to longer form stories. A first feature length VR action movie is planned for release in 2017 by cinema motion-seat developer D-Box and The Virtual Reality Company, a producer which counts Steven Spielberg as a shareholder.
To mention just two of many significant investments in this space, VR display maker HTC recently earmarked $100m for content, and Disney invested $65m in 360-video camera maker and producer Jaunt VR.
Facebook arguably kicked off the current surge with its $2bn purchase of Oculus Rift in 2014. Since then, Facebook and Google have developed an ecosystem for VR from capture to distribution. Google will next launch Daydream, an advanced operating system for Android complete with Daydream-ready phones, and motion controllers; while Facebook has its own 17-camera rig design pending. Google is also making a cinema-style VR camera with IMAX – which is launching a number of physical VR cinemas this year.
Broadcasters have spent the year road-testing VR on everything from documentaries to talk shows with most of the development focussed on live events.
“Currently, there is a joint industry initiative to make the technology work and drive uptake by enticing customers to the platform with free content,” says Futuresource Consulting analyst Carl Hibbert. “As soon as consumer payment becomes a core component, rights will become a major issue – whether that’s sports, concerts or other types of event.”
Sponsors are already investing. Automotive brand Lexus sponsored The Open Golf produced by US VR live stream specialist NextVR for Fox Sports. NextVR plans to test subscription and single view pay-per-view models this year, mostly around live music events. “2016 is a year of audience building. We are not going to put a paywall in the way of audience aggregation,” says co-founder Dave Cole.
VR opens up new opportunities in advertising on multiple fronts. “An entire industry is growing around promotional VR experiences,” notes White. “VR offers the opportunity for brands to touch consumers in ways which were previously unthinkable. With YouTube and Facebook both offering platforms for 360-video, more conventional agencies will see new channels opening up for immersive advertisements.”
Outside of gaming and entertainment VR has a future in all manner of industries from flight simulation to architectural fly-throughs. Applications in education including teaching via virtual classrooms and providing digitized campus tours to prospective students. VR is also making exciting strides in the healthcare market. Indeed, the global AR and VR healthcare market is poised to grow at a CAGR of around 17.8% in the next five years to reach $1.45bn by 2020 [Research and Markets]. Earlier this year the first live broadcast by VR of a surgical operation was streamed from St Bart's hospital in London.
Tuesday, 12 July 2016
Sony, Ang Lee among heavyweights invited to explore Douglas Trumbull's Magi Pod cinema format
Screen Daily
EXCLUSIVE: Private demonstrations of the exhibition format to take place in the US this summer.
Director Ang Lee and Sony Pictures are among Hollywood heavyweights invited to attend private demonstrations of VFX guru Douglas Trumball’s (Blade Runner) anticipated pre-fabricated exhibition format Magi Pod.
The filmmaker, inventor and former vice chair of IMAX, has built a prototype screening room at his studio in Massachusetts and is inviting creatives, studios and exhibitors including Sony and Ang Lee to private demonstrations of the high-tech experimental format over the summer.
“We’ve spent a year building a prototype theatre and I am just at the stage to start individual screenings,” he said. “Sony and Ang Lee are invited as are a lot of important media companies to witness the only option for the future of cinema other than a million dollar projection system.”
According to the three-time Oscar winner, Sony will review the Magi Pod concept for the release of Ang Lee’s Billy Lynn’s Long Halftime Walk in November. However he feels the time frame may be too tight for Billy Lynn and may be best served by later high frame rate titles.
“I also want to invite [Avatar director] Jim Cameron and [producer] Jon Landau and Steven Speilberg,” Trumbull said.
Trumbull presented the system to Lee prior to shooting Billy Lynn, an act which Trumbull says convinced the director to make he project in the high end specification, which will pose a projection challenge if it is to screen as the director intends it to be seen.
“When Ang came here and saw [Trumbull’s 3D 4K 120fps short film UFOTOG] he became very excited about shooting Billy Lynn that way,” Trumbull said.
“One of the outcomes was that he basically told the studio that if he couldn’t shoot it this way then he wouldn’t make it at all which was a bold and brave thing to have done.
“Sony felt they had a safety net to go ahead and shoot [Billy Lynn] this way provided they had options to release the movie at any frame-rate from 120 to 60 or 48 down to 24fps.”
The studio were unavailable for comment.
Global rollout?
Trumbull’s business model anticipates a mass global rollout for the pre-built modules and the industry veteran says he needs a major cinema chain like AMC or studio like Sony to back it.
“We’re aiming to start a whole new company to develop and market it. This could be aligned with any large media company.”
Trumbull says that the Magi Pod is not only capable of playing back content in 4K, 3D and 120 frames a second - which is the ultimate technical specification for Lee’s picture - but will do so at a fraction of the cost of conventional and premium large format (PLF) theatres.
“Our prototype is unlike any movie theatre ever seen,” Trumbull says. “It is not a rectangular box and it doesn’t have a flat screen. It’s more like a holodeck, or ovoid, which envelopes the audience. This gives a giant screen experience in a relatively small space and on a modest budget.”
It is claimed that each Magi Pod would cost a sixth of the cost per seat of an “equivalent PLF experience” and less than half the build cost of a conventional theatre.
Savings are based on a pre-fabricated construction in which each ‘Pod’ would be shipped in its entirety to a site where it would replace an existing small to mid-size multiplex screen. Installation would take a week, rather than many years or months, Trumbull claims.
“There will be significant install and real estate savings since there is no need to hire an architect, or building contractor or go through local planning codes,” he explains.
“While giant screens have about 50ft of space above each person, which is not an optimal use of volumetric space, our system maximises a 20ft ceiling high capacity and a 100 degree wide field of view over just 1200 sq ft.”
In addition, the system would use Trumbull’s patented technique to deploy a single projector to show a 3D 4K 120fps picture rather than the dual projectors which have been the only way to show this high end specification to date.
The DCI standard would, though, have to be revised in order to show 4K 3D 120fps content.
“We are confident that there is no technical impediment to being able to do it,” said Trumbull. “It just requires more bandwidth, more storage media and more Terabytes of data.”
“We need to redesign the movie theatre”
The Magi Pod comprises a Christie Mirage 4K 25 projector throwing 14 ft lamberts, 32 channel surround Christe Vive audio, a hemespherically curved ‘Torus’ screen made by Stewart Filmscreen and a seating system manufactured by Irwin. Each pod will seat 70 people.
“In order to create an experience for consumers which is vastly different from the convenience of downloads or streaming to computers or TV, we need to redesign what the movie theatre could be,” he argues.
“Initiatives like Dolby Cinema (which combines laser projection with immersive audio) are a great direction to go but [installation] is hugely expensive. Magi Pods are a paradigm shift in movie experience and business.”
Trumbull pioneered the 70mm 60fps exhibition format Showscan in the 1980s and has been developing the MAGI filmmaking process for a decade.
“Solving the bottleneck in getting immersive cinema experiences to a wider audience was a completely unexpected result of what we’d been doing,” he adds.
Friday, 1 July 2016
Action Man: Profile Garrett Brown
British Cinematographer
Rocky
Balboa ascending the steps of the Philadelphia Museum of Art in
Rocky,
Danny Torrance riding his tricycle through the cavernous halls of the
Overlook Hotel in The Shining,
Imperial storm troopers rocketing between trees on the forest moon of
Endor in Star Wars: Episode VI –
Return of the Jedi.
What
do these movie moments have in common? They were all filmed on the
Steadicam® and all shot by the Steadicam’s inventor, Garrett
Brown.
The
Steadicam introduced a brand new vocabulary of camera movement into
motion pictures and won Brown his first Academy Award in 1978. In
1985, the Washington Post,
with only a slight hint of hyperbole, described the Steadicam as “the
biggest thing since Technicolor”.
Yet
the Steadicam was only the beginning of Brown’s innovations in
camera-control technology. DiveCam put the viewer beside Olympic
divers from their leap off the springboard to their plunge into the
pool. And the MobyCam moved the audience underneath the water in sync
with the athletes in competitive swimming events. In 2006, Brown
received another Academy Award for Scientific and Technical
Achievement for SkyCam, the aerial camera system which has become a
staple of sports stadium broadcasts.
Looking
at Brown’s early career, you would hardly imagine that he would
revolutionise cinematography. He left college to pursue a calling in
folk singing, even recording for MGM, but quit when The Beatles came
along, he quips. With no job skills he ended up despairingly selling
VWs. “I'd always loved movies and my wife agreed to keep working
while I learned moviemaking by reading all the outdated film books in
the Philly library,” he says.
Fast
forward through bit-part employment as agency copywriter, commercial
director and production start-up, complete with an 800lb ‘Fearless
Panoram’ dolly to move his Bolex.
The
Eureka moment has passed into legend. Fed-up with the cumbersome
Fearless, Brown launched a project to isolate his handheld self from
the camera. In 1972, he began experiments and had a functional object
a year later.
“Even
my big early versions worked astonishingly well, even though they
were way too clumsy and burdensome to be commercially successful,”
he recalls. “I finally went into a hotel for a week and looked at
all the drawings over and over and forced myself to come up with a
smaller, lighter version that could actually handle 35mm movie
cameras. And the marvellous result was that, unlike most inventions,
mine could be demonstrated without giving away how it worked. I could
show the results — a reel of impossible shots — and just blow
away anybody in Hollywood who knew what was possible and what wasn't,
but give them no clue how it was done.”
The
demo reel included shots of a friend swimming the length of a pool
and his wife (then girlfriend) running across a park and up and down
Philly Art Museum steps. Rocky
director John Avildsen got hold of a copy and called Brown up to
recreate the scene with Sylvester Stallone. The same year, 1975,
Brown was hired to shoot Steadicam scenes in Bound For
Glory for Hal Ashby and Marathon
Man for John Schlesinger.
Having
devised the Steadicam as a humble means to “rid himself of my big
crusty old dolly”, Brown confesses to being astounded by his
invention's present ubiquity and usefulness.
The
somewhat rigid language of old linear moves, literally ‘on rails’,
has given way, he admits, to a flowing vernacular “that transports
movie narratives and more closely resembles the way humans - with our
astonishing internal stabilisers - actually perceive our lives.”
Tracking
shots were part of the lexicon of cinema long before 1975 of course.
Directors like Orson Welles and the classic three and half minute
opening sequence to Touch Of Evil
(1958) or Alfred Hitchcock's experimental black comedy Rope
(1948) composed of several single 10-minute film reels, have always
sought to push the boundaries of cinematic time and space.
The
Steadicam though freed the cinematographer to plot ever more complex
and fluid compositions.
“My
wife will nudge me in the middle of a particularly great Steadicam
shot and it's still a thrill,” says Brown. “There are so many
brilliant practitioners and it truly is an instrument, rather than
just a stabiliser. It’s simply an elegant way to move an object in
space, with a mass and weightlessness that could never be
accomplished by hand. You guide it with your fingertips and the
result is a really graceful, beautiful move. At its best, it's like a
ballet for the lens. Of course it’s not curing cancer or ending
WWII, but it's still extraordinarily useful and an immense amount of
fun.”
A
perfect example of immersive film-making (from Steadicam operator
Larry McConkey) is the three-minute shot in Martin Scorsese’s
Goodfellas (1990) in
which Ray Liotta’s mobster leads his date into a nightclub through
the exclusive back entrance, along winding corridors, through a busy
kitchen and to a VIP table. While showcasing the supreme command
Scorsese has over cinematic technique, the shot also invites the
audience into the continuous hustle and bustle of the mobster's
world.
Feature-length
films like Alexander Sokurov’s Russian Ark (2002)
have been filmed in one take, choreographed (after having to restart
three times) by DP and Steadicam operator Tillman BĂĽttner to render
the finished piece more like a ballet.
“I
can’t say I am necessarily enthralled with ‘one-ers’ unless
they’re both sensible and valuable – nobody pays any attention to
cuts, after all,” says Brown. “But the freedom to get the lens
exactly where it’s wanted, to carry on up steps and over doorways
in French curves that would drive a dolly crew berserk, remains
completely seductive.”
There
are so many great Steadicam shots, so asking Brown to select a
personal top ten is like asking someone to choose which child they
prefer. “I’ve just re-watched Joe Wright’s Pride And
Prejudice (2005) and Simon Baker
made some ravishingly beautiful and narratively perfect Steadicam
sequences,” he says. “And The Revenant
(2015) was astoundingly vital and gripping. Alejandro Iñárritu
designed, and Scotty Sakamoto operated, some of my favourite
sequences of all time.”
Brown,
a member of American Society of Cinematographers, contributed to
numerous features including Reds (1981),
One From The Heart (1981),
The King of Comedy (1982),
Indiana Jones And The Temple Of Doom
(1984), Casino (1995)
and Bulworth (1998)
before retiring from shooting in 2004 to concentrate on refining an
arsenal of camera stabilisation supports for which he holds 50
patents. FlyCam is a high-speed point-to-point system; GoCam is a
speedy miniature rail system and SuperFlyCam is an ultra light 35mm
wire-borne flyer.
Brown's
most recently released invention and his all-time favourite, was not
however a commercial success. Tango is a miniature crane perched on a
Steadicam arm that permits floor-to-ceiling shooting and
“marvellously smooth” traverses. “You stroll along, panning and
tilting with a camera-less ‘master’ sled in one hand and a
pantographically controlled ‘slave’ sled on the far end, and the
little lightweight camera perfectly follows your intended moves,”
he describes.
“In
the old days I used to take out all of my camera inventions and shoot
impressively with them to jump-start sales. Since I retired from
shooting, unhappily there has been no champion for Tango. However,
I’m confident it will be revived eventually. It’s too good and
too exciting and is huge fun to operate.”
When
Brown started out, the technology for smooth camera movement were
dollies, cranes and camera cars, all land bound. “Aerials [via
helicopter] came with fierce propwash and needed lots of space,” he
says. “Now, gyro technology lets even minuscule platforms yield
eerily smooth shots; and though much of the operating is ‘legato’
to say the least, and thus a bit dreamlike, that will certainly
improve.”
Filmmaking
via drone has taken Hollywood by storm, much in the way Steadicam
once did. Does Brown think UAVs might also affect the language of
cinema in time?
“Humans
unconsciously ‘operate’ their eyeballs with fierce authority, so
even though drones may show us startling vistas, their ‘effect’
is often relatively druggy and tame,” he says. “Eventually pilots
and operators will acquire the rapid and precise panning/tilting
chops that are a given with Steadicam, and failsafe drones will
finally come into their own as narrative tools.”
Brown
believes that gyro-stabilising and remote-control, and even
autonomously operating technology, are here to stay and will only
become more and more astounding “until of course, we take it all
for granted!” But he's savvy enough to realise that even Skycam
will eventually be displaced “by harmless clouds of nearly
invisible drones that swarm around football squads, each assigned to
a hapless player whose only defence will be a badminton racquet!”
Brown
still has several unreleased inventions, which may yet revolutionise
Steadicam operation and continue to provide the most visceral control
of both moves and framing.
“I
learned long ago to only attempt what I personally want. What still
interests me are the fundamentals – how we perceive moving images,
for example. The externals of camera manipulation, rather than the
internal particulars. I also think we should help people understand
that inventing is something that almost any of us might do. You don’t
necessarily have to be a technical soul; you just have to really want
something and to be motivated enough to chase it with a little money
and a lot of thought.”
Subscribe to:
Posts (Atom)