´óÏó´«Ã½

Trialling live music performances in immersive 3D worlds: MAX-R

Using broadcast cameras to capture content that can be delivered directly to users for experiencing in 3D environments - such as video game worlds.

Published: 13 September 2024
  • Fiona Rivera (PhD)

    Fiona Rivera (PhD)

    Senior R&D Development Producer
  • Graham Thomas (MA PhD CEng FIET)

    Graham Thomas (MA PhD CEng FIET)

    Head of Applied Research, Production
  • Bruce Weir (PhD)

    Bruce Weir (PhD)

    Senior Engineer

´óÏó´«Ã½ Research & Development has been at the forefront of applications of augmented reality (AR) and virtual reality (VR) in broadcasting since the early days of virtual studios, bringing some of the magic that this technology enables to programme production. One example from this work is the sports graphics system that is used by the ´óÏó´«Ã½ and many other broadcasters around the world for post-match analysis.

We are now looking at capturing content that can be delivered directly to users for experiencing in 3D environments such as games, rather than just passively watching conventional video.

2D or not 2D? Exploring different approaches

One strand of this work involves investigating the use of volumetric capture technology. This is ideal for intimate performances of one or two artists under controlled lighting conditions and allows participants to view the performer from any location.

We are also researching approaches using broadcast cameras as a part of our work in . This approach may be better suited for a larger number of performers, with complex stage and lighting conditions, at the expense of a less 3D feel and a more limited range of viewpoints.

To give the audiences who are participating ('players') the illusion that the performer is in the space with them, our aim was to produce a realistic effect using video billboards (video projected onto 2D planes). The main disadvantage of video billboards is that they are flat, and the perspective doesn’t change as the player views them from different angles, breaking the illusion of 3D. To counteract this problem, we are filming the performer(s) simultaneously from multiple angles and swapping between the video feeds depending on the position of the player.

Illustration of the camera set up.

We recently completed a live capture trial with four forward-thinking music artists who were willing to play live for us, be streamed into our 3D test worlds, and allow us to film to the performances for subsequent development of the technology, and for further trials as part of our MAX-R work.

Trial with live music performances

We collaborated with studios and to conduct our first technical trial and media shoot of music artists. They performed and were streamed live into a 3D space hosted by our project partner . The trial was directed and produced by ´óÏó´«Ã½ R&D’s Fiona Rivera with Spencer Marsden as technical producer. Three solo artists and one band were enlisted and performed with full stage lighting, dynamic graphics displayed on an LED wall backdrop, with some occasional haze. Collaborating with Production Park gave us the opportunity to use real live event stage technology for the artists. This let us to test our capture and streaming solution for potential integration with 3D environments under realistic yet controllable performance conditions, and to film media for further enhancements to our approach.

A behind the scenes look at the live performance capture technical trial.

We used three locked off cameras (fixed position and settings) pointing centrally, and to either side of the stage for our trial. Although not essential, we also used a roaming camera to capture closeups that could optionally be displayed on large virtual video screens placed to either side of the virtual stage, as would be the case in music festivals.

Time lapse video showing the rigging (setting up of the stage, LED wall, lights and cameras).

To accurately reflect the viewing angles within the 3D world, we need to know the physical location of the video cameras relative to the stage. To achieve this, we created some calibration software that matches points in the video images to known points in the 3D model of the performance space. When at least four points are set, the software can then work out the position and field of view of that video camera.

Screenshot of the calibration software matching points in the video to a 3D model of the space.
Example of the calibration software being used to match points in the video to an accurate 3D model of the performance space.

Our colleague Bruce Weir also developed a Blueprint Visual Scripting Class customised for a test environment in that can read in the camera calibration information described above and accurately project video billboards into a 3D environment. We chose Unreal Engine because of its compatibility with Improbable’s metaverse platform. An adaptation of this Blueprint was also integrated into a testbed environment on their platform using a scale 3D model replica of the stage to be used in our trial.

A screenshot of the 3D model of the stage used in the trial.
3D model of the stage used for the trial.

We used a 7m x 5.7m stage set, with lights and LED wall rigged from a truss construction. The LED wall was a set of 10 x 5 Roe Black Pearl panels, driven by a Brompton Tessera S4 running at low latency, with Resolume software controlling video playback. We also tested using Unreal Engine as a triggered content generator with one of Production Park's projects. Additional LED wall graphics were provided by our graphics team, and in some cases, the artists’ teams.

Lights were programmed and controlled to be bold, and synchronised with the musical content, not only highlighting the artist but also helping define the stage dimensions. As haze is used extensively in live events, it was also an ideal opportunity to see how it affected the capture pipeline.

An image of lighting and smoke effects on an empty stage - a locked off camera is pointing at the empty stage.
Testing out haze and lighting during setup.
Music artist TWST performing on the stage viewed from the lighting desk.
Music artist TWST performing on the stage viewed from the lighting desk.

The sound was prioritised for artist on stage monitoring or in-ear where preferred, with some sub bass units to keep it vibey and familiar onstage. As there was no in-person audience in the studio, we did not rig a public address (PA) system, this also helped keep operational audio levels manageable for the crew.

The technical trial

We aimed to use our multi-camera setup to stream live performances into two stripped-back 3D environments. We wanted to check the practicalities of our capture set up, along with what improvements we might need to make to the capture and streaming process before launching user trials with purpose-built 3D environments reflecting a live music event.

Music artist KYDN performing on stage overseen by the streaming and graphics crew.
Music artist KYDN performing on stage overseen by the streaming and graphics crew.

While filming live, we merged four video feeds into a single quad image for streaming to our Unreal application and simultaneously to our partner Improbable’s 3D testbed. This was largely due to trying to keep steaming bandwidth down to a minimum so as not to cause undue latency in the 3D testbeds, and to ensure that all video feeds remained synchronised with the audio. The quad included video feeds from the three locked off cameras for the virtual video billboards of the performers, designed to be switched based on position of the users / players (top left, top right, and lower left). The lower right of the quad was used for the feed from close ups provided by the roaming camera offering optional media for supplementary side screens.

A view of a screen showing four separate cameras feeds of the band New Wounds performing in a quad view.
Example of four cameras feeding into the quad views for streaming to the 3D applications (Band: New Wounds)

Conclusion

We successfully streamed into our basic test environments with the video quad sections from three locked-off cameras showing the performers on the virtual stage. The video feed from the lower right of the quad supplied the roaming camera shots for the virtual side screens. Our test users (who joined the virtual event online through Improbable’s platform) were able to move around the environment as avatars while watching and listening to the performers. Although not included as a focus of the trial, the users could also interact with each other through audio and text and express themselves through some dance and clapping movements.

We plan further explorations of the technical side of capturing and streaming and to improve the transitions between camera views within the 3D spaces. We also plan to develop a 3D environment complete with interactive functionalities for user trials of live immersive music events.

Screenshot of the virtual scene - the music artist is on stage and a single viewer is in the audience with their name - 'Emma Y' floating above their head.
A player joins our test as an avatar and watches the performance in Improbable's test environment.
Another screenshot of the same virtual scene - this version shows a higher definition and better quality image - the music artist is on stage and a single 'Robocop'-style avatar watches from the audience.
A player watches a music performer in our own Unreal Engine test environment.
The team involved in the Max-R trial shoot.

Acknowledgements

We would like to thank the music artists - Badliana, KYDN, New Wounds, and TWST - for their willingness to perform live for our trial, and to their teams who supported them.

We also would like to thank our hosts Production Park and crew provided by Academy of Live Technology, including Production Park leads Ian Caballero, Connor Stead, and Aby Cohen MA tutor, Deryck Jones (LED wall tech), Joshua Warnes (LED wall tech), Euan Macphail (sound), Reece Burdon (kit), Tom Stokes (behind the scenes photography), James Ward (KYDN graphics), Alexander Chun (lighting tech).

Our thanks also go to Alex Landen and the team at Improbable for supporting and hosting the technical trial. Images of the live studio that are used throughout this article are courtesy of Tom Stokes.

The ´óÏó´«Ã½ MAX-R on set crew included Fiona Rivera (trial director and lead producer), Sonal Tandon (DoP, TWST segment director), Thom Hetherington (production management, camera op), Spenser Marsden (technical producer), Jack Reynolds (streaming and sound), Maxine Glancy (streaming & artist liaison), and Alia Sheikh (lighting and graphics). Our technical team included Bruce Weir who developed the capture and calibration approaches, with Paul Golds supplying expertise in Unreal Engine Blueprints and environments, and Peter Mills in streaming setups. Thanks also go to our ´óÏó´«Ã½ colleague Thadeous Matthews for Badliana graphics.

MAX-R is a 30-month international collaboration co-funded by Horizon Europe and Innovate UK, Grant agreement ID: 101070072.

Rebuild Page

The page will automatically reload. You may need to reload again if the build takes longer than expected.

Useful links

Theme toggler

Select a theme and theme mode and click "Load theme" to load in your theme combination.

Theme:
Theme Mode: