Sunday’s clash between the NFL’s Atlanta Falcons and Jacksonville Jaguars took place in London’s Wembley Stadium, but that wasn’t the only iconic venue where viewers could see the two teams face off. The NFL also simulcast a version of the game on a much smaller scale: inside Andy’s room from Pixar’s “Toy Story.” 

Toy-like representations of the players competed in a real-time, computer-animated version of the game, as characters like Woody, Buzz Lightyear, Bo Peep, Rex and Forky watched from the sidelines. “Toy Story Funday Football” was the result of a first-time collaboration between the NFL, ESPN’s Creative Studio, Beyond Sports, a Sony-owned AI company, and Pixar, which provided its character and environment assets. Meanwhile, Silver Spoon Animation helped create a motion captured, animated trio of color commentators.

Viewers watching the animated collisions on Sunday were probably wondering the same thing: How did this all come together? 

Jason Donati, an award-winning animator and teaching professor of art and design at Northeastern University, says the technology to make this possible has been in place for a few years. However, this is one of the first times all this cutting-edge technology –– the NFL’s stat tracker, Beyond Sports’ active tracking system and the Unity game engine –– has been used in concert.

“You’ve seen pieces of it, but you’ve not really seen it all together, and certainly not all together for something in real time or really close to real time,” Donati says.

ESPN/Disney

It all starts with the NFL’s Next Gen Stats, a statistic tracking tool powered by Amazon Web Services that can monitor players’ locations, movements and speeds in real time. Each NFL player has a motion-tracking chip embedded in their gear that helps the NFL and its coaches track performance. Amazon also integrates that data into its Thursday night football broadcasts on Prime Video.

Donati says ESPN also used state-of-the-art limb sensor motion-tracking data collected by Beyond Sports’s AI active tracking system to create its animated players. 

Donati says technology like this is a massive step forward for animation and motion capture.

“Think back to ‘The Lord of the Rings’ where we’ve got the motion-capture suits and the little [motion-tracking] markers on them,” Donti says. “That certainly wasn’t in real time. You’re capturing that data, and you’re delivering it to animators to apply onto a CG [computer generated] character.”

“Now, with the Hawk-Eye technology, that can happen in real time without any of the trackers,” Donati continues. “From triangulating camera angles, we can place limbs in space and, in real time, capture that data and feed it into a source.”



LEAVE A REPLY

Please enter your comment!
Please enter your name here