Affiliate Disclosure
If you buy through our links, we may get a commission. Read our ethics policy.

Apple filing takes Podcasts to the next level

A recently published filing discovered by AppleInsider reveals work by Apple's chief software architect to advance the Podcast beyond its static form and into a live interactive presentation medium suitable for use by educational institutes and businesses for their daily presentations.

"Podcasts of classroom lectures and other presentations typically require manual editing to switch the focus between the video feed of [an] instructor and the slides (or other contents) being presented," Bertrand Serlet, Senior Vice President of Software Engineering at Apple, wrote in the 15-page filing. "In a school or enterprise where many presentations take place daily, editing podcasts require a dedicated person, which can be prohibitive. "

To solve this problem, Serlet proposes has proposed an automated content capture and processing system where a live camera feed of a presenter can be automatically merged with a Keynote or PowerPoint presentation to form an entertaining and dynamic podcast that lets the viewer watch the presenter's slides as well as the presenter.

In one example outlined in the filing, the content capture system provides a video stream (Stream A) and an Keynote presentation stream (Stream B) to a recording agent such as a Mac running specialized Podcast creation software. The recording agent then blends the two feeds together based on certain cues and sends the combined feed to a syndication server that would then distribute the video wirelessly as a Podcast to any number of authorized Macs, iPods or iPhones.

Serlet also explained that syndication server could include an automated content creation application that applies one or more operations on the Streams A and/or B to create new content, such as transitions, effects, titles, graphics, audio, narration, avatars, animations, and so forth.

"For example, a content stream (e.g., Stream B) output by the application can be shown as background (e.g., full screen mode) with a small picture in picture (PIP) window overlying the background for showing the video camera output (e.g., Stream A)," he wrote. "If a slide in Stream B does not change (e.g., the "trigger event") for a predetermined interval of time (e.g., 15 seconds), then Stream A can be operated on (e.g., scaled to full screen on the display). A virtual zoom (e.g., Ken Burns effect) or other effect can be applied to Stream A for a close-up of the instructor or other object (e.g., an audience member) in the environment (e.g., a classroom, lecture hall, studio)."

The Apple executive also explained that trigger events can be captured from the actual presentation environment using, for example, the capture system, including patterns of activity of the instructor giving a presentation and/or of the reaction of an audience watching the presentation.

"The instructor could make certain gestures, or movements (e.g., captured by the video camera), speak certain words, commands or phrases (e.g., captured by a microphone as an audio snippet) or take long pauses before speaking, all of which can generate events in Stream A that can be used to trigger operations," he wrote.

"In one exemplary scenario, the video of the instructor could be shown in full screen as a default. But if the capture system detects that the instructor has turned his back to the audience to read a slide of the presentation, such action can be detected in the video stream and used to apply one or more operations on Stream A or Stream B, including zooming Stream B so that the slide being read by the instructor is presented to the viewer in full screen."

Podcast Patent Example

Throughout the filing, Serlet outlined examples of several other potential trigger events, such as the movement of a presentation pointer (e.g., a laser pointer) which could then be captured and detected as an event by an "event detector." For instance, the direction of the laser pointer to a slide can indicate that the instructor is talking about a particular area of the slide. Therefore, in one implementation, an operation can be to show the slide to the viewer.

"The movement of a laser pointer can be detected in the video stream using AVSR software or other known pattern matching algorithms that can isolate the laser's red dot on a pixel device and track its motion (e.g., centroiding)," he added. "If a red dot is detected, then slides can be switched or other operations performed on the video or application streams. Alternatively, a laser pointer can emit a signal (e.g., radio frequency, infrared) when activated that can be received by a suitable receiver (e.g., a wireless transceiver) in the capture system and used to initiate one or more operations.

In some other implementations, a detection of a change of state in a stream is used to determine what is captured from the stream and presented in the final media file or podcast. For instance, the instructors transition to a new slide can cause a switch back from a camera feed of the instructor to a slide. When a new slide is presented by the instructor, the application stream containing the slide would be shown first as a default configuration, and then switched to the video stream showing the instructor, respectively, after a first predetermined period of time has expired. In other implementations, after a second predetermined interval of time has expired, the streams can be switched back to the default configuration.

Taking his next-generation podcast concept a step further, Serlet went on to say that the capture system could conceivably include a video camera that can follow the instructor as he moves about the environment. The cameras could be moved by human operator or automatically using known location detection technology. The camera location information could then be used to trigger an operation on a stream and/or determine what is captured and presented in the final media file or podcast.

It should be noted that Serlet's concept one of at least three Podcast enhancements proposed by Apple employees in recent patent filings, none of which have come to fruition as of yet. Others include personalized on-demand podcasts and Podmaps.



45 Comments

SpamSandwich 32917 comments · 19 Years

Could make slide presentations more interesting by allowing more on-the-fly interactivity. Sounds like a patent that Steve suggested...

kasper 840 comments · 22 Years

Quote:
Originally Posted by SpamSandwich

Could make slide presentations more interesting by allowing more on-the-fly interactivity. Sounds like a patent that Steve suggested...

It's very similar to Steve's setups for Keynote presentations at Macworld, WWDC, etc.

It's interesting.

K

mh71 43 comments · 17 Years

Color me crazy, but this looks a lot like the patent suits that Apple has been hit with lately. Mainly, that the concept is not really patent worthy. They are taking existing technologies (broadcasting, video streaming and podcasting) and repackaging them in a shiney new product. There is just nothing that is intrinsically new.

The software that the patent describes seems like it should be copywritable material. I could certainly understand that. It would be an original work ( or a derivative of say, FCP, podcast producer and Keynote).

I understand that Apple has to patent this, or else someone else will. Years will go by, and a faceless "intelectual property right" firm will file a suit in the eastern district of TX claiming that Apple is "willfully and purposefully" infringing on their patents and causing them irreprable damage. You know, we've seen the script before.

Ah, just some ramblings against the current state of patent law.

richardlol 5 comments · 16 Years

Doesn't the Screenflow product already do most of this? It's easy to capture video of you and what is going on on your screen, switch back and forth, both at once, zoom in, call out areas, etc.

macinthe408 1050 comments · 18 Years

Wow, look at the analog volume slider on that iPod!