INTERACTIVE CONTENT CREATION
First Claim
1. A method for audio content creation, comprising:
- defining a plurality of three-dimensional collision volumes, the plurality of three-dimensional collision zones being at least partially different three-dimensional spaces;
assigning a base track of music to each of the plurality of three-dimensional collision volumes;
receiving a depth image including depth data representing distances from an origin to objects within a scene;
processing the depth data to determine an instance where the depth data indicates a human user in the scene;
tracking movement of the human user in the instance where the depth data indicates a human user in the scene to determine interaction of the human user with one or more of the plurality of collision volumes; and
automatically changing audio content that is played upon interaction of the user with the one or more of the plurality of collision volumes.
2 Assignments
0 Petitions
Accused Products
Abstract
An audio/visual system (e.g., such as an entertainment console or other computing device) plays a base audio track, such as a portion of a pre-recorded song or notes from one or more instruments. Using a depth camera or other sensor, the system automatically detects that a user (or a portion of the user) enters a first collision volume of a plurality of collision volumes. Each collision volume of the plurality of collision volumes is associated with a different audio stem. In one example, an audio stem is a sound from a subset of instruments playing a song, a portion of a vocal track for a song, or notes from one or more instruments. In response to automatically detecting that the user (or a portion of the user) entered the first collision volume, the appropriate audio stem associated with the first collision volume is added to the base audio track or removed from the base audio track.
-
Citations
20 Claims
-
1. A method for audio content creation, comprising:
-
defining a plurality of three-dimensional collision volumes, the plurality of three-dimensional collision zones being at least partially different three-dimensional spaces; assigning a base track of music to each of the plurality of three-dimensional collision volumes; receiving a depth image including depth data representing distances from an origin to objects within a scene; processing the depth data to determine an instance where the depth data indicates a human user in the scene; tracking movement of the human user in the instance where the depth data indicates a human user in the scene to determine interaction of the human user with one or more of the plurality of collision volumes; and automatically changing audio content that is played upon interaction of the user with the one or more of the plurality of collision volumes. - View Dependent Claims (2, 3, 4, 5, 6, 7, 8, 9, 10, 11, 12, 13, 16, 17)
-
-
14. An apparatus that creates audio content, comprising:
-
a depth camera for producing depth data indicative of distances to objects in a scene; a display interface; an audio interface; and a processor in communication with the depth camera, display interface and audio interface, the processor configured to process the depth data from the depth camera to determine the presence and movement of a human user in the scene, the processor further configured to play a first base audio track, and detect first predefined movement of the user from a plurality of predefined movements based on data from the depth camera, each predefined movement is associated with a different audio stem, the processor adds a first audio stem to the base track in response to detecting that the user performed the first predefined movement, the first audio stem corresponds to the first predefined movement. - View Dependent Claims (15)
-
-
18. One or more processor readable storage devices storing processor readable code thereon, the processor readable code for programming one or more processors to perform a method comprising:
-
obtaining depth data representing distances to points in a scene; detecting a human user in the scene from analysis of the depth data; defining a plurality of three dimensional movement zones and a plurality of collision volumes in the scene; defining one or more base tracks for the plurality of movement zones; detecting when a body part of the user enters one or more of the movement zones from the depth data indicating a three-dimensional position of the body part; detecting when a body part of the user enters one or more of the collision volumes from the depth data indicating a three-dimensional position of the body part; identifying audio stems for a set of collision volumes for each movement zone; and creating code based on the defined one or more base tracks for the plurality of movement zones and the identified audio stems for the set of collision volumes for each zone, the code capable of configuring a computing device to play the one or base tracks depending on which zone the user is positioned, the code capable of configuring the computing device to add or subtract audio stems based on the user intersecting corresponding collision volumes. - View Dependent Claims (19, 20)
-
Specification